diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,78033 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9999230828397816, + "eval_steps": 500, + "global_step": 13000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 7.691716021844474e-05, + "learning_rate": 0.0029999999562067186, + "loss": 1.9092, + "step": 1 + }, + { + "epoch": 0.00015383432043688947, + "learning_rate": 0.0029999998248268762, + "loss": 1.9814, + "step": 2 + }, + { + "epoch": 0.0002307514806553342, + "learning_rate": 0.002999999605860481, + "loss": 1.7023, + "step": 3 + }, + { + "epoch": 0.00030766864087377894, + "learning_rate": 0.0029999992993075455, + "loss": 1.9262, + "step": 4 + }, + { + "epoch": 0.00038458580109222365, + "learning_rate": 0.002999998905168087, + "loss": 2.1237, + "step": 5 + }, + { + "epoch": 0.0004615029613106684, + "learning_rate": 0.0029999984234421297, + "loss": 1.6514, + "step": 6 + }, + { + "epoch": 0.0005384201215291131, + "learning_rate": 0.0029999978541297013, + "loss": 2.2443, + "step": 7 + }, + { + "epoch": 0.0006153372817475579, + "learning_rate": 0.0029999971972308344, + "loss": 1.872, + "step": 8 + }, + { + "epoch": 0.0006922544419660026, + "learning_rate": 0.0029999964527455685, + "loss": 1.8875, + "step": 9 + }, + { + "epoch": 0.0007691716021844473, + "learning_rate": 0.002999995620673947, + "loss": 2.0612, + "step": 10 + }, + { + "epoch": 0.0008460887624028921, + "learning_rate": 0.0029999947010160167, + "loss": 1.9884, + "step": 11 + }, + { + "epoch": 0.0009230059226213368, + "learning_rate": 0.0029999936937718332, + "loss": 2.2815, + "step": 12 + }, + { + "epoch": 0.0009999230828397815, + "learning_rate": 0.0029999925989414545, + "loss": 1.6098, + "step": 13 + }, + { + "epoch": 0.0010768402430582262, + "learning_rate": 0.0029999914165249448, + "loss": 1.8513, + "step": 14 + }, + { + "epoch": 0.001153757403276671, + "learning_rate": 0.002999990146522373, + "loss": 1.9658, + "step": 15 + }, + { + "epoch": 0.0012306745634951158, + "learning_rate": 0.002999988788933813, + "loss": 1.6379, + "step": 16 + }, + { + "epoch": 0.0013075917237135605, + "learning_rate": 0.0029999873437593447, + "loss": 1.7758, + "step": 17 + }, + { + "epoch": 0.0013845088839320053, + "learning_rate": 0.0029999858109990524, + "loss": 1.7943, + "step": 18 + }, + { + "epoch": 0.00146142604415045, + "learning_rate": 0.002999984190653025, + "loss": 1.3816, + "step": 19 + }, + { + "epoch": 0.0015383432043688946, + "learning_rate": 0.0029999824827213572, + "loss": 1.7113, + "step": 20 + }, + { + "epoch": 0.0016152603645873394, + "learning_rate": 0.0029999806872041497, + "loss": 1.6993, + "step": 21 + }, + { + "epoch": 0.0016921775248057841, + "learning_rate": 0.0029999788041015062, + "loss": 1.3472, + "step": 22 + }, + { + "epoch": 0.0017690946850242289, + "learning_rate": 0.002999976833413537, + "loss": 1.7283, + "step": 23 + }, + { + "epoch": 0.0018460118452426736, + "learning_rate": 0.002999974775140357, + "loss": 1.0342, + "step": 24 + }, + { + "epoch": 0.0019229290054611184, + "learning_rate": 0.0029999726292820874, + "loss": 1.3413, + "step": 25 + }, + { + "epoch": 0.001999846165679563, + "learning_rate": 0.0029999703958388524, + "loss": 1.3192, + "step": 26 + }, + { + "epoch": 0.002076763325898008, + "learning_rate": 0.0029999680748107826, + "loss": 1.4627, + "step": 27 + }, + { + "epoch": 0.0021536804861164525, + "learning_rate": 0.002999965666198014, + "loss": 1.2075, + "step": 28 + }, + { + "epoch": 0.0022305976463348975, + "learning_rate": 0.0029999631700006863, + "loss": 1.3928, + "step": 29 + }, + { + "epoch": 0.002307514806553342, + "learning_rate": 0.0029999605862189465, + "loss": 1.3366, + "step": 30 + }, + { + "epoch": 0.002384431966771787, + "learning_rate": 0.002999957914852945, + "loss": 1.2492, + "step": 31 + }, + { + "epoch": 0.0024613491269902315, + "learning_rate": 0.002999955155902837, + "loss": 1.5065, + "step": 32 + }, + { + "epoch": 0.002538266287208676, + "learning_rate": 0.0029999523093687847, + "loss": 1.3048, + "step": 33 + }, + { + "epoch": 0.002615183447427121, + "learning_rate": 0.002999949375250953, + "loss": 1.209, + "step": 34 + }, + { + "epoch": 0.0026921006076455656, + "learning_rate": 0.002999946353549515, + "loss": 1.457, + "step": 35 + }, + { + "epoch": 0.0027690177678640106, + "learning_rate": 0.002999943244264646, + "loss": 1.0777, + "step": 36 + }, + { + "epoch": 0.002845934928082455, + "learning_rate": 0.0029999400473965277, + "loss": 1.1821, + "step": 37 + }, + { + "epoch": 0.0029228520883009, + "learning_rate": 0.0029999367629453466, + "loss": 0.8748, + "step": 38 + }, + { + "epoch": 0.0029997692485193446, + "learning_rate": 0.0029999333909112953, + "loss": 1.3999, + "step": 39 + }, + { + "epoch": 0.003076686408737789, + "learning_rate": 0.0029999299312945694, + "loss": 1.1431, + "step": 40 + }, + { + "epoch": 0.003153603568956234, + "learning_rate": 0.0029999263840953722, + "loss": 1.2118, + "step": 41 + }, + { + "epoch": 0.0032305207291746787, + "learning_rate": 0.00299992274931391, + "loss": 1.1694, + "step": 42 + }, + { + "epoch": 0.0033074378893931237, + "learning_rate": 0.0029999190269503952, + "loss": 1.5292, + "step": 43 + }, + { + "epoch": 0.0033843550496115682, + "learning_rate": 0.0029999152170050456, + "loss": 0.9456, + "step": 44 + }, + { + "epoch": 0.0034612722098300132, + "learning_rate": 0.002999911319478083, + "loss": 1.3324, + "step": 45 + }, + { + "epoch": 0.0035381893700484578, + "learning_rate": 0.0029999073343697355, + "loss": 1.2328, + "step": 46 + }, + { + "epoch": 0.0036151065302669027, + "learning_rate": 0.002999903261680235, + "loss": 0.9419, + "step": 47 + }, + { + "epoch": 0.0036920236904853473, + "learning_rate": 0.0029998991014098205, + "loss": 0.8992, + "step": 48 + }, + { + "epoch": 0.003768940850703792, + "learning_rate": 0.0029998948535587345, + "loss": 1.1284, + "step": 49 + }, + { + "epoch": 0.003845858010922237, + "learning_rate": 0.0029998905181272237, + "loss": 1.3362, + "step": 50 + }, + { + "epoch": 0.003922775171140681, + "learning_rate": 0.0029998860951155437, + "loss": 1.3353, + "step": 51 + }, + { + "epoch": 0.003999692331359126, + "learning_rate": 0.0029998815845239508, + "loss": 1.1782, + "step": 52 + }, + { + "epoch": 0.004076609491577571, + "learning_rate": 0.002999876986352709, + "loss": 1.4899, + "step": 53 + }, + { + "epoch": 0.004153526651796016, + "learning_rate": 0.002999872300602087, + "loss": 1.2823, + "step": 54 + }, + { + "epoch": 0.00423044381201446, + "learning_rate": 0.0029998675272723584, + "loss": 1.1416, + "step": 55 + }, + { + "epoch": 0.004307360972232905, + "learning_rate": 0.0029998626663638017, + "loss": 1.4207, + "step": 56 + }, + { + "epoch": 0.0043842781324513495, + "learning_rate": 0.0029998577178767007, + "loss": 1.2175, + "step": 57 + }, + { + "epoch": 0.004461195292669795, + "learning_rate": 0.0029998526818113443, + "loss": 1.1006, + "step": 58 + }, + { + "epoch": 0.0045381124528882395, + "learning_rate": 0.002999847558168027, + "loss": 1.2269, + "step": 59 + }, + { + "epoch": 0.004615029613106684, + "learning_rate": 0.002999842346947047, + "loss": 1.5156, + "step": 60 + }, + { + "epoch": 0.0046919467733251286, + "learning_rate": 0.00299983704814871, + "loss": 1.0262, + "step": 61 + }, + { + "epoch": 0.004768863933543574, + "learning_rate": 0.0029998316617733246, + "loss": 1.5848, + "step": 62 + }, + { + "epoch": 0.0048457810937620185, + "learning_rate": 0.002999826187821205, + "loss": 1.5708, + "step": 63 + }, + { + "epoch": 0.004922698253980463, + "learning_rate": 0.002999820626292671, + "loss": 1.1905, + "step": 64 + }, + { + "epoch": 0.004999615414198908, + "learning_rate": 0.0029998149771880483, + "loss": 1.4874, + "step": 65 + }, + { + "epoch": 0.005076532574417352, + "learning_rate": 0.002999809240507666, + "loss": 1.3132, + "step": 66 + }, + { + "epoch": 0.0051534497346357976, + "learning_rate": 0.0029998034162518586, + "loss": 1.2571, + "step": 67 + }, + { + "epoch": 0.005230366894854242, + "learning_rate": 0.0029997975044209666, + "loss": 1.3234, + "step": 68 + }, + { + "epoch": 0.005307284055072687, + "learning_rate": 0.002999791505015336, + "loss": 1.2275, + "step": 69 + }, + { + "epoch": 0.005384201215291131, + "learning_rate": 0.0029997854180353156, + "loss": 1.2061, + "step": 70 + }, + { + "epoch": 0.005461118375509577, + "learning_rate": 0.002999779243481262, + "loss": 1.1499, + "step": 71 + }, + { + "epoch": 0.005538035535728021, + "learning_rate": 0.0029997729813535353, + "loss": 1.4395, + "step": 72 + }, + { + "epoch": 0.005614952695946466, + "learning_rate": 0.0029997666316525008, + "loss": 1.416, + "step": 73 + }, + { + "epoch": 0.00569186985616491, + "learning_rate": 0.00299976019437853, + "loss": 1.0341, + "step": 74 + }, + { + "epoch": 0.005768787016383355, + "learning_rate": 0.002999753669531999, + "loss": 1.5975, + "step": 75 + }, + { + "epoch": 0.0058457041766018, + "learning_rate": 0.0029997470571132873, + "loss": 0.9277, + "step": 76 + }, + { + "epoch": 0.005922621336820245, + "learning_rate": 0.002999740357122782, + "loss": 0.894, + "step": 77 + }, + { + "epoch": 0.005999538497038689, + "learning_rate": 0.0029997335695608747, + "loss": 1.4544, + "step": 78 + }, + { + "epoch": 0.006076455657257134, + "learning_rate": 0.002999726694427961, + "loss": 0.9724, + "step": 79 + }, + { + "epoch": 0.006153372817475578, + "learning_rate": 0.0029997197317244433, + "loss": 1.4934, + "step": 80 + }, + { + "epoch": 0.006230289977694024, + "learning_rate": 0.0029997126814507264, + "loss": 1.193, + "step": 81 + }, + { + "epoch": 0.006307207137912468, + "learning_rate": 0.0029997055436072236, + "loss": 0.8429, + "step": 82 + }, + { + "epoch": 0.006384124298130913, + "learning_rate": 0.0029996983181943514, + "loss": 1.2726, + "step": 83 + }, + { + "epoch": 0.006461041458349357, + "learning_rate": 0.0029996910052125314, + "loss": 1.218, + "step": 84 + }, + { + "epoch": 0.006537958618567803, + "learning_rate": 0.00299968360466219, + "loss": 0.9128, + "step": 85 + }, + { + "epoch": 0.006614875778786247, + "learning_rate": 0.002999676116543761, + "loss": 1.2736, + "step": 86 + }, + { + "epoch": 0.006691792939004692, + "learning_rate": 0.00299966854085768, + "loss": 1.0608, + "step": 87 + }, + { + "epoch": 0.0067687100992231365, + "learning_rate": 0.00299966087760439, + "loss": 1.0455, + "step": 88 + }, + { + "epoch": 0.006845627259441581, + "learning_rate": 0.0029996531267843384, + "loss": 1.3783, + "step": 89 + }, + { + "epoch": 0.0069225444196600264, + "learning_rate": 0.002999645288397978, + "loss": 1.0357, + "step": 90 + }, + { + "epoch": 0.006999461579878471, + "learning_rate": 0.0029996373624457664, + "loss": 1.2053, + "step": 91 + }, + { + "epoch": 0.0070763787400969155, + "learning_rate": 0.002999629348928166, + "loss": 1.098, + "step": 92 + }, + { + "epoch": 0.00715329590031536, + "learning_rate": 0.0029996212478456455, + "loss": 1.2888, + "step": 93 + }, + { + "epoch": 0.0072302130605338055, + "learning_rate": 0.002999613059198677, + "loss": 1.1579, + "step": 94 + }, + { + "epoch": 0.00730713022075225, + "learning_rate": 0.002999604782987739, + "loss": 1.2183, + "step": 95 + }, + { + "epoch": 0.007384047380970695, + "learning_rate": 0.0029995964192133157, + "loss": 0.9659, + "step": 96 + }, + { + "epoch": 0.007460964541189139, + "learning_rate": 0.002999587967875894, + "loss": 0.9437, + "step": 97 + }, + { + "epoch": 0.007537881701407584, + "learning_rate": 0.002999579428975968, + "loss": 1.0836, + "step": 98 + }, + { + "epoch": 0.007614798861626029, + "learning_rate": 0.002999570802514037, + "loss": 1.0699, + "step": 99 + }, + { + "epoch": 0.007691716021844474, + "learning_rate": 0.002999562088490603, + "loss": 0.7358, + "step": 100 + }, + { + "epoch": 0.007768633182062918, + "learning_rate": 0.0029995532869061766, + "loss": 1.0536, + "step": 101 + }, + { + "epoch": 0.007845550342281363, + "learning_rate": 0.002999544397761271, + "loss": 0.7006, + "step": 102 + }, + { + "epoch": 0.007922467502499808, + "learning_rate": 0.002999535421056405, + "loss": 0.8425, + "step": 103 + }, + { + "epoch": 0.007999384662718252, + "learning_rate": 0.0029995263567921026, + "loss": 1.2017, + "step": 104 + }, + { + "epoch": 0.008076301822936697, + "learning_rate": 0.0029995172049688936, + "loss": 0.9685, + "step": 105 + }, + { + "epoch": 0.008153218983155143, + "learning_rate": 0.0029995079655873126, + "loss": 1.0908, + "step": 106 + }, + { + "epoch": 0.008230136143373586, + "learning_rate": 0.0029994986386478986, + "loss": 0.7359, + "step": 107 + }, + { + "epoch": 0.008307053303592032, + "learning_rate": 0.0029994892241511965, + "loss": 1.1188, + "step": 108 + }, + { + "epoch": 0.008383970463810475, + "learning_rate": 0.0029994797220977555, + "loss": 1.5312, + "step": 109 + }, + { + "epoch": 0.00846088762402892, + "learning_rate": 0.002999470132488131, + "loss": 0.888, + "step": 110 + }, + { + "epoch": 0.008537804784247366, + "learning_rate": 0.002999460455322883, + "loss": 0.8843, + "step": 111 + }, + { + "epoch": 0.00861472194446581, + "learning_rate": 0.0029994506906025763, + "loss": 1.1051, + "step": 112 + }, + { + "epoch": 0.008691639104684255, + "learning_rate": 0.0029994408383277804, + "loss": 1.2141, + "step": 113 + }, + { + "epoch": 0.008768556264902699, + "learning_rate": 0.002999430898499072, + "loss": 1.2885, + "step": 114 + }, + { + "epoch": 0.008845473425121144, + "learning_rate": 0.0029994208711170305, + "loss": 1.3617, + "step": 115 + }, + { + "epoch": 0.00892239058533959, + "learning_rate": 0.0029994107561822415, + "loss": 1.0363, + "step": 116 + }, + { + "epoch": 0.008999307745558034, + "learning_rate": 0.002999400553695296, + "loss": 1.1184, + "step": 117 + }, + { + "epoch": 0.009076224905776479, + "learning_rate": 0.0029993902636567898, + "loss": 0.8976, + "step": 118 + }, + { + "epoch": 0.009153142065994924, + "learning_rate": 0.002999379886067323, + "loss": 1.283, + "step": 119 + }, + { + "epoch": 0.009230059226213368, + "learning_rate": 0.0029993694209275025, + "loss": 1.3905, + "step": 120 + }, + { + "epoch": 0.009306976386431813, + "learning_rate": 0.0029993588682379383, + "loss": 1.2637, + "step": 121 + }, + { + "epoch": 0.009383893546650257, + "learning_rate": 0.0029993482279992477, + "loss": 1.0531, + "step": 122 + }, + { + "epoch": 0.009460810706868703, + "learning_rate": 0.002999337500212051, + "loss": 1.2387, + "step": 123 + }, + { + "epoch": 0.009537727867087148, + "learning_rate": 0.0029993266848769755, + "loss": 1.2568, + "step": 124 + }, + { + "epoch": 0.009614645027305592, + "learning_rate": 0.0029993157819946524, + "loss": 1.0881, + "step": 125 + }, + { + "epoch": 0.009691562187524037, + "learning_rate": 0.002999304791565718, + "loss": 1.0521, + "step": 126 + }, + { + "epoch": 0.00976847934774248, + "learning_rate": 0.002999293713590814, + "loss": 1.1498, + "step": 127 + }, + { + "epoch": 0.009845396507960926, + "learning_rate": 0.0029992825480705883, + "loss": 0.8994, + "step": 128 + }, + { + "epoch": 0.009922313668179372, + "learning_rate": 0.0029992712950056917, + "loss": 1.3625, + "step": 129 + }, + { + "epoch": 0.009999230828397815, + "learning_rate": 0.0029992599543967814, + "loss": 1.3445, + "step": 130 + }, + { + "epoch": 0.01007614798861626, + "learning_rate": 0.00299924852624452, + "loss": 1.3456, + "step": 131 + }, + { + "epoch": 0.010153065148834704, + "learning_rate": 0.002999237010549575, + "loss": 1.2097, + "step": 132 + }, + { + "epoch": 0.01022998230905315, + "learning_rate": 0.0029992254073126184, + "loss": 1.2986, + "step": 133 + }, + { + "epoch": 0.010306899469271595, + "learning_rate": 0.0029992137165343277, + "loss": 1.3062, + "step": 134 + }, + { + "epoch": 0.010383816629490039, + "learning_rate": 0.0029992019382153855, + "loss": 1.0305, + "step": 135 + }, + { + "epoch": 0.010460733789708484, + "learning_rate": 0.00299919007235648, + "loss": 1.1904, + "step": 136 + }, + { + "epoch": 0.010537650949926928, + "learning_rate": 0.0029991781189583036, + "loss": 1.1767, + "step": 137 + }, + { + "epoch": 0.010614568110145373, + "learning_rate": 0.0029991660780215542, + "loss": 1.3267, + "step": 138 + }, + { + "epoch": 0.010691485270363819, + "learning_rate": 0.0029991539495469355, + "loss": 1.2611, + "step": 139 + }, + { + "epoch": 0.010768402430582262, + "learning_rate": 0.0029991417335351552, + "loss": 1.0256, + "step": 140 + }, + { + "epoch": 0.010845319590800708, + "learning_rate": 0.002999129429986927, + "loss": 1.1199, + "step": 141 + }, + { + "epoch": 0.010922236751019153, + "learning_rate": 0.002999117038902968, + "loss": 1.138, + "step": 142 + }, + { + "epoch": 0.010999153911237597, + "learning_rate": 0.002999104560284004, + "loss": 1.232, + "step": 143 + }, + { + "epoch": 0.011076071071456042, + "learning_rate": 0.0029990919941307614, + "loss": 1.1461, + "step": 144 + }, + { + "epoch": 0.011152988231674486, + "learning_rate": 0.0029990793404439755, + "loss": 1.1026, + "step": 145 + }, + { + "epoch": 0.011229905391892931, + "learning_rate": 0.0029990665992243844, + "loss": 0.9884, + "step": 146 + }, + { + "epoch": 0.011306822552111377, + "learning_rate": 0.0029990537704727327, + "loss": 1.3524, + "step": 147 + }, + { + "epoch": 0.01138373971232982, + "learning_rate": 0.0029990408541897685, + "loss": 1.2749, + "step": 148 + }, + { + "epoch": 0.011460656872548266, + "learning_rate": 0.002999027850376247, + "loss": 1.2295, + "step": 149 + }, + { + "epoch": 0.01153757403276671, + "learning_rate": 0.0029990147590329267, + "loss": 1.1709, + "step": 150 + }, + { + "epoch": 0.011614491192985155, + "learning_rate": 0.0029990015801605726, + "loss": 0.7045, + "step": 151 + }, + { + "epoch": 0.0116914083532036, + "learning_rate": 0.002998988313759954, + "loss": 1.0454, + "step": 152 + }, + { + "epoch": 0.011768325513422044, + "learning_rate": 0.0029989749598318453, + "loss": 1.2119, + "step": 153 + }, + { + "epoch": 0.01184524267364049, + "learning_rate": 0.0029989615183770267, + "loss": 1.4649, + "step": 154 + }, + { + "epoch": 0.011922159833858933, + "learning_rate": 0.002998947989396283, + "loss": 1.183, + "step": 155 + }, + { + "epoch": 0.011999076994077379, + "learning_rate": 0.002998934372890404, + "loss": 1.307, + "step": 156 + }, + { + "epoch": 0.012075994154295824, + "learning_rate": 0.0029989206688601847, + "loss": 0.9439, + "step": 157 + }, + { + "epoch": 0.012152911314514268, + "learning_rate": 0.0029989068773064254, + "loss": 1.4585, + "step": 158 + }, + { + "epoch": 0.012229828474732713, + "learning_rate": 0.0029988929982299314, + "loss": 1.0362, + "step": 159 + }, + { + "epoch": 0.012306745634951157, + "learning_rate": 0.0029988790316315132, + "loss": 0.9236, + "step": 160 + }, + { + "epoch": 0.012383662795169602, + "learning_rate": 0.0029988649775119862, + "loss": 1.1296, + "step": 161 + }, + { + "epoch": 0.012460579955388048, + "learning_rate": 0.002998850835872171, + "loss": 0.7674, + "step": 162 + }, + { + "epoch": 0.012537497115606491, + "learning_rate": 0.0029988366067128934, + "loss": 1.3065, + "step": 163 + }, + { + "epoch": 0.012614414275824937, + "learning_rate": 0.0029988222900349847, + "loss": 0.9165, + "step": 164 + }, + { + "epoch": 0.012691331436043382, + "learning_rate": 0.00299880788583928, + "loss": 1.0541, + "step": 165 + }, + { + "epoch": 0.012768248596261826, + "learning_rate": 0.002998793394126621, + "loss": 1.2432, + "step": 166 + }, + { + "epoch": 0.012845165756480271, + "learning_rate": 0.0029987788148978537, + "loss": 1.2695, + "step": 167 + }, + { + "epoch": 0.012922082916698715, + "learning_rate": 0.002998764148153829, + "loss": 1.6526, + "step": 168 + }, + { + "epoch": 0.01299900007691716, + "learning_rate": 0.0029987493938954048, + "loss": 1.4513, + "step": 169 + }, + { + "epoch": 0.013075917237135606, + "learning_rate": 0.0029987345521234406, + "loss": 1.1943, + "step": 170 + }, + { + "epoch": 0.01315283439735405, + "learning_rate": 0.002998719622838804, + "loss": 1.2793, + "step": 171 + }, + { + "epoch": 0.013229751557572495, + "learning_rate": 0.002998704606042367, + "loss": 1.2412, + "step": 172 + }, + { + "epoch": 0.013306668717790938, + "learning_rate": 0.002998689501735006, + "loss": 1.5079, + "step": 173 + }, + { + "epoch": 0.013383585878009384, + "learning_rate": 0.0029986743099176035, + "loss": 1.124, + "step": 174 + }, + { + "epoch": 0.01346050303822783, + "learning_rate": 0.0029986590305910456, + "loss": 1.4621, + "step": 175 + }, + { + "epoch": 0.013537420198446273, + "learning_rate": 0.0029986436637562254, + "loss": 0.8467, + "step": 176 + }, + { + "epoch": 0.013614337358664718, + "learning_rate": 0.0029986282094140396, + "loss": 1.0726, + "step": 177 + }, + { + "epoch": 0.013691254518883162, + "learning_rate": 0.002998612667565391, + "loss": 0.8901, + "step": 178 + }, + { + "epoch": 0.013768171679101607, + "learning_rate": 0.002998597038211187, + "loss": 1.2866, + "step": 179 + }, + { + "epoch": 0.013845088839320053, + "learning_rate": 0.0029985813213523394, + "loss": 1.2543, + "step": 180 + }, + { + "epoch": 0.013922005999538497, + "learning_rate": 0.0029985655169897673, + "loss": 0.9437, + "step": 181 + }, + { + "epoch": 0.013998923159756942, + "learning_rate": 0.002998549625124393, + "loss": 1.0591, + "step": 182 + }, + { + "epoch": 0.014075840319975386, + "learning_rate": 0.0029985336457571438, + "loss": 1.4138, + "step": 183 + }, + { + "epoch": 0.014152757480193831, + "learning_rate": 0.002998517578888954, + "loss": 1.6227, + "step": 184 + }, + { + "epoch": 0.014229674640412276, + "learning_rate": 0.0029985014245207608, + "loss": 1.4723, + "step": 185 + }, + { + "epoch": 0.01430659180063072, + "learning_rate": 0.0029984851826535073, + "loss": 1.3542, + "step": 186 + }, + { + "epoch": 0.014383508960849166, + "learning_rate": 0.0029984688532881426, + "loss": 1.76, + "step": 187 + }, + { + "epoch": 0.014460426121067611, + "learning_rate": 0.00299845243642562, + "loss": 0.858, + "step": 188 + }, + { + "epoch": 0.014537343281286055, + "learning_rate": 0.002998435932066898, + "loss": 1.4137, + "step": 189 + }, + { + "epoch": 0.0146142604415045, + "learning_rate": 0.00299841934021294, + "loss": 1.4808, + "step": 190 + }, + { + "epoch": 0.014691177601722944, + "learning_rate": 0.0029984026608647157, + "loss": 1.1505, + "step": 191 + }, + { + "epoch": 0.01476809476194139, + "learning_rate": 0.0029983858940231976, + "loss": 1.2201, + "step": 192 + }, + { + "epoch": 0.014845011922159835, + "learning_rate": 0.0029983690396893662, + "loss": 1.1977, + "step": 193 + }, + { + "epoch": 0.014921929082378278, + "learning_rate": 0.0029983520978642047, + "loss": 1.2986, + "step": 194 + }, + { + "epoch": 0.014998846242596724, + "learning_rate": 0.0029983350685487027, + "loss": 0.8975, + "step": 195 + }, + { + "epoch": 0.015075763402815167, + "learning_rate": 0.002998317951743854, + "loss": 1.3165, + "step": 196 + }, + { + "epoch": 0.015152680563033613, + "learning_rate": 0.0029983007474506597, + "loss": 1.2725, + "step": 197 + }, + { + "epoch": 0.015229597723252058, + "learning_rate": 0.0029982834556701228, + "loss": 1.3455, + "step": 198 + }, + { + "epoch": 0.015306514883470502, + "learning_rate": 0.0029982660764032534, + "loss": 1.0963, + "step": 199 + }, + { + "epoch": 0.015383432043688947, + "learning_rate": 0.0029982486096510664, + "loss": 1.0346, + "step": 200 + }, + { + "epoch": 0.015460349203907391, + "learning_rate": 0.0029982310554145814, + "loss": 1.281, + "step": 201 + }, + { + "epoch": 0.015537266364125836, + "learning_rate": 0.0029982134136948236, + "loss": 1.6145, + "step": 202 + }, + { + "epoch": 0.015614183524344282, + "learning_rate": 0.0029981956844928235, + "loss": 0.9711, + "step": 203 + }, + { + "epoch": 0.015691100684562725, + "learning_rate": 0.0029981778678096157, + "loss": 0.9053, + "step": 204 + }, + { + "epoch": 0.01576801784478117, + "learning_rate": 0.002998159963646241, + "loss": 1.4662, + "step": 205 + }, + { + "epoch": 0.015844935004999616, + "learning_rate": 0.002998141972003745, + "loss": 1.0082, + "step": 206 + }, + { + "epoch": 0.01592185216521806, + "learning_rate": 0.0029981238928831777, + "loss": 1.4221, + "step": 207 + }, + { + "epoch": 0.015998769325436504, + "learning_rate": 0.002998105726285595, + "loss": 0.7058, + "step": 208 + }, + { + "epoch": 0.01607568648565495, + "learning_rate": 0.0029980874722120576, + "loss": 1.0257, + "step": 209 + }, + { + "epoch": 0.016152603645873394, + "learning_rate": 0.0029980691306636313, + "loss": 1.1321, + "step": 210 + }, + { + "epoch": 0.01622952080609184, + "learning_rate": 0.0029980507016413877, + "loss": 1.0979, + "step": 211 + }, + { + "epoch": 0.016306437966310285, + "learning_rate": 0.0029980321851464015, + "loss": 1.3522, + "step": 212 + }, + { + "epoch": 0.016383355126528727, + "learning_rate": 0.0029980135811797557, + "loss": 0.8686, + "step": 213 + }, + { + "epoch": 0.016460272286747173, + "learning_rate": 0.002997994889742535, + "loss": 1.5654, + "step": 214 + }, + { + "epoch": 0.016537189446965618, + "learning_rate": 0.002997976110835832, + "loss": 0.8115, + "step": 215 + }, + { + "epoch": 0.016614106607184063, + "learning_rate": 0.0029979572444607428, + "loss": 1.0477, + "step": 216 + }, + { + "epoch": 0.01669102376740251, + "learning_rate": 0.0029979382906183687, + "loss": 0.9833, + "step": 217 + }, + { + "epoch": 0.01676794092762095, + "learning_rate": 0.002997919249309816, + "loss": 1.4087, + "step": 218 + }, + { + "epoch": 0.016844858087839396, + "learning_rate": 0.0029979001205361986, + "loss": 0.9413, + "step": 219 + }, + { + "epoch": 0.01692177524805784, + "learning_rate": 0.0029978809042986313, + "loss": 1.3196, + "step": 220 + }, + { + "epoch": 0.016998692408276287, + "learning_rate": 0.0029978616005982375, + "loss": 1.3242, + "step": 221 + }, + { + "epoch": 0.017075609568494732, + "learning_rate": 0.0029978422094361435, + "loss": 1.1892, + "step": 222 + }, + { + "epoch": 0.017152526728713174, + "learning_rate": 0.0029978227308134816, + "loss": 1.1711, + "step": 223 + }, + { + "epoch": 0.01722944388893162, + "learning_rate": 0.0029978031647313894, + "loss": 1.1237, + "step": 224 + }, + { + "epoch": 0.017306361049150065, + "learning_rate": 0.0029977835111910096, + "loss": 1.2171, + "step": 225 + }, + { + "epoch": 0.01738327820936851, + "learning_rate": 0.0029977637701934902, + "loss": 1.4833, + "step": 226 + }, + { + "epoch": 0.017460195369586956, + "learning_rate": 0.002997743941739982, + "loss": 1.383, + "step": 227 + }, + { + "epoch": 0.017537112529805398, + "learning_rate": 0.0029977240258316456, + "loss": 1.2232, + "step": 228 + }, + { + "epoch": 0.017614029690023843, + "learning_rate": 0.0029977040224696415, + "loss": 1.2125, + "step": 229 + }, + { + "epoch": 0.01769094685024229, + "learning_rate": 0.002997683931655139, + "loss": 1.0057, + "step": 230 + }, + { + "epoch": 0.017767864010460734, + "learning_rate": 0.0029976637533893112, + "loss": 1.2638, + "step": 231 + }, + { + "epoch": 0.01784478117067918, + "learning_rate": 0.0029976434876733352, + "loss": 1.3071, + "step": 232 + }, + { + "epoch": 0.01792169833089762, + "learning_rate": 0.0029976231345083958, + "loss": 1.1912, + "step": 233 + }, + { + "epoch": 0.017998615491116067, + "learning_rate": 0.0029976026938956808, + "loss": 1.2945, + "step": 234 + }, + { + "epoch": 0.018075532651334512, + "learning_rate": 0.0029975821658363836, + "loss": 1.1906, + "step": 235 + }, + { + "epoch": 0.018152449811552958, + "learning_rate": 0.0029975615503317027, + "loss": 1.329, + "step": 236 + }, + { + "epoch": 0.018229366971771403, + "learning_rate": 0.0029975408473828426, + "loss": 1.1752, + "step": 237 + }, + { + "epoch": 0.01830628413198985, + "learning_rate": 0.0029975200569910113, + "loss": 1.157, + "step": 238 + }, + { + "epoch": 0.01838320129220829, + "learning_rate": 0.0029974991791574235, + "loss": 1.0557, + "step": 239 + }, + { + "epoch": 0.018460118452426736, + "learning_rate": 0.002997478213883298, + "loss": 0.9176, + "step": 240 + }, + { + "epoch": 0.01853703561264518, + "learning_rate": 0.0029974571611698585, + "loss": 1.3192, + "step": 241 + }, + { + "epoch": 0.018613952772863627, + "learning_rate": 0.002997436021018335, + "loss": 1.2865, + "step": 242 + }, + { + "epoch": 0.018690869933082072, + "learning_rate": 0.0029974147934299623, + "loss": 1.2836, + "step": 243 + }, + { + "epoch": 0.018767787093300514, + "learning_rate": 0.002997393478405978, + "loss": 1.161, + "step": 244 + }, + { + "epoch": 0.01884470425351896, + "learning_rate": 0.002997372075947629, + "loss": 0.9845, + "step": 245 + }, + { + "epoch": 0.018921621413737405, + "learning_rate": 0.0029973505860561634, + "loss": 1.3582, + "step": 246 + }, + { + "epoch": 0.01899853857395585, + "learning_rate": 0.002997329008732837, + "loss": 1.3781, + "step": 247 + }, + { + "epoch": 0.019075455734174296, + "learning_rate": 0.0029973073439789086, + "loss": 1.0758, + "step": 248 + }, + { + "epoch": 0.019152372894392738, + "learning_rate": 0.0029972855917956447, + "loss": 1.1509, + "step": 249 + }, + { + "epoch": 0.019229290054611183, + "learning_rate": 0.0029972637521843144, + "loss": 1.2729, + "step": 250 + }, + { + "epoch": 0.01930620721482963, + "learning_rate": 0.002997241825146193, + "loss": 1.3322, + "step": 251 + }, + { + "epoch": 0.019383124375048074, + "learning_rate": 0.0029972198106825616, + "loss": 1.2791, + "step": 252 + }, + { + "epoch": 0.01946004153526652, + "learning_rate": 0.0029971977087947044, + "loss": 1.4703, + "step": 253 + }, + { + "epoch": 0.01953695869548496, + "learning_rate": 0.002997175519483913, + "loss": 1.2515, + "step": 254 + }, + { + "epoch": 0.019613875855703407, + "learning_rate": 0.0029971532427514828, + "loss": 1.1393, + "step": 255 + }, + { + "epoch": 0.019690793015921852, + "learning_rate": 0.002997130878598714, + "loss": 1.3116, + "step": 256 + }, + { + "epoch": 0.019767710176140298, + "learning_rate": 0.0029971084270269135, + "loss": 1.0154, + "step": 257 + }, + { + "epoch": 0.019844627336358743, + "learning_rate": 0.002997085888037391, + "loss": 1.0423, + "step": 258 + }, + { + "epoch": 0.019921544496577185, + "learning_rate": 0.002997063261631464, + "loss": 1.0003, + "step": 259 + }, + { + "epoch": 0.01999846165679563, + "learning_rate": 0.0029970405478104522, + "loss": 1.2568, + "step": 260 + }, + { + "epoch": 0.020075378817014076, + "learning_rate": 0.0029970177465756835, + "loss": 1.1609, + "step": 261 + }, + { + "epoch": 0.02015229597723252, + "learning_rate": 0.002996994857928488, + "loss": 1.55, + "step": 262 + }, + { + "epoch": 0.020229213137450967, + "learning_rate": 0.0029969718818702025, + "loss": 1.1516, + "step": 263 + }, + { + "epoch": 0.02030613029766941, + "learning_rate": 0.0029969488184021687, + "loss": 1.3809, + "step": 264 + }, + { + "epoch": 0.020383047457887854, + "learning_rate": 0.002996925667525733, + "loss": 1.4538, + "step": 265 + }, + { + "epoch": 0.0204599646181063, + "learning_rate": 0.0029969024292422486, + "loss": 0.8137, + "step": 266 + }, + { + "epoch": 0.020536881778324745, + "learning_rate": 0.0029968791035530702, + "loss": 1.0519, + "step": 267 + }, + { + "epoch": 0.02061379893854319, + "learning_rate": 0.0029968556904595616, + "loss": 1.4587, + "step": 268 + }, + { + "epoch": 0.020690716098761632, + "learning_rate": 0.0029968321899630887, + "loss": 1.2783, + "step": 269 + }, + { + "epoch": 0.020767633258980078, + "learning_rate": 0.0029968086020650247, + "loss": 1.0779, + "step": 270 + }, + { + "epoch": 0.020844550419198523, + "learning_rate": 0.0029967849267667464, + "loss": 1.0817, + "step": 271 + }, + { + "epoch": 0.02092146757941697, + "learning_rate": 0.0029967611640696365, + "loss": 1.3202, + "step": 272 + }, + { + "epoch": 0.020998384739635414, + "learning_rate": 0.0029967373139750822, + "loss": 1.3649, + "step": 273 + }, + { + "epoch": 0.021075301899853856, + "learning_rate": 0.002996713376484476, + "loss": 1.101, + "step": 274 + }, + { + "epoch": 0.0211522190600723, + "learning_rate": 0.002996689351599216, + "loss": 1.0024, + "step": 275 + }, + { + "epoch": 0.021229136220290747, + "learning_rate": 0.0029966652393207055, + "loss": 1.2042, + "step": 276 + }, + { + "epoch": 0.021306053380509192, + "learning_rate": 0.002996641039650351, + "loss": 0.9547, + "step": 277 + }, + { + "epoch": 0.021382970540727637, + "learning_rate": 0.002996616752589567, + "loss": 1.2709, + "step": 278 + }, + { + "epoch": 0.02145988770094608, + "learning_rate": 0.0029965923781397714, + "loss": 1.166, + "step": 279 + }, + { + "epoch": 0.021536804861164525, + "learning_rate": 0.0029965679163023863, + "loss": 1.1005, + "step": 280 + }, + { + "epoch": 0.02161372202138297, + "learning_rate": 0.0029965433670788417, + "loss": 1.1991, + "step": 281 + }, + { + "epoch": 0.021690639181601416, + "learning_rate": 0.0029965187304705695, + "loss": 1.2865, + "step": 282 + }, + { + "epoch": 0.02176755634181986, + "learning_rate": 0.0029964940064790098, + "loss": 1.2444, + "step": 283 + }, + { + "epoch": 0.021844473502038306, + "learning_rate": 0.0029964691951056043, + "loss": 1.1044, + "step": 284 + }, + { + "epoch": 0.02192139066225675, + "learning_rate": 0.0029964442963518042, + "loss": 1.3341, + "step": 285 + }, + { + "epoch": 0.021998307822475194, + "learning_rate": 0.002996419310219061, + "loss": 0.9102, + "step": 286 + }, + { + "epoch": 0.02207522498269364, + "learning_rate": 0.0029963942367088356, + "loss": 1.6223, + "step": 287 + }, + { + "epoch": 0.022152142142912085, + "learning_rate": 0.002996369075822591, + "loss": 1.2945, + "step": 288 + }, + { + "epoch": 0.02222905930313053, + "learning_rate": 0.0029963438275617963, + "loss": 1.1997, + "step": 289 + }, + { + "epoch": 0.022305976463348972, + "learning_rate": 0.002996318491927926, + "loss": 1.1263, + "step": 290 + }, + { + "epoch": 0.022382893623567417, + "learning_rate": 0.00299629306892246, + "loss": 1.1537, + "step": 291 + }, + { + "epoch": 0.022459810783785863, + "learning_rate": 0.002996267558546882, + "loss": 1.0203, + "step": 292 + }, + { + "epoch": 0.022536727944004308, + "learning_rate": 0.0029962419608026815, + "loss": 1.1749, + "step": 293 + }, + { + "epoch": 0.022613645104222754, + "learning_rate": 0.0029962162756913538, + "loss": 0.9919, + "step": 294 + }, + { + "epoch": 0.022690562264441196, + "learning_rate": 0.002996190503214399, + "loss": 1.3889, + "step": 295 + }, + { + "epoch": 0.02276747942465964, + "learning_rate": 0.0029961646433733207, + "loss": 1.2516, + "step": 296 + }, + { + "epoch": 0.022844396584878086, + "learning_rate": 0.00299613869616963, + "loss": 1.1091, + "step": 297 + }, + { + "epoch": 0.022921313745096532, + "learning_rate": 0.0029961126616048417, + "loss": 0.9692, + "step": 298 + }, + { + "epoch": 0.022998230905314977, + "learning_rate": 0.0029960865396804755, + "loss": 1.6793, + "step": 299 + }, + { + "epoch": 0.02307514806553342, + "learning_rate": 0.002996060330398057, + "loss": 1.5065, + "step": 300 + }, + { + "epoch": 0.023152065225751865, + "learning_rate": 0.0029960340337591173, + "loss": 1.0508, + "step": 301 + }, + { + "epoch": 0.02322898238597031, + "learning_rate": 0.0029960076497651906, + "loss": 1.2244, + "step": 302 + }, + { + "epoch": 0.023305899546188755, + "learning_rate": 0.0029959811784178187, + "loss": 1.2929, + "step": 303 + }, + { + "epoch": 0.0233828167064072, + "learning_rate": 0.0029959546197185463, + "loss": 1.0615, + "step": 304 + }, + { + "epoch": 0.023459733866625643, + "learning_rate": 0.002995927973668925, + "loss": 1.2689, + "step": 305 + }, + { + "epoch": 0.023536651026844088, + "learning_rate": 0.0029959012402705097, + "loss": 1.1925, + "step": 306 + }, + { + "epoch": 0.023613568187062534, + "learning_rate": 0.0029958744195248623, + "loss": 1.0969, + "step": 307 + }, + { + "epoch": 0.02369048534728098, + "learning_rate": 0.0029958475114335486, + "loss": 1.2156, + "step": 308 + }, + { + "epoch": 0.023767402507499424, + "learning_rate": 0.0029958205159981398, + "loss": 1.3017, + "step": 309 + }, + { + "epoch": 0.023844319667717866, + "learning_rate": 0.002995793433220212, + "loss": 1.0581, + "step": 310 + }, + { + "epoch": 0.023921236827936312, + "learning_rate": 0.002995766263101347, + "loss": 1.2094, + "step": 311 + }, + { + "epoch": 0.023998153988154757, + "learning_rate": 0.002995739005643131, + "loss": 0.9861, + "step": 312 + }, + { + "epoch": 0.024075071148373203, + "learning_rate": 0.0029957116608471555, + "loss": 0.9938, + "step": 313 + }, + { + "epoch": 0.024151988308591648, + "learning_rate": 0.002995684228715017, + "loss": 1.0069, + "step": 314 + }, + { + "epoch": 0.02422890546881009, + "learning_rate": 0.0029956567092483182, + "loss": 1.0306, + "step": 315 + }, + { + "epoch": 0.024305822629028535, + "learning_rate": 0.0029956291024486655, + "loss": 1.2565, + "step": 316 + }, + { + "epoch": 0.02438273978924698, + "learning_rate": 0.0029956014083176706, + "loss": 1.054, + "step": 317 + }, + { + "epoch": 0.024459656949465426, + "learning_rate": 0.0029955736268569505, + "loss": 0.9882, + "step": 318 + }, + { + "epoch": 0.02453657410968387, + "learning_rate": 0.002995545758068128, + "loss": 1.4163, + "step": 319 + }, + { + "epoch": 0.024613491269902314, + "learning_rate": 0.0029955178019528297, + "loss": 0.9077, + "step": 320 + }, + { + "epoch": 0.02469040843012076, + "learning_rate": 0.002995489758512689, + "loss": 1.0501, + "step": 321 + }, + { + "epoch": 0.024767325590339204, + "learning_rate": 0.002995461627749342, + "loss": 1.2069, + "step": 322 + }, + { + "epoch": 0.02484424275055765, + "learning_rate": 0.0029954334096644326, + "loss": 1.3046, + "step": 323 + }, + { + "epoch": 0.024921159910776095, + "learning_rate": 0.0029954051042596075, + "loss": 1.3116, + "step": 324 + }, + { + "epoch": 0.02499807707099454, + "learning_rate": 0.0029953767115365204, + "loss": 1.2408, + "step": 325 + }, + { + "epoch": 0.025074994231212983, + "learning_rate": 0.002995348231496828, + "loss": 0.8897, + "step": 326 + }, + { + "epoch": 0.025151911391431428, + "learning_rate": 0.0029953196641421946, + "loss": 1.9307, + "step": 327 + }, + { + "epoch": 0.025228828551649873, + "learning_rate": 0.002995291009474287, + "loss": 1.0036, + "step": 328 + }, + { + "epoch": 0.02530574571186832, + "learning_rate": 0.0029952622674947796, + "loss": 1.1598, + "step": 329 + }, + { + "epoch": 0.025382662872086764, + "learning_rate": 0.0029952334382053495, + "loss": 0.8737, + "step": 330 + }, + { + "epoch": 0.025459580032305206, + "learning_rate": 0.0029952045216076807, + "loss": 1.431, + "step": 331 + }, + { + "epoch": 0.02553649719252365, + "learning_rate": 0.0029951755177034622, + "loss": 1.4716, + "step": 332 + }, + { + "epoch": 0.025613414352742097, + "learning_rate": 0.0029951464264943864, + "loss": 0.9831, + "step": 333 + }, + { + "epoch": 0.025690331512960542, + "learning_rate": 0.0029951172479821533, + "loss": 1.1124, + "step": 334 + }, + { + "epoch": 0.025767248673178988, + "learning_rate": 0.002995087982168465, + "loss": 1.0785, + "step": 335 + }, + { + "epoch": 0.02584416583339743, + "learning_rate": 0.0029950586290550317, + "loss": 1.1973, + "step": 336 + }, + { + "epoch": 0.025921082993615875, + "learning_rate": 0.0029950291886435667, + "loss": 1.3794, + "step": 337 + }, + { + "epoch": 0.02599800015383432, + "learning_rate": 0.00299499966093579, + "loss": 1.0184, + "step": 338 + }, + { + "epoch": 0.026074917314052766, + "learning_rate": 0.002994970045933424, + "loss": 1.3075, + "step": 339 + }, + { + "epoch": 0.02615183447427121, + "learning_rate": 0.0029949403436382, + "loss": 1.416, + "step": 340 + }, + { + "epoch": 0.026228751634489653, + "learning_rate": 0.0029949105540518508, + "loss": 1.4782, + "step": 341 + }, + { + "epoch": 0.0263056687947081, + "learning_rate": 0.0029948806771761165, + "loss": 1.1191, + "step": 342 + }, + { + "epoch": 0.026382585954926544, + "learning_rate": 0.0029948507130127416, + "loss": 0.8146, + "step": 343 + }, + { + "epoch": 0.02645950311514499, + "learning_rate": 0.0029948206615634755, + "loss": 2.0228, + "step": 344 + }, + { + "epoch": 0.026536420275363435, + "learning_rate": 0.0029947905228300733, + "loss": 1.329, + "step": 345 + }, + { + "epoch": 0.026613337435581877, + "learning_rate": 0.0029947602968142945, + "loss": 1.0044, + "step": 346 + }, + { + "epoch": 0.026690254595800322, + "learning_rate": 0.002994729983517904, + "loss": 1.1451, + "step": 347 + }, + { + "epoch": 0.026767171756018768, + "learning_rate": 0.002994699582942673, + "loss": 1.4733, + "step": 348 + }, + { + "epoch": 0.026844088916237213, + "learning_rate": 0.002994669095090374, + "loss": 0.8037, + "step": 349 + }, + { + "epoch": 0.02692100607645566, + "learning_rate": 0.00299463851996279, + "loss": 1.0593, + "step": 350 + }, + { + "epoch": 0.0269979232366741, + "learning_rate": 0.0029946078575617046, + "loss": 1.276, + "step": 351 + }, + { + "epoch": 0.027074840396892546, + "learning_rate": 0.002994577107888909, + "loss": 1.3483, + "step": 352 + }, + { + "epoch": 0.02715175755711099, + "learning_rate": 0.0029945462709461985, + "loss": 0.9461, + "step": 353 + }, + { + "epoch": 0.027228674717329437, + "learning_rate": 0.0029945153467353736, + "loss": 1.2791, + "step": 354 + }, + { + "epoch": 0.027305591877547882, + "learning_rate": 0.00299448433525824, + "loss": 0.7401, + "step": 355 + }, + { + "epoch": 0.027382509037766324, + "learning_rate": 0.0029944532365166086, + "loss": 1.539, + "step": 356 + }, + { + "epoch": 0.02745942619798477, + "learning_rate": 0.002994422050512295, + "loss": 1.0192, + "step": 357 + }, + { + "epoch": 0.027536343358203215, + "learning_rate": 0.002994390777247121, + "loss": 1.0759, + "step": 358 + }, + { + "epoch": 0.02761326051842166, + "learning_rate": 0.0029943594167229116, + "loss": 0.9437, + "step": 359 + }, + { + "epoch": 0.027690177678640106, + "learning_rate": 0.0029943279689414984, + "loss": 1.3718, + "step": 360 + }, + { + "epoch": 0.027767094838858548, + "learning_rate": 0.002994296433904718, + "loss": 1.1736, + "step": 361 + }, + { + "epoch": 0.027844011999076993, + "learning_rate": 0.0029942648116144115, + "loss": 1.116, + "step": 362 + }, + { + "epoch": 0.02792092915929544, + "learning_rate": 0.002994233102072425, + "loss": 1.3106, + "step": 363 + }, + { + "epoch": 0.027997846319513884, + "learning_rate": 0.002994201305280611, + "loss": 1.15, + "step": 364 + }, + { + "epoch": 0.02807476347973233, + "learning_rate": 0.0029941694212408253, + "loss": 1.177, + "step": 365 + }, + { + "epoch": 0.02815168063995077, + "learning_rate": 0.0029941374499549297, + "loss": 1.2402, + "step": 366 + }, + { + "epoch": 0.028228597800169217, + "learning_rate": 0.002994105391424792, + "loss": 1.4125, + "step": 367 + }, + { + "epoch": 0.028305514960387662, + "learning_rate": 0.0029940732456522822, + "loss": 1.2672, + "step": 368 + }, + { + "epoch": 0.028382432120606108, + "learning_rate": 0.0029940410126392796, + "loss": 1.5322, + "step": 369 + }, + { + "epoch": 0.028459349280824553, + "learning_rate": 0.0029940086923876644, + "loss": 1.2714, + "step": 370 + }, + { + "epoch": 0.028536266441043, + "learning_rate": 0.0029939762848993252, + "loss": 1.2146, + "step": 371 + }, + { + "epoch": 0.02861318360126144, + "learning_rate": 0.0029939437901761535, + "loss": 1.4489, + "step": 372 + }, + { + "epoch": 0.028690100761479886, + "learning_rate": 0.002993911208220047, + "loss": 0.878, + "step": 373 + }, + { + "epoch": 0.02876701792169833, + "learning_rate": 0.002993878539032908, + "loss": 0.9767, + "step": 374 + }, + { + "epoch": 0.028843935081916777, + "learning_rate": 0.0029938457826166444, + "loss": 1.1914, + "step": 375 + }, + { + "epoch": 0.028920852242135222, + "learning_rate": 0.0029938129389731685, + "loss": 1.1226, + "step": 376 + }, + { + "epoch": 0.028997769402353664, + "learning_rate": 0.0029937800081043984, + "loss": 1.3933, + "step": 377 + }, + { + "epoch": 0.02907468656257211, + "learning_rate": 0.002993746990012257, + "loss": 1.2843, + "step": 378 + }, + { + "epoch": 0.029151603722790555, + "learning_rate": 0.002993713884698672, + "loss": 1.5428, + "step": 379 + }, + { + "epoch": 0.029228520883009, + "learning_rate": 0.0029936806921655765, + "loss": 1.1047, + "step": 380 + }, + { + "epoch": 0.029305438043227446, + "learning_rate": 0.002993647412414909, + "loss": 1.5107, + "step": 381 + }, + { + "epoch": 0.029382355203445887, + "learning_rate": 0.002993614045448612, + "loss": 1.3802, + "step": 382 + }, + { + "epoch": 0.029459272363664333, + "learning_rate": 0.0029935805912686343, + "loss": 1.1889, + "step": 383 + }, + { + "epoch": 0.02953618952388278, + "learning_rate": 0.0029935470498769295, + "loss": 1.0533, + "step": 384 + }, + { + "epoch": 0.029613106684101224, + "learning_rate": 0.0029935134212754557, + "loss": 1.1599, + "step": 385 + }, + { + "epoch": 0.02969002384431967, + "learning_rate": 0.0029934797054661775, + "loss": 0.9938, + "step": 386 + }, + { + "epoch": 0.02976694100453811, + "learning_rate": 0.0029934459024510618, + "loss": 1.2057, + "step": 387 + }, + { + "epoch": 0.029843858164756557, + "learning_rate": 0.002993412012232084, + "loss": 0.9993, + "step": 388 + }, + { + "epoch": 0.029920775324975002, + "learning_rate": 0.002993378034811222, + "loss": 1.4387, + "step": 389 + }, + { + "epoch": 0.029997692485193447, + "learning_rate": 0.002993343970190461, + "loss": 1.2548, + "step": 390 + }, + { + "epoch": 0.030074609645411893, + "learning_rate": 0.0029933098183717883, + "loss": 1.747, + "step": 391 + }, + { + "epoch": 0.030151526805630335, + "learning_rate": 0.0029932755793571995, + "loss": 1.5008, + "step": 392 + }, + { + "epoch": 0.03022844396584878, + "learning_rate": 0.002993241253148693, + "loss": 1.0202, + "step": 393 + }, + { + "epoch": 0.030305361126067226, + "learning_rate": 0.002993206839748274, + "loss": 1.0718, + "step": 394 + }, + { + "epoch": 0.03038227828628567, + "learning_rate": 0.0029931723391579514, + "loss": 1.2228, + "step": 395 + }, + { + "epoch": 0.030459195446504116, + "learning_rate": 0.0029931377513797396, + "loss": 1.4326, + "step": 396 + }, + { + "epoch": 0.03053611260672256, + "learning_rate": 0.0029931030764156585, + "loss": 1.293, + "step": 397 + }, + { + "epoch": 0.030613029766941004, + "learning_rate": 0.0029930683142677323, + "loss": 1.514, + "step": 398 + }, + { + "epoch": 0.03068994692715945, + "learning_rate": 0.0029930334649379917, + "loss": 1.1175, + "step": 399 + }, + { + "epoch": 0.030766864087377895, + "learning_rate": 0.002992998528428471, + "loss": 1.0261, + "step": 400 + }, + { + "epoch": 0.03084378124759634, + "learning_rate": 0.0029929635047412097, + "loss": 1.2312, + "step": 401 + }, + { + "epoch": 0.030920698407814782, + "learning_rate": 0.002992928393878254, + "loss": 1.1255, + "step": 402 + }, + { + "epoch": 0.030997615568033227, + "learning_rate": 0.0029928931958416536, + "loss": 1.2353, + "step": 403 + }, + { + "epoch": 0.031074532728251673, + "learning_rate": 0.002992857910633463, + "loss": 1.3461, + "step": 404 + }, + { + "epoch": 0.031151449888470118, + "learning_rate": 0.002992822538255744, + "loss": 1.3139, + "step": 405 + }, + { + "epoch": 0.031228367048688564, + "learning_rate": 0.0029927870787105604, + "loss": 1.2285, + "step": 406 + }, + { + "epoch": 0.031305284208907005, + "learning_rate": 0.002992751531999984, + "loss": 1.4545, + "step": 407 + }, + { + "epoch": 0.03138220136912545, + "learning_rate": 0.00299271589812609, + "loss": 1.2949, + "step": 408 + }, + { + "epoch": 0.031459118529343896, + "learning_rate": 0.002992680177090959, + "loss": 1.0288, + "step": 409 + }, + { + "epoch": 0.03153603568956234, + "learning_rate": 0.0029926443688966763, + "loss": 0.8704, + "step": 410 + }, + { + "epoch": 0.03161295284978079, + "learning_rate": 0.0029926084735453335, + "loss": 1.3113, + "step": 411 + }, + { + "epoch": 0.03168987000999923, + "learning_rate": 0.002992572491039027, + "loss": 0.9521, + "step": 412 + }, + { + "epoch": 0.03176678717021768, + "learning_rate": 0.002992536421379857, + "loss": 1.1472, + "step": 413 + }, + { + "epoch": 0.03184370433043612, + "learning_rate": 0.0029925002645699297, + "loss": 1.3, + "step": 414 + }, + { + "epoch": 0.03192062149065456, + "learning_rate": 0.002992464020611356, + "loss": 1.5984, + "step": 415 + }, + { + "epoch": 0.03199753865087301, + "learning_rate": 0.0029924276895062535, + "loss": 0.7235, + "step": 416 + }, + { + "epoch": 0.03207445581109145, + "learning_rate": 0.0029923912712567424, + "loss": 1.4046, + "step": 417 + }, + { + "epoch": 0.0321513729713099, + "learning_rate": 0.00299235476586495, + "loss": 1.1582, + "step": 418 + }, + { + "epoch": 0.032228290131528343, + "learning_rate": 0.002992318173333007, + "loss": 1.2819, + "step": 419 + }, + { + "epoch": 0.03230520729174679, + "learning_rate": 0.002992281493663051, + "loss": 1.0412, + "step": 420 + }, + { + "epoch": 0.032382124451965234, + "learning_rate": 0.0029922447268572235, + "loss": 1.2739, + "step": 421 + }, + { + "epoch": 0.03245904161218368, + "learning_rate": 0.0029922078729176706, + "loss": 1.2815, + "step": 422 + }, + { + "epoch": 0.032535958772402125, + "learning_rate": 0.0029921709318465454, + "loss": 1.4791, + "step": 423 + }, + { + "epoch": 0.03261287593262057, + "learning_rate": 0.002992133903646004, + "loss": 1.1519, + "step": 424 + }, + { + "epoch": 0.03268979309283901, + "learning_rate": 0.0029920967883182094, + "loss": 1.0397, + "step": 425 + }, + { + "epoch": 0.032766710253057454, + "learning_rate": 0.002992059585865328, + "loss": 0.7153, + "step": 426 + }, + { + "epoch": 0.0328436274132759, + "learning_rate": 0.0029920222962895324, + "loss": 1.1698, + "step": 427 + }, + { + "epoch": 0.032920544573494345, + "learning_rate": 0.0029919849195930003, + "loss": 1.0705, + "step": 428 + }, + { + "epoch": 0.03299746173371279, + "learning_rate": 0.0029919474557779137, + "loss": 1.2043, + "step": 429 + }, + { + "epoch": 0.033074378893931236, + "learning_rate": 0.00299190990484646, + "loss": 1.1393, + "step": 430 + }, + { + "epoch": 0.03315129605414968, + "learning_rate": 0.0029918722668008323, + "loss": 1.285, + "step": 431 + }, + { + "epoch": 0.03322821321436813, + "learning_rate": 0.002991834541643228, + "loss": 1.0071, + "step": 432 + }, + { + "epoch": 0.03330513037458657, + "learning_rate": 0.00299179672937585, + "loss": 1.197, + "step": 433 + }, + { + "epoch": 0.03338204753480502, + "learning_rate": 0.0029917588300009063, + "loss": 1.0339, + "step": 434 + }, + { + "epoch": 0.03345896469502346, + "learning_rate": 0.00299172084352061, + "loss": 1.3188, + "step": 435 + }, + { + "epoch": 0.0335358818552419, + "learning_rate": 0.002991682769937179, + "loss": 1.0145, + "step": 436 + }, + { + "epoch": 0.03361279901546035, + "learning_rate": 0.0029916446092528364, + "loss": 0.8821, + "step": 437 + }, + { + "epoch": 0.03368971617567879, + "learning_rate": 0.0029916063614698106, + "loss": 0.9893, + "step": 438 + }, + { + "epoch": 0.03376663333589724, + "learning_rate": 0.002991568026590335, + "loss": 1.1368, + "step": 439 + }, + { + "epoch": 0.03384355049611568, + "learning_rate": 0.0029915296046166477, + "loss": 0.9755, + "step": 440 + }, + { + "epoch": 0.03392046765633413, + "learning_rate": 0.0029914910955509922, + "loss": 1.1123, + "step": 441 + }, + { + "epoch": 0.033997384816552574, + "learning_rate": 0.0029914524993956177, + "loss": 1.1482, + "step": 442 + }, + { + "epoch": 0.03407430197677102, + "learning_rate": 0.002991413816152777, + "loss": 0.9763, + "step": 443 + }, + { + "epoch": 0.034151219136989465, + "learning_rate": 0.0029913750458247287, + "loss": 0.8, + "step": 444 + }, + { + "epoch": 0.03422813629720791, + "learning_rate": 0.002991336188413738, + "loss": 1.1179, + "step": 445 + }, + { + "epoch": 0.03430505345742635, + "learning_rate": 0.002991297243922073, + "loss": 1.185, + "step": 446 + }, + { + "epoch": 0.034381970617644794, + "learning_rate": 0.0029912582123520076, + "loss": 1.2302, + "step": 447 + }, + { + "epoch": 0.03445888777786324, + "learning_rate": 0.0029912190937058207, + "loss": 1.2877, + "step": 448 + }, + { + "epoch": 0.034535804938081685, + "learning_rate": 0.0029911798879857974, + "loss": 0.9684, + "step": 449 + }, + { + "epoch": 0.03461272209830013, + "learning_rate": 0.0029911405951942257, + "loss": 1.5201, + "step": 450 + }, + { + "epoch": 0.034689639258518576, + "learning_rate": 0.002991101215333401, + "loss": 1.19, + "step": 451 + }, + { + "epoch": 0.03476655641873702, + "learning_rate": 0.002991061748405623, + "loss": 1.2372, + "step": 452 + }, + { + "epoch": 0.03484347357895547, + "learning_rate": 0.0029910221944131946, + "loss": 1.1227, + "step": 453 + }, + { + "epoch": 0.03492039073917391, + "learning_rate": 0.0029909825533584265, + "loss": 0.8888, + "step": 454 + }, + { + "epoch": 0.03499730789939236, + "learning_rate": 0.0029909428252436338, + "loss": 1.14, + "step": 455 + }, + { + "epoch": 0.035074225059610796, + "learning_rate": 0.002990903010071135, + "loss": 1.1346, + "step": 456 + }, + { + "epoch": 0.03515114221982924, + "learning_rate": 0.0029908631078432563, + "loss": 1.4103, + "step": 457 + }, + { + "epoch": 0.03522805938004769, + "learning_rate": 0.0029908231185623263, + "loss": 1.0122, + "step": 458 + }, + { + "epoch": 0.03530497654026613, + "learning_rate": 0.0029907830422306815, + "loss": 1.1954, + "step": 459 + }, + { + "epoch": 0.03538189370048458, + "learning_rate": 0.0029907428788506604, + "loss": 1.2236, + "step": 460 + }, + { + "epoch": 0.03545881086070302, + "learning_rate": 0.0029907026284246097, + "loss": 0.8695, + "step": 461 + }, + { + "epoch": 0.03553572802092147, + "learning_rate": 0.0029906622909548785, + "loss": 0.7571, + "step": 462 + }, + { + "epoch": 0.035612645181139914, + "learning_rate": 0.0029906218664438227, + "loss": 1.1938, + "step": 463 + }, + { + "epoch": 0.03568956234135836, + "learning_rate": 0.0029905813548938027, + "loss": 0.971, + "step": 464 + }, + { + "epoch": 0.035766479501576805, + "learning_rate": 0.002990540756307184, + "loss": 1.3198, + "step": 465 + }, + { + "epoch": 0.03584339666179524, + "learning_rate": 0.002990500070686337, + "loss": 0.9848, + "step": 466 + }, + { + "epoch": 0.03592031382201369, + "learning_rate": 0.0029904592980336373, + "loss": 1.342, + "step": 467 + }, + { + "epoch": 0.035997230982232134, + "learning_rate": 0.002990418438351466, + "loss": 1.4538, + "step": 468 + }, + { + "epoch": 0.03607414814245058, + "learning_rate": 0.0029903774916422094, + "loss": 1.2134, + "step": 469 + }, + { + "epoch": 0.036151065302669025, + "learning_rate": 0.002990336457908257, + "loss": 1.1878, + "step": 470 + }, + { + "epoch": 0.03622798246288747, + "learning_rate": 0.002990295337152006, + "loss": 1.277, + "step": 471 + }, + { + "epoch": 0.036304899623105916, + "learning_rate": 0.0029902541293758576, + "loss": 1.2576, + "step": 472 + }, + { + "epoch": 0.03638181678332436, + "learning_rate": 0.0029902128345822166, + "loss": 0.758, + "step": 473 + }, + { + "epoch": 0.036458733943542806, + "learning_rate": 0.0029901714527734956, + "loss": 1.2159, + "step": 474 + }, + { + "epoch": 0.03653565110376125, + "learning_rate": 0.00299012998395211, + "loss": 1.1898, + "step": 475 + }, + { + "epoch": 0.0366125682639797, + "learning_rate": 0.0029900884281204823, + "loss": 1.3354, + "step": 476 + }, + { + "epoch": 0.036689485424198136, + "learning_rate": 0.0029900467852810382, + "loss": 1.1051, + "step": 477 + }, + { + "epoch": 0.03676640258441658, + "learning_rate": 0.002990005055436209, + "loss": 1.3771, + "step": 478 + }, + { + "epoch": 0.03684331974463503, + "learning_rate": 0.0029899632385884324, + "loss": 1.0696, + "step": 479 + }, + { + "epoch": 0.03692023690485347, + "learning_rate": 0.002989921334740149, + "loss": 1.15, + "step": 480 + }, + { + "epoch": 0.03699715406507192, + "learning_rate": 0.002989879343893806, + "loss": 1.3266, + "step": 481 + }, + { + "epoch": 0.03707407122529036, + "learning_rate": 0.002989837266051856, + "loss": 1.2287, + "step": 482 + }, + { + "epoch": 0.03715098838550881, + "learning_rate": 0.002989795101216755, + "loss": 1.2021, + "step": 483 + }, + { + "epoch": 0.037227905545727254, + "learning_rate": 0.0029897528493909656, + "loss": 1.4613, + "step": 484 + }, + { + "epoch": 0.0373048227059457, + "learning_rate": 0.0029897105105769544, + "loss": 0.9157, + "step": 485 + }, + { + "epoch": 0.037381739866164145, + "learning_rate": 0.002989668084777194, + "loss": 1.4113, + "step": 486 + }, + { + "epoch": 0.03745865702638258, + "learning_rate": 0.0029896255719941618, + "loss": 1.1766, + "step": 487 + }, + { + "epoch": 0.03753557418660103, + "learning_rate": 0.00298958297223034, + "loss": 1.5105, + "step": 488 + }, + { + "epoch": 0.037612491346819474, + "learning_rate": 0.0029895402854882157, + "loss": 1.1155, + "step": 489 + }, + { + "epoch": 0.03768940850703792, + "learning_rate": 0.002989497511770282, + "loss": 1.1206, + "step": 490 + }, + { + "epoch": 0.037766325667256365, + "learning_rate": 0.002989454651079037, + "loss": 0.8207, + "step": 491 + }, + { + "epoch": 0.03784324282747481, + "learning_rate": 0.0029894117034169813, + "loss": 1.1771, + "step": 492 + }, + { + "epoch": 0.037920159987693255, + "learning_rate": 0.0029893686687866247, + "loss": 1.0857, + "step": 493 + }, + { + "epoch": 0.0379970771479117, + "learning_rate": 0.0029893255471904797, + "loss": 1.0518, + "step": 494 + }, + { + "epoch": 0.038073994308130146, + "learning_rate": 0.002989282338631063, + "loss": 1.1607, + "step": 495 + }, + { + "epoch": 0.03815091146834859, + "learning_rate": 0.0029892390431108997, + "loss": 0.9763, + "step": 496 + }, + { + "epoch": 0.03822782862856703, + "learning_rate": 0.0029891956606325158, + "loss": 1.2348, + "step": 497 + }, + { + "epoch": 0.038304745788785476, + "learning_rate": 0.002989152191198445, + "loss": 1.3312, + "step": 498 + }, + { + "epoch": 0.03838166294900392, + "learning_rate": 0.0029891086348112265, + "loss": 1.1876, + "step": 499 + }, + { + "epoch": 0.038458580109222366, + "learning_rate": 0.002989064991473402, + "loss": 1.3925, + "step": 500 + }, + { + "epoch": 0.03853549726944081, + "learning_rate": 0.002989021261187522, + "loss": 1.2061, + "step": 501 + }, + { + "epoch": 0.03861241442965926, + "learning_rate": 0.0029889774439561378, + "loss": 1.2061, + "step": 502 + }, + { + "epoch": 0.0386893315898777, + "learning_rate": 0.002988933539781809, + "loss": 1.178, + "step": 503 + }, + { + "epoch": 0.03876624875009615, + "learning_rate": 0.0029888895486670994, + "loss": 1.0782, + "step": 504 + }, + { + "epoch": 0.038843165910314593, + "learning_rate": 0.002988845470614577, + "loss": 1.1933, + "step": 505 + }, + { + "epoch": 0.03892008307053304, + "learning_rate": 0.002988801305626816, + "loss": 1.0163, + "step": 506 + }, + { + "epoch": 0.03899700023075148, + "learning_rate": 0.0029887570537063955, + "loss": 1.1219, + "step": 507 + }, + { + "epoch": 0.03907391739096992, + "learning_rate": 0.002988712714855899, + "loss": 1.0401, + "step": 508 + }, + { + "epoch": 0.03915083455118837, + "learning_rate": 0.0029886682890779156, + "loss": 1.1159, + "step": 509 + }, + { + "epoch": 0.039227751711406814, + "learning_rate": 0.002988623776375039, + "loss": 0.8955, + "step": 510 + }, + { + "epoch": 0.03930466887162526, + "learning_rate": 0.002988579176749869, + "loss": 1.1181, + "step": 511 + }, + { + "epoch": 0.039381586031843704, + "learning_rate": 0.0029885344902050096, + "loss": 0.998, + "step": 512 + }, + { + "epoch": 0.03945850319206215, + "learning_rate": 0.00298848971674307, + "loss": 0.9474, + "step": 513 + }, + { + "epoch": 0.039535420352280595, + "learning_rate": 0.0029884448563666645, + "loss": 1.3869, + "step": 514 + }, + { + "epoch": 0.03961233751249904, + "learning_rate": 0.0029883999090784125, + "loss": 1.0953, + "step": 515 + }, + { + "epoch": 0.039689254672717486, + "learning_rate": 0.0029883548748809385, + "loss": 1.061, + "step": 516 + }, + { + "epoch": 0.039766171832935925, + "learning_rate": 0.0029883097537768723, + "loss": 1.1206, + "step": 517 + }, + { + "epoch": 0.03984308899315437, + "learning_rate": 0.0029882645457688485, + "loss": 1.0127, + "step": 518 + }, + { + "epoch": 0.039920006153372815, + "learning_rate": 0.002988219250859507, + "loss": 1.5546, + "step": 519 + }, + { + "epoch": 0.03999692331359126, + "learning_rate": 0.0029881738690514924, + "loss": 1.1249, + "step": 520 + }, + { + "epoch": 0.040073840473809706, + "learning_rate": 0.0029881284003474545, + "loss": 1.5855, + "step": 521 + }, + { + "epoch": 0.04015075763402815, + "learning_rate": 0.002988082844750049, + "loss": 1.1175, + "step": 522 + }, + { + "epoch": 0.0402276747942466, + "learning_rate": 0.0029880372022619344, + "loss": 1.0745, + "step": 523 + }, + { + "epoch": 0.04030459195446504, + "learning_rate": 0.002987991472885777, + "loss": 1.1992, + "step": 524 + }, + { + "epoch": 0.04038150911468349, + "learning_rate": 0.0029879456566242473, + "loss": 1.2495, + "step": 525 + }, + { + "epoch": 0.04045842627490193, + "learning_rate": 0.0029878997534800197, + "loss": 1.0105, + "step": 526 + }, + { + "epoch": 0.04053534343512038, + "learning_rate": 0.0029878537634557747, + "loss": 1.2723, + "step": 527 + }, + { + "epoch": 0.04061226059533882, + "learning_rate": 0.0029878076865541984, + "loss": 1.4241, + "step": 528 + }, + { + "epoch": 0.04068917775555726, + "learning_rate": 0.00298776152277798, + "loss": 0.9104, + "step": 529 + }, + { + "epoch": 0.04076609491577571, + "learning_rate": 0.0029877152721298162, + "loss": 1.346, + "step": 530 + }, + { + "epoch": 0.04084301207599415, + "learning_rate": 0.002987668934612407, + "loss": 1.1694, + "step": 531 + }, + { + "epoch": 0.0409199292362126, + "learning_rate": 0.0029876225102284585, + "loss": 0.8494, + "step": 532 + }, + { + "epoch": 0.040996846396431044, + "learning_rate": 0.0029875759989806813, + "loss": 0.9455, + "step": 533 + }, + { + "epoch": 0.04107376355664949, + "learning_rate": 0.002987529400871791, + "loss": 1.3369, + "step": 534 + }, + { + "epoch": 0.041150680716867935, + "learning_rate": 0.002987482715904509, + "loss": 1.1992, + "step": 535 + }, + { + "epoch": 0.04122759787708638, + "learning_rate": 0.0029874359440815604, + "loss": 1.2606, + "step": 536 + }, + { + "epoch": 0.041304515037304826, + "learning_rate": 0.0029873890854056778, + "loss": 1.5976, + "step": 537 + }, + { + "epoch": 0.041381432197523264, + "learning_rate": 0.002987342139879596, + "loss": 1.2291, + "step": 538 + }, + { + "epoch": 0.04145834935774171, + "learning_rate": 0.002987295107506056, + "loss": 0.8531, + "step": 539 + }, + { + "epoch": 0.041535266517960155, + "learning_rate": 0.002987247988287805, + "loss": 1.3273, + "step": 540 + }, + { + "epoch": 0.0416121836781786, + "learning_rate": 0.002987200782227594, + "loss": 1.0323, + "step": 541 + }, + { + "epoch": 0.041689100838397046, + "learning_rate": 0.0029871534893281797, + "loss": 0.9876, + "step": 542 + }, + { + "epoch": 0.04176601799861549, + "learning_rate": 0.002987106109592323, + "loss": 1.7456, + "step": 543 + }, + { + "epoch": 0.04184293515883394, + "learning_rate": 0.0029870586430227907, + "loss": 0.9066, + "step": 544 + }, + { + "epoch": 0.04191985231905238, + "learning_rate": 0.0029870110896223546, + "loss": 1.4167, + "step": 545 + }, + { + "epoch": 0.04199676947927083, + "learning_rate": 0.0029869634493937913, + "loss": 1.4998, + "step": 546 + }, + { + "epoch": 0.04207368663948927, + "learning_rate": 0.002986915722339882, + "loss": 1.2597, + "step": 547 + }, + { + "epoch": 0.04215060379970771, + "learning_rate": 0.0029868679084634146, + "loss": 1.2115, + "step": 548 + }, + { + "epoch": 0.04222752095992616, + "learning_rate": 0.0029868200077671804, + "loss": 1.027, + "step": 549 + }, + { + "epoch": 0.0423044381201446, + "learning_rate": 0.0029867720202539764, + "loss": 1.1086, + "step": 550 + }, + { + "epoch": 0.04238135528036305, + "learning_rate": 0.0029867239459266046, + "loss": 1.2673, + "step": 551 + }, + { + "epoch": 0.04245827244058149, + "learning_rate": 0.002986675784787872, + "loss": 1.4465, + "step": 552 + }, + { + "epoch": 0.04253518960079994, + "learning_rate": 0.0029866275368405915, + "loss": 1.1503, + "step": 553 + }, + { + "epoch": 0.042612106761018384, + "learning_rate": 0.0029865792020875794, + "loss": 1.2056, + "step": 554 + }, + { + "epoch": 0.04268902392123683, + "learning_rate": 0.0029865307805316582, + "loss": 0.8459, + "step": 555 + }, + { + "epoch": 0.042765941081455275, + "learning_rate": 0.0029864822721756557, + "loss": 0.896, + "step": 556 + }, + { + "epoch": 0.04284285824167372, + "learning_rate": 0.0029864336770224044, + "loss": 1.189, + "step": 557 + }, + { + "epoch": 0.04291977540189216, + "learning_rate": 0.0029863849950747413, + "loss": 1.0664, + "step": 558 + }, + { + "epoch": 0.042996692562110604, + "learning_rate": 0.002986336226335509, + "loss": 1.0806, + "step": 559 + }, + { + "epoch": 0.04307360972232905, + "learning_rate": 0.002986287370807556, + "loss": 0.7105, + "step": 560 + }, + { + "epoch": 0.043150526882547495, + "learning_rate": 0.002986238428493734, + "loss": 1.3178, + "step": 561 + }, + { + "epoch": 0.04322744404276594, + "learning_rate": 0.0029861893993969016, + "loss": 1.0987, + "step": 562 + }, + { + "epoch": 0.043304361202984386, + "learning_rate": 0.002986140283519921, + "loss": 1.0808, + "step": 563 + }, + { + "epoch": 0.04338127836320283, + "learning_rate": 0.0029860910808656607, + "loss": 1.2433, + "step": 564 + }, + { + "epoch": 0.04345819552342128, + "learning_rate": 0.0029860417914369932, + "loss": 1.2721, + "step": 565 + }, + { + "epoch": 0.04353511268363972, + "learning_rate": 0.0029859924152367966, + "loss": 1.0626, + "step": 566 + }, + { + "epoch": 0.04361202984385817, + "learning_rate": 0.0029859429522679547, + "loss": 1.2338, + "step": 567 + }, + { + "epoch": 0.04368894700407661, + "learning_rate": 0.002985893402533355, + "loss": 1.0487, + "step": 568 + }, + { + "epoch": 0.04376586416429505, + "learning_rate": 0.002985843766035891, + "loss": 1.0549, + "step": 569 + }, + { + "epoch": 0.0438427813245135, + "learning_rate": 0.002985794042778461, + "loss": 0.9051, + "step": 570 + }, + { + "epoch": 0.04391969848473194, + "learning_rate": 0.002985744232763968, + "loss": 0.9448, + "step": 571 + }, + { + "epoch": 0.04399661564495039, + "learning_rate": 0.002985694335995321, + "loss": 1.314, + "step": 572 + }, + { + "epoch": 0.04407353280516883, + "learning_rate": 0.002985644352475434, + "loss": 1.0038, + "step": 573 + }, + { + "epoch": 0.04415044996538728, + "learning_rate": 0.0029855942822072244, + "loss": 1.2588, + "step": 574 + }, + { + "epoch": 0.044227367125605724, + "learning_rate": 0.0029855441251936165, + "loss": 1.1826, + "step": 575 + }, + { + "epoch": 0.04430428428582417, + "learning_rate": 0.0029854938814375385, + "loss": 1.1935, + "step": 576 + }, + { + "epoch": 0.044381201446042615, + "learning_rate": 0.0029854435509419253, + "loss": 1.3885, + "step": 577 + }, + { + "epoch": 0.04445811860626106, + "learning_rate": 0.0029853931337097146, + "loss": 1.5294, + "step": 578 + }, + { + "epoch": 0.0445350357664795, + "learning_rate": 0.0029853426297438515, + "loss": 1.0981, + "step": 579 + }, + { + "epoch": 0.044611952926697944, + "learning_rate": 0.0029852920390472833, + "loss": 1.5897, + "step": 580 + }, + { + "epoch": 0.04468887008691639, + "learning_rate": 0.0029852413616229655, + "loss": 1.2229, + "step": 581 + }, + { + "epoch": 0.044765787247134835, + "learning_rate": 0.002985190597473857, + "loss": 1.0862, + "step": 582 + }, + { + "epoch": 0.04484270440735328, + "learning_rate": 0.002985139746602921, + "loss": 1.0314, + "step": 583 + }, + { + "epoch": 0.044919621567571726, + "learning_rate": 0.0029850888090131276, + "loss": 1.0007, + "step": 584 + }, + { + "epoch": 0.04499653872779017, + "learning_rate": 0.002985037784707451, + "loss": 1.0844, + "step": 585 + }, + { + "epoch": 0.045073455888008616, + "learning_rate": 0.002984986673688871, + "loss": 0.95, + "step": 586 + }, + { + "epoch": 0.04515037304822706, + "learning_rate": 0.002984935475960371, + "loss": 0.7464, + "step": 587 + }, + { + "epoch": 0.04522729020844551, + "learning_rate": 0.002984884191524941, + "loss": 1.3654, + "step": 588 + }, + { + "epoch": 0.045304207368663946, + "learning_rate": 0.0029848328203855756, + "loss": 1.3091, + "step": 589 + }, + { + "epoch": 0.04538112452888239, + "learning_rate": 0.0029847813625452743, + "loss": 1.201, + "step": 590 + }, + { + "epoch": 0.04545804168910084, + "learning_rate": 0.002984729818007042, + "loss": 0.9744, + "step": 591 + }, + { + "epoch": 0.04553495884931928, + "learning_rate": 0.002984678186773888, + "loss": 1.5352, + "step": 592 + }, + { + "epoch": 0.04561187600953773, + "learning_rate": 0.002984626468848828, + "loss": 1.1746, + "step": 593 + }, + { + "epoch": 0.04568879316975617, + "learning_rate": 0.0029845746642348805, + "loss": 1.044, + "step": 594 + }, + { + "epoch": 0.04576571032997462, + "learning_rate": 0.0029845227729350717, + "loss": 0.9655, + "step": 595 + }, + { + "epoch": 0.045842627490193064, + "learning_rate": 0.0029844707949524306, + "loss": 1.1739, + "step": 596 + }, + { + "epoch": 0.04591954465041151, + "learning_rate": 0.0029844187302899934, + "loss": 1.2033, + "step": 597 + }, + { + "epoch": 0.045996461810629954, + "learning_rate": 0.002984366578950799, + "loss": 1.0512, + "step": 598 + }, + { + "epoch": 0.04607337897084839, + "learning_rate": 0.002984314340937893, + "loss": 1.1152, + "step": 599 + }, + { + "epoch": 0.04615029613106684, + "learning_rate": 0.002984262016254326, + "loss": 1.351, + "step": 600 + }, + { + "epoch": 0.046227213291285284, + "learning_rate": 0.0029842096049031526, + "loss": 1.1306, + "step": 601 + }, + { + "epoch": 0.04630413045150373, + "learning_rate": 0.002984157106887434, + "loss": 1.548, + "step": 602 + }, + { + "epoch": 0.046381047611722175, + "learning_rate": 0.002984104522210235, + "loss": 1.4748, + "step": 603 + }, + { + "epoch": 0.04645796477194062, + "learning_rate": 0.0029840518508746266, + "loss": 1.2217, + "step": 604 + }, + { + "epoch": 0.046534881932159065, + "learning_rate": 0.0029839990928836836, + "loss": 1.1253, + "step": 605 + }, + { + "epoch": 0.04661179909237751, + "learning_rate": 0.002983946248240487, + "loss": 0.9298, + "step": 606 + }, + { + "epoch": 0.046688716252595956, + "learning_rate": 0.0029838933169481227, + "loss": 0.8226, + "step": 607 + }, + { + "epoch": 0.0467656334128144, + "learning_rate": 0.0029838402990096805, + "loss": 0.8816, + "step": 608 + }, + { + "epoch": 0.04684255057303285, + "learning_rate": 0.0029837871944282577, + "loss": 1.0678, + "step": 609 + }, + { + "epoch": 0.046919467733251286, + "learning_rate": 0.0029837340032069534, + "loss": 0.8852, + "step": 610 + }, + { + "epoch": 0.04699638489346973, + "learning_rate": 0.0029836807253488746, + "loss": 1.4652, + "step": 611 + }, + { + "epoch": 0.047073302053688176, + "learning_rate": 0.0029836273608571324, + "loss": 1.445, + "step": 612 + }, + { + "epoch": 0.04715021921390662, + "learning_rate": 0.0029835739097348424, + "loss": 1.1648, + "step": 613 + }, + { + "epoch": 0.04722713637412507, + "learning_rate": 0.002983520371985125, + "loss": 1.0133, + "step": 614 + }, + { + "epoch": 0.04730405353434351, + "learning_rate": 0.0029834667476111073, + "loss": 1.1974, + "step": 615 + }, + { + "epoch": 0.04738097069456196, + "learning_rate": 0.0029834130366159202, + "loss": 1.2651, + "step": 616 + }, + { + "epoch": 0.0474578878547804, + "learning_rate": 0.0029833592390027, + "loss": 0.8177, + "step": 617 + }, + { + "epoch": 0.04753480501499885, + "learning_rate": 0.0029833053547745885, + "loss": 1.5305, + "step": 618 + }, + { + "epoch": 0.047611722175217294, + "learning_rate": 0.0029832513839347305, + "loss": 1.1941, + "step": 619 + }, + { + "epoch": 0.04768863933543573, + "learning_rate": 0.002983197326486279, + "loss": 1.0049, + "step": 620 + }, + { + "epoch": 0.04776555649565418, + "learning_rate": 0.0029831431824323892, + "loss": 1.4915, + "step": 621 + }, + { + "epoch": 0.047842473655872624, + "learning_rate": 0.002983088951776224, + "loss": 0.9474, + "step": 622 + }, + { + "epoch": 0.04791939081609107, + "learning_rate": 0.0029830346345209487, + "loss": 1.2237, + "step": 623 + }, + { + "epoch": 0.047996307976309514, + "learning_rate": 0.0029829802306697356, + "loss": 1.4277, + "step": 624 + }, + { + "epoch": 0.04807322513652796, + "learning_rate": 0.0029829257402257613, + "loss": 1.0073, + "step": 625 + }, + { + "epoch": 0.048150142296746405, + "learning_rate": 0.0029828711631922082, + "loss": 0.9298, + "step": 626 + }, + { + "epoch": 0.04822705945696485, + "learning_rate": 0.0029828164995722617, + "loss": 1.5419, + "step": 627 + }, + { + "epoch": 0.048303976617183296, + "learning_rate": 0.002982761749369115, + "loss": 1.1724, + "step": 628 + }, + { + "epoch": 0.04838089377740174, + "learning_rate": 0.002982706912585964, + "loss": 1.0517, + "step": 629 + }, + { + "epoch": 0.04845781093762018, + "learning_rate": 0.0029826519892260117, + "loss": 1.3097, + "step": 630 + }, + { + "epoch": 0.048534728097838625, + "learning_rate": 0.002982596979292464, + "loss": 1.1801, + "step": 631 + }, + { + "epoch": 0.04861164525805707, + "learning_rate": 0.0029825418827885337, + "loss": 1.2174, + "step": 632 + }, + { + "epoch": 0.048688562418275516, + "learning_rate": 0.002982486699717438, + "loss": 0.9513, + "step": 633 + }, + { + "epoch": 0.04876547957849396, + "learning_rate": 0.0029824314300823987, + "loss": 1.3247, + "step": 634 + }, + { + "epoch": 0.04884239673871241, + "learning_rate": 0.002982376073886643, + "loss": 1.0623, + "step": 635 + }, + { + "epoch": 0.04891931389893085, + "learning_rate": 0.002982320631133404, + "loss": 1.3911, + "step": 636 + }, + { + "epoch": 0.0489962310591493, + "learning_rate": 0.0029822651018259184, + "loss": 1.2937, + "step": 637 + }, + { + "epoch": 0.04907314821936774, + "learning_rate": 0.002982209485967428, + "loss": 1.1114, + "step": 638 + }, + { + "epoch": 0.04915006537958619, + "learning_rate": 0.002982153783561182, + "loss": 1.2588, + "step": 639 + }, + { + "epoch": 0.04922698253980463, + "learning_rate": 0.0029820979946104313, + "loss": 1.2962, + "step": 640 + }, + { + "epoch": 0.04930389970002307, + "learning_rate": 0.0029820421191184346, + "loss": 1.272, + "step": 641 + }, + { + "epoch": 0.04938081686024152, + "learning_rate": 0.002981986157088453, + "loss": 1.178, + "step": 642 + }, + { + "epoch": 0.04945773402045996, + "learning_rate": 0.002981930108523756, + "loss": 1.3278, + "step": 643 + }, + { + "epoch": 0.04953465118067841, + "learning_rate": 0.0029818739734276153, + "loss": 0.9838, + "step": 644 + }, + { + "epoch": 0.049611568340896854, + "learning_rate": 0.0029818177518033096, + "loss": 1.325, + "step": 645 + }, + { + "epoch": 0.0496884855011153, + "learning_rate": 0.0029817614436541203, + "loss": 1.1888, + "step": 646 + }, + { + "epoch": 0.049765402661333745, + "learning_rate": 0.0029817050489833363, + "loss": 1.5704, + "step": 647 + }, + { + "epoch": 0.04984231982155219, + "learning_rate": 0.00298164856779425, + "loss": 0.7675, + "step": 648 + }, + { + "epoch": 0.049919236981770636, + "learning_rate": 0.00298159200009016, + "loss": 1.1223, + "step": 649 + }, + { + "epoch": 0.04999615414198908, + "learning_rate": 0.0029815353458743686, + "loss": 1.1525, + "step": 650 + }, + { + "epoch": 0.05007307130220752, + "learning_rate": 0.002981478605150185, + "loss": 1.3927, + "step": 651 + }, + { + "epoch": 0.050149988462425965, + "learning_rate": 0.0029814217779209214, + "loss": 1.1732, + "step": 652 + }, + { + "epoch": 0.05022690562264441, + "learning_rate": 0.0029813648641898965, + "loss": 1.3517, + "step": 653 + }, + { + "epoch": 0.050303822782862856, + "learning_rate": 0.002981307863960433, + "loss": 0.8532, + "step": 654 + }, + { + "epoch": 0.0503807399430813, + "learning_rate": 0.0029812507772358594, + "loss": 1.2538, + "step": 655 + }, + { + "epoch": 0.05045765710329975, + "learning_rate": 0.0029811936040195094, + "loss": 1.4504, + "step": 656 + }, + { + "epoch": 0.05053457426351819, + "learning_rate": 0.0029811363443147214, + "loss": 1.0522, + "step": 657 + }, + { + "epoch": 0.05061149142373664, + "learning_rate": 0.0029810789981248387, + "loss": 1.2034, + "step": 658 + }, + { + "epoch": 0.05068840858395508, + "learning_rate": 0.0029810215654532094, + "loss": 1.0663, + "step": 659 + }, + { + "epoch": 0.05076532574417353, + "learning_rate": 0.002980964046303188, + "loss": 1.4473, + "step": 660 + }, + { + "epoch": 0.05084224290439197, + "learning_rate": 0.0029809064406781318, + "loss": 1.0033, + "step": 661 + }, + { + "epoch": 0.05091916006461041, + "learning_rate": 0.002980848748581405, + "loss": 1.2814, + "step": 662 + }, + { + "epoch": 0.05099607722482886, + "learning_rate": 0.002980790970016377, + "loss": 1.2509, + "step": 663 + }, + { + "epoch": 0.0510729943850473, + "learning_rate": 0.002980733104986421, + "loss": 1.04, + "step": 664 + }, + { + "epoch": 0.05114991154526575, + "learning_rate": 0.002980675153494916, + "loss": 0.9383, + "step": 665 + }, + { + "epoch": 0.051226828705484194, + "learning_rate": 0.0029806171155452453, + "loss": 1.2609, + "step": 666 + }, + { + "epoch": 0.05130374586570264, + "learning_rate": 0.0029805589911407982, + "loss": 1.2434, + "step": 667 + }, + { + "epoch": 0.051380663025921085, + "learning_rate": 0.0029805007802849687, + "loss": 1.0247, + "step": 668 + }, + { + "epoch": 0.05145758018613953, + "learning_rate": 0.0029804424829811558, + "loss": 1.3384, + "step": 669 + }, + { + "epoch": 0.051534497346357976, + "learning_rate": 0.0029803840992327635, + "loss": 0.8823, + "step": 670 + }, + { + "epoch": 0.051611414506576414, + "learning_rate": 0.0029803256290432, + "loss": 0.763, + "step": 671 + }, + { + "epoch": 0.05168833166679486, + "learning_rate": 0.0029802670724158814, + "loss": 1.2358, + "step": 672 + }, + { + "epoch": 0.051765248827013305, + "learning_rate": 0.0029802084293542247, + "loss": 1.425, + "step": 673 + }, + { + "epoch": 0.05184216598723175, + "learning_rate": 0.002980149699861656, + "loss": 1.3391, + "step": 674 + }, + { + "epoch": 0.051919083147450196, + "learning_rate": 0.0029800908839416034, + "loss": 1.4196, + "step": 675 + }, + { + "epoch": 0.05199600030766864, + "learning_rate": 0.002980031981597502, + "loss": 1.3749, + "step": 676 + }, + { + "epoch": 0.052072917467887087, + "learning_rate": 0.0029799729928327904, + "loss": 1.0857, + "step": 677 + }, + { + "epoch": 0.05214983462810553, + "learning_rate": 0.002979913917650913, + "loss": 0.8182, + "step": 678 + }, + { + "epoch": 0.05222675178832398, + "learning_rate": 0.00297985475605532, + "loss": 1.1418, + "step": 679 + }, + { + "epoch": 0.05230366894854242, + "learning_rate": 0.0029797955080494653, + "loss": 1.7806, + "step": 680 + }, + { + "epoch": 0.05238058610876086, + "learning_rate": 0.002979736173636809, + "loss": 0.9753, + "step": 681 + }, + { + "epoch": 0.05245750326897931, + "learning_rate": 0.0029796767528208153, + "loss": 1.2917, + "step": 682 + }, + { + "epoch": 0.05253442042919775, + "learning_rate": 0.002979617245604954, + "loss": 0.8755, + "step": 683 + }, + { + "epoch": 0.0526113375894162, + "learning_rate": 0.0029795576519926993, + "loss": 1.1086, + "step": 684 + }, + { + "epoch": 0.05268825474963464, + "learning_rate": 0.0029794979719875315, + "loss": 0.9721, + "step": 685 + }, + { + "epoch": 0.05276517190985309, + "learning_rate": 0.0029794382055929355, + "loss": 0.8475, + "step": 686 + }, + { + "epoch": 0.052842089070071534, + "learning_rate": 0.0029793783528124004, + "loss": 1.1931, + "step": 687 + }, + { + "epoch": 0.05291900623028998, + "learning_rate": 0.0029793184136494223, + "loss": 1.8459, + "step": 688 + }, + { + "epoch": 0.052995923390508425, + "learning_rate": 0.0029792583881074995, + "loss": 1.4024, + "step": 689 + }, + { + "epoch": 0.05307284055072687, + "learning_rate": 0.002979198276190138, + "loss": 1.2084, + "step": 690 + }, + { + "epoch": 0.05314975771094531, + "learning_rate": 0.0029791380779008478, + "loss": 1.2549, + "step": 691 + }, + { + "epoch": 0.053226674871163754, + "learning_rate": 0.0029790777932431436, + "loss": 1.1673, + "step": 692 + }, + { + "epoch": 0.0533035920313822, + "learning_rate": 0.0029790174222205455, + "loss": 0.9272, + "step": 693 + }, + { + "epoch": 0.053380509191600645, + "learning_rate": 0.0029789569648365787, + "loss": 0.9013, + "step": 694 + }, + { + "epoch": 0.05345742635181909, + "learning_rate": 0.0029788964210947736, + "loss": 0.8581, + "step": 695 + }, + { + "epoch": 0.053534343512037535, + "learning_rate": 0.002978835790998665, + "loss": 0.8097, + "step": 696 + }, + { + "epoch": 0.05361126067225598, + "learning_rate": 0.0029787750745517935, + "loss": 1.4727, + "step": 697 + }, + { + "epoch": 0.053688177832474426, + "learning_rate": 0.002978714271757704, + "loss": 1.0355, + "step": 698 + }, + { + "epoch": 0.05376509499269287, + "learning_rate": 0.0029786533826199474, + "loss": 0.8615, + "step": 699 + }, + { + "epoch": 0.05384201215291132, + "learning_rate": 0.0029785924071420787, + "loss": 1.2511, + "step": 700 + }, + { + "epoch": 0.05391892931312976, + "learning_rate": 0.002978531345327658, + "loss": 1.4, + "step": 701 + }, + { + "epoch": 0.0539958464733482, + "learning_rate": 0.0029784701971802517, + "loss": 0.9655, + "step": 702 + }, + { + "epoch": 0.054072763633566646, + "learning_rate": 0.0029784089627034297, + "loss": 1.2346, + "step": 703 + }, + { + "epoch": 0.05414968079378509, + "learning_rate": 0.002978347641900767, + "loss": 1.3287, + "step": 704 + }, + { + "epoch": 0.05422659795400354, + "learning_rate": 0.002978286234775845, + "loss": 1.1252, + "step": 705 + }, + { + "epoch": 0.05430351511422198, + "learning_rate": 0.00297822474133225, + "loss": 0.9429, + "step": 706 + }, + { + "epoch": 0.05438043227444043, + "learning_rate": 0.002978163161573571, + "loss": 1.5489, + "step": 707 + }, + { + "epoch": 0.054457349434658874, + "learning_rate": 0.0029781014955034047, + "loss": 1.1839, + "step": 708 + }, + { + "epoch": 0.05453426659487732, + "learning_rate": 0.0029780397431253517, + "loss": 1.1866, + "step": 709 + }, + { + "epoch": 0.054611183755095764, + "learning_rate": 0.0029779779044430176, + "loss": 1.4301, + "step": 710 + }, + { + "epoch": 0.05468810091531421, + "learning_rate": 0.002977915979460013, + "loss": 1.018, + "step": 711 + }, + { + "epoch": 0.05476501807553265, + "learning_rate": 0.002977853968179955, + "loss": 1.1193, + "step": 712 + }, + { + "epoch": 0.054841935235751094, + "learning_rate": 0.002977791870606463, + "loss": 1.4326, + "step": 713 + }, + { + "epoch": 0.05491885239596954, + "learning_rate": 0.0029777296867431638, + "loss": 1.0494, + "step": 714 + }, + { + "epoch": 0.054995769556187984, + "learning_rate": 0.0029776674165936873, + "loss": 0.8348, + "step": 715 + }, + { + "epoch": 0.05507268671640643, + "learning_rate": 0.0029776050601616716, + "loss": 1.4864, + "step": 716 + }, + { + "epoch": 0.055149603876624875, + "learning_rate": 0.0029775426174507556, + "loss": 1.1649, + "step": 717 + }, + { + "epoch": 0.05522652103684332, + "learning_rate": 0.0029774800884645866, + "loss": 1.0731, + "step": 718 + }, + { + "epoch": 0.055303438197061766, + "learning_rate": 0.0029774174732068157, + "loss": 1.1337, + "step": 719 + }, + { + "epoch": 0.05538035535728021, + "learning_rate": 0.0029773547716810986, + "loss": 1.435, + "step": 720 + }, + { + "epoch": 0.05545727251749866, + "learning_rate": 0.0029772919838910966, + "loss": 1.1594, + "step": 721 + }, + { + "epoch": 0.055534189677717095, + "learning_rate": 0.002977229109840476, + "loss": 0.7081, + "step": 722 + }, + { + "epoch": 0.05561110683793554, + "learning_rate": 0.0029771661495329085, + "loss": 0.9894, + "step": 723 + }, + { + "epoch": 0.055688023998153986, + "learning_rate": 0.0029771031029720697, + "loss": 1.0422, + "step": 724 + }, + { + "epoch": 0.05576494115837243, + "learning_rate": 0.002977039970161642, + "loss": 1.4302, + "step": 725 + }, + { + "epoch": 0.05584185831859088, + "learning_rate": 0.0029769767511053104, + "loss": 1.08, + "step": 726 + }, + { + "epoch": 0.05591877547880932, + "learning_rate": 0.0029769134458067672, + "loss": 1.3204, + "step": 727 + }, + { + "epoch": 0.05599569263902777, + "learning_rate": 0.002976850054269709, + "loss": 1.1759, + "step": 728 + }, + { + "epoch": 0.05607260979924621, + "learning_rate": 0.002976786576497837, + "loss": 1.091, + "step": 729 + }, + { + "epoch": 0.05614952695946466, + "learning_rate": 0.002976723012494857, + "loss": 1.3387, + "step": 730 + }, + { + "epoch": 0.056226444119683104, + "learning_rate": 0.002976659362264482, + "loss": 1.2249, + "step": 731 + }, + { + "epoch": 0.05630336127990154, + "learning_rate": 0.0029765956258104276, + "loss": 1.3977, + "step": 732 + }, + { + "epoch": 0.05638027844011999, + "learning_rate": 0.0029765318031364156, + "loss": 1.3686, + "step": 733 + }, + { + "epoch": 0.05645719560033843, + "learning_rate": 0.002976467894246173, + "loss": 1.0985, + "step": 734 + }, + { + "epoch": 0.05653411276055688, + "learning_rate": 0.0029764038991434314, + "loss": 0.868, + "step": 735 + }, + { + "epoch": 0.056611029920775324, + "learning_rate": 0.0029763398178319278, + "loss": 1.3035, + "step": 736 + }, + { + "epoch": 0.05668794708099377, + "learning_rate": 0.002976275650315403, + "loss": 1.1757, + "step": 737 + }, + { + "epoch": 0.056764864241212215, + "learning_rate": 0.0029762113965976045, + "loss": 0.9565, + "step": 738 + }, + { + "epoch": 0.05684178140143066, + "learning_rate": 0.0029761470566822845, + "loss": 1.1045, + "step": 739 + }, + { + "epoch": 0.056918698561649106, + "learning_rate": 0.0029760826305731992, + "loss": 1.0485, + "step": 740 + }, + { + "epoch": 0.05699561572186755, + "learning_rate": 0.0029760181182741105, + "loss": 0.9465, + "step": 741 + }, + { + "epoch": 0.057072532882086, + "learning_rate": 0.002975953519788786, + "loss": 1.6061, + "step": 742 + }, + { + "epoch": 0.057149450042304435, + "learning_rate": 0.0029758888351209974, + "loss": 1.2262, + "step": 743 + }, + { + "epoch": 0.05722636720252288, + "learning_rate": 0.002975824064274521, + "loss": 1.2773, + "step": 744 + }, + { + "epoch": 0.057303284362741326, + "learning_rate": 0.00297575920725314, + "loss": 1.3701, + "step": 745 + }, + { + "epoch": 0.05738020152295977, + "learning_rate": 0.0029756942640606406, + "loss": 1.3194, + "step": 746 + }, + { + "epoch": 0.05745711868317822, + "learning_rate": 0.0029756292347008152, + "loss": 1.332, + "step": 747 + }, + { + "epoch": 0.05753403584339666, + "learning_rate": 0.002975564119177461, + "loss": 1.3512, + "step": 748 + }, + { + "epoch": 0.05761095300361511, + "learning_rate": 0.00297549891749438, + "loss": 1.2467, + "step": 749 + }, + { + "epoch": 0.05768787016383355, + "learning_rate": 0.0029754336296553795, + "loss": 1.335, + "step": 750 + }, + { + "epoch": 0.057764787324052, + "learning_rate": 0.002975368255664272, + "loss": 1.0811, + "step": 751 + }, + { + "epoch": 0.057841704484270444, + "learning_rate": 0.0029753027955248743, + "loss": 1.0731, + "step": 752 + }, + { + "epoch": 0.05791862164448888, + "learning_rate": 0.0029752372492410084, + "loss": 0.8023, + "step": 753 + }, + { + "epoch": 0.05799553880470733, + "learning_rate": 0.002975171616816503, + "loss": 1.5936, + "step": 754 + }, + { + "epoch": 0.05807245596492577, + "learning_rate": 0.0029751058982551886, + "loss": 1.0659, + "step": 755 + }, + { + "epoch": 0.05814937312514422, + "learning_rate": 0.0029750400935609034, + "loss": 1.4928, + "step": 756 + }, + { + "epoch": 0.058226290285362664, + "learning_rate": 0.0029749742027374904, + "loss": 1.2801, + "step": 757 + }, + { + "epoch": 0.05830320744558111, + "learning_rate": 0.002974908225788796, + "loss": 1.7553, + "step": 758 + }, + { + "epoch": 0.058380124605799555, + "learning_rate": 0.0029748421627186737, + "loss": 1.2631, + "step": 759 + }, + { + "epoch": 0.058457041766018, + "learning_rate": 0.0029747760135309803, + "loss": 1.1292, + "step": 760 + }, + { + "epoch": 0.058533958926236446, + "learning_rate": 0.0029747097782295787, + "loss": 1.4035, + "step": 761 + }, + { + "epoch": 0.05861087608645489, + "learning_rate": 0.002974643456818336, + "loss": 0.9446, + "step": 762 + }, + { + "epoch": 0.05868779324667333, + "learning_rate": 0.0029745770493011256, + "loss": 0.7152, + "step": 763 + }, + { + "epoch": 0.058764710406891775, + "learning_rate": 0.002974510555681824, + "loss": 1.262, + "step": 764 + }, + { + "epoch": 0.05884162756711022, + "learning_rate": 0.0029744439759643144, + "loss": 1.2302, + "step": 765 + }, + { + "epoch": 0.058918544727328666, + "learning_rate": 0.002974377310152484, + "loss": 1.1512, + "step": 766 + }, + { + "epoch": 0.05899546188754711, + "learning_rate": 0.0029743105582502264, + "loss": 1.1798, + "step": 767 + }, + { + "epoch": 0.05907237904776556, + "learning_rate": 0.0029742437202614387, + "loss": 1.1354, + "step": 768 + }, + { + "epoch": 0.059149296207984, + "learning_rate": 0.0029741767961900243, + "loss": 1.3438, + "step": 769 + }, + { + "epoch": 0.05922621336820245, + "learning_rate": 0.0029741097860398894, + "loss": 1.4128, + "step": 770 + }, + { + "epoch": 0.05930313052842089, + "learning_rate": 0.0029740426898149487, + "loss": 1.2916, + "step": 771 + }, + { + "epoch": 0.05938004768863934, + "learning_rate": 0.0029739755075191185, + "loss": 0.9568, + "step": 772 + }, + { + "epoch": 0.05945696484885778, + "learning_rate": 0.0029739082391563223, + "loss": 1.5856, + "step": 773 + }, + { + "epoch": 0.05953388200907622, + "learning_rate": 0.0029738408847304883, + "loss": 1.0162, + "step": 774 + }, + { + "epoch": 0.05961079916929467, + "learning_rate": 0.002973773444245549, + "loss": 1.1085, + "step": 775 + }, + { + "epoch": 0.05968771632951311, + "learning_rate": 0.002973705917705442, + "loss": 1.1943, + "step": 776 + }, + { + "epoch": 0.05976463348973156, + "learning_rate": 0.002973638305114111, + "loss": 1.1486, + "step": 777 + }, + { + "epoch": 0.059841550649950004, + "learning_rate": 0.0029735706064755035, + "loss": 1.1634, + "step": 778 + }, + { + "epoch": 0.05991846781016845, + "learning_rate": 0.0029735028217935723, + "loss": 1.5149, + "step": 779 + }, + { + "epoch": 0.059995384970386895, + "learning_rate": 0.002973434951072276, + "loss": 0.8127, + "step": 780 + }, + { + "epoch": 0.06007230213060534, + "learning_rate": 0.0029733669943155776, + "loss": 0.9215, + "step": 781 + }, + { + "epoch": 0.060149219290823785, + "learning_rate": 0.002973298951527444, + "loss": 1.2661, + "step": 782 + }, + { + "epoch": 0.06022613645104223, + "learning_rate": 0.00297323082271185, + "loss": 1.2779, + "step": 783 + }, + { + "epoch": 0.06030305361126067, + "learning_rate": 0.0029731626078727726, + "loss": 1.2186, + "step": 784 + }, + { + "epoch": 0.060379970771479115, + "learning_rate": 0.0029730943070141952, + "loss": 1.2627, + "step": 785 + }, + { + "epoch": 0.06045688793169756, + "learning_rate": 0.002973025920140106, + "loss": 1.4769, + "step": 786 + }, + { + "epoch": 0.060533805091916006, + "learning_rate": 0.0029729574472544987, + "loss": 1.2963, + "step": 787 + }, + { + "epoch": 0.06061072225213445, + "learning_rate": 0.00297288888836137, + "loss": 1.6462, + "step": 788 + }, + { + "epoch": 0.060687639412352896, + "learning_rate": 0.002972820243464725, + "loss": 0.9503, + "step": 789 + }, + { + "epoch": 0.06076455657257134, + "learning_rate": 0.0029727515125685705, + "loss": 1.1223, + "step": 790 + }, + { + "epoch": 0.06084147373278979, + "learning_rate": 0.00297268269567692, + "loss": 0.9932, + "step": 791 + }, + { + "epoch": 0.06091839089300823, + "learning_rate": 0.0029726137927937927, + "loss": 1.1108, + "step": 792 + }, + { + "epoch": 0.06099530805322668, + "learning_rate": 0.002972544803923211, + "loss": 1.1645, + "step": 793 + }, + { + "epoch": 0.06107222521344512, + "learning_rate": 0.002972475729069204, + "loss": 1.0648, + "step": 794 + }, + { + "epoch": 0.06114914237366356, + "learning_rate": 0.0029724065682358043, + "loss": 1.1004, + "step": 795 + }, + { + "epoch": 0.06122605953388201, + "learning_rate": 0.0029723373214270506, + "loss": 1.2627, + "step": 796 + }, + { + "epoch": 0.06130297669410045, + "learning_rate": 0.002972267988646986, + "loss": 1.2579, + "step": 797 + }, + { + "epoch": 0.0613798938543189, + "learning_rate": 0.0029721985698996597, + "loss": 1.2242, + "step": 798 + }, + { + "epoch": 0.061456811014537344, + "learning_rate": 0.0029721290651891245, + "loss": 1.0346, + "step": 799 + }, + { + "epoch": 0.06153372817475579, + "learning_rate": 0.002972059474519439, + "loss": 1.258, + "step": 800 + }, + { + "epoch": 0.061610645334974234, + "learning_rate": 0.002971989797894666, + "loss": 1.3715, + "step": 801 + }, + { + "epoch": 0.06168756249519268, + "learning_rate": 0.0029719200353188753, + "loss": 1.5083, + "step": 802 + }, + { + "epoch": 0.061764479655411125, + "learning_rate": 0.0029718501867961396, + "loss": 0.7944, + "step": 803 + }, + { + "epoch": 0.061841396815629564, + "learning_rate": 0.0029717802523305374, + "loss": 1.3351, + "step": 804 + }, + { + "epoch": 0.06191831397584801, + "learning_rate": 0.0029717102319261526, + "loss": 1.1212, + "step": 805 + }, + { + "epoch": 0.061995231136066455, + "learning_rate": 0.0029716401255870735, + "loss": 1.1144, + "step": 806 + }, + { + "epoch": 0.0620721482962849, + "learning_rate": 0.0029715699333173937, + "loss": 1.0605, + "step": 807 + }, + { + "epoch": 0.062149065456503345, + "learning_rate": 0.002971499655121212, + "loss": 1.3306, + "step": 808 + }, + { + "epoch": 0.06222598261672179, + "learning_rate": 0.002971429291002632, + "loss": 0.7732, + "step": 809 + }, + { + "epoch": 0.062302899776940236, + "learning_rate": 0.0029713588409657618, + "loss": 1.21, + "step": 810 + }, + { + "epoch": 0.06237981693715868, + "learning_rate": 0.0029712883050147156, + "loss": 1.0372, + "step": 811 + }, + { + "epoch": 0.06245673409737713, + "learning_rate": 0.0029712176831536118, + "loss": 1.3111, + "step": 812 + }, + { + "epoch": 0.06253365125759557, + "learning_rate": 0.0029711469753865745, + "loss": 0.9122, + "step": 813 + }, + { + "epoch": 0.06261056841781401, + "learning_rate": 0.002971076181717732, + "loss": 1.1221, + "step": 814 + }, + { + "epoch": 0.06268748557803246, + "learning_rate": 0.002971005302151218, + "loss": 1.3591, + "step": 815 + }, + { + "epoch": 0.0627644027382509, + "learning_rate": 0.0029709343366911716, + "loss": 1.5147, + "step": 816 + }, + { + "epoch": 0.06284131989846935, + "learning_rate": 0.002970863285341736, + "loss": 1.1363, + "step": 817 + }, + { + "epoch": 0.06291823705868779, + "learning_rate": 0.0029707921481070608, + "loss": 0.8343, + "step": 818 + }, + { + "epoch": 0.06299515421890624, + "learning_rate": 0.002970720924991299, + "loss": 1.0197, + "step": 819 + }, + { + "epoch": 0.06307207137912468, + "learning_rate": 0.0029706496159986094, + "loss": 1.3745, + "step": 820 + }, + { + "epoch": 0.06314898853934313, + "learning_rate": 0.0029705782211331567, + "loss": 1.0971, + "step": 821 + }, + { + "epoch": 0.06322590569956157, + "learning_rate": 0.0029705067403991087, + "loss": 1.1213, + "step": 822 + }, + { + "epoch": 0.06330282285978002, + "learning_rate": 0.0029704351738006392, + "loss": 1.0577, + "step": 823 + }, + { + "epoch": 0.06337974001999847, + "learning_rate": 0.0029703635213419277, + "loss": 0.8942, + "step": 824 + }, + { + "epoch": 0.06345665718021691, + "learning_rate": 0.0029702917830271578, + "loss": 1.5873, + "step": 825 + }, + { + "epoch": 0.06353357434043536, + "learning_rate": 0.002970219958860519, + "loss": 1.5163, + "step": 826 + }, + { + "epoch": 0.0636104915006538, + "learning_rate": 0.0029701480488462037, + "loss": 0.8876, + "step": 827 + }, + { + "epoch": 0.06368740866087225, + "learning_rate": 0.002970076052988412, + "loss": 1.1676, + "step": 828 + }, + { + "epoch": 0.06376432582109069, + "learning_rate": 0.002970003971291348, + "loss": 1.2498, + "step": 829 + }, + { + "epoch": 0.06384124298130912, + "learning_rate": 0.002969931803759219, + "loss": 0.9933, + "step": 830 + }, + { + "epoch": 0.06391816014152757, + "learning_rate": 0.0029698595503962408, + "loss": 1.4362, + "step": 831 + }, + { + "epoch": 0.06399507730174601, + "learning_rate": 0.002969787211206632, + "loss": 0.9662, + "step": 832 + }, + { + "epoch": 0.06407199446196446, + "learning_rate": 0.0029697147861946155, + "loss": 0.9095, + "step": 833 + }, + { + "epoch": 0.0641489116221829, + "learning_rate": 0.0029696422753644207, + "loss": 1.1928, + "step": 834 + }, + { + "epoch": 0.06422582878240135, + "learning_rate": 0.0029695696787202823, + "loss": 1.0297, + "step": 835 + }, + { + "epoch": 0.0643027459426198, + "learning_rate": 0.002969496996266439, + "loss": 0.7538, + "step": 836 + }, + { + "epoch": 0.06437966310283824, + "learning_rate": 0.002969424228007134, + "loss": 1.1861, + "step": 837 + }, + { + "epoch": 0.06445658026305669, + "learning_rate": 0.0029693513739466174, + "loss": 1.2194, + "step": 838 + }, + { + "epoch": 0.06453349742327513, + "learning_rate": 0.0029692784340891423, + "loss": 1.1569, + "step": 839 + }, + { + "epoch": 0.06461041458349358, + "learning_rate": 0.0029692054084389688, + "loss": 1.3693, + "step": 840 + }, + { + "epoch": 0.06468733174371202, + "learning_rate": 0.0029691322970003594, + "loss": 1.275, + "step": 841 + }, + { + "epoch": 0.06476424890393047, + "learning_rate": 0.002969059099777585, + "loss": 0.8621, + "step": 842 + }, + { + "epoch": 0.06484116606414891, + "learning_rate": 0.0029689858167749185, + "loss": 1.258, + "step": 843 + }, + { + "epoch": 0.06491808322436736, + "learning_rate": 0.002968912447996639, + "loss": 1.0037, + "step": 844 + }, + { + "epoch": 0.0649950003845858, + "learning_rate": 0.0029688389934470314, + "loss": 1.148, + "step": 845 + }, + { + "epoch": 0.06507191754480425, + "learning_rate": 0.0029687654531303834, + "loss": 1.1333, + "step": 846 + }, + { + "epoch": 0.0651488347050227, + "learning_rate": 0.0029686918270509905, + "loss": 1.1125, + "step": 847 + }, + { + "epoch": 0.06522575186524114, + "learning_rate": 0.0029686181152131508, + "loss": 1.1169, + "step": 848 + }, + { + "epoch": 0.06530266902545959, + "learning_rate": 0.002968544317621169, + "loss": 1.3645, + "step": 849 + }, + { + "epoch": 0.06537958618567802, + "learning_rate": 0.0029684704342793538, + "loss": 1.1042, + "step": 850 + }, + { + "epoch": 0.06545650334589646, + "learning_rate": 0.0029683964651920198, + "loss": 1.1865, + "step": 851 + }, + { + "epoch": 0.06553342050611491, + "learning_rate": 0.0029683224103634856, + "loss": 1.224, + "step": 852 + }, + { + "epoch": 0.06561033766633335, + "learning_rate": 0.002968248269798076, + "loss": 1.6174, + "step": 853 + }, + { + "epoch": 0.0656872548265518, + "learning_rate": 0.0029681740435001194, + "loss": 1.5593, + "step": 854 + }, + { + "epoch": 0.06576417198677025, + "learning_rate": 0.00296809973147395, + "loss": 1.2401, + "step": 855 + }, + { + "epoch": 0.06584108914698869, + "learning_rate": 0.002968025333723908, + "loss": 1.0987, + "step": 856 + }, + { + "epoch": 0.06591800630720714, + "learning_rate": 0.0029679508502543366, + "loss": 1.2755, + "step": 857 + }, + { + "epoch": 0.06599492346742558, + "learning_rate": 0.002967876281069585, + "loss": 1.6386, + "step": 858 + }, + { + "epoch": 0.06607184062764403, + "learning_rate": 0.0029678016261740076, + "loss": 1.5125, + "step": 859 + }, + { + "epoch": 0.06614875778786247, + "learning_rate": 0.002967726885571964, + "loss": 1.4287, + "step": 860 + }, + { + "epoch": 0.06622567494808092, + "learning_rate": 0.002967652059267817, + "loss": 1.1767, + "step": 861 + }, + { + "epoch": 0.06630259210829936, + "learning_rate": 0.002967577147265937, + "loss": 1.3266, + "step": 862 + }, + { + "epoch": 0.06637950926851781, + "learning_rate": 0.0029675021495706986, + "loss": 1.2988, + "step": 863 + }, + { + "epoch": 0.06645642642873625, + "learning_rate": 0.00296742706618648, + "loss": 0.8827, + "step": 864 + }, + { + "epoch": 0.0665333435889547, + "learning_rate": 0.002967351897117665, + "loss": 1.3811, + "step": 865 + }, + { + "epoch": 0.06661026074917314, + "learning_rate": 0.0029672766423686444, + "loss": 1.1978, + "step": 866 + }, + { + "epoch": 0.06668717790939159, + "learning_rate": 0.002967201301943811, + "loss": 1.3499, + "step": 867 + }, + { + "epoch": 0.06676409506961004, + "learning_rate": 0.0029671258758475643, + "loss": 1.174, + "step": 868 + }, + { + "epoch": 0.06684101222982848, + "learning_rate": 0.0029670503640843093, + "loss": 1.0905, + "step": 869 + }, + { + "epoch": 0.06691792939004693, + "learning_rate": 0.002966974766658454, + "loss": 1.262, + "step": 870 + }, + { + "epoch": 0.06699484655026536, + "learning_rate": 0.0029668990835744137, + "loss": 0.9186, + "step": 871 + }, + { + "epoch": 0.0670717637104838, + "learning_rate": 0.0029668233148366072, + "loss": 0.9934, + "step": 872 + }, + { + "epoch": 0.06714868087070225, + "learning_rate": 0.002966747460449458, + "loss": 1.2823, + "step": 873 + }, + { + "epoch": 0.0672255980309207, + "learning_rate": 0.002966671520417397, + "loss": 1.0951, + "step": 874 + }, + { + "epoch": 0.06730251519113914, + "learning_rate": 0.0029665954947448565, + "loss": 1.2555, + "step": 875 + }, + { + "epoch": 0.06737943235135758, + "learning_rate": 0.002966519383436277, + "loss": 1.0027, + "step": 876 + }, + { + "epoch": 0.06745634951157603, + "learning_rate": 0.0029664431864961024, + "loss": 1.2581, + "step": 877 + }, + { + "epoch": 0.06753326667179448, + "learning_rate": 0.0029663669039287814, + "loss": 1.1946, + "step": 878 + }, + { + "epoch": 0.06761018383201292, + "learning_rate": 0.0029662905357387692, + "loss": 1.1934, + "step": 879 + }, + { + "epoch": 0.06768710099223137, + "learning_rate": 0.002966214081930524, + "loss": 1.2912, + "step": 880 + }, + { + "epoch": 0.06776401815244981, + "learning_rate": 0.002966137542508511, + "loss": 1.4963, + "step": 881 + }, + { + "epoch": 0.06784093531266826, + "learning_rate": 0.0029660609174771983, + "loss": 1.047, + "step": 882 + }, + { + "epoch": 0.0679178524728867, + "learning_rate": 0.0029659842068410617, + "loss": 1.3658, + "step": 883 + }, + { + "epoch": 0.06799476963310515, + "learning_rate": 0.0029659074106045784, + "loss": 1.5428, + "step": 884 + }, + { + "epoch": 0.0680716867933236, + "learning_rate": 0.0029658305287722344, + "loss": 1.0194, + "step": 885 + }, + { + "epoch": 0.06814860395354204, + "learning_rate": 0.002965753561348518, + "loss": 1.2364, + "step": 886 + }, + { + "epoch": 0.06822552111376048, + "learning_rate": 0.0029656765083379234, + "loss": 1.0498, + "step": 887 + }, + { + "epoch": 0.06830243827397893, + "learning_rate": 0.00296559936974495, + "loss": 1.0967, + "step": 888 + }, + { + "epoch": 0.06837935543419738, + "learning_rate": 0.002965522145574102, + "loss": 1.9147, + "step": 889 + }, + { + "epoch": 0.06845627259441582, + "learning_rate": 0.0029654448358298886, + "loss": 1.3101, + "step": 890 + }, + { + "epoch": 0.06853318975463425, + "learning_rate": 0.0029653674405168244, + "loss": 0.8805, + "step": 891 + }, + { + "epoch": 0.0686101069148527, + "learning_rate": 0.0029652899596394277, + "loss": 1.0904, + "step": 892 + }, + { + "epoch": 0.06868702407507114, + "learning_rate": 0.0029652123932022234, + "loss": 0.8722, + "step": 893 + }, + { + "epoch": 0.06876394123528959, + "learning_rate": 0.00296513474120974, + "loss": 1.304, + "step": 894 + }, + { + "epoch": 0.06884085839550803, + "learning_rate": 0.0029650570036665125, + "loss": 0.849, + "step": 895 + }, + { + "epoch": 0.06891777555572648, + "learning_rate": 0.0029649791805770797, + "loss": 0.9992, + "step": 896 + }, + { + "epoch": 0.06899469271594492, + "learning_rate": 0.002964901271945986, + "loss": 1.0795, + "step": 897 + }, + { + "epoch": 0.06907160987616337, + "learning_rate": 0.00296482327777778, + "loss": 1.304, + "step": 898 + }, + { + "epoch": 0.06914852703638182, + "learning_rate": 0.0029647451980770166, + "loss": 1.0488, + "step": 899 + }, + { + "epoch": 0.06922544419660026, + "learning_rate": 0.0029646670328482543, + "loss": 1.2691, + "step": 900 + }, + { + "epoch": 0.0693023613568187, + "learning_rate": 0.0029645887820960574, + "loss": 1.3094, + "step": 901 + }, + { + "epoch": 0.06937927851703715, + "learning_rate": 0.002964510445824995, + "loss": 1.2892, + "step": 902 + }, + { + "epoch": 0.0694561956772556, + "learning_rate": 0.002964432024039642, + "loss": 1.2423, + "step": 903 + }, + { + "epoch": 0.06953311283747404, + "learning_rate": 0.0029643535167445767, + "loss": 1.2492, + "step": 904 + }, + { + "epoch": 0.06961002999769249, + "learning_rate": 0.002964274923944383, + "loss": 1.3041, + "step": 905 + }, + { + "epoch": 0.06968694715791093, + "learning_rate": 0.002964196245643651, + "loss": 1.0715, + "step": 906 + }, + { + "epoch": 0.06976386431812938, + "learning_rate": 0.0029641174818469743, + "loss": 1.0728, + "step": 907 + }, + { + "epoch": 0.06984078147834782, + "learning_rate": 0.0029640386325589514, + "loss": 1.042, + "step": 908 + }, + { + "epoch": 0.06991769863856627, + "learning_rate": 0.0029639596977841876, + "loss": 1.4161, + "step": 909 + }, + { + "epoch": 0.06999461579878472, + "learning_rate": 0.0029638806775272907, + "loss": 1.4001, + "step": 910 + }, + { + "epoch": 0.07007153295900316, + "learning_rate": 0.0029638015717928755, + "loss": 1.355, + "step": 911 + }, + { + "epoch": 0.07014845011922159, + "learning_rate": 0.0029637223805855616, + "loss": 1.1305, + "step": 912 + }, + { + "epoch": 0.07022536727944004, + "learning_rate": 0.002963643103909972, + "loss": 1.1159, + "step": 913 + }, + { + "epoch": 0.07030228443965848, + "learning_rate": 0.002963563741770736, + "loss": 1.3449, + "step": 914 + }, + { + "epoch": 0.07037920159987693, + "learning_rate": 0.002963484294172488, + "loss": 1.338, + "step": 915 + }, + { + "epoch": 0.07045611876009537, + "learning_rate": 0.002963404761119867, + "loss": 1.2533, + "step": 916 + }, + { + "epoch": 0.07053303592031382, + "learning_rate": 0.002963325142617517, + "loss": 1.0319, + "step": 917 + }, + { + "epoch": 0.07060995308053226, + "learning_rate": 0.0029632454386700866, + "loss": 1.2467, + "step": 918 + }, + { + "epoch": 0.07068687024075071, + "learning_rate": 0.0029631656492822302, + "loss": 1.2088, + "step": 919 + }, + { + "epoch": 0.07076378740096916, + "learning_rate": 0.0029630857744586065, + "loss": 1.0837, + "step": 920 + }, + { + "epoch": 0.0708407045611876, + "learning_rate": 0.00296300581420388, + "loss": 1.2227, + "step": 921 + }, + { + "epoch": 0.07091762172140605, + "learning_rate": 0.002962925768522719, + "loss": 0.9956, + "step": 922 + }, + { + "epoch": 0.07099453888162449, + "learning_rate": 0.002962845637419798, + "loss": 1.1343, + "step": 923 + }, + { + "epoch": 0.07107145604184294, + "learning_rate": 0.0029627654208997955, + "loss": 1.0709, + "step": 924 + }, + { + "epoch": 0.07114837320206138, + "learning_rate": 0.0029626851189673955, + "loss": 1.1055, + "step": 925 + }, + { + "epoch": 0.07122529036227983, + "learning_rate": 0.002962604731627287, + "loss": 1.0965, + "step": 926 + }, + { + "epoch": 0.07130220752249827, + "learning_rate": 0.0029625242588841643, + "loss": 0.9764, + "step": 927 + }, + { + "epoch": 0.07137912468271672, + "learning_rate": 0.002962443700742726, + "loss": 1.503, + "step": 928 + }, + { + "epoch": 0.07145604184293516, + "learning_rate": 0.002962363057207675, + "loss": 1.1844, + "step": 929 + }, + { + "epoch": 0.07153295900315361, + "learning_rate": 0.002962282328283722, + "loss": 1.3772, + "step": 930 + }, + { + "epoch": 0.07160987616337205, + "learning_rate": 0.0029622015139755794, + "loss": 1.0582, + "step": 931 + }, + { + "epoch": 0.07168679332359049, + "learning_rate": 0.002962120614287967, + "loss": 1.126, + "step": 932 + }, + { + "epoch": 0.07176371048380893, + "learning_rate": 0.0029620396292256075, + "loss": 1.2031, + "step": 933 + }, + { + "epoch": 0.07184062764402738, + "learning_rate": 0.0029619585587932306, + "loss": 1.3518, + "step": 934 + }, + { + "epoch": 0.07191754480424582, + "learning_rate": 0.00296187740299557, + "loss": 0.9528, + "step": 935 + }, + { + "epoch": 0.07199446196446427, + "learning_rate": 0.0029617961618373647, + "loss": 1.2507, + "step": 936 + }, + { + "epoch": 0.07207137912468271, + "learning_rate": 0.002961714835323357, + "loss": 1.2534, + "step": 937 + }, + { + "epoch": 0.07214829628490116, + "learning_rate": 0.0029616334234582975, + "loss": 1.1605, + "step": 938 + }, + { + "epoch": 0.0722252134451196, + "learning_rate": 0.0029615519262469387, + "loss": 1.1648, + "step": 939 + }, + { + "epoch": 0.07230213060533805, + "learning_rate": 0.00296147034369404, + "loss": 1.0868, + "step": 940 + }, + { + "epoch": 0.0723790477655565, + "learning_rate": 0.0029613886758043653, + "loss": 1.1404, + "step": 941 + }, + { + "epoch": 0.07245596492577494, + "learning_rate": 0.002961306922582682, + "loss": 1.1598, + "step": 942 + }, + { + "epoch": 0.07253288208599339, + "learning_rate": 0.002961225084033765, + "loss": 0.9223, + "step": 943 + }, + { + "epoch": 0.07260979924621183, + "learning_rate": 0.002961143160162392, + "loss": 1.2376, + "step": 944 + }, + { + "epoch": 0.07268671640643028, + "learning_rate": 0.002961061150973347, + "loss": 1.1766, + "step": 945 + }, + { + "epoch": 0.07276363356664872, + "learning_rate": 0.0029609790564714194, + "loss": 1.0539, + "step": 946 + }, + { + "epoch": 0.07284055072686717, + "learning_rate": 0.002960896876661402, + "loss": 1.3398, + "step": 947 + }, + { + "epoch": 0.07291746788708561, + "learning_rate": 0.002960814611548093, + "loss": 1.0852, + "step": 948 + }, + { + "epoch": 0.07299438504730406, + "learning_rate": 0.0029607322611362973, + "loss": 1.0945, + "step": 949 + }, + { + "epoch": 0.0730713022075225, + "learning_rate": 0.0029606498254308214, + "loss": 1.0968, + "step": 950 + }, + { + "epoch": 0.07314821936774095, + "learning_rate": 0.0029605673044364807, + "loss": 0.8015, + "step": 951 + }, + { + "epoch": 0.0732251365279594, + "learning_rate": 0.0029604846981580927, + "loss": 1.1704, + "step": 952 + }, + { + "epoch": 0.07330205368817783, + "learning_rate": 0.002960402006600481, + "loss": 1.2461, + "step": 953 + }, + { + "epoch": 0.07337897084839627, + "learning_rate": 0.0029603192297684745, + "loss": 1.2593, + "step": 954 + }, + { + "epoch": 0.07345588800861472, + "learning_rate": 0.002960236367666906, + "loss": 0.9799, + "step": 955 + }, + { + "epoch": 0.07353280516883316, + "learning_rate": 0.0029601534203006143, + "loss": 1.098, + "step": 956 + }, + { + "epoch": 0.07360972232905161, + "learning_rate": 0.002960070387674442, + "loss": 1.2879, + "step": 957 + }, + { + "epoch": 0.07368663948927005, + "learning_rate": 0.0029599872697932393, + "loss": 1.3842, + "step": 958 + }, + { + "epoch": 0.0737635566494885, + "learning_rate": 0.002959904066661857, + "loss": 1.1728, + "step": 959 + }, + { + "epoch": 0.07384047380970694, + "learning_rate": 0.0029598207782851557, + "loss": 1.1599, + "step": 960 + }, + { + "epoch": 0.07391739096992539, + "learning_rate": 0.0029597374046679975, + "loss": 1.1854, + "step": 961 + }, + { + "epoch": 0.07399430813014383, + "learning_rate": 0.0029596539458152512, + "loss": 1.1199, + "step": 962 + }, + { + "epoch": 0.07407122529036228, + "learning_rate": 0.0029595704017317897, + "loss": 1.3585, + "step": 963 + }, + { + "epoch": 0.07414814245058073, + "learning_rate": 0.002959486772422491, + "loss": 1.2377, + "step": 964 + }, + { + "epoch": 0.07422505961079917, + "learning_rate": 0.0029594030578922385, + "loss": 1.1028, + "step": 965 + }, + { + "epoch": 0.07430197677101762, + "learning_rate": 0.002959319258145921, + "loss": 1.1915, + "step": 966 + }, + { + "epoch": 0.07437889393123606, + "learning_rate": 0.002959235373188431, + "loss": 0.9092, + "step": 967 + }, + { + "epoch": 0.07445581109145451, + "learning_rate": 0.0029591514030246667, + "loss": 0.9983, + "step": 968 + }, + { + "epoch": 0.07453272825167295, + "learning_rate": 0.002959067347659531, + "loss": 1.3972, + "step": 969 + }, + { + "epoch": 0.0746096454118914, + "learning_rate": 0.002958983207097933, + "loss": 1.1719, + "step": 970 + }, + { + "epoch": 0.07468656257210984, + "learning_rate": 0.0029588989813447843, + "loss": 1.009, + "step": 971 + }, + { + "epoch": 0.07476347973232829, + "learning_rate": 0.0029588146704050042, + "loss": 1.3163, + "step": 972 + }, + { + "epoch": 0.07484039689254672, + "learning_rate": 0.002958730274283515, + "loss": 1.0569, + "step": 973 + }, + { + "epoch": 0.07491731405276517, + "learning_rate": 0.0029586457929852445, + "loss": 1.647, + "step": 974 + }, + { + "epoch": 0.07499423121298361, + "learning_rate": 0.0029585612265151263, + "loss": 1.0079, + "step": 975 + }, + { + "epoch": 0.07507114837320206, + "learning_rate": 0.0029584765748780976, + "loss": 1.3997, + "step": 976 + }, + { + "epoch": 0.0751480655334205, + "learning_rate": 0.002958391838079102, + "loss": 0.9782, + "step": 977 + }, + { + "epoch": 0.07522498269363895, + "learning_rate": 0.0029583070161230875, + "loss": 0.8678, + "step": 978 + }, + { + "epoch": 0.07530189985385739, + "learning_rate": 0.002958222109015006, + "loss": 1.0651, + "step": 979 + }, + { + "epoch": 0.07537881701407584, + "learning_rate": 0.002958137116759816, + "loss": 1.3772, + "step": 980 + }, + { + "epoch": 0.07545573417429428, + "learning_rate": 0.00295805203936248, + "loss": 1.3194, + "step": 981 + }, + { + "epoch": 0.07553265133451273, + "learning_rate": 0.0029579668768279664, + "loss": 1.5207, + "step": 982 + }, + { + "epoch": 0.07560956849473117, + "learning_rate": 0.0029578816291612464, + "loss": 1.2981, + "step": 983 + }, + { + "epoch": 0.07568648565494962, + "learning_rate": 0.0029577962963672996, + "loss": 1.4814, + "step": 984 + }, + { + "epoch": 0.07576340281516807, + "learning_rate": 0.002957710878451107, + "loss": 1.1017, + "step": 985 + }, + { + "epoch": 0.07584031997538651, + "learning_rate": 0.0029576253754176576, + "loss": 1.1641, + "step": 986 + }, + { + "epoch": 0.07591723713560496, + "learning_rate": 0.0029575397872719433, + "loss": 1.505, + "step": 987 + }, + { + "epoch": 0.0759941542958234, + "learning_rate": 0.002957454114018962, + "loss": 1.3036, + "step": 988 + }, + { + "epoch": 0.07607107145604185, + "learning_rate": 0.0029573683556637156, + "loss": 0.8347, + "step": 989 + }, + { + "epoch": 0.07614798861626029, + "learning_rate": 0.0029572825122112126, + "loss": 1.1208, + "step": 990 + }, + { + "epoch": 0.07622490577647874, + "learning_rate": 0.0029571965836664646, + "loss": 0.9651, + "step": 991 + }, + { + "epoch": 0.07630182293669718, + "learning_rate": 0.00295711057003449, + "loss": 1.2866, + "step": 992 + }, + { + "epoch": 0.07637874009691561, + "learning_rate": 0.0029570244713203095, + "loss": 1.2155, + "step": 993 + }, + { + "epoch": 0.07645565725713406, + "learning_rate": 0.0029569382875289526, + "loss": 1.4266, + "step": 994 + }, + { + "epoch": 0.0765325744173525, + "learning_rate": 0.0029568520186654505, + "loss": 1.0505, + "step": 995 + }, + { + "epoch": 0.07660949157757095, + "learning_rate": 0.0029567656647348406, + "loss": 1.3778, + "step": 996 + }, + { + "epoch": 0.0766864087377894, + "learning_rate": 0.0029566792257421657, + "loss": 1.2593, + "step": 997 + }, + { + "epoch": 0.07676332589800784, + "learning_rate": 0.002956592701692472, + "loss": 1.1632, + "step": 998 + }, + { + "epoch": 0.07684024305822629, + "learning_rate": 0.002956506092590813, + "loss": 1.1734, + "step": 999 + }, + { + "epoch": 0.07691716021844473, + "learning_rate": 0.002956419398442245, + "loss": 1.1082, + "step": 1000 + }, + { + "epoch": 0.07699407737866318, + "learning_rate": 0.0029563326192518307, + "loss": 1.7411, + "step": 1001 + }, + { + "epoch": 0.07707099453888162, + "learning_rate": 0.0029562457550246367, + "loss": 1.5031, + "step": 1002 + }, + { + "epoch": 0.07714791169910007, + "learning_rate": 0.0029561588057657358, + "loss": 0.7835, + "step": 1003 + }, + { + "epoch": 0.07722482885931851, + "learning_rate": 0.0029560717714802046, + "loss": 1.5095, + "step": 1004 + }, + { + "epoch": 0.07730174601953696, + "learning_rate": 0.002955984652173125, + "loss": 1.4018, + "step": 1005 + }, + { + "epoch": 0.0773786631797554, + "learning_rate": 0.002955897447849584, + "loss": 1.3244, + "step": 1006 + }, + { + "epoch": 0.07745558033997385, + "learning_rate": 0.0029558101585146746, + "loss": 1.0778, + "step": 1007 + }, + { + "epoch": 0.0775324975001923, + "learning_rate": 0.002955722784173492, + "loss": 1.5623, + "step": 1008 + }, + { + "epoch": 0.07760941466041074, + "learning_rate": 0.002955635324831139, + "loss": 0.9722, + "step": 1009 + }, + { + "epoch": 0.07768633182062919, + "learning_rate": 0.0029555477804927223, + "loss": 0.7629, + "step": 1010 + }, + { + "epoch": 0.07776324898084763, + "learning_rate": 0.002955460151163354, + "loss": 1.4302, + "step": 1011 + }, + { + "epoch": 0.07784016614106608, + "learning_rate": 0.002955372436848151, + "loss": 1.8574, + "step": 1012 + }, + { + "epoch": 0.07791708330128452, + "learning_rate": 0.002955284637552234, + "loss": 1.0181, + "step": 1013 + }, + { + "epoch": 0.07799400046150295, + "learning_rate": 0.00295519675328073, + "loss": 1.1129, + "step": 1014 + }, + { + "epoch": 0.0780709176217214, + "learning_rate": 0.002955108784038772, + "loss": 1.1438, + "step": 1015 + }, + { + "epoch": 0.07814783478193985, + "learning_rate": 0.002955020729831495, + "loss": 1.0609, + "step": 1016 + }, + { + "epoch": 0.07822475194215829, + "learning_rate": 0.0029549325906640415, + "loss": 1.307, + "step": 1017 + }, + { + "epoch": 0.07830166910237674, + "learning_rate": 0.002954844366541558, + "loss": 0.9569, + "step": 1018 + }, + { + "epoch": 0.07837858626259518, + "learning_rate": 0.002954756057469195, + "loss": 0.955, + "step": 1019 + }, + { + "epoch": 0.07845550342281363, + "learning_rate": 0.00295466766345211, + "loss": 1.2528, + "step": 1020 + }, + { + "epoch": 0.07853242058303207, + "learning_rate": 0.002954579184495464, + "loss": 1.4933, + "step": 1021 + }, + { + "epoch": 0.07860933774325052, + "learning_rate": 0.002954490620604424, + "loss": 1.0781, + "step": 1022 + }, + { + "epoch": 0.07868625490346896, + "learning_rate": 0.0029544019717841602, + "loss": 0.9809, + "step": 1023 + }, + { + "epoch": 0.07876317206368741, + "learning_rate": 0.00295431323803985, + "loss": 0.8879, + "step": 1024 + }, + { + "epoch": 0.07884008922390585, + "learning_rate": 0.0029542244193766737, + "loss": 0.7664, + "step": 1025 + }, + { + "epoch": 0.0789170063841243, + "learning_rate": 0.0029541355157998187, + "loss": 0.8898, + "step": 1026 + }, + { + "epoch": 0.07899392354434275, + "learning_rate": 0.002954046527314475, + "loss": 1.248, + "step": 1027 + }, + { + "epoch": 0.07907084070456119, + "learning_rate": 0.0029539574539258393, + "loss": 1.0316, + "step": 1028 + }, + { + "epoch": 0.07914775786477964, + "learning_rate": 0.0029538682956391127, + "loss": 0.968, + "step": 1029 + }, + { + "epoch": 0.07922467502499808, + "learning_rate": 0.0029537790524595004, + "loss": 1.1932, + "step": 1030 + }, + { + "epoch": 0.07930159218521653, + "learning_rate": 0.0029536897243922153, + "loss": 1.1205, + "step": 1031 + }, + { + "epoch": 0.07937850934543497, + "learning_rate": 0.0029536003114424714, + "loss": 1.2685, + "step": 1032 + }, + { + "epoch": 0.07945542650565342, + "learning_rate": 0.0029535108136154904, + "loss": 1.2933, + "step": 1033 + }, + { + "epoch": 0.07953234366587185, + "learning_rate": 0.002953421230916499, + "loss": 1.051, + "step": 1034 + }, + { + "epoch": 0.0796092608260903, + "learning_rate": 0.0029533315633507263, + "loss": 1.0509, + "step": 1035 + }, + { + "epoch": 0.07968617798630874, + "learning_rate": 0.0029532418109234093, + "loss": 0.8834, + "step": 1036 + }, + { + "epoch": 0.07976309514652719, + "learning_rate": 0.0029531519736397884, + "loss": 1.2259, + "step": 1037 + }, + { + "epoch": 0.07984001230674563, + "learning_rate": 0.0029530620515051094, + "loss": 1.0971, + "step": 1038 + }, + { + "epoch": 0.07991692946696408, + "learning_rate": 0.0029529720445246226, + "loss": 1.0709, + "step": 1039 + }, + { + "epoch": 0.07999384662718252, + "learning_rate": 0.002952881952703584, + "loss": 1.0895, + "step": 1040 + }, + { + "epoch": 0.08007076378740097, + "learning_rate": 0.0029527917760472544, + "loss": 1.3178, + "step": 1041 + }, + { + "epoch": 0.08014768094761941, + "learning_rate": 0.0029527015145608985, + "loss": 1.3647, + "step": 1042 + }, + { + "epoch": 0.08022459810783786, + "learning_rate": 0.0029526111682497873, + "loss": 0.541, + "step": 1043 + }, + { + "epoch": 0.0803015152680563, + "learning_rate": 0.002952520737119196, + "loss": 1.0206, + "step": 1044 + }, + { + "epoch": 0.08037843242827475, + "learning_rate": 0.0029524302211744054, + "loss": 1.4368, + "step": 1045 + }, + { + "epoch": 0.0804553495884932, + "learning_rate": 0.0029523396204207, + "loss": 1.3588, + "step": 1046 + }, + { + "epoch": 0.08053226674871164, + "learning_rate": 0.0029522489348633714, + "loss": 1.2697, + "step": 1047 + }, + { + "epoch": 0.08060918390893008, + "learning_rate": 0.0029521581645077133, + "loss": 1.1707, + "step": 1048 + }, + { + "epoch": 0.08068610106914853, + "learning_rate": 0.0029520673093590267, + "loss": 1.2145, + "step": 1049 + }, + { + "epoch": 0.08076301822936698, + "learning_rate": 0.0029519763694226164, + "loss": 1.1222, + "step": 1050 + }, + { + "epoch": 0.08083993538958542, + "learning_rate": 0.0029518853447037933, + "loss": 1.0963, + "step": 1051 + }, + { + "epoch": 0.08091685254980387, + "learning_rate": 0.002951794235207871, + "loss": 1.1351, + "step": 1052 + }, + { + "epoch": 0.08099376971002231, + "learning_rate": 0.002951703040940171, + "loss": 1.0011, + "step": 1053 + }, + { + "epoch": 0.08107068687024076, + "learning_rate": 0.0029516117619060173, + "loss": 1.1586, + "step": 1054 + }, + { + "epoch": 0.08114760403045919, + "learning_rate": 0.0029515203981107397, + "loss": 0.9994, + "step": 1055 + }, + { + "epoch": 0.08122452119067763, + "learning_rate": 0.002951428949559674, + "loss": 1.0701, + "step": 1056 + }, + { + "epoch": 0.08130143835089608, + "learning_rate": 0.0029513374162581586, + "loss": 0.9321, + "step": 1057 + }, + { + "epoch": 0.08137835551111453, + "learning_rate": 0.0029512457982115393, + "loss": 1.3266, + "step": 1058 + }, + { + "epoch": 0.08145527267133297, + "learning_rate": 0.0029511540954251653, + "loss": 1.2703, + "step": 1059 + }, + { + "epoch": 0.08153218983155142, + "learning_rate": 0.002951062307904391, + "loss": 1.1536, + "step": 1060 + }, + { + "epoch": 0.08160910699176986, + "learning_rate": 0.002950970435654577, + "loss": 1.0513, + "step": 1061 + }, + { + "epoch": 0.0816860241519883, + "learning_rate": 0.0029508784786810866, + "loss": 1.0564, + "step": 1062 + }, + { + "epoch": 0.08176294131220675, + "learning_rate": 0.0029507864369892898, + "loss": 1.1673, + "step": 1063 + }, + { + "epoch": 0.0818398584724252, + "learning_rate": 0.002950694310584561, + "loss": 1.2823, + "step": 1064 + }, + { + "epoch": 0.08191677563264364, + "learning_rate": 0.002950602099472279, + "loss": 1.1043, + "step": 1065 + }, + { + "epoch": 0.08199369279286209, + "learning_rate": 0.002950509803657829, + "loss": 0.9982, + "step": 1066 + }, + { + "epoch": 0.08207060995308053, + "learning_rate": 0.0029504174231466005, + "loss": 0.8833, + "step": 1067 + }, + { + "epoch": 0.08214752711329898, + "learning_rate": 0.0029503249579439866, + "loss": 1.4851, + "step": 1068 + }, + { + "epoch": 0.08222444427351742, + "learning_rate": 0.002950232408055387, + "loss": 1.4989, + "step": 1069 + }, + { + "epoch": 0.08230136143373587, + "learning_rate": 0.0029501397734862053, + "loss": 1.3504, + "step": 1070 + }, + { + "epoch": 0.08237827859395432, + "learning_rate": 0.0029500470542418514, + "loss": 1.2241, + "step": 1071 + }, + { + "epoch": 0.08245519575417276, + "learning_rate": 0.0029499542503277384, + "loss": 1.1753, + "step": 1072 + }, + { + "epoch": 0.0825321129143912, + "learning_rate": 0.0029498613617492857, + "loss": 1.2218, + "step": 1073 + }, + { + "epoch": 0.08260903007460965, + "learning_rate": 0.0029497683885119173, + "loss": 0.9293, + "step": 1074 + }, + { + "epoch": 0.08268594723482808, + "learning_rate": 0.0029496753306210615, + "loss": 1.1212, + "step": 1075 + }, + { + "epoch": 0.08276286439504653, + "learning_rate": 0.002949582188082153, + "loss": 1.2674, + "step": 1076 + }, + { + "epoch": 0.08283978155526497, + "learning_rate": 0.002949488960900629, + "loss": 1.1751, + "step": 1077 + }, + { + "epoch": 0.08291669871548342, + "learning_rate": 0.0029493956490819345, + "loss": 1.1667, + "step": 1078 + }, + { + "epoch": 0.08299361587570186, + "learning_rate": 0.002949302252631517, + "loss": 1.0874, + "step": 1079 + }, + { + "epoch": 0.08307053303592031, + "learning_rate": 0.002949208771554831, + "loss": 1.1288, + "step": 1080 + }, + { + "epoch": 0.08314745019613876, + "learning_rate": 0.002949115205857334, + "loss": 0.9737, + "step": 1081 + }, + { + "epoch": 0.0832243673563572, + "learning_rate": 0.00294902155554449, + "loss": 1.3464, + "step": 1082 + }, + { + "epoch": 0.08330128451657565, + "learning_rate": 0.002948927820621768, + "loss": 1.1628, + "step": 1083 + }, + { + "epoch": 0.08337820167679409, + "learning_rate": 0.00294883400109464, + "loss": 1.159, + "step": 1084 + }, + { + "epoch": 0.08345511883701254, + "learning_rate": 0.0029487400969685847, + "loss": 1.2899, + "step": 1085 + }, + { + "epoch": 0.08353203599723098, + "learning_rate": 0.0029486461082490856, + "loss": 1.4605, + "step": 1086 + }, + { + "epoch": 0.08360895315744943, + "learning_rate": 0.00294855203494163, + "loss": 1.0857, + "step": 1087 + }, + { + "epoch": 0.08368587031766787, + "learning_rate": 0.0029484578770517115, + "loss": 1.1297, + "step": 1088 + }, + { + "epoch": 0.08376278747788632, + "learning_rate": 0.0029483636345848285, + "loss": 1.0929, + "step": 1089 + }, + { + "epoch": 0.08383970463810476, + "learning_rate": 0.002948269307546483, + "loss": 1.1001, + "step": 1090 + }, + { + "epoch": 0.08391662179832321, + "learning_rate": 0.0029481748959421833, + "loss": 1.2571, + "step": 1091 + }, + { + "epoch": 0.08399353895854166, + "learning_rate": 0.0029480803997774425, + "loss": 1.1372, + "step": 1092 + }, + { + "epoch": 0.0840704561187601, + "learning_rate": 0.0029479858190577775, + "loss": 1.5051, + "step": 1093 + }, + { + "epoch": 0.08414737327897855, + "learning_rate": 0.002947891153788712, + "loss": 1.0785, + "step": 1094 + }, + { + "epoch": 0.08422429043919699, + "learning_rate": 0.002947796403975773, + "loss": 1.3196, + "step": 1095 + }, + { + "epoch": 0.08430120759941542, + "learning_rate": 0.002947701569624493, + "loss": 0.9348, + "step": 1096 + }, + { + "epoch": 0.08437812475963387, + "learning_rate": 0.0029476066507404092, + "loss": 1.3281, + "step": 1097 + }, + { + "epoch": 0.08445504191985231, + "learning_rate": 0.002947511647329065, + "loss": 1.1641, + "step": 1098 + }, + { + "epoch": 0.08453195908007076, + "learning_rate": 0.0029474165593960065, + "loss": 1.2422, + "step": 1099 + }, + { + "epoch": 0.0846088762402892, + "learning_rate": 0.0029473213869467873, + "loss": 0.9473, + "step": 1100 + }, + { + "epoch": 0.08468579340050765, + "learning_rate": 0.0029472261299869633, + "loss": 1.2083, + "step": 1101 + }, + { + "epoch": 0.0847627105607261, + "learning_rate": 0.002947130788522098, + "loss": 1.1972, + "step": 1102 + }, + { + "epoch": 0.08483962772094454, + "learning_rate": 0.0029470353625577574, + "loss": 1.3026, + "step": 1103 + }, + { + "epoch": 0.08491654488116299, + "learning_rate": 0.002946939852099514, + "loss": 1.2338, + "step": 1104 + }, + { + "epoch": 0.08499346204138143, + "learning_rate": 0.002946844257152945, + "loss": 1.2036, + "step": 1105 + }, + { + "epoch": 0.08507037920159988, + "learning_rate": 0.0029467485777236314, + "loss": 0.9109, + "step": 1106 + }, + { + "epoch": 0.08514729636181832, + "learning_rate": 0.002946652813817161, + "loss": 1.0046, + "step": 1107 + }, + { + "epoch": 0.08522421352203677, + "learning_rate": 0.0029465569654391244, + "loss": 1.2388, + "step": 1108 + }, + { + "epoch": 0.08530113068225521, + "learning_rate": 0.00294646103259512, + "loss": 1.1558, + "step": 1109 + }, + { + "epoch": 0.08537804784247366, + "learning_rate": 0.0029463650152907477, + "loss": 0.8834, + "step": 1110 + }, + { + "epoch": 0.0854549650026921, + "learning_rate": 0.002946268913531615, + "loss": 1.1247, + "step": 1111 + }, + { + "epoch": 0.08553188216291055, + "learning_rate": 0.0029461727273233334, + "loss": 0.9333, + "step": 1112 + }, + { + "epoch": 0.085608799323129, + "learning_rate": 0.0029460764566715193, + "loss": 1.2425, + "step": 1113 + }, + { + "epoch": 0.08568571648334744, + "learning_rate": 0.0029459801015817933, + "loss": 0.9309, + "step": 1114 + }, + { + "epoch": 0.08576263364356589, + "learning_rate": 0.0029458836620597823, + "loss": 1.3583, + "step": 1115 + }, + { + "epoch": 0.08583955080378432, + "learning_rate": 0.0029457871381111172, + "loss": 0.9936, + "step": 1116 + }, + { + "epoch": 0.08591646796400276, + "learning_rate": 0.002945690529741435, + "loss": 1.1197, + "step": 1117 + }, + { + "epoch": 0.08599338512422121, + "learning_rate": 0.002945593836956376, + "loss": 0.935, + "step": 1118 + }, + { + "epoch": 0.08607030228443965, + "learning_rate": 0.002945497059761586, + "loss": 1.1282, + "step": 1119 + }, + { + "epoch": 0.0861472194446581, + "learning_rate": 0.002945400198162716, + "loss": 1.2256, + "step": 1120 + }, + { + "epoch": 0.08622413660487654, + "learning_rate": 0.0029453032521654225, + "loss": 1.156, + "step": 1121 + }, + { + "epoch": 0.08630105376509499, + "learning_rate": 0.0029452062217753656, + "loss": 1.3209, + "step": 1122 + }, + { + "epoch": 0.08637797092531344, + "learning_rate": 0.0029451091069982115, + "loss": 1.1248, + "step": 1123 + }, + { + "epoch": 0.08645488808553188, + "learning_rate": 0.0029450119078396303, + "loss": 1.3843, + "step": 1124 + }, + { + "epoch": 0.08653180524575033, + "learning_rate": 0.002944914624305298, + "loss": 1.3654, + "step": 1125 + }, + { + "epoch": 0.08660872240596877, + "learning_rate": 0.0029448172564008946, + "loss": 1.0795, + "step": 1126 + }, + { + "epoch": 0.08668563956618722, + "learning_rate": 0.0029447198041321065, + "loss": 1.3927, + "step": 1127 + }, + { + "epoch": 0.08676255672640566, + "learning_rate": 0.002944622267504623, + "loss": 0.9433, + "step": 1128 + }, + { + "epoch": 0.08683947388662411, + "learning_rate": 0.00294452464652414, + "loss": 0.9105, + "step": 1129 + }, + { + "epoch": 0.08691639104684255, + "learning_rate": 0.0029444269411963565, + "loss": 1.0921, + "step": 1130 + }, + { + "epoch": 0.086993308207061, + "learning_rate": 0.0029443291515269796, + "loss": 1.4, + "step": 1131 + }, + { + "epoch": 0.08707022536727944, + "learning_rate": 0.002944231277521718, + "loss": 1.5666, + "step": 1132 + }, + { + "epoch": 0.08714714252749789, + "learning_rate": 0.002944133319186287, + "loss": 1.0999, + "step": 1133 + }, + { + "epoch": 0.08722405968771633, + "learning_rate": 0.002944035276526406, + "loss": 1.1352, + "step": 1134 + }, + { + "epoch": 0.08730097684793478, + "learning_rate": 0.002943937149547801, + "loss": 1.3064, + "step": 1135 + }, + { + "epoch": 0.08737789400815323, + "learning_rate": 0.0029438389382562004, + "loss": 1.0006, + "step": 1136 + }, + { + "epoch": 0.08745481116837166, + "learning_rate": 0.00294374064265734, + "loss": 1.2764, + "step": 1137 + }, + { + "epoch": 0.0875317283285901, + "learning_rate": 0.0029436422627569583, + "loss": 1.3404, + "step": 1138 + }, + { + "epoch": 0.08760864548880855, + "learning_rate": 0.002943543798560801, + "loss": 1.3129, + "step": 1139 + }, + { + "epoch": 0.087685562649027, + "learning_rate": 0.0029434452500746162, + "loss": 1.1383, + "step": 1140 + }, + { + "epoch": 0.08776247980924544, + "learning_rate": 0.002943346617304159, + "loss": 1.2198, + "step": 1141 + }, + { + "epoch": 0.08783939696946388, + "learning_rate": 0.0029432479002551894, + "loss": 1.0944, + "step": 1142 + }, + { + "epoch": 0.08791631412968233, + "learning_rate": 0.00294314909893347, + "loss": 1.2941, + "step": 1143 + }, + { + "epoch": 0.08799323128990078, + "learning_rate": 0.0029430502133447707, + "loss": 1.426, + "step": 1144 + }, + { + "epoch": 0.08807014845011922, + "learning_rate": 0.002942951243494866, + "loss": 1.2258, + "step": 1145 + }, + { + "epoch": 0.08814706561033767, + "learning_rate": 0.002942852189389534, + "loss": 1.2612, + "step": 1146 + }, + { + "epoch": 0.08822398277055611, + "learning_rate": 0.002942753051034559, + "loss": 1.0856, + "step": 1147 + }, + { + "epoch": 0.08830089993077456, + "learning_rate": 0.0029426538284357297, + "loss": 1.2527, + "step": 1148 + }, + { + "epoch": 0.088377817090993, + "learning_rate": 0.00294255452159884, + "loss": 1.212, + "step": 1149 + }, + { + "epoch": 0.08845473425121145, + "learning_rate": 0.0029424551305296887, + "loss": 0.6959, + "step": 1150 + }, + { + "epoch": 0.08853165141142989, + "learning_rate": 0.0029423556552340786, + "loss": 1.2818, + "step": 1151 + }, + { + "epoch": 0.08860856857164834, + "learning_rate": 0.0029422560957178185, + "loss": 0.7935, + "step": 1152 + }, + { + "epoch": 0.08868548573186678, + "learning_rate": 0.0029421564519867224, + "loss": 1.2913, + "step": 1153 + }, + { + "epoch": 0.08876240289208523, + "learning_rate": 0.0029420567240466076, + "loss": 0.7006, + "step": 1154 + }, + { + "epoch": 0.08883932005230367, + "learning_rate": 0.002941956911903298, + "loss": 1.3979, + "step": 1155 + }, + { + "epoch": 0.08891623721252212, + "learning_rate": 0.0029418570155626215, + "loss": 0.8993, + "step": 1156 + }, + { + "epoch": 0.08899315437274055, + "learning_rate": 0.002941757035030412, + "loss": 1.0958, + "step": 1157 + }, + { + "epoch": 0.089070071532959, + "learning_rate": 0.002941656970312505, + "loss": 1.0649, + "step": 1158 + }, + { + "epoch": 0.08914698869317744, + "learning_rate": 0.0029415568214147467, + "loss": 1.014, + "step": 1159 + }, + { + "epoch": 0.08922390585339589, + "learning_rate": 0.002941456588342982, + "loss": 1.2312, + "step": 1160 + }, + { + "epoch": 0.08930082301361433, + "learning_rate": 0.0029413562711030658, + "loss": 1.2211, + "step": 1161 + }, + { + "epoch": 0.08937774017383278, + "learning_rate": 0.002941255869700854, + "loss": 1.1923, + "step": 1162 + }, + { + "epoch": 0.08945465733405122, + "learning_rate": 0.00294115538414221, + "loss": 1.2505, + "step": 1163 + }, + { + "epoch": 0.08953157449426967, + "learning_rate": 0.0029410548144330017, + "loss": 0.8847, + "step": 1164 + }, + { + "epoch": 0.08960849165448811, + "learning_rate": 0.0029409541605791008, + "loss": 1.242, + "step": 1165 + }, + { + "epoch": 0.08968540881470656, + "learning_rate": 0.0029408534225863845, + "loss": 1.1686, + "step": 1166 + }, + { + "epoch": 0.089762325974925, + "learning_rate": 0.002940752600460736, + "loss": 0.9645, + "step": 1167 + }, + { + "epoch": 0.08983924313514345, + "learning_rate": 0.0029406516942080403, + "loss": 1.215, + "step": 1168 + }, + { + "epoch": 0.0899161602953619, + "learning_rate": 0.0029405507038341916, + "loss": 1.117, + "step": 1169 + }, + { + "epoch": 0.08999307745558034, + "learning_rate": 0.002940449629345086, + "loss": 1.1072, + "step": 1170 + }, + { + "epoch": 0.09006999461579879, + "learning_rate": 0.002940348470746625, + "loss": 1.0377, + "step": 1171 + }, + { + "epoch": 0.09014691177601723, + "learning_rate": 0.002940247228044716, + "loss": 0.8495, + "step": 1172 + }, + { + "epoch": 0.09022382893623568, + "learning_rate": 0.00294014590124527, + "loss": 1.0416, + "step": 1173 + }, + { + "epoch": 0.09030074609645412, + "learning_rate": 0.002940044490354205, + "loss": 1.1154, + "step": 1174 + }, + { + "epoch": 0.09037766325667257, + "learning_rate": 0.0029399429953774403, + "loss": 1.1448, + "step": 1175 + }, + { + "epoch": 0.09045458041689101, + "learning_rate": 0.0029398414163209034, + "loss": 1.407, + "step": 1176 + }, + { + "epoch": 0.09053149757710946, + "learning_rate": 0.0029397397531905262, + "loss": 1.0101, + "step": 1177 + }, + { + "epoch": 0.09060841473732789, + "learning_rate": 0.002939638005992244, + "loss": 0.9946, + "step": 1178 + }, + { + "epoch": 0.09068533189754634, + "learning_rate": 0.002939536174731999, + "loss": 1.0516, + "step": 1179 + }, + { + "epoch": 0.09076224905776478, + "learning_rate": 0.0029394342594157357, + "loss": 1.7657, + "step": 1180 + }, + { + "epoch": 0.09083916621798323, + "learning_rate": 0.002939332260049406, + "loss": 1.4462, + "step": 1181 + }, + { + "epoch": 0.09091608337820167, + "learning_rate": 0.0029392301766389657, + "loss": 1.2188, + "step": 1182 + }, + { + "epoch": 0.09099300053842012, + "learning_rate": 0.0029391280091903755, + "loss": 0.9906, + "step": 1183 + }, + { + "epoch": 0.09106991769863856, + "learning_rate": 0.002939025757709601, + "loss": 1.3368, + "step": 1184 + }, + { + "epoch": 0.09114683485885701, + "learning_rate": 0.0029389234222026127, + "loss": 1.2898, + "step": 1185 + }, + { + "epoch": 0.09122375201907545, + "learning_rate": 0.0029388210026753863, + "loss": 1.1105, + "step": 1186 + }, + { + "epoch": 0.0913006691792939, + "learning_rate": 0.002938718499133902, + "loss": 1.0106, + "step": 1187 + }, + { + "epoch": 0.09137758633951235, + "learning_rate": 0.0029386159115841453, + "loss": 1.1443, + "step": 1188 + }, + { + "epoch": 0.09145450349973079, + "learning_rate": 0.0029385132400321055, + "loss": 1.2002, + "step": 1189 + }, + { + "epoch": 0.09153142065994924, + "learning_rate": 0.0029384104844837793, + "loss": 1.2615, + "step": 1190 + }, + { + "epoch": 0.09160833782016768, + "learning_rate": 0.0029383076449451657, + "loss": 1.1345, + "step": 1191 + }, + { + "epoch": 0.09168525498038613, + "learning_rate": 0.0029382047214222694, + "loss": 1.4055, + "step": 1192 + }, + { + "epoch": 0.09176217214060457, + "learning_rate": 0.0029381017139211004, + "loss": 1.4962, + "step": 1193 + }, + { + "epoch": 0.09183908930082302, + "learning_rate": 0.002937998622447674, + "loss": 1.1178, + "step": 1194 + }, + { + "epoch": 0.09191600646104146, + "learning_rate": 0.0029378954470080094, + "loss": 1.0983, + "step": 1195 + }, + { + "epoch": 0.09199292362125991, + "learning_rate": 0.0029377921876081307, + "loss": 0.9965, + "step": 1196 + }, + { + "epoch": 0.09206984078147835, + "learning_rate": 0.002937688844254068, + "loss": 1.1954, + "step": 1197 + }, + { + "epoch": 0.09214675794169679, + "learning_rate": 0.002937585416951856, + "loss": 1.6167, + "step": 1198 + }, + { + "epoch": 0.09222367510191523, + "learning_rate": 0.0029374819057075324, + "loss": 1.2936, + "step": 1199 + }, + { + "epoch": 0.09230059226213368, + "learning_rate": 0.0029373783105271427, + "loss": 0.9425, + "step": 1200 + }, + { + "epoch": 0.09237750942235212, + "learning_rate": 0.002937274631416735, + "loss": 1.5481, + "step": 1201 + }, + { + "epoch": 0.09245442658257057, + "learning_rate": 0.002937170868382364, + "loss": 1.1663, + "step": 1202 + }, + { + "epoch": 0.09253134374278901, + "learning_rate": 0.002937067021430088, + "loss": 1.0448, + "step": 1203 + }, + { + "epoch": 0.09260826090300746, + "learning_rate": 0.0029369630905659713, + "loss": 1.3691, + "step": 1204 + }, + { + "epoch": 0.0926851780632259, + "learning_rate": 0.002936859075796082, + "loss": 1.1063, + "step": 1205 + }, + { + "epoch": 0.09276209522344435, + "learning_rate": 0.0029367549771264936, + "loss": 1.2637, + "step": 1206 + }, + { + "epoch": 0.0928390123836628, + "learning_rate": 0.002936650794563285, + "loss": 1.298, + "step": 1207 + }, + { + "epoch": 0.09291592954388124, + "learning_rate": 0.0029365465281125393, + "loss": 1.0207, + "step": 1208 + }, + { + "epoch": 0.09299284670409969, + "learning_rate": 0.002936442177780344, + "loss": 0.8481, + "step": 1209 + }, + { + "epoch": 0.09306976386431813, + "learning_rate": 0.0029363377435727934, + "loss": 1.1319, + "step": 1210 + }, + { + "epoch": 0.09314668102453658, + "learning_rate": 0.0029362332254959855, + "loss": 1.1559, + "step": 1211 + }, + { + "epoch": 0.09322359818475502, + "learning_rate": 0.002936128623556022, + "loss": 1.5641, + "step": 1212 + }, + { + "epoch": 0.09330051534497347, + "learning_rate": 0.002936023937759012, + "loss": 1.5606, + "step": 1213 + }, + { + "epoch": 0.09337743250519191, + "learning_rate": 0.0029359191681110675, + "loss": 1.3077, + "step": 1214 + }, + { + "epoch": 0.09345434966541036, + "learning_rate": 0.0029358143146183064, + "loss": 1.3394, + "step": 1215 + }, + { + "epoch": 0.0935312668256288, + "learning_rate": 0.002935709377286851, + "loss": 1.0419, + "step": 1216 + }, + { + "epoch": 0.09360818398584725, + "learning_rate": 0.002935604356122829, + "loss": 1.3052, + "step": 1217 + }, + { + "epoch": 0.0936851011460657, + "learning_rate": 0.002935499251132372, + "loss": 1.0781, + "step": 1218 + }, + { + "epoch": 0.09376201830628413, + "learning_rate": 0.0029353940623216184, + "loss": 1.1302, + "step": 1219 + }, + { + "epoch": 0.09383893546650257, + "learning_rate": 0.002935288789696709, + "loss": 1.2968, + "step": 1220 + }, + { + "epoch": 0.09391585262672102, + "learning_rate": 0.0029351834332637912, + "loss": 1.3628, + "step": 1221 + }, + { + "epoch": 0.09399276978693946, + "learning_rate": 0.002935077993029018, + "loss": 1.2263, + "step": 1222 + }, + { + "epoch": 0.09406968694715791, + "learning_rate": 0.002934972468998544, + "loss": 1.4475, + "step": 1223 + }, + { + "epoch": 0.09414660410737635, + "learning_rate": 0.0029348668611785325, + "loss": 1.2735, + "step": 1224 + }, + { + "epoch": 0.0942235212675948, + "learning_rate": 0.0029347611695751496, + "loss": 1.124, + "step": 1225 + }, + { + "epoch": 0.09430043842781324, + "learning_rate": 0.0029346553941945667, + "loss": 1.5155, + "step": 1226 + }, + { + "epoch": 0.09437735558803169, + "learning_rate": 0.0029345495350429603, + "loss": 1.2345, + "step": 1227 + }, + { + "epoch": 0.09445427274825013, + "learning_rate": 0.002934443592126512, + "loss": 1.0955, + "step": 1228 + }, + { + "epoch": 0.09453118990846858, + "learning_rate": 0.002934337565451406, + "loss": 1.3493, + "step": 1229 + }, + { + "epoch": 0.09460810706868703, + "learning_rate": 0.002934231455023836, + "loss": 1.1968, + "step": 1230 + }, + { + "epoch": 0.09468502422890547, + "learning_rate": 0.0029341252608499962, + "loss": 1.3625, + "step": 1231 + }, + { + "epoch": 0.09476194138912392, + "learning_rate": 0.0029340189829360876, + "loss": 1.2965, + "step": 1232 + }, + { + "epoch": 0.09483885854934236, + "learning_rate": 0.0029339126212883167, + "loss": 1.1667, + "step": 1233 + }, + { + "epoch": 0.0949157757095608, + "learning_rate": 0.002933806175912893, + "loss": 1.4469, + "step": 1234 + }, + { + "epoch": 0.09499269286977925, + "learning_rate": 0.0029336996468160323, + "loss": 0.9501, + "step": 1235 + }, + { + "epoch": 0.0950696100299977, + "learning_rate": 0.0029335930340039554, + "loss": 1.132, + "step": 1236 + }, + { + "epoch": 0.09514652719021614, + "learning_rate": 0.002933486337482887, + "loss": 1.2609, + "step": 1237 + }, + { + "epoch": 0.09522344435043459, + "learning_rate": 0.0029333795572590573, + "loss": 1.0248, + "step": 1238 + }, + { + "epoch": 0.09530036151065302, + "learning_rate": 0.0029332726933387018, + "loss": 1.0056, + "step": 1239 + }, + { + "epoch": 0.09537727867087147, + "learning_rate": 0.0029331657457280593, + "loss": 1.3739, + "step": 1240 + }, + { + "epoch": 0.09545419583108991, + "learning_rate": 0.002933058714433376, + "loss": 1.1242, + "step": 1241 + }, + { + "epoch": 0.09553111299130836, + "learning_rate": 0.0029329515994609007, + "loss": 1.238, + "step": 1242 + }, + { + "epoch": 0.0956080301515268, + "learning_rate": 0.002932844400816888, + "loss": 1.7216, + "step": 1243 + }, + { + "epoch": 0.09568494731174525, + "learning_rate": 0.0029327371185075973, + "loss": 1.3232, + "step": 1244 + }, + { + "epoch": 0.09576186447196369, + "learning_rate": 0.0029326297525392934, + "loss": 0.8693, + "step": 1245 + }, + { + "epoch": 0.09583878163218214, + "learning_rate": 0.002932522302918245, + "loss": 1.1943, + "step": 1246 + }, + { + "epoch": 0.09591569879240058, + "learning_rate": 0.0029324147696507268, + "loss": 1.289, + "step": 1247 + }, + { + "epoch": 0.09599261595261903, + "learning_rate": 0.002932307152743017, + "loss": 0.98, + "step": 1248 + }, + { + "epoch": 0.09606953311283747, + "learning_rate": 0.0029321994522013995, + "loss": 1.4621, + "step": 1249 + }, + { + "epoch": 0.09614645027305592, + "learning_rate": 0.002932091668032164, + "loss": 1.1523, + "step": 1250 + }, + { + "epoch": 0.09622336743327436, + "learning_rate": 0.0029319838002416034, + "loss": 1.1686, + "step": 1251 + }, + { + "epoch": 0.09630028459349281, + "learning_rate": 0.0029318758488360163, + "loss": 1.2622, + "step": 1252 + }, + { + "epoch": 0.09637720175371126, + "learning_rate": 0.002931767813821706, + "loss": 1.2453, + "step": 1253 + }, + { + "epoch": 0.0964541189139297, + "learning_rate": 0.002931659695204981, + "loss": 1.4202, + "step": 1254 + }, + { + "epoch": 0.09653103607414815, + "learning_rate": 0.002931551492992154, + "loss": 1.1503, + "step": 1255 + }, + { + "epoch": 0.09660795323436659, + "learning_rate": 0.002931443207189544, + "loss": 1.1302, + "step": 1256 + }, + { + "epoch": 0.09668487039458504, + "learning_rate": 0.002931334837803473, + "loss": 1.2203, + "step": 1257 + }, + { + "epoch": 0.09676178755480348, + "learning_rate": 0.002931226384840269, + "loss": 1.2188, + "step": 1258 + }, + { + "epoch": 0.09683870471502193, + "learning_rate": 0.002931117848306265, + "loss": 1.0731, + "step": 1259 + }, + { + "epoch": 0.09691562187524036, + "learning_rate": 0.002931009228207798, + "loss": 1.2517, + "step": 1260 + }, + { + "epoch": 0.0969925390354588, + "learning_rate": 0.0029309005245512117, + "loss": 1.1722, + "step": 1261 + }, + { + "epoch": 0.09706945619567725, + "learning_rate": 0.0029307917373428516, + "loss": 1.2239, + "step": 1262 + }, + { + "epoch": 0.0971463733558957, + "learning_rate": 0.002930682866589071, + "loss": 0.8266, + "step": 1263 + }, + { + "epoch": 0.09722329051611414, + "learning_rate": 0.002930573912296227, + "loss": 1.2538, + "step": 1264 + }, + { + "epoch": 0.09730020767633259, + "learning_rate": 0.002930464874470681, + "loss": 1.2797, + "step": 1265 + }, + { + "epoch": 0.09737712483655103, + "learning_rate": 0.0029303557531188008, + "loss": 1.3732, + "step": 1266 + }, + { + "epoch": 0.09745404199676948, + "learning_rate": 0.0029302465482469574, + "loss": 1.1501, + "step": 1267 + }, + { + "epoch": 0.09753095915698792, + "learning_rate": 0.0029301372598615275, + "loss": 1.3214, + "step": 1268 + }, + { + "epoch": 0.09760787631720637, + "learning_rate": 0.002930027887968892, + "loss": 0.9647, + "step": 1269 + }, + { + "epoch": 0.09768479347742481, + "learning_rate": 0.002929918432575438, + "loss": 1.1907, + "step": 1270 + }, + { + "epoch": 0.09776171063764326, + "learning_rate": 0.002929808893687556, + "loss": 0.8753, + "step": 1271 + }, + { + "epoch": 0.0978386277978617, + "learning_rate": 0.0029296992713116433, + "loss": 1.0652, + "step": 1272 + }, + { + "epoch": 0.09791554495808015, + "learning_rate": 0.0029295895654541004, + "loss": 0.9914, + "step": 1273 + }, + { + "epoch": 0.0979924621182986, + "learning_rate": 0.0029294797761213326, + "loss": 1.0565, + "step": 1274 + }, + { + "epoch": 0.09806937927851704, + "learning_rate": 0.0029293699033197504, + "loss": 1.1092, + "step": 1275 + }, + { + "epoch": 0.09814629643873549, + "learning_rate": 0.0029292599470557706, + "loss": 1.1121, + "step": 1276 + }, + { + "epoch": 0.09822321359895393, + "learning_rate": 0.0029291499073358126, + "loss": 1.0894, + "step": 1277 + }, + { + "epoch": 0.09830013075917238, + "learning_rate": 0.002929039784166302, + "loss": 0.7796, + "step": 1278 + }, + { + "epoch": 0.09837704791939082, + "learning_rate": 0.0029289295775536696, + "loss": 1.0671, + "step": 1279 + }, + { + "epoch": 0.09845396507960925, + "learning_rate": 0.0029288192875043494, + "loss": 0.8831, + "step": 1280 + }, + { + "epoch": 0.0985308822398277, + "learning_rate": 0.0029287089140247824, + "loss": 0.9901, + "step": 1281 + }, + { + "epoch": 0.09860779940004614, + "learning_rate": 0.0029285984571214125, + "loss": 1.1922, + "step": 1282 + }, + { + "epoch": 0.09868471656026459, + "learning_rate": 0.0029284879168006896, + "loss": 1.2796, + "step": 1283 + }, + { + "epoch": 0.09876163372048304, + "learning_rate": 0.0029283772930690692, + "loss": 0.9585, + "step": 1284 + }, + { + "epoch": 0.09883855088070148, + "learning_rate": 0.0029282665859330097, + "loss": 1.2213, + "step": 1285 + }, + { + "epoch": 0.09891546804091993, + "learning_rate": 0.002928155795398976, + "loss": 1.2803, + "step": 1286 + }, + { + "epoch": 0.09899238520113837, + "learning_rate": 0.0029280449214734365, + "loss": 1.4555, + "step": 1287 + }, + { + "epoch": 0.09906930236135682, + "learning_rate": 0.002927933964162866, + "loss": 1.3525, + "step": 1288 + }, + { + "epoch": 0.09914621952157526, + "learning_rate": 0.0029278229234737433, + "loss": 1.4479, + "step": 1289 + }, + { + "epoch": 0.09922313668179371, + "learning_rate": 0.002927711799412552, + "loss": 0.7878, + "step": 1290 + }, + { + "epoch": 0.09930005384201215, + "learning_rate": 0.002927600591985781, + "loss": 0.7441, + "step": 1291 + }, + { + "epoch": 0.0993769710022306, + "learning_rate": 0.0029274893011999236, + "loss": 1.5054, + "step": 1292 + }, + { + "epoch": 0.09945388816244904, + "learning_rate": 0.0029273779270614776, + "loss": 1.2258, + "step": 1293 + }, + { + "epoch": 0.09953080532266749, + "learning_rate": 0.002927266469576947, + "loss": 1.2355, + "step": 1294 + }, + { + "epoch": 0.09960772248288594, + "learning_rate": 0.0029271549287528406, + "loss": 0.9092, + "step": 1295 + }, + { + "epoch": 0.09968463964310438, + "learning_rate": 0.00292704330459567, + "loss": 1.1802, + "step": 1296 + }, + { + "epoch": 0.09976155680332283, + "learning_rate": 0.002926931597111954, + "loss": 1.2831, + "step": 1297 + }, + { + "epoch": 0.09983847396354127, + "learning_rate": 0.0029268198063082144, + "loss": 1.1867, + "step": 1298 + }, + { + "epoch": 0.09991539112375972, + "learning_rate": 0.0029267079321909795, + "loss": 1.1651, + "step": 1299 + }, + { + "epoch": 0.09999230828397816, + "learning_rate": 0.0029265959747667813, + "loss": 1.1498, + "step": 1300 + }, + { + "epoch": 0.1000692254441966, + "learning_rate": 0.002926483934042158, + "loss": 1.1721, + "step": 1301 + }, + { + "epoch": 0.10014614260441504, + "learning_rate": 0.0029263718100236507, + "loss": 1.0412, + "step": 1302 + }, + { + "epoch": 0.10022305976463348, + "learning_rate": 0.0029262596027178067, + "loss": 1.5262, + "step": 1303 + }, + { + "epoch": 0.10029997692485193, + "learning_rate": 0.0029261473121311786, + "loss": 1.2988, + "step": 1304 + }, + { + "epoch": 0.10037689408507038, + "learning_rate": 0.002926034938270322, + "loss": 0.9277, + "step": 1305 + }, + { + "epoch": 0.10045381124528882, + "learning_rate": 0.0029259224811418, + "loss": 1.4692, + "step": 1306 + }, + { + "epoch": 0.10053072840550727, + "learning_rate": 0.0029258099407521777, + "loss": 1.1353, + "step": 1307 + }, + { + "epoch": 0.10060764556572571, + "learning_rate": 0.0029256973171080274, + "loss": 1.0995, + "step": 1308 + }, + { + "epoch": 0.10068456272594416, + "learning_rate": 0.0029255846102159244, + "loss": 1.1535, + "step": 1309 + }, + { + "epoch": 0.1007614798861626, + "learning_rate": 0.002925471820082451, + "loss": 1.4002, + "step": 1310 + }, + { + "epoch": 0.10083839704638105, + "learning_rate": 0.0029253589467141917, + "loss": 1.2245, + "step": 1311 + }, + { + "epoch": 0.1009153142065995, + "learning_rate": 0.0029252459901177386, + "loss": 1.2784, + "step": 1312 + }, + { + "epoch": 0.10099223136681794, + "learning_rate": 0.0029251329502996866, + "loss": 1.3721, + "step": 1313 + }, + { + "epoch": 0.10106914852703638, + "learning_rate": 0.0029250198272666363, + "loss": 1.288, + "step": 1314 + }, + { + "epoch": 0.10114606568725483, + "learning_rate": 0.002924906621025193, + "loss": 1.091, + "step": 1315 + }, + { + "epoch": 0.10122298284747328, + "learning_rate": 0.0029247933315819674, + "loss": 0.844, + "step": 1316 + }, + { + "epoch": 0.10129990000769172, + "learning_rate": 0.0029246799589435738, + "loss": 0.9985, + "step": 1317 + }, + { + "epoch": 0.10137681716791017, + "learning_rate": 0.0029245665031166334, + "loss": 1.2714, + "step": 1318 + }, + { + "epoch": 0.10145373432812861, + "learning_rate": 0.00292445296410777, + "loss": 1.3008, + "step": 1319 + }, + { + "epoch": 0.10153065148834706, + "learning_rate": 0.0029243393419236133, + "loss": 1.233, + "step": 1320 + }, + { + "epoch": 0.10160756864856549, + "learning_rate": 0.002924225636570798, + "loss": 0.8218, + "step": 1321 + }, + { + "epoch": 0.10168448580878393, + "learning_rate": 0.0029241118480559636, + "loss": 1.3125, + "step": 1322 + }, + { + "epoch": 0.10176140296900238, + "learning_rate": 0.002923997976385754, + "loss": 1.3914, + "step": 1323 + }, + { + "epoch": 0.10183832012922082, + "learning_rate": 0.0029238840215668185, + "loss": 1.1458, + "step": 1324 + }, + { + "epoch": 0.10191523728943927, + "learning_rate": 0.0029237699836058115, + "loss": 1.2458, + "step": 1325 + }, + { + "epoch": 0.10199215444965772, + "learning_rate": 0.0029236558625093906, + "loss": 1.1755, + "step": 1326 + }, + { + "epoch": 0.10206907160987616, + "learning_rate": 0.002923541658284221, + "loss": 1.1569, + "step": 1327 + }, + { + "epoch": 0.1021459887700946, + "learning_rate": 0.00292342737093697, + "loss": 1.127, + "step": 1328 + }, + { + "epoch": 0.10222290593031305, + "learning_rate": 0.0029233130004743114, + "loss": 1.0872, + "step": 1329 + }, + { + "epoch": 0.1022998230905315, + "learning_rate": 0.0029231985469029232, + "loss": 1.2195, + "step": 1330 + }, + { + "epoch": 0.10237674025074994, + "learning_rate": 0.002923084010229489, + "loss": 1.3099, + "step": 1331 + }, + { + "epoch": 0.10245365741096839, + "learning_rate": 0.002922969390460696, + "loss": 1.4398, + "step": 1332 + }, + { + "epoch": 0.10253057457118683, + "learning_rate": 0.0029228546876032373, + "loss": 0.9453, + "step": 1333 + }, + { + "epoch": 0.10260749173140528, + "learning_rate": 0.002922739901663811, + "loss": 1.2617, + "step": 1334 + }, + { + "epoch": 0.10268440889162372, + "learning_rate": 0.0029226250326491185, + "loss": 1.1086, + "step": 1335 + }, + { + "epoch": 0.10276132605184217, + "learning_rate": 0.002922510080565868, + "loss": 1.055, + "step": 1336 + }, + { + "epoch": 0.10283824321206061, + "learning_rate": 0.0029223950454207713, + "loss": 1.0562, + "step": 1337 + }, + { + "epoch": 0.10291516037227906, + "learning_rate": 0.002922279927220546, + "loss": 0.9574, + "step": 1338 + }, + { + "epoch": 0.1029920775324975, + "learning_rate": 0.002922164725971913, + "loss": 0.9332, + "step": 1339 + }, + { + "epoch": 0.10306899469271595, + "learning_rate": 0.0029220494416815996, + "loss": 1.2528, + "step": 1340 + }, + { + "epoch": 0.1031459118529344, + "learning_rate": 0.002921934074356337, + "loss": 1.23, + "step": 1341 + }, + { + "epoch": 0.10322282901315283, + "learning_rate": 0.002921818624002862, + "loss": 1.1342, + "step": 1342 + }, + { + "epoch": 0.10329974617337127, + "learning_rate": 0.002921703090627916, + "loss": 1.157, + "step": 1343 + }, + { + "epoch": 0.10337666333358972, + "learning_rate": 0.002921587474238245, + "loss": 0.9476, + "step": 1344 + }, + { + "epoch": 0.10345358049380816, + "learning_rate": 0.0029214717748405994, + "loss": 1.3534, + "step": 1345 + }, + { + "epoch": 0.10353049765402661, + "learning_rate": 0.0029213559924417356, + "loss": 0.888, + "step": 1346 + }, + { + "epoch": 0.10360741481424506, + "learning_rate": 0.002921240127048414, + "loss": 0.6896, + "step": 1347 + }, + { + "epoch": 0.1036843319744635, + "learning_rate": 0.0029211241786674005, + "loss": 1.3125, + "step": 1348 + }, + { + "epoch": 0.10376124913468195, + "learning_rate": 0.0029210081473054652, + "loss": 1.1898, + "step": 1349 + }, + { + "epoch": 0.10383816629490039, + "learning_rate": 0.002920892032969383, + "loss": 1.3149, + "step": 1350 + }, + { + "epoch": 0.10391508345511884, + "learning_rate": 0.0029207758356659336, + "loss": 1.0003, + "step": 1351 + }, + { + "epoch": 0.10399200061533728, + "learning_rate": 0.0029206595554019037, + "loss": 1.3983, + "step": 1352 + }, + { + "epoch": 0.10406891777555573, + "learning_rate": 0.002920543192184081, + "loss": 1.2183, + "step": 1353 + }, + { + "epoch": 0.10414583493577417, + "learning_rate": 0.002920426746019261, + "loss": 0.9984, + "step": 1354 + }, + { + "epoch": 0.10422275209599262, + "learning_rate": 0.002920310216914243, + "loss": 1.2361, + "step": 1355 + }, + { + "epoch": 0.10429966925621106, + "learning_rate": 0.002920193604875831, + "loss": 1.258, + "step": 1356 + }, + { + "epoch": 0.10437658641642951, + "learning_rate": 0.0029200769099108344, + "loss": 0.6998, + "step": 1357 + }, + { + "epoch": 0.10445350357664795, + "learning_rate": 0.0029199601320260673, + "loss": 1.3332, + "step": 1358 + }, + { + "epoch": 0.1045304207368664, + "learning_rate": 0.002919843271228348, + "loss": 1.24, + "step": 1359 + }, + { + "epoch": 0.10460733789708485, + "learning_rate": 0.0029197263275245007, + "loss": 1.3677, + "step": 1360 + }, + { + "epoch": 0.10468425505730329, + "learning_rate": 0.0029196093009213535, + "loss": 1.3422, + "step": 1361 + }, + { + "epoch": 0.10476117221752172, + "learning_rate": 0.0029194921914257395, + "loss": 0.9152, + "step": 1362 + }, + { + "epoch": 0.10483808937774017, + "learning_rate": 0.002919374999044497, + "loss": 1.4853, + "step": 1363 + }, + { + "epoch": 0.10491500653795861, + "learning_rate": 0.002919257723784469, + "loss": 0.9956, + "step": 1364 + }, + { + "epoch": 0.10499192369817706, + "learning_rate": 0.0029191403656525037, + "loss": 1.1749, + "step": 1365 + }, + { + "epoch": 0.1050688408583955, + "learning_rate": 0.0029190229246554536, + "loss": 1.02, + "step": 1366 + }, + { + "epoch": 0.10514575801861395, + "learning_rate": 0.002918905400800176, + "loss": 1.0859, + "step": 1367 + }, + { + "epoch": 0.1052226751788324, + "learning_rate": 0.0029187877940935326, + "loss": 1.2528, + "step": 1368 + }, + { + "epoch": 0.10529959233905084, + "learning_rate": 0.002918670104542392, + "loss": 1.1351, + "step": 1369 + }, + { + "epoch": 0.10537650949926929, + "learning_rate": 0.0029185523321536254, + "loss": 0.9711, + "step": 1370 + }, + { + "epoch": 0.10545342665948773, + "learning_rate": 0.0029184344769341095, + "loss": 1.2338, + "step": 1371 + }, + { + "epoch": 0.10553034381970618, + "learning_rate": 0.0029183165388907265, + "loss": 0.9078, + "step": 1372 + }, + { + "epoch": 0.10560726097992462, + "learning_rate": 0.0029181985180303624, + "loss": 0.9939, + "step": 1373 + }, + { + "epoch": 0.10568417814014307, + "learning_rate": 0.002918080414359909, + "loss": 0.8116, + "step": 1374 + }, + { + "epoch": 0.10576109530036151, + "learning_rate": 0.0029179622278862625, + "loss": 1.3753, + "step": 1375 + }, + { + "epoch": 0.10583801246057996, + "learning_rate": 0.0029178439586163233, + "loss": 1.0247, + "step": 1376 + }, + { + "epoch": 0.1059149296207984, + "learning_rate": 0.002917725606556998, + "loss": 1.6761, + "step": 1377 + }, + { + "epoch": 0.10599184678101685, + "learning_rate": 0.0029176071717151968, + "loss": 0.9708, + "step": 1378 + }, + { + "epoch": 0.1060687639412353, + "learning_rate": 0.0029174886540978356, + "loss": 1.2245, + "step": 1379 + }, + { + "epoch": 0.10614568110145374, + "learning_rate": 0.0029173700537118348, + "loss": 1.2728, + "step": 1380 + }, + { + "epoch": 0.10622259826167219, + "learning_rate": 0.002917251370564119, + "loss": 1.1181, + "step": 1381 + }, + { + "epoch": 0.10629951542189062, + "learning_rate": 0.002917132604661619, + "loss": 0.8526, + "step": 1382 + }, + { + "epoch": 0.10637643258210906, + "learning_rate": 0.002917013756011269, + "loss": 1.185, + "step": 1383 + }, + { + "epoch": 0.10645334974232751, + "learning_rate": 0.002916894824620009, + "loss": 1.3072, + "step": 1384 + }, + { + "epoch": 0.10653026690254595, + "learning_rate": 0.0029167758104947845, + "loss": 1.5782, + "step": 1385 + }, + { + "epoch": 0.1066071840627644, + "learning_rate": 0.0029166567136425427, + "loss": 1.4001, + "step": 1386 + }, + { + "epoch": 0.10668410122298284, + "learning_rate": 0.0029165375340702393, + "loss": 0.7133, + "step": 1387 + }, + { + "epoch": 0.10676101838320129, + "learning_rate": 0.002916418271784833, + "loss": 1.149, + "step": 1388 + }, + { + "epoch": 0.10683793554341973, + "learning_rate": 0.0029162989267932883, + "loss": 1.1076, + "step": 1389 + }, + { + "epoch": 0.10691485270363818, + "learning_rate": 0.0029161794991025723, + "loss": 1.0589, + "step": 1390 + }, + { + "epoch": 0.10699176986385663, + "learning_rate": 0.00291605998871966, + "loss": 0.9592, + "step": 1391 + }, + { + "epoch": 0.10706868702407507, + "learning_rate": 0.002915940395651529, + "loss": 1.1013, + "step": 1392 + }, + { + "epoch": 0.10714560418429352, + "learning_rate": 0.002915820719905163, + "loss": 1.2462, + "step": 1393 + }, + { + "epoch": 0.10722252134451196, + "learning_rate": 0.002915700961487549, + "loss": 0.9328, + "step": 1394 + }, + { + "epoch": 0.10729943850473041, + "learning_rate": 0.0029155811204056806, + "loss": 1.3122, + "step": 1395 + }, + { + "epoch": 0.10737635566494885, + "learning_rate": 0.0029154611966665557, + "loss": 1.1443, + "step": 1396 + }, + { + "epoch": 0.1074532728251673, + "learning_rate": 0.002915341190277176, + "loss": 0.9814, + "step": 1397 + }, + { + "epoch": 0.10753018998538574, + "learning_rate": 0.0029152211012445496, + "loss": 1.0536, + "step": 1398 + }, + { + "epoch": 0.10760710714560419, + "learning_rate": 0.002915100929575688, + "loss": 1.1837, + "step": 1399 + }, + { + "epoch": 0.10768402430582263, + "learning_rate": 0.002914980675277608, + "loss": 1.275, + "step": 1400 + }, + { + "epoch": 0.10776094146604108, + "learning_rate": 0.002914860338357332, + "loss": 1.1234, + "step": 1401 + }, + { + "epoch": 0.10783785862625953, + "learning_rate": 0.002914739918821886, + "loss": 1.1729, + "step": 1402 + }, + { + "epoch": 0.10791477578647796, + "learning_rate": 0.002914619416678302, + "loss": 0.9903, + "step": 1403 + }, + { + "epoch": 0.1079916929466964, + "learning_rate": 0.002914498831933616, + "loss": 1.1748, + "step": 1404 + }, + { + "epoch": 0.10806861010691485, + "learning_rate": 0.002914378164594869, + "loss": 1.4379, + "step": 1405 + }, + { + "epoch": 0.10814552726713329, + "learning_rate": 0.002914257414669107, + "loss": 0.8967, + "step": 1406 + }, + { + "epoch": 0.10822244442735174, + "learning_rate": 0.0029141365821633808, + "loss": 0.9649, + "step": 1407 + }, + { + "epoch": 0.10829936158757018, + "learning_rate": 0.0029140156670847454, + "loss": 1.4865, + "step": 1408 + }, + { + "epoch": 0.10837627874778863, + "learning_rate": 0.0029138946694402617, + "loss": 0.5231, + "step": 1409 + }, + { + "epoch": 0.10845319590800707, + "learning_rate": 0.002913773589236995, + "loss": 1.3454, + "step": 1410 + }, + { + "epoch": 0.10853011306822552, + "learning_rate": 0.0029136524264820147, + "loss": 1.2543, + "step": 1411 + }, + { + "epoch": 0.10860703022844397, + "learning_rate": 0.0029135311811823955, + "loss": 1.2049, + "step": 1412 + }, + { + "epoch": 0.10868394738866241, + "learning_rate": 0.0029134098533452183, + "loss": 1.1168, + "step": 1413 + }, + { + "epoch": 0.10876086454888086, + "learning_rate": 0.0029132884429775665, + "loss": 0.8141, + "step": 1414 + }, + { + "epoch": 0.1088377817090993, + "learning_rate": 0.00291316695008653, + "loss": 1.2728, + "step": 1415 + }, + { + "epoch": 0.10891469886931775, + "learning_rate": 0.0029130453746792017, + "loss": 0.832, + "step": 1416 + }, + { + "epoch": 0.10899161602953619, + "learning_rate": 0.0029129237167626817, + "loss": 1.3316, + "step": 1417 + }, + { + "epoch": 0.10906853318975464, + "learning_rate": 0.0029128019763440733, + "loss": 1.3856, + "step": 1418 + }, + { + "epoch": 0.10914545034997308, + "learning_rate": 0.0029126801534304853, + "loss": 1.1343, + "step": 1419 + }, + { + "epoch": 0.10922236751019153, + "learning_rate": 0.002912558248029031, + "loss": 0.9837, + "step": 1420 + }, + { + "epoch": 0.10929928467040997, + "learning_rate": 0.002912436260146828, + "loss": 1.0679, + "step": 1421 + }, + { + "epoch": 0.10937620183062842, + "learning_rate": 0.002912314189791, + "loss": 1.2493, + "step": 1422 + }, + { + "epoch": 0.10945311899084685, + "learning_rate": 0.0029121920369686747, + "loss": 0.9199, + "step": 1423 + }, + { + "epoch": 0.1095300361510653, + "learning_rate": 0.0029120698016869844, + "loss": 1.4005, + "step": 1424 + }, + { + "epoch": 0.10960695331128374, + "learning_rate": 0.002911947483953067, + "loss": 1.2883, + "step": 1425 + }, + { + "epoch": 0.10968387047150219, + "learning_rate": 0.0029118250837740643, + "loss": 1.1278, + "step": 1426 + }, + { + "epoch": 0.10976078763172063, + "learning_rate": 0.0029117026011571236, + "loss": 1.1546, + "step": 1427 + }, + { + "epoch": 0.10983770479193908, + "learning_rate": 0.002911580036109397, + "loss": 1.0478, + "step": 1428 + }, + { + "epoch": 0.10991462195215752, + "learning_rate": 0.0029114573886380407, + "loss": 1.1066, + "step": 1429 + }, + { + "epoch": 0.10999153911237597, + "learning_rate": 0.0029113346587502165, + "loss": 1.1094, + "step": 1430 + }, + { + "epoch": 0.11006845627259441, + "learning_rate": 0.002911211846453091, + "loss": 0.882, + "step": 1431 + }, + { + "epoch": 0.11014537343281286, + "learning_rate": 0.0029110889517538346, + "loss": 1.0255, + "step": 1432 + }, + { + "epoch": 0.1102222905930313, + "learning_rate": 0.002910965974659624, + "loss": 1.2265, + "step": 1433 + }, + { + "epoch": 0.11029920775324975, + "learning_rate": 0.00291084291517764, + "loss": 1.204, + "step": 1434 + }, + { + "epoch": 0.1103761249134682, + "learning_rate": 0.0029107197733150673, + "loss": 1.0782, + "step": 1435 + }, + { + "epoch": 0.11045304207368664, + "learning_rate": 0.002910596549079097, + "loss": 0.9163, + "step": 1436 + }, + { + "epoch": 0.11052995923390509, + "learning_rate": 0.002910473242476924, + "loss": 1.0607, + "step": 1437 + }, + { + "epoch": 0.11060687639412353, + "learning_rate": 0.002910349853515749, + "loss": 1.1778, + "step": 1438 + }, + { + "epoch": 0.11068379355434198, + "learning_rate": 0.0029102263822027756, + "loss": 1.3244, + "step": 1439 + }, + { + "epoch": 0.11076071071456042, + "learning_rate": 0.002910102828545214, + "loss": 1.2257, + "step": 1440 + }, + { + "epoch": 0.11083762787477887, + "learning_rate": 0.0029099791925502787, + "loss": 1.2334, + "step": 1441 + }, + { + "epoch": 0.11091454503499731, + "learning_rate": 0.0029098554742251893, + "loss": 1.2848, + "step": 1442 + }, + { + "epoch": 0.11099146219521576, + "learning_rate": 0.002909731673577169, + "loss": 0.6049, + "step": 1443 + }, + { + "epoch": 0.11106837935543419, + "learning_rate": 0.002909607790613448, + "loss": 0.5912, + "step": 1444 + }, + { + "epoch": 0.11114529651565264, + "learning_rate": 0.002909483825341258, + "loss": 0.964, + "step": 1445 + }, + { + "epoch": 0.11122221367587108, + "learning_rate": 0.0029093597777678387, + "loss": 0.9893, + "step": 1446 + }, + { + "epoch": 0.11129913083608953, + "learning_rate": 0.0029092356479004334, + "loss": 1.3322, + "step": 1447 + }, + { + "epoch": 0.11137604799630797, + "learning_rate": 0.00290911143574629, + "loss": 1.5149, + "step": 1448 + }, + { + "epoch": 0.11145296515652642, + "learning_rate": 0.0029089871413126613, + "loss": 1.0509, + "step": 1449 + }, + { + "epoch": 0.11152988231674486, + "learning_rate": 0.002908862764606805, + "loss": 1.0904, + "step": 1450 + }, + { + "epoch": 0.11160679947696331, + "learning_rate": 0.0029087383056359832, + "loss": 1.0722, + "step": 1451 + }, + { + "epoch": 0.11168371663718175, + "learning_rate": 0.002908613764407464, + "loss": 1.4654, + "step": 1452 + }, + { + "epoch": 0.1117606337974002, + "learning_rate": 0.0029084891409285186, + "loss": 0.9479, + "step": 1453 + }, + { + "epoch": 0.11183755095761864, + "learning_rate": 0.002908364435206425, + "loss": 1.4954, + "step": 1454 + }, + { + "epoch": 0.11191446811783709, + "learning_rate": 0.002908239647248464, + "loss": 1.2428, + "step": 1455 + }, + { + "epoch": 0.11199138527805554, + "learning_rate": 0.0029081147770619225, + "loss": 0.7893, + "step": 1456 + }, + { + "epoch": 0.11206830243827398, + "learning_rate": 0.0029079898246540917, + "loss": 1.233, + "step": 1457 + }, + { + "epoch": 0.11214521959849243, + "learning_rate": 0.0029078647900322675, + "loss": 1.2769, + "step": 1458 + }, + { + "epoch": 0.11222213675871087, + "learning_rate": 0.002907739673203751, + "loss": 0.9771, + "step": 1459 + }, + { + "epoch": 0.11229905391892932, + "learning_rate": 0.0029076144741758477, + "loss": 1.0689, + "step": 1460 + }, + { + "epoch": 0.11237597107914776, + "learning_rate": 0.0029074891929558684, + "loss": 1.2342, + "step": 1461 + }, + { + "epoch": 0.11245288823936621, + "learning_rate": 0.0029073638295511284, + "loss": 1.099, + "step": 1462 + }, + { + "epoch": 0.11252980539958465, + "learning_rate": 0.0029072383839689477, + "loss": 1.1543, + "step": 1463 + }, + { + "epoch": 0.11260672255980309, + "learning_rate": 0.002907112856216651, + "loss": 1.1191, + "step": 1464 + }, + { + "epoch": 0.11268363972002153, + "learning_rate": 0.0029069872463015674, + "loss": 1.0351, + "step": 1465 + }, + { + "epoch": 0.11276055688023998, + "learning_rate": 0.002906861554231033, + "loss": 1.2029, + "step": 1466 + }, + { + "epoch": 0.11283747404045842, + "learning_rate": 0.0029067357800123864, + "loss": 1.2132, + "step": 1467 + }, + { + "epoch": 0.11291439120067687, + "learning_rate": 0.002906609923652972, + "loss": 0.8952, + "step": 1468 + }, + { + "epoch": 0.11299130836089531, + "learning_rate": 0.002906483985160137, + "loss": 1.1924, + "step": 1469 + }, + { + "epoch": 0.11306822552111376, + "learning_rate": 0.0029063579645412373, + "loss": 0.6819, + "step": 1470 + }, + { + "epoch": 0.1131451426813322, + "learning_rate": 0.0029062318618036298, + "loss": 1.1736, + "step": 1471 + }, + { + "epoch": 0.11322205984155065, + "learning_rate": 0.0029061056769546785, + "loss": 1.297, + "step": 1472 + }, + { + "epoch": 0.1132989770017691, + "learning_rate": 0.0029059794100017512, + "loss": 1.1583, + "step": 1473 + }, + { + "epoch": 0.11337589416198754, + "learning_rate": 0.002905853060952221, + "loss": 1.4155, + "step": 1474 + }, + { + "epoch": 0.11345281132220598, + "learning_rate": 0.0029057266298134657, + "loss": 1.332, + "step": 1475 + }, + { + "epoch": 0.11352972848242443, + "learning_rate": 0.0029056001165928672, + "loss": 1.2283, + "step": 1476 + }, + { + "epoch": 0.11360664564264288, + "learning_rate": 0.002905473521297813, + "loss": 0.999, + "step": 1477 + }, + { + "epoch": 0.11368356280286132, + "learning_rate": 0.002905346843935695, + "loss": 1.3447, + "step": 1478 + }, + { + "epoch": 0.11376047996307977, + "learning_rate": 0.0029052200845139102, + "loss": 1.4668, + "step": 1479 + }, + { + "epoch": 0.11383739712329821, + "learning_rate": 0.0029050932430398605, + "loss": 1.0701, + "step": 1480 + }, + { + "epoch": 0.11391431428351666, + "learning_rate": 0.0029049663195209515, + "loss": 1.2138, + "step": 1481 + }, + { + "epoch": 0.1139912314437351, + "learning_rate": 0.0029048393139645954, + "loss": 1.2256, + "step": 1482 + }, + { + "epoch": 0.11406814860395355, + "learning_rate": 0.0029047122263782072, + "loss": 1.254, + "step": 1483 + }, + { + "epoch": 0.114145065764172, + "learning_rate": 0.0029045850567692086, + "loss": 0.9742, + "step": 1484 + }, + { + "epoch": 0.11422198292439042, + "learning_rate": 0.0029044578051450245, + "loss": 1.1314, + "step": 1485 + }, + { + "epoch": 0.11429890008460887, + "learning_rate": 0.002904330471513085, + "loss": 0.9521, + "step": 1486 + }, + { + "epoch": 0.11437581724482732, + "learning_rate": 0.002904203055880826, + "loss": 1.2706, + "step": 1487 + }, + { + "epoch": 0.11445273440504576, + "learning_rate": 0.0029040755582556877, + "loss": 0.8659, + "step": 1488 + }, + { + "epoch": 0.1145296515652642, + "learning_rate": 0.002903947978645114, + "loss": 1.2421, + "step": 1489 + }, + { + "epoch": 0.11460656872548265, + "learning_rate": 0.0029038203170565543, + "loss": 1.1359, + "step": 1490 + }, + { + "epoch": 0.1146834858857011, + "learning_rate": 0.0029036925734974637, + "loss": 0.8697, + "step": 1491 + }, + { + "epoch": 0.11476040304591954, + "learning_rate": 0.0029035647479753006, + "loss": 1.4948, + "step": 1492 + }, + { + "epoch": 0.11483732020613799, + "learning_rate": 0.0029034368404975293, + "loss": 1.2606, + "step": 1493 + }, + { + "epoch": 0.11491423736635643, + "learning_rate": 0.002903308851071618, + "loss": 0.9511, + "step": 1494 + }, + { + "epoch": 0.11499115452657488, + "learning_rate": 0.0029031807797050405, + "loss": 1.1019, + "step": 1495 + }, + { + "epoch": 0.11506807168679332, + "learning_rate": 0.002903052626405275, + "loss": 1.3968, + "step": 1496 + }, + { + "epoch": 0.11514498884701177, + "learning_rate": 0.002902924391179804, + "loss": 0.9133, + "step": 1497 + }, + { + "epoch": 0.11522190600723022, + "learning_rate": 0.0029027960740361162, + "loss": 1.3496, + "step": 1498 + }, + { + "epoch": 0.11529882316744866, + "learning_rate": 0.002902667674981703, + "loss": 0.7661, + "step": 1499 + }, + { + "epoch": 0.1153757403276671, + "learning_rate": 0.0029025391940240633, + "loss": 1.001, + "step": 1500 + }, + { + "epoch": 0.11545265748788555, + "learning_rate": 0.0029024106311706984, + "loss": 1.0003, + "step": 1501 + }, + { + "epoch": 0.115529574648104, + "learning_rate": 0.0029022819864291147, + "loss": 1.336, + "step": 1502 + }, + { + "epoch": 0.11560649180832244, + "learning_rate": 0.0029021532598068243, + "loss": 1.4581, + "step": 1503 + }, + { + "epoch": 0.11568340896854089, + "learning_rate": 0.0029020244513113438, + "loss": 1.194, + "step": 1504 + }, + { + "epoch": 0.11576032612875932, + "learning_rate": 0.002901895560950195, + "loss": 0.9644, + "step": 1505 + }, + { + "epoch": 0.11583724328897776, + "learning_rate": 0.0029017665887309026, + "loss": 1.1786, + "step": 1506 + }, + { + "epoch": 0.11591416044919621, + "learning_rate": 0.002901637534660999, + "loss": 1.0125, + "step": 1507 + }, + { + "epoch": 0.11599107760941466, + "learning_rate": 0.0029015083987480182, + "loss": 1.2137, + "step": 1508 + }, + { + "epoch": 0.1160679947696331, + "learning_rate": 0.002901379180999502, + "loss": 1.1543, + "step": 1509 + }, + { + "epoch": 0.11614491192985155, + "learning_rate": 0.0029012498814229943, + "loss": 1.0397, + "step": 1510 + }, + { + "epoch": 0.11622182909006999, + "learning_rate": 0.0029011205000260463, + "loss": 1.5104, + "step": 1511 + }, + { + "epoch": 0.11629874625028844, + "learning_rate": 0.002900991036816212, + "loss": 1.2308, + "step": 1512 + }, + { + "epoch": 0.11637566341050688, + "learning_rate": 0.0029008614918010506, + "loss": 1.1723, + "step": 1513 + }, + { + "epoch": 0.11645258057072533, + "learning_rate": 0.0029007318649881273, + "loss": 1.0011, + "step": 1514 + }, + { + "epoch": 0.11652949773094377, + "learning_rate": 0.0029006021563850104, + "loss": 1.0574, + "step": 1515 + }, + { + "epoch": 0.11660641489116222, + "learning_rate": 0.0029004723659992733, + "loss": 1.2372, + "step": 1516 + }, + { + "epoch": 0.11668333205138066, + "learning_rate": 0.002900342493838496, + "loss": 1.1541, + "step": 1517 + }, + { + "epoch": 0.11676024921159911, + "learning_rate": 0.0029002125399102607, + "loss": 1.0897, + "step": 1518 + }, + { + "epoch": 0.11683716637181756, + "learning_rate": 0.002900082504222156, + "loss": 1.3226, + "step": 1519 + }, + { + "epoch": 0.116914083532036, + "learning_rate": 0.0028999523867817745, + "loss": 1.8294, + "step": 1520 + }, + { + "epoch": 0.11699100069225445, + "learning_rate": 0.002899822187596715, + "loss": 1.4568, + "step": 1521 + }, + { + "epoch": 0.11706791785247289, + "learning_rate": 0.002899691906674578, + "loss": 1.2607, + "step": 1522 + }, + { + "epoch": 0.11714483501269134, + "learning_rate": 0.0028995615440229722, + "loss": 1.2988, + "step": 1523 + }, + { + "epoch": 0.11722175217290978, + "learning_rate": 0.0028994310996495096, + "loss": 1.024, + "step": 1524 + }, + { + "epoch": 0.11729866933312823, + "learning_rate": 0.0028993005735618066, + "loss": 1.2166, + "step": 1525 + }, + { + "epoch": 0.11737558649334666, + "learning_rate": 0.0028991699657674847, + "loss": 1.3122, + "step": 1526 + }, + { + "epoch": 0.1174525036535651, + "learning_rate": 0.00289903927627417, + "loss": 1.1074, + "step": 1527 + }, + { + "epoch": 0.11752942081378355, + "learning_rate": 0.0028989085050894945, + "loss": 1.1349, + "step": 1528 + }, + { + "epoch": 0.117606337974002, + "learning_rate": 0.002898777652221093, + "loss": 0.9543, + "step": 1529 + }, + { + "epoch": 0.11768325513422044, + "learning_rate": 0.002898646717676607, + "loss": 1.1482, + "step": 1530 + }, + { + "epoch": 0.11776017229443889, + "learning_rate": 0.0028985157014636814, + "loss": 1.4885, + "step": 1531 + }, + { + "epoch": 0.11783708945465733, + "learning_rate": 0.002898384603589966, + "loss": 1.0584, + "step": 1532 + }, + { + "epoch": 0.11791400661487578, + "learning_rate": 0.0028982534240631166, + "loss": 1.3822, + "step": 1533 + }, + { + "epoch": 0.11799092377509422, + "learning_rate": 0.0028981221628907924, + "loss": 1.074, + "step": 1534 + }, + { + "epoch": 0.11806784093531267, + "learning_rate": 0.002897990820080659, + "loss": 1.5324, + "step": 1535 + }, + { + "epoch": 0.11814475809553111, + "learning_rate": 0.0028978593956403833, + "loss": 0.8464, + "step": 1536 + }, + { + "epoch": 0.11822167525574956, + "learning_rate": 0.0028977278895776413, + "loss": 1.0793, + "step": 1537 + }, + { + "epoch": 0.118298592415968, + "learning_rate": 0.002897596301900111, + "loss": 1.324, + "step": 1538 + }, + { + "epoch": 0.11837550957618645, + "learning_rate": 0.0028974646326154763, + "loss": 1.1548, + "step": 1539 + }, + { + "epoch": 0.1184524267364049, + "learning_rate": 0.002897332881731425, + "loss": 1.3605, + "step": 1540 + }, + { + "epoch": 0.11852934389662334, + "learning_rate": 0.0028972010492556506, + "loss": 1.192, + "step": 1541 + }, + { + "epoch": 0.11860626105684179, + "learning_rate": 0.002897069135195851, + "loss": 0.6856, + "step": 1542 + }, + { + "epoch": 0.11868317821706023, + "learning_rate": 0.0028969371395597284, + "loss": 1.0507, + "step": 1543 + }, + { + "epoch": 0.11876009537727868, + "learning_rate": 0.0028968050623549907, + "loss": 1.3073, + "step": 1544 + }, + { + "epoch": 0.11883701253749712, + "learning_rate": 0.002896672903589349, + "loss": 1.2736, + "step": 1545 + }, + { + "epoch": 0.11891392969771555, + "learning_rate": 0.0028965406632705214, + "loss": 1.3829, + "step": 1546 + }, + { + "epoch": 0.118990846857934, + "learning_rate": 0.002896408341406229, + "loss": 1.2508, + "step": 1547 + }, + { + "epoch": 0.11906776401815244, + "learning_rate": 0.0028962759380041984, + "loss": 1.176, + "step": 1548 + }, + { + "epoch": 0.11914468117837089, + "learning_rate": 0.00289614345307216, + "loss": 1.1064, + "step": 1549 + }, + { + "epoch": 0.11922159833858934, + "learning_rate": 0.002896010886617851, + "loss": 1.2934, + "step": 1550 + }, + { + "epoch": 0.11929851549880778, + "learning_rate": 0.002895878238649011, + "loss": 0.9442, + "step": 1551 + }, + { + "epoch": 0.11937543265902623, + "learning_rate": 0.002895745509173386, + "loss": 1.0226, + "step": 1552 + }, + { + "epoch": 0.11945234981924467, + "learning_rate": 0.002895612698198726, + "loss": 1.0044, + "step": 1553 + }, + { + "epoch": 0.11952926697946312, + "learning_rate": 0.0028954798057327863, + "loss": 1.4746, + "step": 1554 + }, + { + "epoch": 0.11960618413968156, + "learning_rate": 0.0028953468317833267, + "loss": 1.0465, + "step": 1555 + }, + { + "epoch": 0.11968310129990001, + "learning_rate": 0.002895213776358111, + "loss": 0.9802, + "step": 1556 + }, + { + "epoch": 0.11976001846011845, + "learning_rate": 0.002895080639464909, + "loss": 0.9997, + "step": 1557 + }, + { + "epoch": 0.1198369356203369, + "learning_rate": 0.0028949474211114943, + "loss": 0.9378, + "step": 1558 + }, + { + "epoch": 0.11991385278055534, + "learning_rate": 0.002894814121305646, + "loss": 1.0979, + "step": 1559 + }, + { + "epoch": 0.11999076994077379, + "learning_rate": 0.0028946807400551475, + "loss": 1.1207, + "step": 1560 + }, + { + "epoch": 0.12006768710099223, + "learning_rate": 0.0028945472773677875, + "loss": 1.6021, + "step": 1561 + }, + { + "epoch": 0.12014460426121068, + "learning_rate": 0.0028944137332513583, + "loss": 1.1666, + "step": 1562 + }, + { + "epoch": 0.12022152142142913, + "learning_rate": 0.0028942801077136574, + "loss": 1.1044, + "step": 1563 + }, + { + "epoch": 0.12029843858164757, + "learning_rate": 0.0028941464007624885, + "loss": 1.0011, + "step": 1564 + }, + { + "epoch": 0.12037535574186602, + "learning_rate": 0.002894012612405659, + "loss": 1.0823, + "step": 1565 + }, + { + "epoch": 0.12045227290208446, + "learning_rate": 0.0028938787426509792, + "loss": 1.0855, + "step": 1566 + }, + { + "epoch": 0.1205291900623029, + "learning_rate": 0.002893744791506267, + "loss": 1.0098, + "step": 1567 + }, + { + "epoch": 0.12060610722252134, + "learning_rate": 0.0028936107589793447, + "loss": 1.3746, + "step": 1568 + }, + { + "epoch": 0.12068302438273978, + "learning_rate": 0.0028934766450780373, + "loss": 1.0474, + "step": 1569 + }, + { + "epoch": 0.12075994154295823, + "learning_rate": 0.0028933424498101762, + "loss": 0.7657, + "step": 1570 + }, + { + "epoch": 0.12083685870317667, + "learning_rate": 0.0028932081731835976, + "loss": 1.2252, + "step": 1571 + }, + { + "epoch": 0.12091377586339512, + "learning_rate": 0.0028930738152061417, + "loss": 1.1658, + "step": 1572 + }, + { + "epoch": 0.12099069302361357, + "learning_rate": 0.002892939375885654, + "loss": 1.1699, + "step": 1573 + }, + { + "epoch": 0.12106761018383201, + "learning_rate": 0.002892804855229984, + "loss": 1.1706, + "step": 1574 + }, + { + "epoch": 0.12114452734405046, + "learning_rate": 0.0028926702532469876, + "loss": 1.1747, + "step": 1575 + }, + { + "epoch": 0.1212214445042689, + "learning_rate": 0.0028925355699445233, + "loss": 1.2883, + "step": 1576 + }, + { + "epoch": 0.12129836166448735, + "learning_rate": 0.002892400805330456, + "loss": 1.0269, + "step": 1577 + }, + { + "epoch": 0.12137527882470579, + "learning_rate": 0.0028922659594126546, + "loss": 1.1479, + "step": 1578 + }, + { + "epoch": 0.12145219598492424, + "learning_rate": 0.002892131032198992, + "loss": 1.1261, + "step": 1579 + }, + { + "epoch": 0.12152911314514268, + "learning_rate": 0.0028919960236973487, + "loss": 1.2145, + "step": 1580 + }, + { + "epoch": 0.12160603030536113, + "learning_rate": 0.0028918609339156067, + "loss": 1.2964, + "step": 1581 + }, + { + "epoch": 0.12168294746557957, + "learning_rate": 0.002891725762861654, + "loss": 0.8897, + "step": 1582 + }, + { + "epoch": 0.12175986462579802, + "learning_rate": 0.0028915905105433835, + "loss": 1.0423, + "step": 1583 + }, + { + "epoch": 0.12183678178601647, + "learning_rate": 0.0028914551769686925, + "loss": 1.2013, + "step": 1584 + }, + { + "epoch": 0.12191369894623491, + "learning_rate": 0.0028913197621454846, + "loss": 1.2252, + "step": 1585 + }, + { + "epoch": 0.12199061610645336, + "learning_rate": 0.0028911842660816647, + "loss": 1.0297, + "step": 1586 + }, + { + "epoch": 0.12206753326667179, + "learning_rate": 0.0028910486887851463, + "loss": 1.1314, + "step": 1587 + }, + { + "epoch": 0.12214445042689023, + "learning_rate": 0.002890913030263845, + "loss": 1.1527, + "step": 1588 + }, + { + "epoch": 0.12222136758710868, + "learning_rate": 0.0028907772905256823, + "loss": 1.4608, + "step": 1589 + }, + { + "epoch": 0.12229828474732712, + "learning_rate": 0.0028906414695785846, + "loss": 1.0999, + "step": 1590 + }, + { + "epoch": 0.12237520190754557, + "learning_rate": 0.0028905055674304813, + "loss": 1.2517, + "step": 1591 + }, + { + "epoch": 0.12245211906776401, + "learning_rate": 0.0028903695840893093, + "loss": 1.0303, + "step": 1592 + }, + { + "epoch": 0.12252903622798246, + "learning_rate": 0.0028902335195630084, + "loss": 1.153, + "step": 1593 + }, + { + "epoch": 0.1226059533882009, + "learning_rate": 0.002890097373859523, + "loss": 1.199, + "step": 1594 + }, + { + "epoch": 0.12268287054841935, + "learning_rate": 0.0028899611469868037, + "loss": 1.2185, + "step": 1595 + }, + { + "epoch": 0.1227597877086378, + "learning_rate": 0.002889824838952804, + "loss": 0.7877, + "step": 1596 + }, + { + "epoch": 0.12283670486885624, + "learning_rate": 0.002889688449765484, + "loss": 1.426, + "step": 1597 + }, + { + "epoch": 0.12291362202907469, + "learning_rate": 0.0028895519794328067, + "loss": 1.2392, + "step": 1598 + }, + { + "epoch": 0.12299053918929313, + "learning_rate": 0.0028894154279627414, + "loss": 1.1887, + "step": 1599 + }, + { + "epoch": 0.12306745634951158, + "learning_rate": 0.0028892787953632613, + "loss": 0.842, + "step": 1600 + }, + { + "epoch": 0.12314437350973002, + "learning_rate": 0.0028891420816423443, + "loss": 1.1089, + "step": 1601 + }, + { + "epoch": 0.12322129066994847, + "learning_rate": 0.0028890052868079737, + "loss": 1.207, + "step": 1602 + }, + { + "epoch": 0.12329820783016691, + "learning_rate": 0.0028888684108681366, + "loss": 1.1223, + "step": 1603 + }, + { + "epoch": 0.12337512499038536, + "learning_rate": 0.0028887314538308256, + "loss": 1.1937, + "step": 1604 + }, + { + "epoch": 0.1234520421506038, + "learning_rate": 0.0028885944157040376, + "loss": 1.0141, + "step": 1605 + }, + { + "epoch": 0.12352895931082225, + "learning_rate": 0.002888457296495774, + "loss": 1.2021, + "step": 1606 + }, + { + "epoch": 0.1236058764710407, + "learning_rate": 0.0028883200962140426, + "loss": 1.2874, + "step": 1607 + }, + { + "epoch": 0.12368279363125913, + "learning_rate": 0.0028881828148668537, + "loss": 1.3708, + "step": 1608 + }, + { + "epoch": 0.12375971079147757, + "learning_rate": 0.0028880454524622236, + "loss": 1.1228, + "step": 1609 + }, + { + "epoch": 0.12383662795169602, + "learning_rate": 0.002887908009008173, + "loss": 0.7563, + "step": 1610 + }, + { + "epoch": 0.12391354511191446, + "learning_rate": 0.0028877704845127273, + "loss": 1.0775, + "step": 1611 + }, + { + "epoch": 0.12399046227213291, + "learning_rate": 0.0028876328789839167, + "loss": 0.855, + "step": 1612 + }, + { + "epoch": 0.12406737943235135, + "learning_rate": 0.002887495192429776, + "loss": 1.0989, + "step": 1613 + }, + { + "epoch": 0.1241442965925698, + "learning_rate": 0.002887357424858345, + "loss": 1.3884, + "step": 1614 + }, + { + "epoch": 0.12422121375278825, + "learning_rate": 0.0028872195762776683, + "loss": 0.7945, + "step": 1615 + }, + { + "epoch": 0.12429813091300669, + "learning_rate": 0.0028870816466957942, + "loss": 1.2136, + "step": 1616 + }, + { + "epoch": 0.12437504807322514, + "learning_rate": 0.0028869436361207775, + "loss": 1.1295, + "step": 1617 + }, + { + "epoch": 0.12445196523344358, + "learning_rate": 0.0028868055445606766, + "loss": 1.2467, + "step": 1618 + }, + { + "epoch": 0.12452888239366203, + "learning_rate": 0.0028866673720235546, + "loss": 1.2908, + "step": 1619 + }, + { + "epoch": 0.12460579955388047, + "learning_rate": 0.0028865291185174798, + "loss": 1.2547, + "step": 1620 + }, + { + "epoch": 0.12468271671409892, + "learning_rate": 0.0028863907840505246, + "loss": 1.543, + "step": 1621 + }, + { + "epoch": 0.12475963387431736, + "learning_rate": 0.0028862523686307663, + "loss": 1.0831, + "step": 1622 + }, + { + "epoch": 0.12483655103453581, + "learning_rate": 0.002886113872266288, + "loss": 1.2436, + "step": 1623 + }, + { + "epoch": 0.12491346819475425, + "learning_rate": 0.0028859752949651762, + "loss": 1.5413, + "step": 1624 + }, + { + "epoch": 0.1249903853549727, + "learning_rate": 0.002885836636735522, + "loss": 1.1666, + "step": 1625 + }, + { + "epoch": 0.12506730251519113, + "learning_rate": 0.0028856978975854222, + "loss": 1.1694, + "step": 1626 + }, + { + "epoch": 0.1251442196754096, + "learning_rate": 0.0028855590775229783, + "loss": 0.9153, + "step": 1627 + }, + { + "epoch": 0.12522113683562802, + "learning_rate": 0.002885420176556296, + "loss": 1.4449, + "step": 1628 + }, + { + "epoch": 0.12529805399584648, + "learning_rate": 0.0028852811946934857, + "loss": 1.1671, + "step": 1629 + }, + { + "epoch": 0.1253749711560649, + "learning_rate": 0.002885142131942663, + "loss": 1.253, + "step": 1630 + }, + { + "epoch": 0.12545188831628337, + "learning_rate": 0.0028850029883119473, + "loss": 1.1113, + "step": 1631 + }, + { + "epoch": 0.1255288054765018, + "learning_rate": 0.0028848637638094637, + "loss": 0.9715, + "step": 1632 + }, + { + "epoch": 0.12560572263672026, + "learning_rate": 0.0028847244584433414, + "loss": 0.9045, + "step": 1633 + }, + { + "epoch": 0.1256826397969387, + "learning_rate": 0.0028845850722217154, + "loss": 1.2711, + "step": 1634 + }, + { + "epoch": 0.12575955695715715, + "learning_rate": 0.0028844456051527237, + "loss": 0.9732, + "step": 1635 + }, + { + "epoch": 0.12583647411737559, + "learning_rate": 0.0028843060572445105, + "loss": 1.3341, + "step": 1636 + }, + { + "epoch": 0.12591339127759404, + "learning_rate": 0.0028841664285052236, + "loss": 0.8971, + "step": 1637 + }, + { + "epoch": 0.12599030843781248, + "learning_rate": 0.0028840267189430167, + "loss": 1.0697, + "step": 1638 + }, + { + "epoch": 0.1260672255980309, + "learning_rate": 0.002883886928566047, + "loss": 0.9462, + "step": 1639 + }, + { + "epoch": 0.12614414275824937, + "learning_rate": 0.0028837470573824775, + "loss": 0.4884, + "step": 1640 + }, + { + "epoch": 0.1262210599184678, + "learning_rate": 0.0028836071054004755, + "loss": 1.8954, + "step": 1641 + }, + { + "epoch": 0.12629797707868626, + "learning_rate": 0.0028834670726282123, + "loss": 0.9863, + "step": 1642 + }, + { + "epoch": 0.1263748942389047, + "learning_rate": 0.002883326959073865, + "loss": 1.015, + "step": 1643 + }, + { + "epoch": 0.12645181139912315, + "learning_rate": 0.0028831867647456143, + "loss": 1.1744, + "step": 1644 + }, + { + "epoch": 0.12652872855934158, + "learning_rate": 0.002883046489651648, + "loss": 0.8641, + "step": 1645 + }, + { + "epoch": 0.12660564571956004, + "learning_rate": 0.0028829061338001547, + "loss": 1.0122, + "step": 1646 + }, + { + "epoch": 0.12668256287977847, + "learning_rate": 0.0028827656971993315, + "loss": 1.2379, + "step": 1647 + }, + { + "epoch": 0.12675948003999693, + "learning_rate": 0.0028826251798573783, + "loss": 0.9072, + "step": 1648 + }, + { + "epoch": 0.12683639720021536, + "learning_rate": 0.0028824845817825005, + "loss": 0.8084, + "step": 1649 + }, + { + "epoch": 0.12691331436043382, + "learning_rate": 0.002882343902982906, + "loss": 1.1727, + "step": 1650 + }, + { + "epoch": 0.12699023152065225, + "learning_rate": 0.0028822031434668108, + "loss": 1.1908, + "step": 1651 + }, + { + "epoch": 0.1270671486808707, + "learning_rate": 0.0028820623032424337, + "loss": 1.2055, + "step": 1652 + }, + { + "epoch": 0.12714406584108914, + "learning_rate": 0.0028819213823179984, + "loss": 1.3214, + "step": 1653 + }, + { + "epoch": 0.1272209830013076, + "learning_rate": 0.002881780380701733, + "loss": 1.1241, + "step": 1654 + }, + { + "epoch": 0.12729790016152603, + "learning_rate": 0.0028816392984018714, + "loss": 0.9759, + "step": 1655 + }, + { + "epoch": 0.1273748173217445, + "learning_rate": 0.0028814981354266514, + "loss": 1.2301, + "step": 1656 + }, + { + "epoch": 0.12745173448196292, + "learning_rate": 0.0028813568917843155, + "loss": 1.2305, + "step": 1657 + }, + { + "epoch": 0.12752865164218138, + "learning_rate": 0.0028812155674831103, + "loss": 0.9721, + "step": 1658 + }, + { + "epoch": 0.12760556880239982, + "learning_rate": 0.0028810741625312896, + "loss": 1.0428, + "step": 1659 + }, + { + "epoch": 0.12768248596261825, + "learning_rate": 0.002880932676937109, + "loss": 1.1266, + "step": 1660 + }, + { + "epoch": 0.1277594031228367, + "learning_rate": 0.00288079111070883, + "loss": 1.5436, + "step": 1661 + }, + { + "epoch": 0.12783632028305514, + "learning_rate": 0.002880649463854719, + "loss": 1.0405, + "step": 1662 + }, + { + "epoch": 0.1279132374432736, + "learning_rate": 0.002880507736383047, + "loss": 1.3491, + "step": 1663 + }, + { + "epoch": 0.12799015460349203, + "learning_rate": 0.0028803659283020895, + "loss": 1.2567, + "step": 1664 + }, + { + "epoch": 0.1280670717637105, + "learning_rate": 0.0028802240396201273, + "loss": 1.1244, + "step": 1665 + }, + { + "epoch": 0.12814398892392892, + "learning_rate": 0.0028800820703454447, + "loss": 1.061, + "step": 1666 + }, + { + "epoch": 0.12822090608414738, + "learning_rate": 0.0028799400204863315, + "loss": 1.0545, + "step": 1667 + }, + { + "epoch": 0.1282978232443658, + "learning_rate": 0.0028797978900510822, + "loss": 0.9163, + "step": 1668 + }, + { + "epoch": 0.12837474040458427, + "learning_rate": 0.0028796556790479967, + "loss": 1.3919, + "step": 1669 + }, + { + "epoch": 0.1284516575648027, + "learning_rate": 0.0028795133874853778, + "loss": 1.1179, + "step": 1670 + }, + { + "epoch": 0.12852857472502116, + "learning_rate": 0.0028793710153715347, + "loss": 1.1639, + "step": 1671 + }, + { + "epoch": 0.1286054918852396, + "learning_rate": 0.0028792285627147808, + "loss": 1.3275, + "step": 1672 + }, + { + "epoch": 0.12868240904545805, + "learning_rate": 0.002879086029523433, + "loss": 1.1799, + "step": 1673 + }, + { + "epoch": 0.12875932620567648, + "learning_rate": 0.0028789434158058153, + "loss": 1.23, + "step": 1674 + }, + { + "epoch": 0.12883624336589494, + "learning_rate": 0.0028788007215702544, + "loss": 1.0948, + "step": 1675 + }, + { + "epoch": 0.12891316052611337, + "learning_rate": 0.002878657946825082, + "loss": 1.1994, + "step": 1676 + }, + { + "epoch": 0.12899007768633183, + "learning_rate": 0.0028785150915786356, + "loss": 1.4292, + "step": 1677 + }, + { + "epoch": 0.12906699484655026, + "learning_rate": 0.0028783721558392566, + "loss": 1.3145, + "step": 1678 + }, + { + "epoch": 0.1291439120067687, + "learning_rate": 0.00287822913961529, + "loss": 1.2676, + "step": 1679 + }, + { + "epoch": 0.12922082916698716, + "learning_rate": 0.0028780860429150883, + "loss": 1.4908, + "step": 1680 + }, + { + "epoch": 0.1292977463272056, + "learning_rate": 0.0028779428657470063, + "loss": 1.2185, + "step": 1681 + }, + { + "epoch": 0.12937466348742405, + "learning_rate": 0.002877799608119404, + "loss": 1.4401, + "step": 1682 + }, + { + "epoch": 0.12945158064764248, + "learning_rate": 0.002877656270040647, + "loss": 1.0149, + "step": 1683 + }, + { + "epoch": 0.12952849780786094, + "learning_rate": 0.0028775128515191053, + "loss": 0.9654, + "step": 1684 + }, + { + "epoch": 0.12960541496807937, + "learning_rate": 0.0028773693525631513, + "loss": 1.1157, + "step": 1685 + }, + { + "epoch": 0.12968233212829783, + "learning_rate": 0.002877225773181166, + "loss": 0.9185, + "step": 1686 + }, + { + "epoch": 0.12975924928851626, + "learning_rate": 0.0028770821133815325, + "loss": 1.1889, + "step": 1687 + }, + { + "epoch": 0.12983616644873472, + "learning_rate": 0.002876938373172639, + "loss": 1.491, + "step": 1688 + }, + { + "epoch": 0.12991308360895315, + "learning_rate": 0.002876794552562879, + "loss": 1.3061, + "step": 1689 + }, + { + "epoch": 0.1299900007691716, + "learning_rate": 0.00287665065156065, + "loss": 1.2791, + "step": 1690 + }, + { + "epoch": 0.13006691792939004, + "learning_rate": 0.0028765066701743553, + "loss": 1.3298, + "step": 1691 + }, + { + "epoch": 0.1301438350896085, + "learning_rate": 0.002876362608412401, + "loss": 1.3656, + "step": 1692 + }, + { + "epoch": 0.13022075224982693, + "learning_rate": 0.0028762184662832, + "loss": 1.0401, + "step": 1693 + }, + { + "epoch": 0.1302976694100454, + "learning_rate": 0.0028760742437951682, + "loss": 1.1322, + "step": 1694 + }, + { + "epoch": 0.13037458657026382, + "learning_rate": 0.002875929940956727, + "loss": 1.0038, + "step": 1695 + }, + { + "epoch": 0.13045150373048228, + "learning_rate": 0.002875785557776303, + "loss": 0.963, + "step": 1696 + }, + { + "epoch": 0.1305284208907007, + "learning_rate": 0.0028756410942623266, + "loss": 0.9686, + "step": 1697 + }, + { + "epoch": 0.13060533805091917, + "learning_rate": 0.002875496550423233, + "loss": 1.0261, + "step": 1698 + }, + { + "epoch": 0.1306822552111376, + "learning_rate": 0.0028753519262674616, + "loss": 1.0338, + "step": 1699 + }, + { + "epoch": 0.13075917237135604, + "learning_rate": 0.0028752072218034587, + "loss": 1.2176, + "step": 1700 + }, + { + "epoch": 0.1308360895315745, + "learning_rate": 0.002875062437039672, + "loss": 1.1064, + "step": 1701 + }, + { + "epoch": 0.13091300669179293, + "learning_rate": 0.0028749175719845576, + "loss": 1.1406, + "step": 1702 + }, + { + "epoch": 0.1309899238520114, + "learning_rate": 0.002874772626646573, + "loss": 1.3195, + "step": 1703 + }, + { + "epoch": 0.13106684101222982, + "learning_rate": 0.0028746276010341815, + "loss": 0.963, + "step": 1704 + }, + { + "epoch": 0.13114375817244828, + "learning_rate": 0.0028744824951558523, + "loss": 0.9756, + "step": 1705 + }, + { + "epoch": 0.1312206753326667, + "learning_rate": 0.002874337309020057, + "loss": 0.9419, + "step": 1706 + }, + { + "epoch": 0.13129759249288517, + "learning_rate": 0.002874192042635275, + "loss": 0.9682, + "step": 1707 + }, + { + "epoch": 0.1313745096531036, + "learning_rate": 0.0028740466960099865, + "loss": 1.1678, + "step": 1708 + }, + { + "epoch": 0.13145142681332206, + "learning_rate": 0.00287390126915268, + "loss": 1.7509, + "step": 1709 + }, + { + "epoch": 0.1315283439735405, + "learning_rate": 0.0028737557620718466, + "loss": 0.9172, + "step": 1710 + }, + { + "epoch": 0.13160526113375895, + "learning_rate": 0.0028736101747759823, + "loss": 1.3299, + "step": 1711 + }, + { + "epoch": 0.13168217829397738, + "learning_rate": 0.0028734645072735886, + "loss": 1.1382, + "step": 1712 + }, + { + "epoch": 0.13175909545419584, + "learning_rate": 0.00287331875957317, + "loss": 1.2588, + "step": 1713 + }, + { + "epoch": 0.13183601261441427, + "learning_rate": 0.002873172931683239, + "loss": 0.9852, + "step": 1714 + }, + { + "epoch": 0.13191292977463273, + "learning_rate": 0.002873027023612309, + "loss": 1.3062, + "step": 1715 + }, + { + "epoch": 0.13198984693485116, + "learning_rate": 0.0028728810353689, + "loss": 0.9516, + "step": 1716 + }, + { + "epoch": 0.13206676409506962, + "learning_rate": 0.0028727349669615365, + "loss": 1.0101, + "step": 1717 + }, + { + "epoch": 0.13214368125528805, + "learning_rate": 0.0028725888183987478, + "loss": 0.9942, + "step": 1718 + }, + { + "epoch": 0.1322205984155065, + "learning_rate": 0.0028724425896890678, + "loss": 1.1523, + "step": 1719 + }, + { + "epoch": 0.13229751557572494, + "learning_rate": 0.0028722962808410343, + "loss": 1.1402, + "step": 1720 + }, + { + "epoch": 0.13237443273594338, + "learning_rate": 0.002872149891863191, + "loss": 1.0075, + "step": 1721 + }, + { + "epoch": 0.13245134989616184, + "learning_rate": 0.002872003422764085, + "loss": 1.3004, + "step": 1722 + }, + { + "epoch": 0.13252826705638027, + "learning_rate": 0.00287185687355227, + "loss": 1.0971, + "step": 1723 + }, + { + "epoch": 0.13260518421659873, + "learning_rate": 0.002871710244236302, + "loss": 1.4052, + "step": 1724 + }, + { + "epoch": 0.13268210137681716, + "learning_rate": 0.0028715635348247437, + "loss": 1.1598, + "step": 1725 + }, + { + "epoch": 0.13275901853703562, + "learning_rate": 0.0028714167453261605, + "loss": 1.0552, + "step": 1726 + }, + { + "epoch": 0.13283593569725405, + "learning_rate": 0.0028712698757491248, + "loss": 1.0522, + "step": 1727 + }, + { + "epoch": 0.1329128528574725, + "learning_rate": 0.002871122926102212, + "loss": 1.2255, + "step": 1728 + }, + { + "epoch": 0.13298977001769094, + "learning_rate": 0.0028709758963940024, + "loss": 0.9807, + "step": 1729 + }, + { + "epoch": 0.1330666871779094, + "learning_rate": 0.0028708287866330813, + "loss": 1.1592, + "step": 1730 + }, + { + "epoch": 0.13314360433812783, + "learning_rate": 0.0028706815968280387, + "loss": 1.0734, + "step": 1731 + }, + { + "epoch": 0.1332205214983463, + "learning_rate": 0.0028705343269874693, + "loss": 1.2188, + "step": 1732 + }, + { + "epoch": 0.13329743865856472, + "learning_rate": 0.002870386977119972, + "loss": 1.4174, + "step": 1733 + }, + { + "epoch": 0.13337435581878318, + "learning_rate": 0.0028702395472341513, + "loss": 1.3331, + "step": 1734 + }, + { + "epoch": 0.1334512729790016, + "learning_rate": 0.0028700920373386156, + "loss": 1.1743, + "step": 1735 + }, + { + "epoch": 0.13352819013922007, + "learning_rate": 0.0028699444474419773, + "loss": 1.0421, + "step": 1736 + }, + { + "epoch": 0.1336051072994385, + "learning_rate": 0.0028697967775528554, + "loss": 1.0958, + "step": 1737 + }, + { + "epoch": 0.13368202445965696, + "learning_rate": 0.0028696490276798716, + "loss": 1.1462, + "step": 1738 + }, + { + "epoch": 0.1337589416198754, + "learning_rate": 0.0028695011978316543, + "loss": 1.2942, + "step": 1739 + }, + { + "epoch": 0.13383585878009385, + "learning_rate": 0.0028693532880168346, + "loss": 1.1019, + "step": 1740 + }, + { + "epoch": 0.13391277594031228, + "learning_rate": 0.0028692052982440494, + "loss": 1.1589, + "step": 1741 + }, + { + "epoch": 0.13398969310053072, + "learning_rate": 0.0028690572285219396, + "loss": 1.3864, + "step": 1742 + }, + { + "epoch": 0.13406661026074917, + "learning_rate": 0.0028689090788591516, + "loss": 1.0047, + "step": 1743 + }, + { + "epoch": 0.1341435274209676, + "learning_rate": 0.0028687608492643357, + "loss": 1.1152, + "step": 1744 + }, + { + "epoch": 0.13422044458118607, + "learning_rate": 0.0028686125397461475, + "loss": 1.1046, + "step": 1745 + }, + { + "epoch": 0.1342973617414045, + "learning_rate": 0.002868464150313247, + "loss": 1.3783, + "step": 1746 + }, + { + "epoch": 0.13437427890162296, + "learning_rate": 0.0028683156809742975, + "loss": 1.0825, + "step": 1747 + }, + { + "epoch": 0.1344511960618414, + "learning_rate": 0.0028681671317379704, + "loss": 1.1973, + "step": 1748 + }, + { + "epoch": 0.13452811322205985, + "learning_rate": 0.002868018502612938, + "loss": 0.8653, + "step": 1749 + }, + { + "epoch": 0.13460503038227828, + "learning_rate": 0.0028678697936078803, + "loss": 1.3479, + "step": 1750 + }, + { + "epoch": 0.13468194754249674, + "learning_rate": 0.0028677210047314793, + "loss": 0.9689, + "step": 1751 + }, + { + "epoch": 0.13475886470271517, + "learning_rate": 0.0028675721359924236, + "loss": 1.2804, + "step": 1752 + }, + { + "epoch": 0.13483578186293363, + "learning_rate": 0.002867423187399405, + "loss": 1.017, + "step": 1753 + }, + { + "epoch": 0.13491269902315206, + "learning_rate": 0.0028672741589611224, + "loss": 0.9328, + "step": 1754 + }, + { + "epoch": 0.13498961618337052, + "learning_rate": 0.0028671250506862762, + "loss": 1.1762, + "step": 1755 + }, + { + "epoch": 0.13506653334358895, + "learning_rate": 0.002866975862583574, + "loss": 1.1972, + "step": 1756 + }, + { + "epoch": 0.1351434505038074, + "learning_rate": 0.0028668265946617255, + "loss": 1.0895, + "step": 1757 + }, + { + "epoch": 0.13522036766402584, + "learning_rate": 0.0028666772469294484, + "loss": 1.2591, + "step": 1758 + }, + { + "epoch": 0.1352972848242443, + "learning_rate": 0.0028665278193954623, + "loss": 1.5245, + "step": 1759 + }, + { + "epoch": 0.13537420198446273, + "learning_rate": 0.0028663783120684927, + "loss": 1.2559, + "step": 1760 + }, + { + "epoch": 0.13545111914468116, + "learning_rate": 0.0028662287249572697, + "loss": 1.2404, + "step": 1761 + }, + { + "epoch": 0.13552803630489962, + "learning_rate": 0.0028660790580705277, + "loss": 0.9549, + "step": 1762 + }, + { + "epoch": 0.13560495346511806, + "learning_rate": 0.002865929311417005, + "loss": 0.9313, + "step": 1763 + }, + { + "epoch": 0.13568187062533651, + "learning_rate": 0.002865779485005447, + "loss": 1.2338, + "step": 1764 + }, + { + "epoch": 0.13575878778555495, + "learning_rate": 0.0028656295788446013, + "loss": 1.2663, + "step": 1765 + }, + { + "epoch": 0.1358357049457734, + "learning_rate": 0.002865479592943221, + "loss": 1.0019, + "step": 1766 + }, + { + "epoch": 0.13591262210599184, + "learning_rate": 0.0028653295273100648, + "loss": 1.5426, + "step": 1767 + }, + { + "epoch": 0.1359895392662103, + "learning_rate": 0.0028651793819538937, + "loss": 1.3661, + "step": 1768 + }, + { + "epoch": 0.13606645642642873, + "learning_rate": 0.0028650291568834765, + "loss": 1.0231, + "step": 1769 + }, + { + "epoch": 0.1361433735866472, + "learning_rate": 0.002864878852107584, + "loss": 1.4282, + "step": 1770 + }, + { + "epoch": 0.13622029074686562, + "learning_rate": 0.0028647284676349928, + "loss": 1.0069, + "step": 1771 + }, + { + "epoch": 0.13629720790708408, + "learning_rate": 0.002864578003474484, + "loss": 1.207, + "step": 1772 + }, + { + "epoch": 0.1363741250673025, + "learning_rate": 0.0028644274596348438, + "loss": 1.0187, + "step": 1773 + }, + { + "epoch": 0.13645104222752097, + "learning_rate": 0.0028642768361248623, + "loss": 1.1986, + "step": 1774 + }, + { + "epoch": 0.1365279593877394, + "learning_rate": 0.0028641261329533344, + "loss": 1.0283, + "step": 1775 + }, + { + "epoch": 0.13660487654795786, + "learning_rate": 0.00286397535012906, + "loss": 1.2079, + "step": 1776 + }, + { + "epoch": 0.1366817937081763, + "learning_rate": 0.0028638244876608428, + "loss": 0.8502, + "step": 1777 + }, + { + "epoch": 0.13675871086839475, + "learning_rate": 0.002863673545557493, + "loss": 0.8472, + "step": 1778 + }, + { + "epoch": 0.13683562802861318, + "learning_rate": 0.0028635225238278234, + "loss": 1.2772, + "step": 1779 + }, + { + "epoch": 0.13691254518883164, + "learning_rate": 0.002863371422480653, + "loss": 1.1688, + "step": 1780 + }, + { + "epoch": 0.13698946234905007, + "learning_rate": 0.002863220241524804, + "loss": 0.7301, + "step": 1781 + }, + { + "epoch": 0.1370663795092685, + "learning_rate": 0.0028630689809691044, + "loss": 1.0963, + "step": 1782 + }, + { + "epoch": 0.13714329666948696, + "learning_rate": 0.002862917640822387, + "loss": 0.9698, + "step": 1783 + }, + { + "epoch": 0.1372202138297054, + "learning_rate": 0.0028627662210934874, + "loss": 1.1856, + "step": 1784 + }, + { + "epoch": 0.13729713098992385, + "learning_rate": 0.0028626147217912477, + "loss": 0.9043, + "step": 1785 + }, + { + "epoch": 0.13737404815014229, + "learning_rate": 0.002862463142924515, + "loss": 1.1654, + "step": 1786 + }, + { + "epoch": 0.13745096531036075, + "learning_rate": 0.0028623114845021396, + "loss": 1.2161, + "step": 1787 + }, + { + "epoch": 0.13752788247057918, + "learning_rate": 0.0028621597465329765, + "loss": 1.1165, + "step": 1788 + }, + { + "epoch": 0.13760479963079764, + "learning_rate": 0.0028620079290258863, + "loss": 1.0131, + "step": 1789 + }, + { + "epoch": 0.13768171679101607, + "learning_rate": 0.002861856031989733, + "loss": 0.8552, + "step": 1790 + }, + { + "epoch": 0.13775863395123453, + "learning_rate": 0.0028617040554333876, + "loss": 1.0169, + "step": 1791 + }, + { + "epoch": 0.13783555111145296, + "learning_rate": 0.0028615519993657234, + "loss": 1.3486, + "step": 1792 + }, + { + "epoch": 0.13791246827167142, + "learning_rate": 0.002861399863795618, + "loss": 0.9617, + "step": 1793 + }, + { + "epoch": 0.13798938543188985, + "learning_rate": 0.002861247648731957, + "loss": 0.9184, + "step": 1794 + }, + { + "epoch": 0.1380663025921083, + "learning_rate": 0.0028610953541836257, + "loss": 1.1691, + "step": 1795 + }, + { + "epoch": 0.13814321975232674, + "learning_rate": 0.002860942980159519, + "loss": 1.2551, + "step": 1796 + }, + { + "epoch": 0.1382201369125452, + "learning_rate": 0.0028607905266685336, + "loss": 1.2939, + "step": 1797 + }, + { + "epoch": 0.13829705407276363, + "learning_rate": 0.0028606379937195707, + "loss": 1.2948, + "step": 1798 + }, + { + "epoch": 0.1383739712329821, + "learning_rate": 0.0028604853813215376, + "loss": 0.8462, + "step": 1799 + }, + { + "epoch": 0.13845088839320052, + "learning_rate": 0.002860332689483345, + "loss": 1.3494, + "step": 1800 + }, + { + "epoch": 0.13852780555341898, + "learning_rate": 0.0028601799182139086, + "loss": 0.9626, + "step": 1801 + }, + { + "epoch": 0.1386047227136374, + "learning_rate": 0.00286002706752215, + "loss": 1.0662, + "step": 1802 + }, + { + "epoch": 0.13868163987385584, + "learning_rate": 0.002859874137416993, + "loss": 1.2386, + "step": 1803 + }, + { + "epoch": 0.1387585570340743, + "learning_rate": 0.002859721127907368, + "loss": 1.5507, + "step": 1804 + }, + { + "epoch": 0.13883547419429274, + "learning_rate": 0.0028595680390022093, + "loss": 1.2978, + "step": 1805 + }, + { + "epoch": 0.1389123913545112, + "learning_rate": 0.002859414870710456, + "loss": 1.1999, + "step": 1806 + }, + { + "epoch": 0.13898930851472963, + "learning_rate": 0.0028592616230410516, + "loss": 1.1199, + "step": 1807 + }, + { + "epoch": 0.13906622567494809, + "learning_rate": 0.0028591082960029443, + "loss": 0.9053, + "step": 1808 + }, + { + "epoch": 0.13914314283516652, + "learning_rate": 0.0028589548896050874, + "loss": 1.2272, + "step": 1809 + }, + { + "epoch": 0.13922005999538498, + "learning_rate": 0.002858801403856438, + "loss": 1.04, + "step": 1810 + }, + { + "epoch": 0.1392969771556034, + "learning_rate": 0.0028586478387659588, + "loss": 1.0117, + "step": 1811 + }, + { + "epoch": 0.13937389431582187, + "learning_rate": 0.0028584941943426163, + "loss": 1.0784, + "step": 1812 + }, + { + "epoch": 0.1394508114760403, + "learning_rate": 0.002858340470595382, + "loss": 0.914, + "step": 1813 + }, + { + "epoch": 0.13952772863625876, + "learning_rate": 0.002858186667533232, + "loss": 1.0431, + "step": 1814 + }, + { + "epoch": 0.1396046457964772, + "learning_rate": 0.0028580327851651464, + "loss": 1.0706, + "step": 1815 + }, + { + "epoch": 0.13968156295669565, + "learning_rate": 0.0028578788235001117, + "loss": 1.2811, + "step": 1816 + }, + { + "epoch": 0.13975848011691408, + "learning_rate": 0.0028577247825471177, + "loss": 1.2036, + "step": 1817 + }, + { + "epoch": 0.13983539727713254, + "learning_rate": 0.002857570662315158, + "loss": 1.3338, + "step": 1818 + }, + { + "epoch": 0.13991231443735097, + "learning_rate": 0.002857416462813233, + "loss": 1.1993, + "step": 1819 + }, + { + "epoch": 0.13998923159756943, + "learning_rate": 0.0028572621840503454, + "loss": 1.1732, + "step": 1820 + }, + { + "epoch": 0.14006614875778786, + "learning_rate": 0.002857107826035505, + "loss": 1.1578, + "step": 1821 + }, + { + "epoch": 0.14014306591800632, + "learning_rate": 0.0028569533887777237, + "loss": 1.3077, + "step": 1822 + }, + { + "epoch": 0.14021998307822475, + "learning_rate": 0.0028567988722860206, + "loss": 0.9864, + "step": 1823 + }, + { + "epoch": 0.14029690023844318, + "learning_rate": 0.002856644276569417, + "loss": 1.3024, + "step": 1824 + }, + { + "epoch": 0.14037381739866164, + "learning_rate": 0.00285648960163694, + "loss": 1.3467, + "step": 1825 + }, + { + "epoch": 0.14045073455888007, + "learning_rate": 0.002856334847497622, + "loss": 1.1539, + "step": 1826 + }, + { + "epoch": 0.14052765171909853, + "learning_rate": 0.002856180014160498, + "loss": 1.0856, + "step": 1827 + }, + { + "epoch": 0.14060456887931697, + "learning_rate": 0.0028560251016346103, + "loss": 1.2987, + "step": 1828 + }, + { + "epoch": 0.14068148603953542, + "learning_rate": 0.0028558701099290033, + "loss": 1.3895, + "step": 1829 + }, + { + "epoch": 0.14075840319975386, + "learning_rate": 0.0028557150390527276, + "loss": 1.4298, + "step": 1830 + }, + { + "epoch": 0.14083532035997232, + "learning_rate": 0.002855559889014838, + "loss": 1.4291, + "step": 1831 + }, + { + "epoch": 0.14091223752019075, + "learning_rate": 0.0028554046598243937, + "loss": 1.0573, + "step": 1832 + }, + { + "epoch": 0.1409891546804092, + "learning_rate": 0.0028552493514904585, + "loss": 1.2162, + "step": 1833 + }, + { + "epoch": 0.14106607184062764, + "learning_rate": 0.0028550939640221017, + "loss": 0.8612, + "step": 1834 + }, + { + "epoch": 0.1411429890008461, + "learning_rate": 0.002854938497428396, + "loss": 0.9214, + "step": 1835 + }, + { + "epoch": 0.14121990616106453, + "learning_rate": 0.002854782951718419, + "loss": 1.0045, + "step": 1836 + }, + { + "epoch": 0.141296823321283, + "learning_rate": 0.0028546273269012537, + "loss": 1.2892, + "step": 1837 + }, + { + "epoch": 0.14137374048150142, + "learning_rate": 0.002854471622985987, + "loss": 1.2209, + "step": 1838 + }, + { + "epoch": 0.14145065764171988, + "learning_rate": 0.0028543158399817106, + "loss": 1.263, + "step": 1839 + }, + { + "epoch": 0.1415275748019383, + "learning_rate": 0.0028541599778975205, + "loss": 1.2222, + "step": 1840 + }, + { + "epoch": 0.14160449196215677, + "learning_rate": 0.002854004036742519, + "loss": 1.1956, + "step": 1841 + }, + { + "epoch": 0.1416814091223752, + "learning_rate": 0.00285384801652581, + "loss": 1.1176, + "step": 1842 + }, + { + "epoch": 0.14175832628259363, + "learning_rate": 0.0028536919172565045, + "loss": 0.9716, + "step": 1843 + }, + { + "epoch": 0.1418352434428121, + "learning_rate": 0.0028535357389437173, + "loss": 1.0649, + "step": 1844 + }, + { + "epoch": 0.14191216060303052, + "learning_rate": 0.0028533794815965674, + "loss": 0.9432, + "step": 1845 + }, + { + "epoch": 0.14198907776324898, + "learning_rate": 0.0028532231452241797, + "loss": 0.5686, + "step": 1846 + }, + { + "epoch": 0.14206599492346741, + "learning_rate": 0.0028530667298356815, + "loss": 0.9443, + "step": 1847 + }, + { + "epoch": 0.14214291208368587, + "learning_rate": 0.002852910235440207, + "loss": 1.2505, + "step": 1848 + }, + { + "epoch": 0.1422198292439043, + "learning_rate": 0.0028527536620468943, + "loss": 1.3604, + "step": 1849 + }, + { + "epoch": 0.14229674640412276, + "learning_rate": 0.002852597009664885, + "loss": 1.1274, + "step": 1850 + }, + { + "epoch": 0.1423736635643412, + "learning_rate": 0.002852440278303327, + "loss": 1.2805, + "step": 1851 + }, + { + "epoch": 0.14245058072455966, + "learning_rate": 0.0028522834679713715, + "loss": 1.4961, + "step": 1852 + }, + { + "epoch": 0.1425274978847781, + "learning_rate": 0.0028521265786781752, + "loss": 1.5282, + "step": 1853 + }, + { + "epoch": 0.14260441504499655, + "learning_rate": 0.0028519696104328984, + "loss": 1.2071, + "step": 1854 + }, + { + "epoch": 0.14268133220521498, + "learning_rate": 0.0028518125632447074, + "loss": 0.8997, + "step": 1855 + }, + { + "epoch": 0.14275824936543344, + "learning_rate": 0.0028516554371227723, + "loss": 1.5829, + "step": 1856 + }, + { + "epoch": 0.14283516652565187, + "learning_rate": 0.0028514982320762667, + "loss": 1.0073, + "step": 1857 + }, + { + "epoch": 0.14291208368587033, + "learning_rate": 0.0028513409481143713, + "loss": 0.9761, + "step": 1858 + }, + { + "epoch": 0.14298900084608876, + "learning_rate": 0.0028511835852462696, + "loss": 1.5369, + "step": 1859 + }, + { + "epoch": 0.14306591800630722, + "learning_rate": 0.0028510261434811506, + "loss": 1.4548, + "step": 1860 + }, + { + "epoch": 0.14314283516652565, + "learning_rate": 0.0028508686228282064, + "loss": 0.9244, + "step": 1861 + }, + { + "epoch": 0.1432197523267441, + "learning_rate": 0.002850711023296635, + "loss": 1.0911, + "step": 1862 + }, + { + "epoch": 0.14329666948696254, + "learning_rate": 0.0028505533448956407, + "loss": 1.191, + "step": 1863 + }, + { + "epoch": 0.14337358664718097, + "learning_rate": 0.0028503955876344285, + "loss": 1.3685, + "step": 1864 + }, + { + "epoch": 0.14345050380739943, + "learning_rate": 0.00285023775152221, + "loss": 1.1106, + "step": 1865 + }, + { + "epoch": 0.14352742096761786, + "learning_rate": 0.002850079836568203, + "loss": 0.9043, + "step": 1866 + }, + { + "epoch": 0.14360433812783632, + "learning_rate": 0.0028499218427816267, + "loss": 1.0367, + "step": 1867 + }, + { + "epoch": 0.14368125528805475, + "learning_rate": 0.002849763770171707, + "loss": 0.9066, + "step": 1868 + }, + { + "epoch": 0.1437581724482732, + "learning_rate": 0.0028496056187476745, + "loss": 1.5269, + "step": 1869 + }, + { + "epoch": 0.14383508960849165, + "learning_rate": 0.002849447388518763, + "loss": 1.2532, + "step": 1870 + }, + { + "epoch": 0.1439120067687101, + "learning_rate": 0.0028492890794942125, + "loss": 0.57, + "step": 1871 + }, + { + "epoch": 0.14398892392892854, + "learning_rate": 0.0028491306916832663, + "loss": 1.0754, + "step": 1872 + }, + { + "epoch": 0.144065841089147, + "learning_rate": 0.002848972225095173, + "loss": 1.199, + "step": 1873 + }, + { + "epoch": 0.14414275824936543, + "learning_rate": 0.0028488136797391855, + "loss": 1.037, + "step": 1874 + }, + { + "epoch": 0.1442196754095839, + "learning_rate": 0.0028486550556245617, + "loss": 1.0015, + "step": 1875 + }, + { + "epoch": 0.14429659256980232, + "learning_rate": 0.0028484963527605637, + "loss": 1.2186, + "step": 1876 + }, + { + "epoch": 0.14437350973002078, + "learning_rate": 0.002848337571156458, + "loss": 1.2143, + "step": 1877 + }, + { + "epoch": 0.1444504268902392, + "learning_rate": 0.0028481787108215165, + "loss": 1.2161, + "step": 1878 + }, + { + "epoch": 0.14452734405045767, + "learning_rate": 0.0028480197717650154, + "loss": 0.712, + "step": 1879 + }, + { + "epoch": 0.1446042612106761, + "learning_rate": 0.002847860753996234, + "loss": 1.2457, + "step": 1880 + }, + { + "epoch": 0.14468117837089456, + "learning_rate": 0.002847701657524459, + "loss": 1.331, + "step": 1881 + }, + { + "epoch": 0.144758095531113, + "learning_rate": 0.0028475424823589802, + "loss": 1.4052, + "step": 1882 + }, + { + "epoch": 0.14483501269133145, + "learning_rate": 0.0028473832285090908, + "loss": 1.0963, + "step": 1883 + }, + { + "epoch": 0.14491192985154988, + "learning_rate": 0.002847223895984091, + "loss": 1.1275, + "step": 1884 + }, + { + "epoch": 0.1449888470117683, + "learning_rate": 0.0028470644847932835, + "loss": 1.5927, + "step": 1885 + }, + { + "epoch": 0.14506576417198677, + "learning_rate": 0.002846904994945977, + "loss": 1.0431, + "step": 1886 + }, + { + "epoch": 0.1451426813322052, + "learning_rate": 0.002846745426451484, + "loss": 1.37, + "step": 1887 + }, + { + "epoch": 0.14521959849242366, + "learning_rate": 0.0028465857793191223, + "loss": 1.3338, + "step": 1888 + }, + { + "epoch": 0.1452965156526421, + "learning_rate": 0.002846426053558213, + "loss": 0.9425, + "step": 1889 + }, + { + "epoch": 0.14537343281286055, + "learning_rate": 0.0028462662491780836, + "loss": 1.2258, + "step": 1890 + }, + { + "epoch": 0.14545034997307899, + "learning_rate": 0.0028461063661880653, + "loss": 0.8685, + "step": 1891 + }, + { + "epoch": 0.14552726713329744, + "learning_rate": 0.0028459464045974925, + "loss": 1.0388, + "step": 1892 + }, + { + "epoch": 0.14560418429351588, + "learning_rate": 0.0028457863644157067, + "loss": 1.2992, + "step": 1893 + }, + { + "epoch": 0.14568110145373434, + "learning_rate": 0.002845626245652053, + "loss": 1.0715, + "step": 1894 + }, + { + "epoch": 0.14575801861395277, + "learning_rate": 0.00284546604831588, + "loss": 1.5183, + "step": 1895 + }, + { + "epoch": 0.14583493577417123, + "learning_rate": 0.0028453057724165424, + "loss": 1.0368, + "step": 1896 + }, + { + "epoch": 0.14591185293438966, + "learning_rate": 0.0028451454179633986, + "loss": 1.0643, + "step": 1897 + }, + { + "epoch": 0.14598877009460812, + "learning_rate": 0.0028449849849658117, + "loss": 1.4871, + "step": 1898 + }, + { + "epoch": 0.14606568725482655, + "learning_rate": 0.0028448244734331505, + "loss": 0.7954, + "step": 1899 + }, + { + "epoch": 0.146142604415045, + "learning_rate": 0.0028446638833747864, + "loss": 0.9552, + "step": 1900 + }, + { + "epoch": 0.14621952157526344, + "learning_rate": 0.0028445032148000965, + "loss": 1.2515, + "step": 1901 + }, + { + "epoch": 0.1462964387354819, + "learning_rate": 0.002844342467718463, + "loss": 1.0473, + "step": 1902 + }, + { + "epoch": 0.14637335589570033, + "learning_rate": 0.0028441816421392718, + "loss": 0.8347, + "step": 1903 + }, + { + "epoch": 0.1464502730559188, + "learning_rate": 0.0028440207380719135, + "loss": 1.0339, + "step": 1904 + }, + { + "epoch": 0.14652719021613722, + "learning_rate": 0.0028438597555257837, + "loss": 1.388, + "step": 1905 + }, + { + "epoch": 0.14660410737635565, + "learning_rate": 0.002843698694510282, + "loss": 0.9474, + "step": 1906 + }, + { + "epoch": 0.1466810245365741, + "learning_rate": 0.0028435375550348134, + "loss": 1.3463, + "step": 1907 + }, + { + "epoch": 0.14675794169679254, + "learning_rate": 0.002843376337108787, + "loss": 1.4996, + "step": 1908 + }, + { + "epoch": 0.146834858857011, + "learning_rate": 0.002843215040741616, + "loss": 1.5202, + "step": 1909 + }, + { + "epoch": 0.14691177601722943, + "learning_rate": 0.002843053665942719, + "loss": 0.8599, + "step": 1910 + }, + { + "epoch": 0.1469886931774479, + "learning_rate": 0.0028428922127215184, + "loss": 0.9895, + "step": 1911 + }, + { + "epoch": 0.14706561033766632, + "learning_rate": 0.002842730681087442, + "loss": 1.2132, + "step": 1912 + }, + { + "epoch": 0.14714252749788478, + "learning_rate": 0.002842569071049922, + "loss": 1.0408, + "step": 1913 + }, + { + "epoch": 0.14721944465810322, + "learning_rate": 0.002842407382618395, + "loss": 0.9303, + "step": 1914 + }, + { + "epoch": 0.14729636181832167, + "learning_rate": 0.0028422456158023017, + "loss": 1.5299, + "step": 1915 + }, + { + "epoch": 0.1473732789785401, + "learning_rate": 0.0028420837706110876, + "loss": 1.1313, + "step": 1916 + }, + { + "epoch": 0.14745019613875857, + "learning_rate": 0.002841921847054204, + "loss": 1.3461, + "step": 1917 + }, + { + "epoch": 0.147527113298977, + "learning_rate": 0.002841759845141105, + "loss": 1.2145, + "step": 1918 + }, + { + "epoch": 0.14760403045919546, + "learning_rate": 0.0028415977648812503, + "loss": 1.4611, + "step": 1919 + }, + { + "epoch": 0.1476809476194139, + "learning_rate": 0.002841435606284104, + "loss": 0.866, + "step": 1920 + }, + { + "epoch": 0.14775786477963235, + "learning_rate": 0.0028412733693591345, + "loss": 1.0392, + "step": 1921 + }, + { + "epoch": 0.14783478193985078, + "learning_rate": 0.002841111054115815, + "loss": 1.1667, + "step": 1922 + }, + { + "epoch": 0.14791169910006924, + "learning_rate": 0.0028409486605636236, + "loss": 1.4217, + "step": 1923 + }, + { + "epoch": 0.14798861626028767, + "learning_rate": 0.002840786188712042, + "loss": 1.1517, + "step": 1924 + }, + { + "epoch": 0.1480655334205061, + "learning_rate": 0.0028406236385705584, + "loss": 0.9707, + "step": 1925 + }, + { + "epoch": 0.14814245058072456, + "learning_rate": 0.0028404610101486626, + "loss": 1.2975, + "step": 1926 + }, + { + "epoch": 0.148219367740943, + "learning_rate": 0.002840298303455852, + "loss": 1.2556, + "step": 1927 + }, + { + "epoch": 0.14829628490116145, + "learning_rate": 0.0028401355185016263, + "loss": 1.1761, + "step": 1928 + }, + { + "epoch": 0.14837320206137988, + "learning_rate": 0.002839972655295491, + "loss": 0.9831, + "step": 1929 + }, + { + "epoch": 0.14845011922159834, + "learning_rate": 0.002839809713846956, + "loss": 1.1987, + "step": 1930 + }, + { + "epoch": 0.14852703638181677, + "learning_rate": 0.0028396466941655357, + "loss": 1.0997, + "step": 1931 + }, + { + "epoch": 0.14860395354203523, + "learning_rate": 0.002839483596260748, + "loss": 1.0292, + "step": 1932 + }, + { + "epoch": 0.14868087070225366, + "learning_rate": 0.002839320420142118, + "loss": 1.0209, + "step": 1933 + }, + { + "epoch": 0.14875778786247212, + "learning_rate": 0.0028391571658191723, + "loss": 1.2296, + "step": 1934 + }, + { + "epoch": 0.14883470502269056, + "learning_rate": 0.0028389938333014445, + "loss": 1.1078, + "step": 1935 + }, + { + "epoch": 0.14891162218290901, + "learning_rate": 0.002838830422598471, + "loss": 0.9301, + "step": 1936 + }, + { + "epoch": 0.14898853934312745, + "learning_rate": 0.002838666933719794, + "loss": 1.3642, + "step": 1937 + }, + { + "epoch": 0.1490654565033459, + "learning_rate": 0.0028385033666749595, + "loss": 1.2228, + "step": 1938 + }, + { + "epoch": 0.14914237366356434, + "learning_rate": 0.002838339721473518, + "loss": 1.2052, + "step": 1939 + }, + { + "epoch": 0.1492192908237828, + "learning_rate": 0.0028381759981250263, + "loss": 1.1428, + "step": 1940 + }, + { + "epoch": 0.14929620798400123, + "learning_rate": 0.0028380121966390427, + "loss": 1.3212, + "step": 1941 + }, + { + "epoch": 0.1493731251442197, + "learning_rate": 0.0028378483170251327, + "loss": 0.8909, + "step": 1942 + }, + { + "epoch": 0.14945004230443812, + "learning_rate": 0.0028376843592928655, + "loss": 1.0285, + "step": 1943 + }, + { + "epoch": 0.14952695946465658, + "learning_rate": 0.002837520323451814, + "loss": 0.9375, + "step": 1944 + }, + { + "epoch": 0.149603876624875, + "learning_rate": 0.002837356209511557, + "loss": 1.2237, + "step": 1945 + }, + { + "epoch": 0.14968079378509344, + "learning_rate": 0.0028371920174816777, + "loss": 1.1826, + "step": 1946 + }, + { + "epoch": 0.1497577109453119, + "learning_rate": 0.0028370277473717628, + "loss": 1.1488, + "step": 1947 + }, + { + "epoch": 0.14983462810553033, + "learning_rate": 0.0028368633991914037, + "loss": 1.0076, + "step": 1948 + }, + { + "epoch": 0.1499115452657488, + "learning_rate": 0.0028366989729501977, + "loss": 0.9484, + "step": 1949 + }, + { + "epoch": 0.14998846242596722, + "learning_rate": 0.0028365344686577462, + "loss": 1.055, + "step": 1950 + }, + { + "epoch": 0.15006537958618568, + "learning_rate": 0.0028363698863236534, + "loss": 1.0813, + "step": 1951 + }, + { + "epoch": 0.1501422967464041, + "learning_rate": 0.0028362052259575305, + "loss": 1.2597, + "step": 1952 + }, + { + "epoch": 0.15021921390662257, + "learning_rate": 0.002836040487568992, + "loss": 1.1157, + "step": 1953 + }, + { + "epoch": 0.150296131066841, + "learning_rate": 0.002835875671167657, + "loss": 1.2661, + "step": 1954 + }, + { + "epoch": 0.15037304822705946, + "learning_rate": 0.002835710776763149, + "loss": 1.3811, + "step": 1955 + }, + { + "epoch": 0.1504499653872779, + "learning_rate": 0.002835545804365097, + "loss": 1.2374, + "step": 1956 + }, + { + "epoch": 0.15052688254749635, + "learning_rate": 0.0028353807539831335, + "loss": 1.2996, + "step": 1957 + }, + { + "epoch": 0.15060379970771479, + "learning_rate": 0.002835215625626896, + "loss": 1.39, + "step": 1958 + }, + { + "epoch": 0.15068071686793325, + "learning_rate": 0.002835050419306027, + "loss": 0.9618, + "step": 1959 + }, + { + "epoch": 0.15075763402815168, + "learning_rate": 0.0028348851350301725, + "loss": 0.8978, + "step": 1960 + }, + { + "epoch": 0.15083455118837014, + "learning_rate": 0.0028347197728089838, + "loss": 1.002, + "step": 1961 + }, + { + "epoch": 0.15091146834858857, + "learning_rate": 0.0028345543326521163, + "loss": 1.0203, + "step": 1962 + }, + { + "epoch": 0.15098838550880703, + "learning_rate": 0.002834388814569231, + "loss": 1.1797, + "step": 1963 + }, + { + "epoch": 0.15106530266902546, + "learning_rate": 0.0028342232185699914, + "loss": 1.3316, + "step": 1964 + }, + { + "epoch": 0.15114221982924392, + "learning_rate": 0.002834057544664068, + "loss": 0.8223, + "step": 1965 + }, + { + "epoch": 0.15121913698946235, + "learning_rate": 0.002833891792861134, + "loss": 1.2784, + "step": 1966 + }, + { + "epoch": 0.15129605414968078, + "learning_rate": 0.002833725963170868, + "loss": 0.9315, + "step": 1967 + }, + { + "epoch": 0.15137297130989924, + "learning_rate": 0.002833560055602954, + "loss": 0.9381, + "step": 1968 + }, + { + "epoch": 0.15144988847011767, + "learning_rate": 0.0028333940701670774, + "loss": 0.8995, + "step": 1969 + }, + { + "epoch": 0.15152680563033613, + "learning_rate": 0.002833228006872932, + "loss": 1.3886, + "step": 1970 + }, + { + "epoch": 0.15160372279055456, + "learning_rate": 0.0028330618657302136, + "loss": 1.0036, + "step": 1971 + }, + { + "epoch": 0.15168063995077302, + "learning_rate": 0.0028328956467486236, + "loss": 1.2779, + "step": 1972 + }, + { + "epoch": 0.15175755711099145, + "learning_rate": 0.0028327293499378677, + "loss": 1.2057, + "step": 1973 + }, + { + "epoch": 0.1518344742712099, + "learning_rate": 0.002832562975307656, + "loss": 1.0431, + "step": 1974 + }, + { + "epoch": 0.15191139143142834, + "learning_rate": 0.0028323965228677036, + "loss": 1.0746, + "step": 1975 + }, + { + "epoch": 0.1519883085916468, + "learning_rate": 0.00283222999262773, + "loss": 1.0914, + "step": 1976 + }, + { + "epoch": 0.15206522575186524, + "learning_rate": 0.0028320633845974577, + "loss": 1.291, + "step": 1977 + }, + { + "epoch": 0.1521421429120837, + "learning_rate": 0.002831896698786617, + "loss": 1.1502, + "step": 1978 + }, + { + "epoch": 0.15221906007230213, + "learning_rate": 0.00283172993520494, + "loss": 1.2315, + "step": 1979 + }, + { + "epoch": 0.15229597723252059, + "learning_rate": 0.0028315630938621633, + "loss": 1.1341, + "step": 1980 + }, + { + "epoch": 0.15237289439273902, + "learning_rate": 0.0028313961747680307, + "loss": 1.3974, + "step": 1981 + }, + { + "epoch": 0.15244981155295748, + "learning_rate": 0.0028312291779322876, + "loss": 1.0439, + "step": 1982 + }, + { + "epoch": 0.1525267287131759, + "learning_rate": 0.0028310621033646854, + "loss": 1.0077, + "step": 1983 + }, + { + "epoch": 0.15260364587339437, + "learning_rate": 0.00283089495107498, + "loss": 1.6118, + "step": 1984 + }, + { + "epoch": 0.1526805630336128, + "learning_rate": 0.0028307277210729313, + "loss": 0.9113, + "step": 1985 + }, + { + "epoch": 0.15275748019383123, + "learning_rate": 0.002830560413368304, + "loss": 1.3026, + "step": 1986 + }, + { + "epoch": 0.1528343973540497, + "learning_rate": 0.002830393027970868, + "loss": 1.131, + "step": 1987 + }, + { + "epoch": 0.15291131451426812, + "learning_rate": 0.002830225564890396, + "loss": 1.2247, + "step": 1988 + }, + { + "epoch": 0.15298823167448658, + "learning_rate": 0.002830058024136667, + "loss": 1.2465, + "step": 1989 + }, + { + "epoch": 0.153065148834705, + "learning_rate": 0.002829890405719464, + "loss": 1.0329, + "step": 1990 + }, + { + "epoch": 0.15314206599492347, + "learning_rate": 0.0028297227096485736, + "loss": 1.2137, + "step": 1991 + }, + { + "epoch": 0.1532189831551419, + "learning_rate": 0.0028295549359337894, + "loss": 1.1542, + "step": 1992 + }, + { + "epoch": 0.15329590031536036, + "learning_rate": 0.002829387084584906, + "loss": 1.127, + "step": 1993 + }, + { + "epoch": 0.1533728174755788, + "learning_rate": 0.0028292191556117258, + "loss": 0.8747, + "step": 1994 + }, + { + "epoch": 0.15344973463579725, + "learning_rate": 0.002829051149024054, + "loss": 1.1015, + "step": 1995 + }, + { + "epoch": 0.15352665179601568, + "learning_rate": 0.0028288830648317, + "loss": 0.9848, + "step": 1996 + }, + { + "epoch": 0.15360356895623414, + "learning_rate": 0.0028287149030444786, + "loss": 1.3574, + "step": 1997 + }, + { + "epoch": 0.15368048611645257, + "learning_rate": 0.00282854666367221, + "loss": 1.0259, + "step": 1998 + }, + { + "epoch": 0.15375740327667103, + "learning_rate": 0.002828378346724717, + "loss": 1.081, + "step": 1999 + }, + { + "epoch": 0.15383432043688947, + "learning_rate": 0.0028282099522118275, + "loss": 0.9127, + "step": 2000 + }, + { + "epoch": 0.15391123759710792, + "learning_rate": 0.002828041480143375, + "loss": 1.1192, + "step": 2001 + }, + { + "epoch": 0.15398815475732636, + "learning_rate": 0.0028278729305291966, + "loss": 1.2785, + "step": 2002 + }, + { + "epoch": 0.15406507191754482, + "learning_rate": 0.0028277043033791338, + "loss": 1.0425, + "step": 2003 + }, + { + "epoch": 0.15414198907776325, + "learning_rate": 0.0028275355987030333, + "loss": 1.3098, + "step": 2004 + }, + { + "epoch": 0.1542189062379817, + "learning_rate": 0.002827366816510745, + "loss": 1.0905, + "step": 2005 + }, + { + "epoch": 0.15429582339820014, + "learning_rate": 0.0028271979568121252, + "loss": 1.1252, + "step": 2006 + }, + { + "epoch": 0.15437274055841857, + "learning_rate": 0.0028270290196170337, + "loss": 1.1481, + "step": 2007 + }, + { + "epoch": 0.15444965771863703, + "learning_rate": 0.0028268600049353343, + "loss": 1.1522, + "step": 2008 + }, + { + "epoch": 0.15452657487885546, + "learning_rate": 0.002826690912776897, + "loss": 1.1175, + "step": 2009 + }, + { + "epoch": 0.15460349203907392, + "learning_rate": 0.0028265217431515943, + "loss": 0.8834, + "step": 2010 + }, + { + "epoch": 0.15468040919929235, + "learning_rate": 0.0028263524960693044, + "loss": 1.1786, + "step": 2011 + }, + { + "epoch": 0.1547573263595108, + "learning_rate": 0.0028261831715399107, + "loss": 1.085, + "step": 2012 + }, + { + "epoch": 0.15483424351972924, + "learning_rate": 0.0028260137695732984, + "loss": 1.3008, + "step": 2013 + }, + { + "epoch": 0.1549111606799477, + "learning_rate": 0.0028258442901793607, + "loss": 1.0226, + "step": 2014 + }, + { + "epoch": 0.15498807784016613, + "learning_rate": 0.0028256747333679933, + "loss": 1.5447, + "step": 2015 + }, + { + "epoch": 0.1550649950003846, + "learning_rate": 0.0028255050991490963, + "loss": 1.1576, + "step": 2016 + }, + { + "epoch": 0.15514191216060302, + "learning_rate": 0.0028253353875325755, + "loss": 0.6988, + "step": 2017 + }, + { + "epoch": 0.15521882932082148, + "learning_rate": 0.00282516559852834, + "loss": 1.0037, + "step": 2018 + }, + { + "epoch": 0.15529574648103991, + "learning_rate": 0.0028249957321463044, + "loss": 1.1871, + "step": 2019 + }, + { + "epoch": 0.15537266364125837, + "learning_rate": 0.002824825788396387, + "loss": 0.9937, + "step": 2020 + }, + { + "epoch": 0.1554495808014768, + "learning_rate": 0.0028246557672885106, + "loss": 1.0193, + "step": 2021 + }, + { + "epoch": 0.15552649796169526, + "learning_rate": 0.0028244856688326037, + "loss": 1.2292, + "step": 2022 + }, + { + "epoch": 0.1556034151219137, + "learning_rate": 0.0028243154930385986, + "loss": 1.3331, + "step": 2023 + }, + { + "epoch": 0.15568033228213216, + "learning_rate": 0.002824145239916432, + "loss": 1.2292, + "step": 2024 + }, + { + "epoch": 0.1557572494423506, + "learning_rate": 0.002823974909476044, + "loss": 1.4199, + "step": 2025 + }, + { + "epoch": 0.15583416660256905, + "learning_rate": 0.0028238045017273815, + "loss": 1.2479, + "step": 2026 + }, + { + "epoch": 0.15591108376278748, + "learning_rate": 0.0028236340166803947, + "loss": 1.1911, + "step": 2027 + }, + { + "epoch": 0.1559880009230059, + "learning_rate": 0.0028234634543450377, + "loss": 0.9725, + "step": 2028 + }, + { + "epoch": 0.15606491808322437, + "learning_rate": 0.002823292814731271, + "loss": 1.4719, + "step": 2029 + }, + { + "epoch": 0.1561418352434428, + "learning_rate": 0.0028231220978490574, + "loss": 0.932, + "step": 2030 + }, + { + "epoch": 0.15621875240366126, + "learning_rate": 0.002822951303708366, + "loss": 1.1865, + "step": 2031 + }, + { + "epoch": 0.1562956695638797, + "learning_rate": 0.0028227804323191687, + "loss": 1.0766, + "step": 2032 + }, + { + "epoch": 0.15637258672409815, + "learning_rate": 0.0028226094836914435, + "loss": 1.0572, + "step": 2033 + }, + { + "epoch": 0.15644950388431658, + "learning_rate": 0.002822438457835172, + "loss": 1.0668, + "step": 2034 + }, + { + "epoch": 0.15652642104453504, + "learning_rate": 0.002822267354760341, + "loss": 1.2249, + "step": 2035 + }, + { + "epoch": 0.15660333820475347, + "learning_rate": 0.002822096174476941, + "loss": 1.1557, + "step": 2036 + }, + { + "epoch": 0.15668025536497193, + "learning_rate": 0.0028219249169949678, + "loss": 1.1815, + "step": 2037 + }, + { + "epoch": 0.15675717252519036, + "learning_rate": 0.0028217535823244206, + "loss": 0.876, + "step": 2038 + }, + { + "epoch": 0.15683408968540882, + "learning_rate": 0.0028215821704753044, + "loss": 1.365, + "step": 2039 + }, + { + "epoch": 0.15691100684562725, + "learning_rate": 0.002821410681457628, + "loss": 1.1182, + "step": 2040 + }, + { + "epoch": 0.1569879240058457, + "learning_rate": 0.0028212391152814048, + "loss": 0.9862, + "step": 2041 + }, + { + "epoch": 0.15706484116606415, + "learning_rate": 0.0028210674719566526, + "loss": 1.2278, + "step": 2042 + }, + { + "epoch": 0.1571417583262826, + "learning_rate": 0.002820895751493394, + "loss": 1.8615, + "step": 2043 + }, + { + "epoch": 0.15721867548650104, + "learning_rate": 0.002820723953901656, + "loss": 1.2583, + "step": 2044 + }, + { + "epoch": 0.1572955926467195, + "learning_rate": 0.0028205520791914694, + "loss": 0.9442, + "step": 2045 + }, + { + "epoch": 0.15737250980693793, + "learning_rate": 0.0028203801273728713, + "loss": 1.3713, + "step": 2046 + }, + { + "epoch": 0.1574494269671564, + "learning_rate": 0.0028202080984559012, + "loss": 1.2726, + "step": 2047 + }, + { + "epoch": 0.15752634412737482, + "learning_rate": 0.002820035992450604, + "loss": 1.0752, + "step": 2048 + }, + { + "epoch": 0.15760326128759325, + "learning_rate": 0.00281986380936703, + "loss": 1.2511, + "step": 2049 + }, + { + "epoch": 0.1576801784478117, + "learning_rate": 0.002819691549215233, + "loss": 1.3435, + "step": 2050 + }, + { + "epoch": 0.15775709560803014, + "learning_rate": 0.00281951921200527, + "loss": 1.2046, + "step": 2051 + }, + { + "epoch": 0.1578340127682486, + "learning_rate": 0.0028193467977472055, + "loss": 1.1837, + "step": 2052 + }, + { + "epoch": 0.15791092992846703, + "learning_rate": 0.002819174306451107, + "loss": 1.4107, + "step": 2053 + }, + { + "epoch": 0.1579878470886855, + "learning_rate": 0.002819001738127045, + "loss": 1.3604, + "step": 2054 + }, + { + "epoch": 0.15806476424890392, + "learning_rate": 0.002818829092785098, + "loss": 1.2017, + "step": 2055 + }, + { + "epoch": 0.15814168140912238, + "learning_rate": 0.002818656370435345, + "loss": 1.2019, + "step": 2056 + }, + { + "epoch": 0.1582185985693408, + "learning_rate": 0.0028184835710878724, + "loss": 0.9851, + "step": 2057 + }, + { + "epoch": 0.15829551572955927, + "learning_rate": 0.00281831069475277, + "loss": 0.898, + "step": 2058 + }, + { + "epoch": 0.1583724328897777, + "learning_rate": 0.0028181377414401325, + "loss": 0.919, + "step": 2059 + }, + { + "epoch": 0.15844935004999616, + "learning_rate": 0.002817964711160058, + "loss": 1.5117, + "step": 2060 + }, + { + "epoch": 0.1585262672102146, + "learning_rate": 0.0028177916039226506, + "loss": 1.1425, + "step": 2061 + }, + { + "epoch": 0.15860318437043305, + "learning_rate": 0.0028176184197380182, + "loss": 1.2193, + "step": 2062 + }, + { + "epoch": 0.15868010153065149, + "learning_rate": 0.0028174451586162738, + "loss": 0.95, + "step": 2063 + }, + { + "epoch": 0.15875701869086994, + "learning_rate": 0.002817271820567533, + "loss": 1.0872, + "step": 2064 + }, + { + "epoch": 0.15883393585108838, + "learning_rate": 0.002817098405601917, + "loss": 1.1749, + "step": 2065 + }, + { + "epoch": 0.15891085301130684, + "learning_rate": 0.0028169249137295536, + "loss": 1.0426, + "step": 2066 + }, + { + "epoch": 0.15898777017152527, + "learning_rate": 0.0028167513449605715, + "loss": 1.4273, + "step": 2067 + }, + { + "epoch": 0.1590646873317437, + "learning_rate": 0.0028165776993051062, + "loss": 0.7961, + "step": 2068 + }, + { + "epoch": 0.15914160449196216, + "learning_rate": 0.002816403976773297, + "loss": 1.0494, + "step": 2069 + }, + { + "epoch": 0.1592185216521806, + "learning_rate": 0.0028162301773752875, + "loss": 1.3031, + "step": 2070 + }, + { + "epoch": 0.15929543881239905, + "learning_rate": 0.002816056301121226, + "loss": 0.9875, + "step": 2071 + }, + { + "epoch": 0.15937235597261748, + "learning_rate": 0.0028158823480212663, + "loss": 1.3561, + "step": 2072 + }, + { + "epoch": 0.15944927313283594, + "learning_rate": 0.002815708318085564, + "loss": 1.1828, + "step": 2073 + }, + { + "epoch": 0.15952619029305437, + "learning_rate": 0.0028155342113242825, + "loss": 1.1529, + "step": 2074 + }, + { + "epoch": 0.15960310745327283, + "learning_rate": 0.002815360027747587, + "loss": 1.2199, + "step": 2075 + }, + { + "epoch": 0.15968002461349126, + "learning_rate": 0.002815185767365649, + "loss": 1.6978, + "step": 2076 + }, + { + "epoch": 0.15975694177370972, + "learning_rate": 0.0028150114301886426, + "loss": 1.0705, + "step": 2077 + }, + { + "epoch": 0.15983385893392815, + "learning_rate": 0.0028148370162267493, + "loss": 1.4204, + "step": 2078 + }, + { + "epoch": 0.1599107760941466, + "learning_rate": 0.0028146625254901517, + "loss": 1.2823, + "step": 2079 + }, + { + "epoch": 0.15998769325436504, + "learning_rate": 0.0028144879579890393, + "loss": 1.0765, + "step": 2080 + }, + { + "epoch": 0.1600646104145835, + "learning_rate": 0.0028143133137336056, + "loss": 1.034, + "step": 2081 + }, + { + "epoch": 0.16014152757480193, + "learning_rate": 0.002814138592734047, + "loss": 1.1956, + "step": 2082 + }, + { + "epoch": 0.1602184447350204, + "learning_rate": 0.0028139637950005667, + "loss": 1.0665, + "step": 2083 + }, + { + "epoch": 0.16029536189523882, + "learning_rate": 0.002813788920543371, + "loss": 1.1069, + "step": 2084 + }, + { + "epoch": 0.16037227905545728, + "learning_rate": 0.002813613969372671, + "loss": 1.3421, + "step": 2085 + }, + { + "epoch": 0.16044919621567572, + "learning_rate": 0.002813438941498682, + "loss": 1.2892, + "step": 2086 + }, + { + "epoch": 0.16052611337589417, + "learning_rate": 0.002813263836931625, + "loss": 1.2271, + "step": 2087 + }, + { + "epoch": 0.1606030305361126, + "learning_rate": 0.002813088655681724, + "loss": 0.9485, + "step": 2088 + }, + { + "epoch": 0.16067994769633104, + "learning_rate": 0.0028129133977592073, + "loss": 1.8525, + "step": 2089 + }, + { + "epoch": 0.1607568648565495, + "learning_rate": 0.002812738063174309, + "loss": 1.0006, + "step": 2090 + }, + { + "epoch": 0.16083378201676793, + "learning_rate": 0.0028125626519372674, + "loss": 1.1479, + "step": 2091 + }, + { + "epoch": 0.1609106991769864, + "learning_rate": 0.002812387164058325, + "loss": 0.7483, + "step": 2092 + }, + { + "epoch": 0.16098761633720482, + "learning_rate": 0.002812211599547728, + "loss": 1.3081, + "step": 2093 + }, + { + "epoch": 0.16106453349742328, + "learning_rate": 0.002812035958415728, + "loss": 1.1483, + "step": 2094 + }, + { + "epoch": 0.1611414506576417, + "learning_rate": 0.0028118602406725815, + "loss": 1.334, + "step": 2095 + }, + { + "epoch": 0.16121836781786017, + "learning_rate": 0.002811684446328548, + "loss": 1.0362, + "step": 2096 + }, + { + "epoch": 0.1612952849780786, + "learning_rate": 0.0028115085753938926, + "loss": 0.8121, + "step": 2097 + }, + { + "epoch": 0.16137220213829706, + "learning_rate": 0.0028113326278788845, + "loss": 1.4638, + "step": 2098 + }, + { + "epoch": 0.1614491192985155, + "learning_rate": 0.002811156603793798, + "loss": 1.3577, + "step": 2099 + }, + { + "epoch": 0.16152603645873395, + "learning_rate": 0.002810980503148911, + "loss": 1.1596, + "step": 2100 + }, + { + "epoch": 0.16160295361895238, + "learning_rate": 0.0028108043259545057, + "loss": 1.3218, + "step": 2101 + }, + { + "epoch": 0.16167987077917084, + "learning_rate": 0.0028106280722208703, + "loss": 1.068, + "step": 2102 + }, + { + "epoch": 0.16175678793938927, + "learning_rate": 0.0028104517419582954, + "loss": 1.2523, + "step": 2103 + }, + { + "epoch": 0.16183370509960773, + "learning_rate": 0.0028102753351770776, + "loss": 1.2487, + "step": 2104 + }, + { + "epoch": 0.16191062225982616, + "learning_rate": 0.0028100988518875176, + "loss": 1.0539, + "step": 2105 + }, + { + "epoch": 0.16198753942004462, + "learning_rate": 0.0028099222920999205, + "loss": 0.9122, + "step": 2106 + }, + { + "epoch": 0.16206445658026306, + "learning_rate": 0.002809745655824595, + "loss": 1.2692, + "step": 2107 + }, + { + "epoch": 0.16214137374048151, + "learning_rate": 0.002809568943071856, + "loss": 1.0734, + "step": 2108 + }, + { + "epoch": 0.16221829090069995, + "learning_rate": 0.002809392153852022, + "loss": 1.3023, + "step": 2109 + }, + { + "epoch": 0.16229520806091838, + "learning_rate": 0.002809215288175415, + "loss": 1.0414, + "step": 2110 + }, + { + "epoch": 0.16237212522113684, + "learning_rate": 0.0028090383460523624, + "loss": 1.6666, + "step": 2111 + }, + { + "epoch": 0.16244904238135527, + "learning_rate": 0.0028088613274931977, + "loss": 1.0372, + "step": 2112 + }, + { + "epoch": 0.16252595954157373, + "learning_rate": 0.0028086842325082554, + "loss": 1.187, + "step": 2113 + }, + { + "epoch": 0.16260287670179216, + "learning_rate": 0.002808507061107877, + "loss": 1.1738, + "step": 2114 + }, + { + "epoch": 0.16267979386201062, + "learning_rate": 0.0028083298133024072, + "loss": 0.9106, + "step": 2115 + }, + { + "epoch": 0.16275671102222905, + "learning_rate": 0.0028081524891021966, + "loss": 1.1799, + "step": 2116 + }, + { + "epoch": 0.1628336281824475, + "learning_rate": 0.002807975088517599, + "loss": 1.3328, + "step": 2117 + }, + { + "epoch": 0.16291054534266594, + "learning_rate": 0.002807797611558972, + "loss": 1.0491, + "step": 2118 + }, + { + "epoch": 0.1629874625028844, + "learning_rate": 0.00280762005823668, + "loss": 1.1215, + "step": 2119 + }, + { + "epoch": 0.16306437966310283, + "learning_rate": 0.0028074424285610905, + "loss": 0.8615, + "step": 2120 + }, + { + "epoch": 0.1631412968233213, + "learning_rate": 0.0028072647225425747, + "loss": 1.1726, + "step": 2121 + }, + { + "epoch": 0.16321821398353972, + "learning_rate": 0.002807086940191509, + "loss": 1.2703, + "step": 2122 + }, + { + "epoch": 0.16329513114375818, + "learning_rate": 0.0028069090815182746, + "loss": 1.3984, + "step": 2123 + }, + { + "epoch": 0.1633720483039766, + "learning_rate": 0.0028067311465332576, + "loss": 0.8942, + "step": 2124 + }, + { + "epoch": 0.16344896546419507, + "learning_rate": 0.0028065531352468465, + "loss": 1.2872, + "step": 2125 + }, + { + "epoch": 0.1635258826244135, + "learning_rate": 0.002806375047669437, + "loss": 1.2614, + "step": 2126 + }, + { + "epoch": 0.16360279978463196, + "learning_rate": 0.002806196883811426, + "loss": 1.0975, + "step": 2127 + }, + { + "epoch": 0.1636797169448504, + "learning_rate": 0.002806018643683218, + "loss": 1.0155, + "step": 2128 + }, + { + "epoch": 0.16375663410506885, + "learning_rate": 0.0028058403272952204, + "loss": 1.3019, + "step": 2129 + }, + { + "epoch": 0.16383355126528729, + "learning_rate": 0.002805661934657845, + "loss": 1.2465, + "step": 2130 + }, + { + "epoch": 0.16391046842550572, + "learning_rate": 0.002805483465781509, + "loss": 1.0104, + "step": 2131 + }, + { + "epoch": 0.16398738558572418, + "learning_rate": 0.0028053049206766324, + "loss": 0.8387, + "step": 2132 + }, + { + "epoch": 0.1640643027459426, + "learning_rate": 0.002805126299353641, + "loss": 1.1378, + "step": 2133 + }, + { + "epoch": 0.16414121990616107, + "learning_rate": 0.002804947601822965, + "loss": 1.3031, + "step": 2134 + }, + { + "epoch": 0.1642181370663795, + "learning_rate": 0.002804768828095039, + "loss": 1.0428, + "step": 2135 + }, + { + "epoch": 0.16429505422659796, + "learning_rate": 0.0028045899781803008, + "loss": 1.4127, + "step": 2136 + }, + { + "epoch": 0.1643719713868164, + "learning_rate": 0.002804411052089194, + "loss": 1.2058, + "step": 2137 + }, + { + "epoch": 0.16444888854703485, + "learning_rate": 0.0028042320498321673, + "loss": 1.3751, + "step": 2138 + }, + { + "epoch": 0.16452580570725328, + "learning_rate": 0.002804052971419671, + "loss": 0.6894, + "step": 2139 + }, + { + "epoch": 0.16460272286747174, + "learning_rate": 0.002803873816862163, + "loss": 1.0824, + "step": 2140 + }, + { + "epoch": 0.16467964002769017, + "learning_rate": 0.002803694586170104, + "loss": 0.9984, + "step": 2141 + }, + { + "epoch": 0.16475655718790863, + "learning_rate": 0.0028035152793539596, + "loss": 1.0274, + "step": 2142 + }, + { + "epoch": 0.16483347434812706, + "learning_rate": 0.0028033358964241993, + "loss": 1.2265, + "step": 2143 + }, + { + "epoch": 0.16491039150834552, + "learning_rate": 0.0028031564373912983, + "loss": 1.3405, + "step": 2144 + }, + { + "epoch": 0.16498730866856395, + "learning_rate": 0.002802976902265734, + "loss": 1.0912, + "step": 2145 + }, + { + "epoch": 0.1650642258287824, + "learning_rate": 0.0028027972910579907, + "loss": 1.3862, + "step": 2146 + }, + { + "epoch": 0.16514114298900084, + "learning_rate": 0.002802617603778556, + "loss": 1.3538, + "step": 2147 + }, + { + "epoch": 0.1652180601492193, + "learning_rate": 0.0028024378404379216, + "loss": 1.4892, + "step": 2148 + }, + { + "epoch": 0.16529497730943774, + "learning_rate": 0.002802258001046585, + "loss": 1.4131, + "step": 2149 + }, + { + "epoch": 0.16537189446965617, + "learning_rate": 0.002802078085615046, + "loss": 1.2851, + "step": 2150 + }, + { + "epoch": 0.16544881162987463, + "learning_rate": 0.0028018980941538103, + "loss": 1.3294, + "step": 2151 + }, + { + "epoch": 0.16552572879009306, + "learning_rate": 0.0028017180266733882, + "loss": 1.2027, + "step": 2152 + }, + { + "epoch": 0.16560264595031152, + "learning_rate": 0.002801537883184294, + "loss": 1.0872, + "step": 2153 + }, + { + "epoch": 0.16567956311052995, + "learning_rate": 0.0028013576636970465, + "loss": 1.2291, + "step": 2154 + }, + { + "epoch": 0.1657564802707484, + "learning_rate": 0.0028011773682221686, + "loss": 1.332, + "step": 2155 + }, + { + "epoch": 0.16583339743096684, + "learning_rate": 0.0028009969967701886, + "loss": 0.7623, + "step": 2156 + }, + { + "epoch": 0.1659103145911853, + "learning_rate": 0.0028008165493516378, + "loss": 1.2742, + "step": 2157 + }, + { + "epoch": 0.16598723175140373, + "learning_rate": 0.002800636025977053, + "loss": 0.6873, + "step": 2158 + }, + { + "epoch": 0.1660641489116222, + "learning_rate": 0.002800455426656975, + "loss": 1.072, + "step": 2159 + }, + { + "epoch": 0.16614106607184062, + "learning_rate": 0.0028002747514019495, + "loss": 1.1598, + "step": 2160 + }, + { + "epoch": 0.16621798323205908, + "learning_rate": 0.002800094000222526, + "loss": 1.0351, + "step": 2161 + }, + { + "epoch": 0.1662949003922775, + "learning_rate": 0.002799913173129259, + "loss": 1.3774, + "step": 2162 + }, + { + "epoch": 0.16637181755249597, + "learning_rate": 0.0027997322701327067, + "loss": 1.1639, + "step": 2163 + }, + { + "epoch": 0.1664487347127144, + "learning_rate": 0.0027995512912434334, + "loss": 1.3926, + "step": 2164 + }, + { + "epoch": 0.16652565187293286, + "learning_rate": 0.0027993702364720054, + "loss": 1.5027, + "step": 2165 + }, + { + "epoch": 0.1666025690331513, + "learning_rate": 0.002799189105828995, + "loss": 1.4819, + "step": 2166 + }, + { + "epoch": 0.16667948619336975, + "learning_rate": 0.002799007899324979, + "loss": 1.1601, + "step": 2167 + }, + { + "epoch": 0.16675640335358818, + "learning_rate": 0.002798826616970538, + "loss": 1.0782, + "step": 2168 + }, + { + "epoch": 0.16683332051380664, + "learning_rate": 0.002798645258776257, + "loss": 1.1136, + "step": 2169 + }, + { + "epoch": 0.16691023767402507, + "learning_rate": 0.002798463824752726, + "loss": 1.3599, + "step": 2170 + }, + { + "epoch": 0.1669871548342435, + "learning_rate": 0.0027982823149105386, + "loss": 1.3108, + "step": 2171 + }, + { + "epoch": 0.16706407199446197, + "learning_rate": 0.0027981007292602945, + "loss": 0.9383, + "step": 2172 + }, + { + "epoch": 0.1671409891546804, + "learning_rate": 0.0027979190678125957, + "loss": 1.1164, + "step": 2173 + }, + { + "epoch": 0.16721790631489886, + "learning_rate": 0.00279773733057805, + "loss": 1.2451, + "step": 2174 + }, + { + "epoch": 0.1672948234751173, + "learning_rate": 0.0027975555175672687, + "loss": 0.9756, + "step": 2175 + }, + { + "epoch": 0.16737174063533575, + "learning_rate": 0.0027973736287908685, + "loss": 1.1213, + "step": 2176 + }, + { + "epoch": 0.16744865779555418, + "learning_rate": 0.0027971916642594704, + "loss": 1.0953, + "step": 2177 + }, + { + "epoch": 0.16752557495577264, + "learning_rate": 0.002797009623983699, + "loss": 1.499, + "step": 2178 + }, + { + "epoch": 0.16760249211599107, + "learning_rate": 0.0027968275079741844, + "loss": 1.0063, + "step": 2179 + }, + { + "epoch": 0.16767940927620953, + "learning_rate": 0.002796645316241559, + "loss": 1.2216, + "step": 2180 + }, + { + "epoch": 0.16775632643642796, + "learning_rate": 0.0027964630487964633, + "loss": 1.3844, + "step": 2181 + }, + { + "epoch": 0.16783324359664642, + "learning_rate": 0.0027962807056495383, + "loss": 1.2136, + "step": 2182 + }, + { + "epoch": 0.16791016075686485, + "learning_rate": 0.002796098286811432, + "loss": 1.3936, + "step": 2183 + }, + { + "epoch": 0.1679870779170833, + "learning_rate": 0.002795915792292797, + "loss": 1.2363, + "step": 2184 + }, + { + "epoch": 0.16806399507730174, + "learning_rate": 0.0027957332221042867, + "loss": 1.2134, + "step": 2185 + }, + { + "epoch": 0.1681409122375202, + "learning_rate": 0.002795550576256564, + "loss": 1.7939, + "step": 2186 + }, + { + "epoch": 0.16821782939773863, + "learning_rate": 0.0027953678547602934, + "loss": 1.2313, + "step": 2187 + }, + { + "epoch": 0.1682947465579571, + "learning_rate": 0.002795185057626143, + "loss": 1.0828, + "step": 2188 + }, + { + "epoch": 0.16837166371817552, + "learning_rate": 0.0027950021848647875, + "loss": 1.3092, + "step": 2189 + }, + { + "epoch": 0.16844858087839398, + "learning_rate": 0.002794819236486905, + "loss": 1.3984, + "step": 2190 + }, + { + "epoch": 0.16852549803861241, + "learning_rate": 0.0027946362125031783, + "loss": 1.1199, + "step": 2191 + }, + { + "epoch": 0.16860241519883085, + "learning_rate": 0.0027944531129242933, + "loss": 0.9542, + "step": 2192 + }, + { + "epoch": 0.1686793323590493, + "learning_rate": 0.002794269937760942, + "loss": 1.0747, + "step": 2193 + }, + { + "epoch": 0.16875624951926774, + "learning_rate": 0.00279408668702382, + "loss": 1.178, + "step": 2194 + }, + { + "epoch": 0.1688331666794862, + "learning_rate": 0.002793903360723628, + "loss": 1.1669, + "step": 2195 + }, + { + "epoch": 0.16891008383970463, + "learning_rate": 0.0027937199588710703, + "loss": 0.9588, + "step": 2196 + }, + { + "epoch": 0.1689870009999231, + "learning_rate": 0.0027935364814768555, + "loss": 1.1285, + "step": 2197 + }, + { + "epoch": 0.16906391816014152, + "learning_rate": 0.0027933529285516985, + "loss": 0.8729, + "step": 2198 + }, + { + "epoch": 0.16914083532035998, + "learning_rate": 0.002793169300106315, + "loss": 1.1388, + "step": 2199 + }, + { + "epoch": 0.1692177524805784, + "learning_rate": 0.002792985596151429, + "loss": 1.0768, + "step": 2200 + }, + { + "epoch": 0.16929466964079687, + "learning_rate": 0.002792801816697766, + "loss": 0.9755, + "step": 2201 + }, + { + "epoch": 0.1693715868010153, + "learning_rate": 0.002792617961756058, + "loss": 0.8531, + "step": 2202 + }, + { + "epoch": 0.16944850396123376, + "learning_rate": 0.00279243403133704, + "loss": 1.5626, + "step": 2203 + }, + { + "epoch": 0.1695254211214522, + "learning_rate": 0.0027922500254514517, + "loss": 1.0973, + "step": 2204 + }, + { + "epoch": 0.16960233828167065, + "learning_rate": 0.002792065944110038, + "loss": 0.9663, + "step": 2205 + }, + { + "epoch": 0.16967925544188908, + "learning_rate": 0.0027918817873235475, + "loss": 1.0623, + "step": 2206 + }, + { + "epoch": 0.16975617260210754, + "learning_rate": 0.0027916975551027325, + "loss": 1.3159, + "step": 2207 + }, + { + "epoch": 0.16983308976232597, + "learning_rate": 0.0027915132474583514, + "loss": 1.1395, + "step": 2208 + }, + { + "epoch": 0.16991000692254443, + "learning_rate": 0.0027913288644011657, + "loss": 1.2279, + "step": 2209 + }, + { + "epoch": 0.16998692408276286, + "learning_rate": 0.0027911444059419413, + "loss": 1.2001, + "step": 2210 + }, + { + "epoch": 0.17006384124298132, + "learning_rate": 0.0027909598720914502, + "loss": 0.9281, + "step": 2211 + }, + { + "epoch": 0.17014075840319975, + "learning_rate": 0.002790775262860466, + "loss": 0.7931, + "step": 2212 + }, + { + "epoch": 0.17021767556341819, + "learning_rate": 0.0027905905782597703, + "loss": 0.9781, + "step": 2213 + }, + { + "epoch": 0.17029459272363665, + "learning_rate": 0.002790405818300145, + "loss": 0.8806, + "step": 2214 + }, + { + "epoch": 0.17037150988385508, + "learning_rate": 0.002790220982992379, + "loss": 1.2641, + "step": 2215 + }, + { + "epoch": 0.17044842704407354, + "learning_rate": 0.002790036072347265, + "loss": 1.3273, + "step": 2216 + }, + { + "epoch": 0.17052534420429197, + "learning_rate": 0.0027898510863756006, + "loss": 0.8347, + "step": 2217 + }, + { + "epoch": 0.17060226136451043, + "learning_rate": 0.002789666025088187, + "loss": 1.4892, + "step": 2218 + }, + { + "epoch": 0.17067917852472886, + "learning_rate": 0.00278948088849583, + "loss": 0.8616, + "step": 2219 + }, + { + "epoch": 0.17075609568494732, + "learning_rate": 0.00278929567660934, + "loss": 1.0284, + "step": 2220 + }, + { + "epoch": 0.17083301284516575, + "learning_rate": 0.002789110389439532, + "loss": 0.8165, + "step": 2221 + }, + { + "epoch": 0.1709099300053842, + "learning_rate": 0.0027889250269972246, + "loss": 0.7955, + "step": 2222 + }, + { + "epoch": 0.17098684716560264, + "learning_rate": 0.0027887395892932415, + "loss": 1.1538, + "step": 2223 + }, + { + "epoch": 0.1710637643258211, + "learning_rate": 0.0027885540763384112, + "loss": 1.2271, + "step": 2224 + }, + { + "epoch": 0.17114068148603953, + "learning_rate": 0.0027883684881435653, + "loss": 1.0206, + "step": 2225 + }, + { + "epoch": 0.171217598646258, + "learning_rate": 0.002788182824719541, + "loss": 1.3821, + "step": 2226 + }, + { + "epoch": 0.17129451580647642, + "learning_rate": 0.0027879970860771783, + "loss": 1.13, + "step": 2227 + }, + { + "epoch": 0.17137143296669488, + "learning_rate": 0.002787811272227324, + "loss": 1.5313, + "step": 2228 + }, + { + "epoch": 0.1714483501269133, + "learning_rate": 0.002787625383180827, + "loss": 1.2021, + "step": 2229 + }, + { + "epoch": 0.17152526728713177, + "learning_rate": 0.002787439418948542, + "loss": 1.143, + "step": 2230 + }, + { + "epoch": 0.1716021844473502, + "learning_rate": 0.002787253379541328, + "loss": 0.9585, + "step": 2231 + }, + { + "epoch": 0.17167910160756863, + "learning_rate": 0.0027870672649700475, + "loss": 0.9756, + "step": 2232 + }, + { + "epoch": 0.1717560187677871, + "learning_rate": 0.0027868810752455678, + "loss": 1.1284, + "step": 2233 + }, + { + "epoch": 0.17183293592800553, + "learning_rate": 0.0027866948103787608, + "loss": 1.5215, + "step": 2234 + }, + { + "epoch": 0.17190985308822399, + "learning_rate": 0.0027865084703805033, + "loss": 0.7911, + "step": 2235 + }, + { + "epoch": 0.17198677024844242, + "learning_rate": 0.0027863220552616754, + "loss": 1.0903, + "step": 2236 + }, + { + "epoch": 0.17206368740866088, + "learning_rate": 0.0027861355650331617, + "loss": 1.073, + "step": 2237 + }, + { + "epoch": 0.1721406045688793, + "learning_rate": 0.0027859489997058523, + "loss": 1.1637, + "step": 2238 + }, + { + "epoch": 0.17221752172909777, + "learning_rate": 0.0027857623592906404, + "loss": 1.2024, + "step": 2239 + }, + { + "epoch": 0.1722944388893162, + "learning_rate": 0.002785575643798425, + "loss": 1.2597, + "step": 2240 + }, + { + "epoch": 0.17237135604953466, + "learning_rate": 0.0027853888532401072, + "loss": 1.2187, + "step": 2241 + }, + { + "epoch": 0.1724482732097531, + "learning_rate": 0.0027852019876265953, + "loss": 1.2744, + "step": 2242 + }, + { + "epoch": 0.17252519036997155, + "learning_rate": 0.0027850150469687996, + "loss": 1.0699, + "step": 2243 + }, + { + "epoch": 0.17260210753018998, + "learning_rate": 0.002784828031277636, + "loss": 1.2331, + "step": 2244 + }, + { + "epoch": 0.17267902469040844, + "learning_rate": 0.002784640940564025, + "loss": 1.2178, + "step": 2245 + }, + { + "epoch": 0.17275594185062687, + "learning_rate": 0.0027844537748388898, + "loss": 1.3137, + "step": 2246 + }, + { + "epoch": 0.17283285901084533, + "learning_rate": 0.002784266534113161, + "loss": 1.3004, + "step": 2247 + }, + { + "epoch": 0.17290977617106376, + "learning_rate": 0.00278407921839777, + "loss": 1.21, + "step": 2248 + }, + { + "epoch": 0.17298669333128222, + "learning_rate": 0.0027838918277036558, + "loss": 1.0841, + "step": 2249 + }, + { + "epoch": 0.17306361049150065, + "learning_rate": 0.0027837043620417595, + "loss": 1.35, + "step": 2250 + }, + { + "epoch": 0.1731405276517191, + "learning_rate": 0.002783516821423028, + "loss": 0.6665, + "step": 2251 + }, + { + "epoch": 0.17321744481193754, + "learning_rate": 0.0027833292058584117, + "loss": 1.0572, + "step": 2252 + }, + { + "epoch": 0.17329436197215597, + "learning_rate": 0.0027831415153588652, + "loss": 1.0276, + "step": 2253 + }, + { + "epoch": 0.17337127913237443, + "learning_rate": 0.0027829537499353485, + "loss": 1.426, + "step": 2254 + }, + { + "epoch": 0.17344819629259287, + "learning_rate": 0.0027827659095988255, + "loss": 1.1894, + "step": 2255 + }, + { + "epoch": 0.17352511345281132, + "learning_rate": 0.0027825779943602635, + "loss": 1.0175, + "step": 2256 + }, + { + "epoch": 0.17360203061302976, + "learning_rate": 0.0027823900042306363, + "loss": 1.2415, + "step": 2257 + }, + { + "epoch": 0.17367894777324822, + "learning_rate": 0.0027822019392209204, + "loss": 1.3457, + "step": 2258 + }, + { + "epoch": 0.17375586493346665, + "learning_rate": 0.0027820137993420967, + "loss": 1.537, + "step": 2259 + }, + { + "epoch": 0.1738327820936851, + "learning_rate": 0.002781825584605151, + "loss": 1.5206, + "step": 2260 + }, + { + "epoch": 0.17390969925390354, + "learning_rate": 0.0027816372950210744, + "loss": 0.7763, + "step": 2261 + }, + { + "epoch": 0.173986616414122, + "learning_rate": 0.002781448930600859, + "loss": 1.408, + "step": 2262 + }, + { + "epoch": 0.17406353357434043, + "learning_rate": 0.002781260491355506, + "loss": 1.8919, + "step": 2263 + }, + { + "epoch": 0.1741404507345589, + "learning_rate": 0.002781071977296018, + "loss": 1.3851, + "step": 2264 + }, + { + "epoch": 0.17421736789477732, + "learning_rate": 0.002780883388433402, + "loss": 0.706, + "step": 2265 + }, + { + "epoch": 0.17429428505499578, + "learning_rate": 0.002780694724778669, + "loss": 1.2864, + "step": 2266 + }, + { + "epoch": 0.1743712022152142, + "learning_rate": 0.0027805059863428372, + "loss": 1.0764, + "step": 2267 + }, + { + "epoch": 0.17444811937543267, + "learning_rate": 0.0027803171731369263, + "loss": 1.3292, + "step": 2268 + }, + { + "epoch": 0.1745250365356511, + "learning_rate": 0.002780128285171961, + "loss": 1.0075, + "step": 2269 + }, + { + "epoch": 0.17460195369586956, + "learning_rate": 0.0027799393224589714, + "loss": 1.3501, + "step": 2270 + }, + { + "epoch": 0.174678870856088, + "learning_rate": 0.0027797502850089905, + "loss": 0.8623, + "step": 2271 + }, + { + "epoch": 0.17475578801630645, + "learning_rate": 0.0027795611728330567, + "loss": 0.828, + "step": 2272 + }, + { + "epoch": 0.17483270517652488, + "learning_rate": 0.002779371985942213, + "loss": 1.1315, + "step": 2273 + }, + { + "epoch": 0.17490962233674331, + "learning_rate": 0.002779182724347505, + "loss": 1.4622, + "step": 2274 + }, + { + "epoch": 0.17498653949696177, + "learning_rate": 0.0027789933880599848, + "loss": 0.8753, + "step": 2275 + }, + { + "epoch": 0.1750634566571802, + "learning_rate": 0.0027788039770907085, + "loss": 1.1484, + "step": 2276 + }, + { + "epoch": 0.17514037381739866, + "learning_rate": 0.0027786144914507347, + "loss": 1.1718, + "step": 2277 + }, + { + "epoch": 0.1752172909776171, + "learning_rate": 0.0027784249311511282, + "loss": 1.1838, + "step": 2278 + }, + { + "epoch": 0.17529420813783556, + "learning_rate": 0.002778235296202957, + "loss": 1.3006, + "step": 2279 + }, + { + "epoch": 0.175371125298054, + "learning_rate": 0.0027780455866172954, + "loss": 1.0545, + "step": 2280 + }, + { + "epoch": 0.17544804245827245, + "learning_rate": 0.00277785580240522, + "loss": 1.2178, + "step": 2281 + }, + { + "epoch": 0.17552495961849088, + "learning_rate": 0.002777665943577813, + "loss": 0.7611, + "step": 2282 + }, + { + "epoch": 0.17560187677870934, + "learning_rate": 0.002777476010146159, + "loss": 1.3487, + "step": 2283 + }, + { + "epoch": 0.17567879393892777, + "learning_rate": 0.00277728600212135, + "loss": 1.2143, + "step": 2284 + }, + { + "epoch": 0.17575571109914623, + "learning_rate": 0.00277709591951448, + "loss": 1.2719, + "step": 2285 + }, + { + "epoch": 0.17583262825936466, + "learning_rate": 0.0027769057623366483, + "loss": 1.5202, + "step": 2286 + }, + { + "epoch": 0.17590954541958312, + "learning_rate": 0.0027767155305989583, + "loss": 1.0031, + "step": 2287 + }, + { + "epoch": 0.17598646257980155, + "learning_rate": 0.0027765252243125184, + "loss": 1.3378, + "step": 2288 + }, + { + "epoch": 0.17606337974002, + "learning_rate": 0.0027763348434884398, + "loss": 1.3234, + "step": 2289 + }, + { + "epoch": 0.17614029690023844, + "learning_rate": 0.00277614438813784, + "loss": 1.1323, + "step": 2290 + }, + { + "epoch": 0.1762172140604569, + "learning_rate": 0.0027759538582718393, + "loss": 1.294, + "step": 2291 + }, + { + "epoch": 0.17629413122067533, + "learning_rate": 0.0027757632539015626, + "loss": 1.2474, + "step": 2292 + }, + { + "epoch": 0.1763710483808938, + "learning_rate": 0.00277557257503814, + "loss": 0.7762, + "step": 2293 + }, + { + "epoch": 0.17644796554111222, + "learning_rate": 0.0027753818216927056, + "loss": 1.5522, + "step": 2294 + }, + { + "epoch": 0.17652488270133065, + "learning_rate": 0.0027751909938763973, + "loss": 1.964, + "step": 2295 + }, + { + "epoch": 0.1766017998615491, + "learning_rate": 0.002775000091600358, + "loss": 1.2278, + "step": 2296 + }, + { + "epoch": 0.17667871702176755, + "learning_rate": 0.002774809114875735, + "loss": 1.1071, + "step": 2297 + }, + { + "epoch": 0.176755634181986, + "learning_rate": 0.0027746180637136784, + "loss": 1.3481, + "step": 2298 + }, + { + "epoch": 0.17683255134220444, + "learning_rate": 0.002774426938125345, + "loss": 0.9673, + "step": 2299 + }, + { + "epoch": 0.1769094685024229, + "learning_rate": 0.0027742357381218947, + "loss": 0.9776, + "step": 2300 + }, + { + "epoch": 0.17698638566264133, + "learning_rate": 0.0027740444637144915, + "loss": 0.8895, + "step": 2301 + }, + { + "epoch": 0.17706330282285979, + "learning_rate": 0.0027738531149143044, + "loss": 1.5286, + "step": 2302 + }, + { + "epoch": 0.17714021998307822, + "learning_rate": 0.0027736616917325065, + "loss": 1.1316, + "step": 2303 + }, + { + "epoch": 0.17721713714329668, + "learning_rate": 0.002773470194180275, + "loss": 1.0178, + "step": 2304 + }, + { + "epoch": 0.1772940543035151, + "learning_rate": 0.0027732786222687913, + "loss": 1.1816, + "step": 2305 + }, + { + "epoch": 0.17737097146373357, + "learning_rate": 0.002773086976009242, + "loss": 1.172, + "step": 2306 + }, + { + "epoch": 0.177447888623952, + "learning_rate": 0.0027728952554128172, + "loss": 1.0449, + "step": 2307 + }, + { + "epoch": 0.17752480578417046, + "learning_rate": 0.0027727034604907125, + "loss": 0.8095, + "step": 2308 + }, + { + "epoch": 0.1776017229443889, + "learning_rate": 0.0027725115912541256, + "loss": 0.899, + "step": 2309 + }, + { + "epoch": 0.17767864010460735, + "learning_rate": 0.0027723196477142612, + "loss": 1.2357, + "step": 2310 + }, + { + "epoch": 0.17775555726482578, + "learning_rate": 0.0027721276298823265, + "loss": 1.2257, + "step": 2311 + }, + { + "epoch": 0.17783247442504424, + "learning_rate": 0.002771935537769534, + "loss": 1.3586, + "step": 2312 + }, + { + "epoch": 0.17790939158526267, + "learning_rate": 0.0027717433713870996, + "loss": 1.4614, + "step": 2313 + }, + { + "epoch": 0.1779863087454811, + "learning_rate": 0.0027715511307462443, + "loss": 1.0453, + "step": 2314 + }, + { + "epoch": 0.17806322590569956, + "learning_rate": 0.002771358815858193, + "loss": 1.1424, + "step": 2315 + }, + { + "epoch": 0.178140143065918, + "learning_rate": 0.002771166426734176, + "loss": 0.7497, + "step": 2316 + }, + { + "epoch": 0.17821706022613645, + "learning_rate": 0.0027709739633854265, + "loss": 1.0717, + "step": 2317 + }, + { + "epoch": 0.17829397738635488, + "learning_rate": 0.002770781425823182, + "loss": 1.2349, + "step": 2318 + }, + { + "epoch": 0.17837089454657334, + "learning_rate": 0.002770588814058687, + "loss": 1.1313, + "step": 2319 + }, + { + "epoch": 0.17844781170679178, + "learning_rate": 0.0027703961281031857, + "loss": 1.3433, + "step": 2320 + }, + { + "epoch": 0.17852472886701024, + "learning_rate": 0.0027702033679679315, + "loss": 1.1871, + "step": 2321 + }, + { + "epoch": 0.17860164602722867, + "learning_rate": 0.0027700105336641782, + "loss": 1.0887, + "step": 2322 + }, + { + "epoch": 0.17867856318744713, + "learning_rate": 0.0027698176252031865, + "loss": 1.3306, + "step": 2323 + }, + { + "epoch": 0.17875548034766556, + "learning_rate": 0.00276962464259622, + "loss": 1.1755, + "step": 2324 + }, + { + "epoch": 0.17883239750788402, + "learning_rate": 0.0027694315858545477, + "loss": 1.2721, + "step": 2325 + }, + { + "epoch": 0.17890931466810245, + "learning_rate": 0.002769238454989442, + "loss": 1.1679, + "step": 2326 + }, + { + "epoch": 0.1789862318283209, + "learning_rate": 0.0027690452500121806, + "loss": 1.0343, + "step": 2327 + }, + { + "epoch": 0.17906314898853934, + "learning_rate": 0.002768851970934044, + "loss": 0.8905, + "step": 2328 + }, + { + "epoch": 0.1791400661487578, + "learning_rate": 0.0027686586177663184, + "loss": 1.1715, + "step": 2329 + }, + { + "epoch": 0.17921698330897623, + "learning_rate": 0.0027684651905202943, + "loss": 1.1536, + "step": 2330 + }, + { + "epoch": 0.1792939004691947, + "learning_rate": 0.0027682716892072656, + "loss": 1.1649, + "step": 2331 + }, + { + "epoch": 0.17937081762941312, + "learning_rate": 0.0027680781138385313, + "loss": 1.1529, + "step": 2332 + }, + { + "epoch": 0.17944773478963158, + "learning_rate": 0.0027678844644253944, + "loss": 0.8628, + "step": 2333 + }, + { + "epoch": 0.17952465194985, + "learning_rate": 0.002767690740979162, + "loss": 0.9465, + "step": 2334 + }, + { + "epoch": 0.17960156911006844, + "learning_rate": 0.0027674969435111463, + "loss": 1.1379, + "step": 2335 + }, + { + "epoch": 0.1796784862702869, + "learning_rate": 0.002767303072032663, + "loss": 1.0461, + "step": 2336 + }, + { + "epoch": 0.17975540343050533, + "learning_rate": 0.0027671091265550323, + "loss": 1.126, + "step": 2337 + }, + { + "epoch": 0.1798323205907238, + "learning_rate": 0.00276691510708958, + "loss": 1.2887, + "step": 2338 + }, + { + "epoch": 0.17990923775094222, + "learning_rate": 0.0027667210136476336, + "loss": 0.9879, + "step": 2339 + }, + { + "epoch": 0.17998615491116068, + "learning_rate": 0.002766526846240527, + "loss": 1.0731, + "step": 2340 + }, + { + "epoch": 0.18006307207137912, + "learning_rate": 0.002766332604879598, + "loss": 1.0854, + "step": 2341 + }, + { + "epoch": 0.18013998923159757, + "learning_rate": 0.0027661382895761886, + "loss": 1.3078, + "step": 2342 + }, + { + "epoch": 0.180216906391816, + "learning_rate": 0.0027659439003416453, + "loss": 1.3538, + "step": 2343 + }, + { + "epoch": 0.18029382355203447, + "learning_rate": 0.0027657494371873172, + "loss": 1.092, + "step": 2344 + }, + { + "epoch": 0.1803707407122529, + "learning_rate": 0.002765554900124561, + "loss": 1.2404, + "step": 2345 + }, + { + "epoch": 0.18044765787247136, + "learning_rate": 0.0027653602891647357, + "loss": 1.1784, + "step": 2346 + }, + { + "epoch": 0.1805245750326898, + "learning_rate": 0.0027651656043192036, + "loss": 1.2711, + "step": 2347 + }, + { + "epoch": 0.18060149219290825, + "learning_rate": 0.002764970845599334, + "loss": 0.9122, + "step": 2348 + }, + { + "epoch": 0.18067840935312668, + "learning_rate": 0.002764776013016498, + "loss": 1.0016, + "step": 2349 + }, + { + "epoch": 0.18075532651334514, + "learning_rate": 0.0027645811065820725, + "loss": 1.0989, + "step": 2350 + }, + { + "epoch": 0.18083224367356357, + "learning_rate": 0.0027643861263074387, + "loss": 1.3739, + "step": 2351 + }, + { + "epoch": 0.18090916083378203, + "learning_rate": 0.002764191072203981, + "loss": 1.2372, + "step": 2352 + }, + { + "epoch": 0.18098607799400046, + "learning_rate": 0.002763995944283089, + "loss": 0.8442, + "step": 2353 + }, + { + "epoch": 0.18106299515421892, + "learning_rate": 0.0027638007425561566, + "loss": 0.8965, + "step": 2354 + }, + { + "epoch": 0.18113991231443735, + "learning_rate": 0.0027636054670345813, + "loss": 0.8469, + "step": 2355 + }, + { + "epoch": 0.18121682947465578, + "learning_rate": 0.002763410117729767, + "loss": 1.1008, + "step": 2356 + }, + { + "epoch": 0.18129374663487424, + "learning_rate": 0.002763214694653118, + "loss": 1.0531, + "step": 2357 + }, + { + "epoch": 0.18137066379509267, + "learning_rate": 0.002763019197816047, + "loss": 0.9141, + "step": 2358 + }, + { + "epoch": 0.18144758095531113, + "learning_rate": 0.0027628236272299687, + "loss": 0.9273, + "step": 2359 + }, + { + "epoch": 0.18152449811552956, + "learning_rate": 0.002762627982906303, + "loss": 1.2542, + "step": 2360 + }, + { + "epoch": 0.18160141527574802, + "learning_rate": 0.0027624322648564727, + "loss": 1.2319, + "step": 2361 + }, + { + "epoch": 0.18167833243596646, + "learning_rate": 0.0027622364730919076, + "loss": 0.8684, + "step": 2362 + }, + { + "epoch": 0.18175524959618491, + "learning_rate": 0.002762040607624039, + "loss": 1.0326, + "step": 2363 + }, + { + "epoch": 0.18183216675640335, + "learning_rate": 0.0027618446684643046, + "loss": 1.1698, + "step": 2364 + }, + { + "epoch": 0.1819090839166218, + "learning_rate": 0.002761648655624144, + "loss": 1.4721, + "step": 2365 + }, + { + "epoch": 0.18198600107684024, + "learning_rate": 0.002761452569115004, + "loss": 1.4152, + "step": 2366 + }, + { + "epoch": 0.1820629182370587, + "learning_rate": 0.002761256408948334, + "loss": 1.2847, + "step": 2367 + }, + { + "epoch": 0.18213983539727713, + "learning_rate": 0.002761060175135588, + "loss": 1.0758, + "step": 2368 + }, + { + "epoch": 0.1822167525574956, + "learning_rate": 0.002760863867688224, + "loss": 1.4763, + "step": 2369 + }, + { + "epoch": 0.18229366971771402, + "learning_rate": 0.002760667486617704, + "loss": 0.9241, + "step": 2370 + }, + { + "epoch": 0.18237058687793248, + "learning_rate": 0.002760471031935497, + "loss": 1.1661, + "step": 2371 + }, + { + "epoch": 0.1824475040381509, + "learning_rate": 0.0027602745036530716, + "loss": 0.9507, + "step": 2372 + }, + { + "epoch": 0.18252442119836937, + "learning_rate": 0.002760077901781905, + "loss": 1.3158, + "step": 2373 + }, + { + "epoch": 0.1826013383585878, + "learning_rate": 0.0027598812263334763, + "loss": 1.2181, + "step": 2374 + }, + { + "epoch": 0.18267825551880623, + "learning_rate": 0.0027596844773192697, + "loss": 1.3014, + "step": 2375 + }, + { + "epoch": 0.1827551726790247, + "learning_rate": 0.0027594876547507737, + "loss": 1.0981, + "step": 2376 + }, + { + "epoch": 0.18283208983924312, + "learning_rate": 0.0027592907586394813, + "loss": 1.2466, + "step": 2377 + }, + { + "epoch": 0.18290900699946158, + "learning_rate": 0.002759093788996889, + "loss": 1.2532, + "step": 2378 + }, + { + "epoch": 0.18298592415968, + "learning_rate": 0.0027588967458344977, + "loss": 1.1043, + "step": 2379 + }, + { + "epoch": 0.18306284131989847, + "learning_rate": 0.002758699629163814, + "loss": 1.0542, + "step": 2380 + }, + { + "epoch": 0.1831397584801169, + "learning_rate": 0.002758502438996347, + "loss": 1.0792, + "step": 2381 + }, + { + "epoch": 0.18321667564033536, + "learning_rate": 0.0027583051753436107, + "loss": 1.3526, + "step": 2382 + }, + { + "epoch": 0.1832935928005538, + "learning_rate": 0.0027581078382171243, + "loss": 0.8125, + "step": 2383 + }, + { + "epoch": 0.18337050996077225, + "learning_rate": 0.0027579104276284094, + "loss": 1.2082, + "step": 2384 + }, + { + "epoch": 0.18344742712099069, + "learning_rate": 0.002757712943588994, + "loss": 1.0168, + "step": 2385 + }, + { + "epoch": 0.18352434428120915, + "learning_rate": 0.002757515386110409, + "loss": 1.36, + "step": 2386 + }, + { + "epoch": 0.18360126144142758, + "learning_rate": 0.002757317755204191, + "loss": 1.3223, + "step": 2387 + }, + { + "epoch": 0.18367817860164604, + "learning_rate": 0.0027571200508818778, + "loss": 1.367, + "step": 2388 + }, + { + "epoch": 0.18375509576186447, + "learning_rate": 0.0027569222731550153, + "loss": 1.4544, + "step": 2389 + }, + { + "epoch": 0.18383201292208293, + "learning_rate": 0.0027567244220351506, + "loss": 1.2253, + "step": 2390 + }, + { + "epoch": 0.18390893008230136, + "learning_rate": 0.002756526497533838, + "loss": 1.4035, + "step": 2391 + }, + { + "epoch": 0.18398584724251982, + "learning_rate": 0.002756328499662633, + "loss": 0.9734, + "step": 2392 + }, + { + "epoch": 0.18406276440273825, + "learning_rate": 0.0027561304284330983, + "loss": 1.2042, + "step": 2393 + }, + { + "epoch": 0.1841396815629567, + "learning_rate": 0.0027559322838567983, + "loss": 1.3424, + "step": 2394 + }, + { + "epoch": 0.18421659872317514, + "learning_rate": 0.002755734065945303, + "loss": 0.8253, + "step": 2395 + }, + { + "epoch": 0.18429351588339357, + "learning_rate": 0.0027555357747101874, + "loss": 1.2893, + "step": 2396 + }, + { + "epoch": 0.18437043304361203, + "learning_rate": 0.00275533741016303, + "loss": 1.082, + "step": 2397 + }, + { + "epoch": 0.18444735020383046, + "learning_rate": 0.002755138972315412, + "loss": 1.2415, + "step": 2398 + }, + { + "epoch": 0.18452426736404892, + "learning_rate": 0.002754940461178922, + "loss": 1.0947, + "step": 2399 + }, + { + "epoch": 0.18460118452426735, + "learning_rate": 0.00275474187676515, + "loss": 1.0091, + "step": 2400 + }, + { + "epoch": 0.1846781016844858, + "learning_rate": 0.0027545432190856922, + "loss": 1.1793, + "step": 2401 + }, + { + "epoch": 0.18475501884470424, + "learning_rate": 0.0027543444881521486, + "loss": 1.3007, + "step": 2402 + }, + { + "epoch": 0.1848319360049227, + "learning_rate": 0.002754145683976123, + "loss": 1.3381, + "step": 2403 + }, + { + "epoch": 0.18490885316514113, + "learning_rate": 0.002753946806569224, + "loss": 1.0557, + "step": 2404 + }, + { + "epoch": 0.1849857703253596, + "learning_rate": 0.002753747855943064, + "loss": 1.3429, + "step": 2405 + }, + { + "epoch": 0.18506268748557803, + "learning_rate": 0.00275354883210926, + "loss": 0.9561, + "step": 2406 + }, + { + "epoch": 0.18513960464579649, + "learning_rate": 0.0027533497350794335, + "loss": 1.0618, + "step": 2407 + }, + { + "epoch": 0.18521652180601492, + "learning_rate": 0.0027531505648652095, + "loss": 1.359, + "step": 2408 + }, + { + "epoch": 0.18529343896623338, + "learning_rate": 0.0027529513214782186, + "loss": 1.0228, + "step": 2409 + }, + { + "epoch": 0.1853703561264518, + "learning_rate": 0.0027527520049300935, + "loss": 1.1886, + "step": 2410 + }, + { + "epoch": 0.18544727328667027, + "learning_rate": 0.0027525526152324736, + "loss": 1.165, + "step": 2411 + }, + { + "epoch": 0.1855241904468887, + "learning_rate": 0.002752353152397001, + "loss": 0.9727, + "step": 2412 + }, + { + "epoch": 0.18560110760710716, + "learning_rate": 0.0027521536164353227, + "loss": 0.8979, + "step": 2413 + }, + { + "epoch": 0.1856780247673256, + "learning_rate": 0.00275195400735909, + "loss": 0.971, + "step": 2414 + }, + { + "epoch": 0.18575494192754405, + "learning_rate": 0.002751754325179958, + "loss": 1.3337, + "step": 2415 + }, + { + "epoch": 0.18583185908776248, + "learning_rate": 0.0027515545699095863, + "loss": 0.9449, + "step": 2416 + }, + { + "epoch": 0.1859087762479809, + "learning_rate": 0.002751354741559639, + "loss": 1.4593, + "step": 2417 + }, + { + "epoch": 0.18598569340819937, + "learning_rate": 0.002751154840141784, + "loss": 1.0102, + "step": 2418 + }, + { + "epoch": 0.1860626105684178, + "learning_rate": 0.002750954865667694, + "loss": 1.1943, + "step": 2419 + }, + { + "epoch": 0.18613952772863626, + "learning_rate": 0.0027507548181490457, + "loss": 1.2264, + "step": 2420 + }, + { + "epoch": 0.1862164448888547, + "learning_rate": 0.0027505546975975202, + "loss": 1.3027, + "step": 2421 + }, + { + "epoch": 0.18629336204907315, + "learning_rate": 0.0027503545040248024, + "loss": 1.3922, + "step": 2422 + }, + { + "epoch": 0.18637027920929158, + "learning_rate": 0.002750154237442582, + "loss": 1.252, + "step": 2423 + }, + { + "epoch": 0.18644719636951004, + "learning_rate": 0.0027499538978625533, + "loss": 1.1408, + "step": 2424 + }, + { + "epoch": 0.18652411352972847, + "learning_rate": 0.0027497534852964135, + "loss": 1.3712, + "step": 2425 + }, + { + "epoch": 0.18660103068994693, + "learning_rate": 0.002749552999755865, + "loss": 1.067, + "step": 2426 + }, + { + "epoch": 0.18667794785016537, + "learning_rate": 0.002749352441252615, + "loss": 1.0006, + "step": 2427 + }, + { + "epoch": 0.18675486501038382, + "learning_rate": 0.0027491518097983733, + "loss": 1.1334, + "step": 2428 + }, + { + "epoch": 0.18683178217060226, + "learning_rate": 0.002748951105404856, + "loss": 0.9525, + "step": 2429 + }, + { + "epoch": 0.18690869933082072, + "learning_rate": 0.0027487503280837823, + "loss": 0.7266, + "step": 2430 + }, + { + "epoch": 0.18698561649103915, + "learning_rate": 0.0027485494778468752, + "loss": 1.1159, + "step": 2431 + }, + { + "epoch": 0.1870625336512576, + "learning_rate": 0.002748348554705863, + "loss": 1.356, + "step": 2432 + }, + { + "epoch": 0.18713945081147604, + "learning_rate": 0.0027481475586724775, + "loss": 1.0046, + "step": 2433 + }, + { + "epoch": 0.1872163679716945, + "learning_rate": 0.0027479464897584554, + "loss": 1.0578, + "step": 2434 + }, + { + "epoch": 0.18729328513191293, + "learning_rate": 0.002747745347975537, + "loss": 0.8953, + "step": 2435 + }, + { + "epoch": 0.1873702022921314, + "learning_rate": 0.002747544133335468, + "loss": 1.1441, + "step": 2436 + }, + { + "epoch": 0.18744711945234982, + "learning_rate": 0.0027473428458499967, + "loss": 1.2491, + "step": 2437 + }, + { + "epoch": 0.18752403661256825, + "learning_rate": 0.002747141485530876, + "loss": 1.2872, + "step": 2438 + }, + { + "epoch": 0.1876009537727867, + "learning_rate": 0.002746940052389865, + "loss": 1.0918, + "step": 2439 + }, + { + "epoch": 0.18767787093300514, + "learning_rate": 0.0027467385464387244, + "loss": 1.0533, + "step": 2440 + }, + { + "epoch": 0.1877547880932236, + "learning_rate": 0.0027465369676892204, + "loss": 1.1003, + "step": 2441 + }, + { + "epoch": 0.18783170525344203, + "learning_rate": 0.0027463353161531244, + "loss": 1.1003, + "step": 2442 + }, + { + "epoch": 0.1879086224136605, + "learning_rate": 0.00274613359184221, + "loss": 1.1075, + "step": 2443 + }, + { + "epoch": 0.18798553957387892, + "learning_rate": 0.002745931794768257, + "loss": 0.9396, + "step": 2444 + }, + { + "epoch": 0.18806245673409738, + "learning_rate": 0.0027457299249430476, + "loss": 1.3618, + "step": 2445 + }, + { + "epoch": 0.18813937389431581, + "learning_rate": 0.0027455279823783694, + "loss": 1.0066, + "step": 2446 + }, + { + "epoch": 0.18821629105453427, + "learning_rate": 0.0027453259670860144, + "loss": 1.2253, + "step": 2447 + }, + { + "epoch": 0.1882932082147527, + "learning_rate": 0.002745123879077779, + "loss": 1.2503, + "step": 2448 + }, + { + "epoch": 0.18837012537497116, + "learning_rate": 0.0027449217183654617, + "loss": 0.852, + "step": 2449 + }, + { + "epoch": 0.1884470425351896, + "learning_rate": 0.0027447194849608684, + "loss": 1.2111, + "step": 2450 + }, + { + "epoch": 0.18852395969540806, + "learning_rate": 0.0027445171788758067, + "loss": 0.8599, + "step": 2451 + }, + { + "epoch": 0.1886008768556265, + "learning_rate": 0.0027443148001220904, + "loss": 1.1196, + "step": 2452 + }, + { + "epoch": 0.18867779401584495, + "learning_rate": 0.002744112348711536, + "loss": 0.9139, + "step": 2453 + }, + { + "epoch": 0.18875471117606338, + "learning_rate": 0.002743909824655965, + "loss": 1.205, + "step": 2454 + }, + { + "epoch": 0.18883162833628184, + "learning_rate": 0.002743707227967203, + "loss": 1.1818, + "step": 2455 + }, + { + "epoch": 0.18890854549650027, + "learning_rate": 0.0027435045586570793, + "loss": 1.8483, + "step": 2456 + }, + { + "epoch": 0.1889854626567187, + "learning_rate": 0.002743301816737429, + "loss": 0.9127, + "step": 2457 + }, + { + "epoch": 0.18906237981693716, + "learning_rate": 0.0027430990022200897, + "loss": 1.0292, + "step": 2458 + }, + { + "epoch": 0.1891392969771556, + "learning_rate": 0.0027428961151169046, + "loss": 0.9004, + "step": 2459 + }, + { + "epoch": 0.18921621413737405, + "learning_rate": 0.0027426931554397193, + "loss": 1.0631, + "step": 2460 + }, + { + "epoch": 0.18929313129759248, + "learning_rate": 0.002742490123200386, + "loss": 1.1435, + "step": 2461 + }, + { + "epoch": 0.18937004845781094, + "learning_rate": 0.0027422870184107594, + "loss": 1.2402, + "step": 2462 + }, + { + "epoch": 0.18944696561802937, + "learning_rate": 0.002742083841082699, + "loss": 1.1791, + "step": 2463 + }, + { + "epoch": 0.18952388277824783, + "learning_rate": 0.002741880591228069, + "loss": 1.0982, + "step": 2464 + }, + { + "epoch": 0.18960079993846626, + "learning_rate": 0.0027416772688587366, + "loss": 1.0543, + "step": 2465 + }, + { + "epoch": 0.18967771709868472, + "learning_rate": 0.0027414738739865744, + "loss": 1.6758, + "step": 2466 + }, + { + "epoch": 0.18975463425890315, + "learning_rate": 0.0027412704066234594, + "loss": 1.0502, + "step": 2467 + }, + { + "epoch": 0.1898315514191216, + "learning_rate": 0.0027410668667812712, + "loss": 1.0975, + "step": 2468 + }, + { + "epoch": 0.18990846857934005, + "learning_rate": 0.0027408632544718954, + "loss": 1.3923, + "step": 2469 + }, + { + "epoch": 0.1899853857395585, + "learning_rate": 0.002740659569707221, + "loss": 1.1496, + "step": 2470 + }, + { + "epoch": 0.19006230289977694, + "learning_rate": 0.0027404558124991415, + "loss": 1.1202, + "step": 2471 + }, + { + "epoch": 0.1901392200599954, + "learning_rate": 0.0027402519828595537, + "loss": 1.1255, + "step": 2472 + }, + { + "epoch": 0.19021613722021383, + "learning_rate": 0.002740048080800361, + "loss": 0.8023, + "step": 2473 + }, + { + "epoch": 0.19029305438043229, + "learning_rate": 0.002739844106333468, + "loss": 1.0702, + "step": 2474 + }, + { + "epoch": 0.19036997154065072, + "learning_rate": 0.0027396400594707853, + "loss": 1.4331, + "step": 2475 + }, + { + "epoch": 0.19044688870086918, + "learning_rate": 0.002739435940224228, + "loss": 0.8715, + "step": 2476 + }, + { + "epoch": 0.1905238058610876, + "learning_rate": 0.0027392317486057147, + "loss": 1.0336, + "step": 2477 + }, + { + "epoch": 0.19060072302130604, + "learning_rate": 0.0027390274846271676, + "loss": 1.3031, + "step": 2478 + }, + { + "epoch": 0.1906776401815245, + "learning_rate": 0.0027388231483005143, + "loss": 1.8521, + "step": 2479 + }, + { + "epoch": 0.19075455734174293, + "learning_rate": 0.0027386187396376866, + "loss": 1.0224, + "step": 2480 + }, + { + "epoch": 0.1908314745019614, + "learning_rate": 0.00273841425865062, + "loss": 1.1065, + "step": 2481 + }, + { + "epoch": 0.19090839166217982, + "learning_rate": 0.0027382097053512537, + "loss": 1.4644, + "step": 2482 + }, + { + "epoch": 0.19098530882239828, + "learning_rate": 0.002738005079751532, + "loss": 1.2938, + "step": 2483 + }, + { + "epoch": 0.1910622259826167, + "learning_rate": 0.0027378003818634045, + "loss": 1.2478, + "step": 2484 + }, + { + "epoch": 0.19113914314283517, + "learning_rate": 0.002737595611698822, + "loss": 1.1379, + "step": 2485 + }, + { + "epoch": 0.1912160603030536, + "learning_rate": 0.0027373907692697422, + "loss": 1.4759, + "step": 2486 + }, + { + "epoch": 0.19129297746327206, + "learning_rate": 0.0027371858545881256, + "loss": 0.9392, + "step": 2487 + }, + { + "epoch": 0.1913698946234905, + "learning_rate": 0.0027369808676659378, + "loss": 1.0521, + "step": 2488 + }, + { + "epoch": 0.19144681178370895, + "learning_rate": 0.002736775808515148, + "loss": 1.143, + "step": 2489 + }, + { + "epoch": 0.19152372894392738, + "learning_rate": 0.0027365706771477296, + "loss": 1.0759, + "step": 2490 + }, + { + "epoch": 0.19160064610414584, + "learning_rate": 0.00273636547357566, + "loss": 1.4833, + "step": 2491 + }, + { + "epoch": 0.19167756326436428, + "learning_rate": 0.002736160197810923, + "loss": 1.121, + "step": 2492 + }, + { + "epoch": 0.19175448042458274, + "learning_rate": 0.0027359548498655036, + "loss": 0.9887, + "step": 2493 + }, + { + "epoch": 0.19183139758480117, + "learning_rate": 0.002735749429751392, + "loss": 1.022, + "step": 2494 + }, + { + "epoch": 0.19190831474501963, + "learning_rate": 0.0027355439374805836, + "loss": 1.4389, + "step": 2495 + }, + { + "epoch": 0.19198523190523806, + "learning_rate": 0.002735338373065077, + "loss": 0.9638, + "step": 2496 + }, + { + "epoch": 0.19206214906545652, + "learning_rate": 0.002735132736516875, + "loss": 1.2374, + "step": 2497 + }, + { + "epoch": 0.19213906622567495, + "learning_rate": 0.002734927027847986, + "loss": 1.0329, + "step": 2498 + }, + { + "epoch": 0.19221598338589338, + "learning_rate": 0.0027347212470704204, + "loss": 0.9132, + "step": 2499 + }, + { + "epoch": 0.19229290054611184, + "learning_rate": 0.0027345153941961946, + "loss": 1.1657, + "step": 2500 + }, + { + "epoch": 0.19236981770633027, + "learning_rate": 0.002734309469237328, + "loss": 1.459, + "step": 2501 + }, + { + "epoch": 0.19244673486654873, + "learning_rate": 0.0027341034722058455, + "loss": 1.1164, + "step": 2502 + }, + { + "epoch": 0.19252365202676716, + "learning_rate": 0.0027338974031137745, + "loss": 1.3067, + "step": 2503 + }, + { + "epoch": 0.19260056918698562, + "learning_rate": 0.002733691261973149, + "loss": 1.2746, + "step": 2504 + }, + { + "epoch": 0.19267748634720405, + "learning_rate": 0.0027334850487960045, + "loss": 0.8684, + "step": 2505 + }, + { + "epoch": 0.1927544035074225, + "learning_rate": 0.0027332787635943826, + "loss": 1.1647, + "step": 2506 + }, + { + "epoch": 0.19283132066764094, + "learning_rate": 0.0027330724063803287, + "loss": 1.6763, + "step": 2507 + }, + { + "epoch": 0.1929082378278594, + "learning_rate": 0.0027328659771658913, + "loss": 0.8732, + "step": 2508 + }, + { + "epoch": 0.19298515498807783, + "learning_rate": 0.0027326594759631254, + "loss": 1.2196, + "step": 2509 + }, + { + "epoch": 0.1930620721482963, + "learning_rate": 0.0027324529027840873, + "loss": 1.0564, + "step": 2510 + }, + { + "epoch": 0.19313898930851472, + "learning_rate": 0.00273224625764084, + "loss": 0.8968, + "step": 2511 + }, + { + "epoch": 0.19321590646873318, + "learning_rate": 0.0027320395405454495, + "loss": 1.2223, + "step": 2512 + }, + { + "epoch": 0.19329282362895162, + "learning_rate": 0.002731832751509986, + "loss": 0.7521, + "step": 2513 + }, + { + "epoch": 0.19336974078917007, + "learning_rate": 0.002731625890546525, + "loss": 1.1269, + "step": 2514 + }, + { + "epoch": 0.1934466579493885, + "learning_rate": 0.0027314189576671446, + "loss": 1.0745, + "step": 2515 + }, + { + "epoch": 0.19352357510960697, + "learning_rate": 0.0027312119528839275, + "loss": 1.1512, + "step": 2516 + }, + { + "epoch": 0.1936004922698254, + "learning_rate": 0.0027310048762089617, + "loss": 1.1137, + "step": 2517 + }, + { + "epoch": 0.19367740943004386, + "learning_rate": 0.002730797727654338, + "loss": 1.2492, + "step": 2518 + }, + { + "epoch": 0.1937543265902623, + "learning_rate": 0.002730590507232152, + "loss": 1.2935, + "step": 2519 + }, + { + "epoch": 0.19383124375048072, + "learning_rate": 0.002730383214954504, + "loss": 1.3263, + "step": 2520 + }, + { + "epoch": 0.19390816091069918, + "learning_rate": 0.002730175850833498, + "loss": 1.1364, + "step": 2521 + }, + { + "epoch": 0.1939850780709176, + "learning_rate": 0.002729968414881242, + "loss": 1.0328, + "step": 2522 + }, + { + "epoch": 0.19406199523113607, + "learning_rate": 0.0027297609071098485, + "loss": 1.3088, + "step": 2523 + }, + { + "epoch": 0.1941389123913545, + "learning_rate": 0.002729553327531434, + "loss": 1.1823, + "step": 2524 + }, + { + "epoch": 0.19421582955157296, + "learning_rate": 0.002729345676158119, + "loss": 1.1516, + "step": 2525 + }, + { + "epoch": 0.1942927467117914, + "learning_rate": 0.0027291379530020293, + "loss": 1.0633, + "step": 2526 + }, + { + "epoch": 0.19436966387200985, + "learning_rate": 0.0027289301580752925, + "loss": 1.0713, + "step": 2527 + }, + { + "epoch": 0.19444658103222828, + "learning_rate": 0.002728722291390044, + "loss": 1.4266, + "step": 2528 + }, + { + "epoch": 0.19452349819244674, + "learning_rate": 0.00272851435295842, + "loss": 0.9635, + "step": 2529 + }, + { + "epoch": 0.19460041535266517, + "learning_rate": 0.0027283063427925624, + "loss": 1.653, + "step": 2530 + }, + { + "epoch": 0.19467733251288363, + "learning_rate": 0.0027280982609046178, + "loss": 0.9572, + "step": 2531 + }, + { + "epoch": 0.19475424967310206, + "learning_rate": 0.0027278901073067353, + "loss": 1.4323, + "step": 2532 + }, + { + "epoch": 0.19483116683332052, + "learning_rate": 0.00272768188201107, + "loss": 1.4142, + "step": 2533 + }, + { + "epoch": 0.19490808399353896, + "learning_rate": 0.00272747358502978, + "loss": 1.173, + "step": 2534 + }, + { + "epoch": 0.19498500115375741, + "learning_rate": 0.002727265216375029, + "loss": 1.0115, + "step": 2535 + }, + { + "epoch": 0.19506191831397585, + "learning_rate": 0.0027270567760589825, + "loss": 1.2721, + "step": 2536 + }, + { + "epoch": 0.1951388354741943, + "learning_rate": 0.0027268482640938115, + "loss": 1.1139, + "step": 2537 + }, + { + "epoch": 0.19521575263441274, + "learning_rate": 0.0027266396804916924, + "loss": 1.286, + "step": 2538 + }, + { + "epoch": 0.19529266979463117, + "learning_rate": 0.0027264310252648034, + "loss": 0.7293, + "step": 2539 + }, + { + "epoch": 0.19536958695484963, + "learning_rate": 0.002726222298425329, + "loss": 1.8487, + "step": 2540 + }, + { + "epoch": 0.19544650411506806, + "learning_rate": 0.002726013499985457, + "loss": 1.2382, + "step": 2541 + }, + { + "epoch": 0.19552342127528652, + "learning_rate": 0.0027258046299573784, + "loss": 1.0999, + "step": 2542 + }, + { + "epoch": 0.19560033843550495, + "learning_rate": 0.0027255956883532905, + "loss": 1.1137, + "step": 2543 + }, + { + "epoch": 0.1956772555957234, + "learning_rate": 0.0027253866751853926, + "loss": 0.9086, + "step": 2544 + }, + { + "epoch": 0.19575417275594184, + "learning_rate": 0.00272517759046589, + "loss": 1.0784, + "step": 2545 + }, + { + "epoch": 0.1958310899161603, + "learning_rate": 0.002724968434206991, + "loss": 1.2894, + "step": 2546 + }, + { + "epoch": 0.19590800707637873, + "learning_rate": 0.0027247592064209084, + "loss": 1.22, + "step": 2547 + }, + { + "epoch": 0.1959849242365972, + "learning_rate": 0.002724549907119859, + "loss": 0.5908, + "step": 2548 + }, + { + "epoch": 0.19606184139681562, + "learning_rate": 0.002724340536316065, + "loss": 0.9553, + "step": 2549 + }, + { + "epoch": 0.19613875855703408, + "learning_rate": 0.0027241310940217507, + "loss": 1.5002, + "step": 2550 + }, + { + "epoch": 0.1962156757172525, + "learning_rate": 0.0027239215802491463, + "loss": 1.4841, + "step": 2551 + }, + { + "epoch": 0.19629259287747097, + "learning_rate": 0.0027237119950104852, + "loss": 1.0263, + "step": 2552 + }, + { + "epoch": 0.1963695100376894, + "learning_rate": 0.002723502338318005, + "loss": 0.772, + "step": 2553 + }, + { + "epoch": 0.19644642719790786, + "learning_rate": 0.002723292610183949, + "loss": 1.1455, + "step": 2554 + }, + { + "epoch": 0.1965233443581263, + "learning_rate": 0.002723082810620562, + "loss": 1.0296, + "step": 2555 + }, + { + "epoch": 0.19660026151834475, + "learning_rate": 0.002722872939640095, + "loss": 1.2724, + "step": 2556 + }, + { + "epoch": 0.19667717867856319, + "learning_rate": 0.002722662997254803, + "loss": 1.2582, + "step": 2557 + }, + { + "epoch": 0.19675409583878165, + "learning_rate": 0.0027224529834769438, + "loss": 1.3994, + "step": 2558 + }, + { + "epoch": 0.19683101299900008, + "learning_rate": 0.002722242898318781, + "loss": 1.1654, + "step": 2559 + }, + { + "epoch": 0.1969079301592185, + "learning_rate": 0.002722032741792582, + "loss": 1.2423, + "step": 2560 + }, + { + "epoch": 0.19698484731943697, + "learning_rate": 0.0027218225139106172, + "loss": 0.9579, + "step": 2561 + }, + { + "epoch": 0.1970617644796554, + "learning_rate": 0.002721612214685163, + "loss": 1.0188, + "step": 2562 + }, + { + "epoch": 0.19713868163987386, + "learning_rate": 0.0027214018441284976, + "loss": 1.2168, + "step": 2563 + }, + { + "epoch": 0.1972155988000923, + "learning_rate": 0.0027211914022529063, + "loss": 0.9237, + "step": 2564 + }, + { + "epoch": 0.19729251596031075, + "learning_rate": 0.002720980889070676, + "loss": 1.4665, + "step": 2565 + }, + { + "epoch": 0.19736943312052918, + "learning_rate": 0.002720770304594099, + "loss": 1.1105, + "step": 2566 + }, + { + "epoch": 0.19744635028074764, + "learning_rate": 0.0027205596488354723, + "loss": 1.3942, + "step": 2567 + }, + { + "epoch": 0.19752326744096607, + "learning_rate": 0.002720348921807095, + "loss": 1.0266, + "step": 2568 + }, + { + "epoch": 0.19760018460118453, + "learning_rate": 0.0027201381235212726, + "loss": 1.2189, + "step": 2569 + }, + { + "epoch": 0.19767710176140296, + "learning_rate": 0.0027199272539903136, + "loss": 1.0854, + "step": 2570 + }, + { + "epoch": 0.19775401892162142, + "learning_rate": 0.0027197163132265307, + "loss": 1.4957, + "step": 2571 + }, + { + "epoch": 0.19783093608183985, + "learning_rate": 0.0027195053012422413, + "loss": 1.0693, + "step": 2572 + }, + { + "epoch": 0.1979078532420583, + "learning_rate": 0.0027192942180497663, + "loss": 0.7972, + "step": 2573 + }, + { + "epoch": 0.19798477040227674, + "learning_rate": 0.0027190830636614313, + "loss": 1.1248, + "step": 2574 + }, + { + "epoch": 0.1980616875624952, + "learning_rate": 0.002718871838089566, + "loss": 1.4307, + "step": 2575 + }, + { + "epoch": 0.19813860472271363, + "learning_rate": 0.002718660541346503, + "loss": 1.168, + "step": 2576 + }, + { + "epoch": 0.1982155218829321, + "learning_rate": 0.0027184491734445816, + "loss": 1.1479, + "step": 2577 + }, + { + "epoch": 0.19829243904315053, + "learning_rate": 0.002718237734396143, + "loss": 1.3318, + "step": 2578 + }, + { + "epoch": 0.19836935620336899, + "learning_rate": 0.0027180262242135333, + "loss": 0.8992, + "step": 2579 + }, + { + "epoch": 0.19844627336358742, + "learning_rate": 0.002717814642909103, + "loss": 1.1816, + "step": 2580 + }, + { + "epoch": 0.19852319052380585, + "learning_rate": 0.0027176029904952066, + "loss": 1.1761, + "step": 2581 + }, + { + "epoch": 0.1986001076840243, + "learning_rate": 0.0027173912669842025, + "loss": 1.6291, + "step": 2582 + }, + { + "epoch": 0.19867702484424274, + "learning_rate": 0.002717179472388454, + "loss": 1.3271, + "step": 2583 + }, + { + "epoch": 0.1987539420044612, + "learning_rate": 0.0027169676067203274, + "loss": 1.136, + "step": 2584 + }, + { + "epoch": 0.19883085916467963, + "learning_rate": 0.0027167556699921936, + "loss": 1.3296, + "step": 2585 + }, + { + "epoch": 0.1989077763248981, + "learning_rate": 0.0027165436622164285, + "loss": 1.1308, + "step": 2586 + }, + { + "epoch": 0.19898469348511652, + "learning_rate": 0.002716331583405411, + "loss": 1.2838, + "step": 2587 + }, + { + "epoch": 0.19906161064533498, + "learning_rate": 0.0027161194335715245, + "loss": 1.2179, + "step": 2588 + }, + { + "epoch": 0.1991385278055534, + "learning_rate": 0.0027159072127271575, + "loss": 1.3697, + "step": 2589 + }, + { + "epoch": 0.19921544496577187, + "learning_rate": 0.0027156949208847004, + "loss": 1.3041, + "step": 2590 + }, + { + "epoch": 0.1992923621259903, + "learning_rate": 0.0027154825580565506, + "loss": 0.8961, + "step": 2591 + }, + { + "epoch": 0.19936927928620876, + "learning_rate": 0.0027152701242551067, + "loss": 1.1366, + "step": 2592 + }, + { + "epoch": 0.1994461964464272, + "learning_rate": 0.0027150576194927747, + "loss": 1.0879, + "step": 2593 + }, + { + "epoch": 0.19952311360664565, + "learning_rate": 0.002714845043781961, + "loss": 1.441, + "step": 2594 + }, + { + "epoch": 0.19960003076686408, + "learning_rate": 0.00271463239713508, + "loss": 1.2458, + "step": 2595 + }, + { + "epoch": 0.19967694792708254, + "learning_rate": 0.002714419679564547, + "loss": 0.899, + "step": 2596 + }, + { + "epoch": 0.19975386508730097, + "learning_rate": 0.0027142068910827833, + "loss": 1.0908, + "step": 2597 + }, + { + "epoch": 0.19983078224751943, + "learning_rate": 0.0027139940317022137, + "loss": 1.3489, + "step": 2598 + }, + { + "epoch": 0.19990769940773787, + "learning_rate": 0.002713781101435268, + "loss": 1.4459, + "step": 2599 + }, + { + "epoch": 0.19998461656795632, + "learning_rate": 0.002713568100294378, + "loss": 1.1666, + "step": 2600 + }, + { + "epoch": 0.20006153372817476, + "learning_rate": 0.0027133550282919822, + "loss": 1.4039, + "step": 2601 + }, + { + "epoch": 0.2001384508883932, + "learning_rate": 0.0027131418854405218, + "loss": 1.2346, + "step": 2602 + }, + { + "epoch": 0.20021536804861165, + "learning_rate": 0.002712928671752442, + "loss": 0.8976, + "step": 2603 + }, + { + "epoch": 0.20029228520883008, + "learning_rate": 0.0027127153872401936, + "loss": 1.128, + "step": 2604 + }, + { + "epoch": 0.20036920236904854, + "learning_rate": 0.0027125020319162295, + "loss": 0.9785, + "step": 2605 + }, + { + "epoch": 0.20044611952926697, + "learning_rate": 0.0027122886057930084, + "loss": 1.1928, + "step": 2606 + }, + { + "epoch": 0.20052303668948543, + "learning_rate": 0.0027120751088829925, + "loss": 0.9113, + "step": 2607 + }, + { + "epoch": 0.20059995384970386, + "learning_rate": 0.002711861541198647, + "loss": 0.9073, + "step": 2608 + }, + { + "epoch": 0.20067687100992232, + "learning_rate": 0.0027116479027524437, + "loss": 1.063, + "step": 2609 + }, + { + "epoch": 0.20075378817014075, + "learning_rate": 0.0027114341935568563, + "loss": 1.3355, + "step": 2610 + }, + { + "epoch": 0.2008307053303592, + "learning_rate": 0.002711220413624364, + "loss": 1.2035, + "step": 2611 + }, + { + "epoch": 0.20090762249057764, + "learning_rate": 0.002711006562967449, + "loss": 1.4746, + "step": 2612 + }, + { + "epoch": 0.2009845396507961, + "learning_rate": 0.002710792641598599, + "loss": 1.3634, + "step": 2613 + }, + { + "epoch": 0.20106145681101453, + "learning_rate": 0.002710578649530305, + "loss": 1.2459, + "step": 2614 + }, + { + "epoch": 0.201138373971233, + "learning_rate": 0.002710364586775062, + "loss": 1.4056, + "step": 2615 + }, + { + "epoch": 0.20121529113145142, + "learning_rate": 0.002710150453345369, + "loss": 1.4892, + "step": 2616 + }, + { + "epoch": 0.20129220829166988, + "learning_rate": 0.00270993624925373, + "loss": 1.1893, + "step": 2617 + }, + { + "epoch": 0.20136912545188831, + "learning_rate": 0.002709721974512653, + "loss": 1.1919, + "step": 2618 + }, + { + "epoch": 0.20144604261210677, + "learning_rate": 0.0027095076291346483, + "loss": 1.3314, + "step": 2619 + }, + { + "epoch": 0.2015229597723252, + "learning_rate": 0.0027092932131322334, + "loss": 1.2867, + "step": 2620 + }, + { + "epoch": 0.20159987693254364, + "learning_rate": 0.0027090787265179267, + "loss": 1.4829, + "step": 2621 + }, + { + "epoch": 0.2016767940927621, + "learning_rate": 0.002708864169304254, + "loss": 0.7613, + "step": 2622 + }, + { + "epoch": 0.20175371125298053, + "learning_rate": 0.0027086495415037416, + "loss": 1.1193, + "step": 2623 + }, + { + "epoch": 0.201830628413199, + "learning_rate": 0.0027084348431289235, + "loss": 1.1944, + "step": 2624 + }, + { + "epoch": 0.20190754557341742, + "learning_rate": 0.0027082200741923353, + "loss": 1.3288, + "step": 2625 + }, + { + "epoch": 0.20198446273363588, + "learning_rate": 0.002708005234706518, + "loss": 1.0715, + "step": 2626 + }, + { + "epoch": 0.2020613798938543, + "learning_rate": 0.002707790324684016, + "loss": 1.121, + "step": 2627 + }, + { + "epoch": 0.20213829705407277, + "learning_rate": 0.002707575344137378, + "loss": 0.808, + "step": 2628 + }, + { + "epoch": 0.2022152142142912, + "learning_rate": 0.0027073602930791575, + "loss": 1.3753, + "step": 2629 + }, + { + "epoch": 0.20229213137450966, + "learning_rate": 0.0027071451715219106, + "loss": 0.9612, + "step": 2630 + }, + { + "epoch": 0.2023690485347281, + "learning_rate": 0.0027069299794781998, + "loss": 1.4398, + "step": 2631 + }, + { + "epoch": 0.20244596569494655, + "learning_rate": 0.0027067147169605892, + "loss": 1.1476, + "step": 2632 + }, + { + "epoch": 0.20252288285516498, + "learning_rate": 0.002706499383981649, + "loss": 1.1654, + "step": 2633 + }, + { + "epoch": 0.20259980001538344, + "learning_rate": 0.002706283980553952, + "loss": 0.8518, + "step": 2634 + }, + { + "epoch": 0.20267671717560187, + "learning_rate": 0.0027060685066900765, + "loss": 1.1888, + "step": 2635 + }, + { + "epoch": 0.20275363433582033, + "learning_rate": 0.0027058529624026037, + "loss": 1.0883, + "step": 2636 + }, + { + "epoch": 0.20283055149603876, + "learning_rate": 0.0027056373477041197, + "loss": 0.9727, + "step": 2637 + }, + { + "epoch": 0.20290746865625722, + "learning_rate": 0.0027054216626072147, + "loss": 1.0949, + "step": 2638 + }, + { + "epoch": 0.20298438581647565, + "learning_rate": 0.0027052059071244828, + "loss": 1.2156, + "step": 2639 + }, + { + "epoch": 0.2030613029766941, + "learning_rate": 0.002704990081268522, + "loss": 1.4779, + "step": 2640 + }, + { + "epoch": 0.20313822013691255, + "learning_rate": 0.002704774185051934, + "loss": 1.258, + "step": 2641 + }, + { + "epoch": 0.20321513729713098, + "learning_rate": 0.0027045582184873256, + "loss": 0.8716, + "step": 2642 + }, + { + "epoch": 0.20329205445734944, + "learning_rate": 0.002704342181587308, + "loss": 1.1194, + "step": 2643 + }, + { + "epoch": 0.20336897161756787, + "learning_rate": 0.0027041260743644953, + "loss": 1.3263, + "step": 2644 + }, + { + "epoch": 0.20344588877778633, + "learning_rate": 0.0027039098968315058, + "loss": 1.1777, + "step": 2645 + }, + { + "epoch": 0.20352280593800476, + "learning_rate": 0.0027036936490009633, + "loss": 1.062, + "step": 2646 + }, + { + "epoch": 0.20359972309822322, + "learning_rate": 0.002703477330885494, + "loss": 1.0762, + "step": 2647 + }, + { + "epoch": 0.20367664025844165, + "learning_rate": 0.0027032609424977286, + "loss": 0.9619, + "step": 2648 + }, + { + "epoch": 0.2037535574186601, + "learning_rate": 0.002703044483850303, + "loss": 0.9424, + "step": 2649 + }, + { + "epoch": 0.20383047457887854, + "learning_rate": 0.0027028279549558562, + "loss": 1.333, + "step": 2650 + }, + { + "epoch": 0.203907391739097, + "learning_rate": 0.002702611355827032, + "loss": 1.2799, + "step": 2651 + }, + { + "epoch": 0.20398430889931543, + "learning_rate": 0.002702394686476477, + "loss": 1.186, + "step": 2652 + }, + { + "epoch": 0.2040612260595339, + "learning_rate": 0.002702177946916843, + "loss": 1.1085, + "step": 2653 + }, + { + "epoch": 0.20413814321975232, + "learning_rate": 0.0027019611371607863, + "loss": 1.0352, + "step": 2654 + }, + { + "epoch": 0.20421506037997078, + "learning_rate": 0.0027017442572209657, + "loss": 1.4193, + "step": 2655 + }, + { + "epoch": 0.2042919775401892, + "learning_rate": 0.0027015273071100457, + "loss": 1.2035, + "step": 2656 + }, + { + "epoch": 0.20436889470040767, + "learning_rate": 0.002701310286840694, + "loss": 1.5892, + "step": 2657 + }, + { + "epoch": 0.2044458118606261, + "learning_rate": 0.0027010931964255827, + "loss": 0.9946, + "step": 2658 + }, + { + "epoch": 0.20452272902084456, + "learning_rate": 0.002700876035877388, + "loss": 1.1391, + "step": 2659 + }, + { + "epoch": 0.204599646181063, + "learning_rate": 0.00270065880520879, + "loss": 0.9078, + "step": 2660 + }, + { + "epoch": 0.20467656334128145, + "learning_rate": 0.002700441504432473, + "loss": 1.0461, + "step": 2661 + }, + { + "epoch": 0.20475348050149988, + "learning_rate": 0.0027002241335611255, + "loss": 0.9501, + "step": 2662 + }, + { + "epoch": 0.20483039766171832, + "learning_rate": 0.00270000669260744, + "loss": 1.4062, + "step": 2663 + }, + { + "epoch": 0.20490731482193678, + "learning_rate": 0.002699789181584113, + "loss": 1.2921, + "step": 2664 + }, + { + "epoch": 0.2049842319821552, + "learning_rate": 0.0026995716005038454, + "loss": 1.4167, + "step": 2665 + }, + { + "epoch": 0.20506114914237367, + "learning_rate": 0.002699353949379342, + "loss": 1.1758, + "step": 2666 + }, + { + "epoch": 0.2051380663025921, + "learning_rate": 0.0026991362282233113, + "loss": 1.2451, + "step": 2667 + }, + { + "epoch": 0.20521498346281056, + "learning_rate": 0.0026989184370484667, + "loss": 1.105, + "step": 2668 + }, + { + "epoch": 0.205291900623029, + "learning_rate": 0.0026987005758675247, + "loss": 0.9573, + "step": 2669 + }, + { + "epoch": 0.20536881778324745, + "learning_rate": 0.002698482644693207, + "loss": 1.0918, + "step": 2670 + }, + { + "epoch": 0.20544573494346588, + "learning_rate": 0.002698264643538239, + "loss": 1.137, + "step": 2671 + }, + { + "epoch": 0.20552265210368434, + "learning_rate": 0.0026980465724153494, + "loss": 1.3692, + "step": 2672 + }, + { + "epoch": 0.20559956926390277, + "learning_rate": 0.002697828431337272, + "loss": 1.614, + "step": 2673 + }, + { + "epoch": 0.20567648642412123, + "learning_rate": 0.0026976102203167445, + "loss": 1.3403, + "step": 2674 + }, + { + "epoch": 0.20575340358433966, + "learning_rate": 0.0026973919393665073, + "loss": 1.0917, + "step": 2675 + }, + { + "epoch": 0.20583032074455812, + "learning_rate": 0.002697173588499307, + "loss": 1.7301, + "step": 2676 + }, + { + "epoch": 0.20590723790477655, + "learning_rate": 0.0026969551677278937, + "loss": 1.4602, + "step": 2677 + }, + { + "epoch": 0.205984155064995, + "learning_rate": 0.0026967366770650207, + "loss": 1.1644, + "step": 2678 + }, + { + "epoch": 0.20606107222521344, + "learning_rate": 0.0026965181165234456, + "loss": 1.2144, + "step": 2679 + }, + { + "epoch": 0.2061379893854319, + "learning_rate": 0.002696299486115931, + "loss": 1.0635, + "step": 2680 + }, + { + "epoch": 0.20621490654565033, + "learning_rate": 0.0026960807858552424, + "loss": 1.1161, + "step": 2681 + }, + { + "epoch": 0.2062918237058688, + "learning_rate": 0.00269586201575415, + "loss": 1.135, + "step": 2682 + }, + { + "epoch": 0.20636874086608722, + "learning_rate": 0.0026956431758254286, + "loss": 1.1538, + "step": 2683 + }, + { + "epoch": 0.20644565802630566, + "learning_rate": 0.002695424266081856, + "loss": 1.3421, + "step": 2684 + }, + { + "epoch": 0.20652257518652412, + "learning_rate": 0.0026952052865362145, + "loss": 0.6206, + "step": 2685 + }, + { + "epoch": 0.20659949234674255, + "learning_rate": 0.0026949862372012906, + "loss": 1.226, + "step": 2686 + }, + { + "epoch": 0.206676409506961, + "learning_rate": 0.002694767118089875, + "loss": 1.194, + "step": 2687 + }, + { + "epoch": 0.20675332666717944, + "learning_rate": 0.0026945479292147623, + "loss": 1.1061, + "step": 2688 + }, + { + "epoch": 0.2068302438273979, + "learning_rate": 0.0026943286705887514, + "loss": 1.2629, + "step": 2689 + }, + { + "epoch": 0.20690716098761633, + "learning_rate": 0.002694109342224644, + "loss": 0.9402, + "step": 2690 + }, + { + "epoch": 0.2069840781478348, + "learning_rate": 0.0026938899441352484, + "loss": 1.1339, + "step": 2691 + }, + { + "epoch": 0.20706099530805322, + "learning_rate": 0.0026936704763333742, + "loss": 1.0795, + "step": 2692 + }, + { + "epoch": 0.20713791246827168, + "learning_rate": 0.002693450938831837, + "loss": 1.2599, + "step": 2693 + }, + { + "epoch": 0.2072148296284901, + "learning_rate": 0.002693231331643456, + "loss": 1.2991, + "step": 2694 + }, + { + "epoch": 0.20729174678870857, + "learning_rate": 0.002693011654781054, + "loss": 0.9029, + "step": 2695 + }, + { + "epoch": 0.207368663948927, + "learning_rate": 0.002692791908257458, + "loss": 1.1686, + "step": 2696 + }, + { + "epoch": 0.20744558110914546, + "learning_rate": 0.0026925720920854992, + "loss": 1.1356, + "step": 2697 + }, + { + "epoch": 0.2075224982693639, + "learning_rate": 0.0026923522062780128, + "loss": 1.0119, + "step": 2698 + }, + { + "epoch": 0.20759941542958235, + "learning_rate": 0.002692132250847839, + "loss": 1.5194, + "step": 2699 + }, + { + "epoch": 0.20767633258980078, + "learning_rate": 0.0026919122258078207, + "loss": 0.8539, + "step": 2700 + }, + { + "epoch": 0.20775324975001924, + "learning_rate": 0.002691692131170805, + "loss": 1.1621, + "step": 2701 + }, + { + "epoch": 0.20783016691023767, + "learning_rate": 0.0026914719669496436, + "loss": 1.0978, + "step": 2702 + }, + { + "epoch": 0.2079070840704561, + "learning_rate": 0.002691251733157193, + "loss": 1.0201, + "step": 2703 + }, + { + "epoch": 0.20798400123067456, + "learning_rate": 0.0026910314298063116, + "loss": 1.3523, + "step": 2704 + }, + { + "epoch": 0.208060918390893, + "learning_rate": 0.002690811056909864, + "loss": 1.374, + "step": 2705 + }, + { + "epoch": 0.20813783555111146, + "learning_rate": 0.0026905906144807177, + "loss": 1.2682, + "step": 2706 + }, + { + "epoch": 0.2082147527113299, + "learning_rate": 0.002690370102531745, + "loss": 1.0675, + "step": 2707 + }, + { + "epoch": 0.20829166987154835, + "learning_rate": 0.0026901495210758207, + "loss": 1.0536, + "step": 2708 + }, + { + "epoch": 0.20836858703176678, + "learning_rate": 0.0026899288701258263, + "loss": 1.2805, + "step": 2709 + }, + { + "epoch": 0.20844550419198524, + "learning_rate": 0.0026897081496946446, + "loss": 0.9618, + "step": 2710 + }, + { + "epoch": 0.20852242135220367, + "learning_rate": 0.002689487359795164, + "loss": 0.9159, + "step": 2711 + }, + { + "epoch": 0.20859933851242213, + "learning_rate": 0.0026892665004402774, + "loss": 0.8268, + "step": 2712 + }, + { + "epoch": 0.20867625567264056, + "learning_rate": 0.00268904557164288, + "loss": 1.5303, + "step": 2713 + }, + { + "epoch": 0.20875317283285902, + "learning_rate": 0.0026888245734158723, + "loss": 0.9274, + "step": 2714 + }, + { + "epoch": 0.20883008999307745, + "learning_rate": 0.0026886035057721593, + "loss": 1.3888, + "step": 2715 + }, + { + "epoch": 0.2089070071532959, + "learning_rate": 0.0026883823687246485, + "loss": 0.9998, + "step": 2716 + }, + { + "epoch": 0.20898392431351434, + "learning_rate": 0.0026881611622862527, + "loss": 1.1488, + "step": 2717 + }, + { + "epoch": 0.2090608414737328, + "learning_rate": 0.0026879398864698887, + "loss": 1.058, + "step": 2718 + }, + { + "epoch": 0.20913775863395123, + "learning_rate": 0.0026877185412884762, + "loss": 1.1346, + "step": 2719 + }, + { + "epoch": 0.2092146757941697, + "learning_rate": 0.002687497126754941, + "loss": 0.9705, + "step": 2720 + }, + { + "epoch": 0.20929159295438812, + "learning_rate": 0.00268727564288221, + "loss": 1.2282, + "step": 2721 + }, + { + "epoch": 0.20936851011460658, + "learning_rate": 0.002687054089683218, + "loss": 1.4496, + "step": 2722 + }, + { + "epoch": 0.209445427274825, + "learning_rate": 0.0026868324671708997, + "loss": 1.2114, + "step": 2723 + }, + { + "epoch": 0.20952234443504344, + "learning_rate": 0.002686610775358197, + "loss": 1.2715, + "step": 2724 + }, + { + "epoch": 0.2095992615952619, + "learning_rate": 0.0026863890142580543, + "loss": 1.1555, + "step": 2725 + }, + { + "epoch": 0.20967617875548034, + "learning_rate": 0.0026861671838834212, + "loss": 1.3641, + "step": 2726 + }, + { + "epoch": 0.2097530959156988, + "learning_rate": 0.0026859452842472493, + "loss": 0.9379, + "step": 2727 + }, + { + "epoch": 0.20983001307591723, + "learning_rate": 0.0026857233153624966, + "loss": 1.3668, + "step": 2728 + }, + { + "epoch": 0.20990693023613569, + "learning_rate": 0.002685501277242124, + "loss": 1.1526, + "step": 2729 + }, + { + "epoch": 0.20998384739635412, + "learning_rate": 0.002685279169899096, + "loss": 1.4391, + "step": 2730 + }, + { + "epoch": 0.21006076455657258, + "learning_rate": 0.002685056993346382, + "loss": 1.3306, + "step": 2731 + }, + { + "epoch": 0.210137681716791, + "learning_rate": 0.0026848347475969554, + "loss": 1.1268, + "step": 2732 + }, + { + "epoch": 0.21021459887700947, + "learning_rate": 0.002684612432663793, + "loss": 1.0123, + "step": 2733 + }, + { + "epoch": 0.2102915160372279, + "learning_rate": 0.002684390048559876, + "loss": 0.7296, + "step": 2734 + }, + { + "epoch": 0.21036843319744636, + "learning_rate": 0.0026841675952981895, + "loss": 1.0813, + "step": 2735 + }, + { + "epoch": 0.2104453503576648, + "learning_rate": 0.002683945072891724, + "loss": 1.1279, + "step": 2736 + }, + { + "epoch": 0.21052226751788325, + "learning_rate": 0.0026837224813534705, + "loss": 1.2752, + "step": 2737 + }, + { + "epoch": 0.21059918467810168, + "learning_rate": 0.002683499820696428, + "loss": 1.0637, + "step": 2738 + }, + { + "epoch": 0.21067610183832014, + "learning_rate": 0.0026832770909335983, + "loss": 1.0091, + "step": 2739 + }, + { + "epoch": 0.21075301899853857, + "learning_rate": 0.0026830542920779858, + "loss": 1.1888, + "step": 2740 + }, + { + "epoch": 0.21082993615875703, + "learning_rate": 0.0026828314241425996, + "loss": 0.8736, + "step": 2741 + }, + { + "epoch": 0.21090685331897546, + "learning_rate": 0.0026826084871404542, + "loss": 1.1327, + "step": 2742 + }, + { + "epoch": 0.21098377047919392, + "learning_rate": 0.002682385481084567, + "loss": 1.0802, + "step": 2743 + }, + { + "epoch": 0.21106068763941235, + "learning_rate": 0.002682162405987959, + "loss": 1.1862, + "step": 2744 + }, + { + "epoch": 0.21113760479963078, + "learning_rate": 0.002681939261863656, + "loss": 1.0259, + "step": 2745 + }, + { + "epoch": 0.21121452195984924, + "learning_rate": 0.002681716048724688, + "loss": 1.4125, + "step": 2746 + }, + { + "epoch": 0.21129143912006768, + "learning_rate": 0.002681492766584088, + "loss": 1.3, + "step": 2747 + }, + { + "epoch": 0.21136835628028613, + "learning_rate": 0.0026812694154548945, + "loss": 1.0045, + "step": 2748 + }, + { + "epoch": 0.21144527344050457, + "learning_rate": 0.0026810459953501486, + "loss": 1.0391, + "step": 2749 + }, + { + "epoch": 0.21152219060072303, + "learning_rate": 0.002680822506282896, + "loss": 1.1679, + "step": 2750 + }, + { + "epoch": 0.21159910776094146, + "learning_rate": 0.002680598948266187, + "loss": 0.96, + "step": 2751 + }, + { + "epoch": 0.21167602492115992, + "learning_rate": 0.0026803753213130747, + "loss": 1.0611, + "step": 2752 + }, + { + "epoch": 0.21175294208137835, + "learning_rate": 0.0026801516254366174, + "loss": 1.3594, + "step": 2753 + }, + { + "epoch": 0.2118298592415968, + "learning_rate": 0.0026799278606498764, + "loss": 0.9369, + "step": 2754 + }, + { + "epoch": 0.21190677640181524, + "learning_rate": 0.0026797040269659184, + "loss": 1.237, + "step": 2755 + }, + { + "epoch": 0.2119836935620337, + "learning_rate": 0.0026794801243978125, + "loss": 0.8864, + "step": 2756 + }, + { + "epoch": 0.21206061072225213, + "learning_rate": 0.002679256152958633, + "loss": 1.373, + "step": 2757 + }, + { + "epoch": 0.2121375278824706, + "learning_rate": 0.002679032112661458, + "loss": 1.2083, + "step": 2758 + }, + { + "epoch": 0.21221444504268902, + "learning_rate": 0.002678808003519369, + "loss": 1.1763, + "step": 2759 + }, + { + "epoch": 0.21229136220290748, + "learning_rate": 0.002678583825545452, + "loss": 1.2056, + "step": 2760 + }, + { + "epoch": 0.2123682793631259, + "learning_rate": 0.0026783595787527974, + "loss": 1.6132, + "step": 2761 + }, + { + "epoch": 0.21244519652334437, + "learning_rate": 0.0026781352631544985, + "loss": 1.074, + "step": 2762 + }, + { + "epoch": 0.2125221136835628, + "learning_rate": 0.0026779108787636542, + "loss": 1.371, + "step": 2763 + }, + { + "epoch": 0.21259903084378123, + "learning_rate": 0.002677686425593366, + "loss": 1.1205, + "step": 2764 + }, + { + "epoch": 0.2126759480039997, + "learning_rate": 0.0026774619036567396, + "loss": 1.0216, + "step": 2765 + }, + { + "epoch": 0.21275286516421812, + "learning_rate": 0.0026772373129668858, + "loss": 1.3456, + "step": 2766 + }, + { + "epoch": 0.21282978232443658, + "learning_rate": 0.0026770126535369185, + "loss": 1.093, + "step": 2767 + }, + { + "epoch": 0.21290669948465502, + "learning_rate": 0.0026767879253799557, + "loss": 1.3328, + "step": 2768 + }, + { + "epoch": 0.21298361664487347, + "learning_rate": 0.00267656312850912, + "loss": 1.1032, + "step": 2769 + }, + { + "epoch": 0.2130605338050919, + "learning_rate": 0.0026763382629375367, + "loss": 1.1695, + "step": 2770 + }, + { + "epoch": 0.21313745096531037, + "learning_rate": 0.002676113328678336, + "loss": 1.0566, + "step": 2771 + }, + { + "epoch": 0.2132143681255288, + "learning_rate": 0.0026758883257446526, + "loss": 1.3098, + "step": 2772 + }, + { + "epoch": 0.21329128528574726, + "learning_rate": 0.002675663254149624, + "loss": 1.3136, + "step": 2773 + }, + { + "epoch": 0.2133682024459657, + "learning_rate": 0.0026754381139063926, + "loss": 1.2384, + "step": 2774 + }, + { + "epoch": 0.21344511960618415, + "learning_rate": 0.002675212905028105, + "loss": 0.8283, + "step": 2775 + }, + { + "epoch": 0.21352203676640258, + "learning_rate": 0.0026749876275279118, + "loss": 1.0777, + "step": 2776 + }, + { + "epoch": 0.21359895392662104, + "learning_rate": 0.002674762281418966, + "loss": 1.1138, + "step": 2777 + }, + { + "epoch": 0.21367587108683947, + "learning_rate": 0.002674536866714426, + "loss": 0.8933, + "step": 2778 + }, + { + "epoch": 0.21375278824705793, + "learning_rate": 0.0026743113834274547, + "loss": 1.2152, + "step": 2779 + }, + { + "epoch": 0.21382970540727636, + "learning_rate": 0.0026740858315712174, + "loss": 1.2671, + "step": 2780 + }, + { + "epoch": 0.21390662256749482, + "learning_rate": 0.002673860211158885, + "loss": 1.1557, + "step": 2781 + }, + { + "epoch": 0.21398353972771325, + "learning_rate": 0.0026736345222036317, + "loss": 0.9709, + "step": 2782 + }, + { + "epoch": 0.2140604568879317, + "learning_rate": 0.002673408764718635, + "loss": 1.2213, + "step": 2783 + }, + { + "epoch": 0.21413737404815014, + "learning_rate": 0.002673182938717078, + "loss": 0.9639, + "step": 2784 + }, + { + "epoch": 0.21421429120836857, + "learning_rate": 0.0026729570442121466, + "loss": 1.2288, + "step": 2785 + }, + { + "epoch": 0.21429120836858703, + "learning_rate": 0.0026727310812170307, + "loss": 1.1893, + "step": 2786 + }, + { + "epoch": 0.21436812552880546, + "learning_rate": 0.0026725050497449256, + "loss": 1.0382, + "step": 2787 + }, + { + "epoch": 0.21444504268902392, + "learning_rate": 0.002672278949809028, + "loss": 1.1415, + "step": 2788 + }, + { + "epoch": 0.21452195984924236, + "learning_rate": 0.0026720527814225407, + "loss": 1.0127, + "step": 2789 + }, + { + "epoch": 0.21459887700946081, + "learning_rate": 0.00267182654459867, + "loss": 1.2729, + "step": 2790 + }, + { + "epoch": 0.21467579416967925, + "learning_rate": 0.0026716002393506262, + "loss": 1.2256, + "step": 2791 + }, + { + "epoch": 0.2147527113298977, + "learning_rate": 0.0026713738656916237, + "loss": 1.4644, + "step": 2792 + }, + { + "epoch": 0.21482962849011614, + "learning_rate": 0.0026711474236348802, + "loss": 0.797, + "step": 2793 + }, + { + "epoch": 0.2149065456503346, + "learning_rate": 0.002670920913193618, + "loss": 1.2909, + "step": 2794 + }, + { + "epoch": 0.21498346281055303, + "learning_rate": 0.0026706943343810636, + "loss": 1.3884, + "step": 2795 + }, + { + "epoch": 0.2150603799707715, + "learning_rate": 0.0026704676872104467, + "loss": 1.1233, + "step": 2796 + }, + { + "epoch": 0.21513729713098992, + "learning_rate": 0.0026702409716950017, + "loss": 1.3233, + "step": 2797 + }, + { + "epoch": 0.21521421429120838, + "learning_rate": 0.002670014187847967, + "loss": 1.0681, + "step": 2798 + }, + { + "epoch": 0.2152911314514268, + "learning_rate": 0.0026697873356825844, + "loss": 1.3902, + "step": 2799 + }, + { + "epoch": 0.21536804861164527, + "learning_rate": 0.0026695604152121005, + "loss": 0.9033, + "step": 2800 + }, + { + "epoch": 0.2154449657718637, + "learning_rate": 0.0026693334264497647, + "loss": 1.2176, + "step": 2801 + }, + { + "epoch": 0.21552188293208216, + "learning_rate": 0.0026691063694088313, + "loss": 1.4218, + "step": 2802 + }, + { + "epoch": 0.2155988000923006, + "learning_rate": 0.0026688792441025588, + "loss": 1.2279, + "step": 2803 + }, + { + "epoch": 0.21567571725251905, + "learning_rate": 0.002668652050544209, + "loss": 1.4316, + "step": 2804 + }, + { + "epoch": 0.21575263441273748, + "learning_rate": 0.002668424788747048, + "loss": 0.9718, + "step": 2805 + }, + { + "epoch": 0.2158295515729559, + "learning_rate": 0.002668197458724346, + "loss": 1.4095, + "step": 2806 + }, + { + "epoch": 0.21590646873317437, + "learning_rate": 0.0026679700604893765, + "loss": 1.143, + "step": 2807 + }, + { + "epoch": 0.2159833858933928, + "learning_rate": 0.002667742594055419, + "loss": 1.2975, + "step": 2808 + }, + { + "epoch": 0.21606030305361126, + "learning_rate": 0.0026675150594357533, + "loss": 0.6969, + "step": 2809 + }, + { + "epoch": 0.2161372202138297, + "learning_rate": 0.0026672874566436673, + "loss": 1.015, + "step": 2810 + }, + { + "epoch": 0.21621413737404815, + "learning_rate": 0.0026670597856924493, + "loss": 1.2185, + "step": 2811 + }, + { + "epoch": 0.21629105453426659, + "learning_rate": 0.0026668320465953947, + "loss": 1.2983, + "step": 2812 + }, + { + "epoch": 0.21636797169448505, + "learning_rate": 0.0026666042393658014, + "loss": 1.0997, + "step": 2813 + }, + { + "epoch": 0.21644488885470348, + "learning_rate": 0.00266637636401697, + "loss": 1.1781, + "step": 2814 + }, + { + "epoch": 0.21652180601492194, + "learning_rate": 0.002666148420562207, + "loss": 1.0728, + "step": 2815 + }, + { + "epoch": 0.21659872317514037, + "learning_rate": 0.002665920409014823, + "loss": 1.3593, + "step": 2816 + }, + { + "epoch": 0.21667564033535883, + "learning_rate": 0.0026656923293881312, + "loss": 0.9013, + "step": 2817 + }, + { + "epoch": 0.21675255749557726, + "learning_rate": 0.002665464181695449, + "loss": 2.0623, + "step": 2818 + }, + { + "epoch": 0.21682947465579572, + "learning_rate": 0.0026652359659500987, + "loss": 0.9032, + "step": 2819 + }, + { + "epoch": 0.21690639181601415, + "learning_rate": 0.0026650076821654062, + "loss": 1.2291, + "step": 2820 + }, + { + "epoch": 0.2169833089762326, + "learning_rate": 0.002664779330354701, + "loss": 1.515, + "step": 2821 + }, + { + "epoch": 0.21706022613645104, + "learning_rate": 0.0026645509105313165, + "loss": 1.4406, + "step": 2822 + }, + { + "epoch": 0.2171371432966695, + "learning_rate": 0.002664322422708591, + "loss": 1.3025, + "step": 2823 + }, + { + "epoch": 0.21721406045688793, + "learning_rate": 0.0026640938668998655, + "loss": 1.0984, + "step": 2824 + }, + { + "epoch": 0.2172909776171064, + "learning_rate": 0.0026638652431184863, + "loss": 1.2292, + "step": 2825 + }, + { + "epoch": 0.21736789477732482, + "learning_rate": 0.0026636365513778024, + "loss": 1.1934, + "step": 2826 + }, + { + "epoch": 0.21744481193754325, + "learning_rate": 0.0026634077916911675, + "loss": 1.1607, + "step": 2827 + }, + { + "epoch": 0.2175217290977617, + "learning_rate": 0.002663178964071939, + "loss": 1.6558, + "step": 2828 + }, + { + "epoch": 0.21759864625798014, + "learning_rate": 0.0026629500685334782, + "loss": 1.4741, + "step": 2829 + }, + { + "epoch": 0.2176755634181986, + "learning_rate": 0.002662721105089152, + "loss": 1.2368, + "step": 2830 + }, + { + "epoch": 0.21775248057841703, + "learning_rate": 0.002662492073752328, + "loss": 0.9789, + "step": 2831 + }, + { + "epoch": 0.2178293977386355, + "learning_rate": 0.0026622629745363803, + "loss": 1.1643, + "step": 2832 + }, + { + "epoch": 0.21790631489885393, + "learning_rate": 0.0026620338074546863, + "loss": 1.3747, + "step": 2833 + }, + { + "epoch": 0.21798323205907238, + "learning_rate": 0.0026618045725206275, + "loss": 0.9163, + "step": 2834 + }, + { + "epoch": 0.21806014921929082, + "learning_rate": 0.0026615752697475886, + "loss": 1.019, + "step": 2835 + }, + { + "epoch": 0.21813706637950928, + "learning_rate": 0.0026613458991489596, + "loss": 0.892, + "step": 2836 + }, + { + "epoch": 0.2182139835397277, + "learning_rate": 0.002661116460738133, + "loss": 1.0551, + "step": 2837 + }, + { + "epoch": 0.21829090069994617, + "learning_rate": 0.0026608869545285056, + "loss": 1.1778, + "step": 2838 + }, + { + "epoch": 0.2183678178601646, + "learning_rate": 0.00266065738053348, + "loss": 1.3514, + "step": 2839 + }, + { + "epoch": 0.21844473502038306, + "learning_rate": 0.0026604277387664594, + "loss": 0.8138, + "step": 2840 + }, + { + "epoch": 0.2185216521806015, + "learning_rate": 0.0026601980292408544, + "loss": 1.3814, + "step": 2841 + }, + { + "epoch": 0.21859856934081995, + "learning_rate": 0.002659968251970077, + "loss": 1.3129, + "step": 2842 + }, + { + "epoch": 0.21867548650103838, + "learning_rate": 0.002659738406967545, + "loss": 1.1769, + "step": 2843 + }, + { + "epoch": 0.21875240366125684, + "learning_rate": 0.0026595084942466784, + "loss": 1.0289, + "step": 2844 + }, + { + "epoch": 0.21882932082147527, + "learning_rate": 0.0026592785138209022, + "loss": 1.1511, + "step": 2845 + }, + { + "epoch": 0.2189062379816937, + "learning_rate": 0.002659048465703646, + "loss": 1.2311, + "step": 2846 + }, + { + "epoch": 0.21898315514191216, + "learning_rate": 0.0026588183499083413, + "loss": 1.0108, + "step": 2847 + }, + { + "epoch": 0.2190600723021306, + "learning_rate": 0.002658588166448426, + "loss": 1.1901, + "step": 2848 + }, + { + "epoch": 0.21913698946234905, + "learning_rate": 0.00265835791533734, + "loss": 0.9988, + "step": 2849 + }, + { + "epoch": 0.21921390662256748, + "learning_rate": 0.0026581275965885283, + "loss": 1.0638, + "step": 2850 + }, + { + "epoch": 0.21929082378278594, + "learning_rate": 0.0026578972102154386, + "loss": 1.2478, + "step": 2851 + }, + { + "epoch": 0.21936774094300437, + "learning_rate": 0.002657666756231525, + "loss": 1.3969, + "step": 2852 + }, + { + "epoch": 0.21944465810322283, + "learning_rate": 0.002657436234650242, + "loss": 1.2669, + "step": 2853 + }, + { + "epoch": 0.21952157526344127, + "learning_rate": 0.002657205645485052, + "loss": 1.2133, + "step": 2854 + }, + { + "epoch": 0.21959849242365972, + "learning_rate": 0.0026569749887494177, + "loss": 0.8658, + "step": 2855 + }, + { + "epoch": 0.21967540958387816, + "learning_rate": 0.002656744264456808, + "loss": 1.2947, + "step": 2856 + }, + { + "epoch": 0.21975232674409662, + "learning_rate": 0.0026565134726206957, + "loss": 1.3169, + "step": 2857 + }, + { + "epoch": 0.21982924390431505, + "learning_rate": 0.002656282613254556, + "loss": 1.1669, + "step": 2858 + }, + { + "epoch": 0.2199061610645335, + "learning_rate": 0.00265605168637187, + "loss": 1.3458, + "step": 2859 + }, + { + "epoch": 0.21998307822475194, + "learning_rate": 0.0026558206919861205, + "loss": 1.2217, + "step": 2860 + }, + { + "epoch": 0.2200599953849704, + "learning_rate": 0.0026555896301107967, + "loss": 1.4341, + "step": 2861 + }, + { + "epoch": 0.22013691254518883, + "learning_rate": 0.0026553585007593893, + "loss": 0.9718, + "step": 2862 + }, + { + "epoch": 0.2202138297054073, + "learning_rate": 0.002655127303945396, + "loss": 1.3726, + "step": 2863 + }, + { + "epoch": 0.22029074686562572, + "learning_rate": 0.002654896039682315, + "loss": 1.0828, + "step": 2864 + }, + { + "epoch": 0.22036766402584418, + "learning_rate": 0.0026546647079836506, + "loss": 1.177, + "step": 2865 + }, + { + "epoch": 0.2204445811860626, + "learning_rate": 0.002654433308862911, + "loss": 1.1054, + "step": 2866 + }, + { + "epoch": 0.22052149834628104, + "learning_rate": 0.002654201842333607, + "loss": 1.4729, + "step": 2867 + }, + { + "epoch": 0.2205984155064995, + "learning_rate": 0.0026539703084092546, + "loss": 1.4224, + "step": 2868 + }, + { + "epoch": 0.22067533266671793, + "learning_rate": 0.0026537387071033734, + "loss": 0.9926, + "step": 2869 + }, + { + "epoch": 0.2207522498269364, + "learning_rate": 0.002653507038429487, + "loss": 1.4223, + "step": 2870 + }, + { + "epoch": 0.22082916698715482, + "learning_rate": 0.0026532753024011216, + "loss": 1.1191, + "step": 2871 + }, + { + "epoch": 0.22090608414737328, + "learning_rate": 0.0026530434990318097, + "loss": 1.4377, + "step": 2872 + }, + { + "epoch": 0.22098300130759171, + "learning_rate": 0.0026528116283350864, + "loss": 1.1086, + "step": 2873 + }, + { + "epoch": 0.22105991846781017, + "learning_rate": 0.002652579690324491, + "loss": 1.0963, + "step": 2874 + }, + { + "epoch": 0.2211368356280286, + "learning_rate": 0.0026523476850135656, + "loss": 1.1156, + "step": 2875 + }, + { + "epoch": 0.22121375278824706, + "learning_rate": 0.0026521156124158585, + "loss": 0.9587, + "step": 2876 + }, + { + "epoch": 0.2212906699484655, + "learning_rate": 0.0026518834725449197, + "loss": 1.466, + "step": 2877 + }, + { + "epoch": 0.22136758710868396, + "learning_rate": 0.0026516512654143048, + "loss": 1.2132, + "step": 2878 + }, + { + "epoch": 0.2214445042689024, + "learning_rate": 0.0026514189910375723, + "loss": 1.0236, + "step": 2879 + }, + { + "epoch": 0.22152142142912085, + "learning_rate": 0.0026511866494282844, + "loss": 1.3139, + "step": 2880 + }, + { + "epoch": 0.22159833858933928, + "learning_rate": 0.0026509542406000087, + "loss": 1.0677, + "step": 2881 + }, + { + "epoch": 0.22167525574955774, + "learning_rate": 0.0026507217645663156, + "loss": 1.5382, + "step": 2882 + }, + { + "epoch": 0.22175217290977617, + "learning_rate": 0.002650489221340779, + "loss": 1.2255, + "step": 2883 + }, + { + "epoch": 0.22182909006999463, + "learning_rate": 0.002650256610936978, + "loss": 1.1327, + "step": 2884 + }, + { + "epoch": 0.22190600723021306, + "learning_rate": 0.0026500239333684945, + "loss": 0.9811, + "step": 2885 + }, + { + "epoch": 0.22198292439043152, + "learning_rate": 0.0026497911886489154, + "loss": 1.2869, + "step": 2886 + }, + { + "epoch": 0.22205984155064995, + "learning_rate": 0.0026495583767918297, + "loss": 1.4157, + "step": 2887 + }, + { + "epoch": 0.22213675871086838, + "learning_rate": 0.0026493254978108335, + "loss": 1.3822, + "step": 2888 + }, + { + "epoch": 0.22221367587108684, + "learning_rate": 0.002649092551719523, + "loss": 1.158, + "step": 2889 + }, + { + "epoch": 0.22229059303130527, + "learning_rate": 0.0026488595385315006, + "loss": 1.387, + "step": 2890 + }, + { + "epoch": 0.22236751019152373, + "learning_rate": 0.0026486264582603726, + "loss": 1.2149, + "step": 2891 + }, + { + "epoch": 0.22244442735174216, + "learning_rate": 0.002648393310919749, + "loss": 1.0277, + "step": 2892 + }, + { + "epoch": 0.22252134451196062, + "learning_rate": 0.002648160096523243, + "loss": 1.1177, + "step": 2893 + }, + { + "epoch": 0.22259826167217905, + "learning_rate": 0.0026479268150844722, + "loss": 1.1437, + "step": 2894 + }, + { + "epoch": 0.2226751788323975, + "learning_rate": 0.0026476934666170584, + "loss": 1.0757, + "step": 2895 + }, + { + "epoch": 0.22275209599261594, + "learning_rate": 0.0026474600511346273, + "loss": 1.2584, + "step": 2896 + }, + { + "epoch": 0.2228290131528344, + "learning_rate": 0.002647226568650808, + "loss": 1.3191, + "step": 2897 + }, + { + "epoch": 0.22290593031305284, + "learning_rate": 0.0026469930191792327, + "loss": 1.0578, + "step": 2898 + }, + { + "epoch": 0.2229828474732713, + "learning_rate": 0.0026467594027335404, + "loss": 1.3145, + "step": 2899 + }, + { + "epoch": 0.22305976463348973, + "learning_rate": 0.0026465257193273714, + "loss": 1.1279, + "step": 2900 + }, + { + "epoch": 0.22313668179370819, + "learning_rate": 0.002646291968974371, + "loss": 1.2705, + "step": 2901 + }, + { + "epoch": 0.22321359895392662, + "learning_rate": 0.002646058151688188, + "loss": 1.0089, + "step": 2902 + }, + { + "epoch": 0.22329051611414508, + "learning_rate": 0.0026458242674824747, + "loss": 1.2941, + "step": 2903 + }, + { + "epoch": 0.2233674332743635, + "learning_rate": 0.0026455903163708884, + "loss": 0.9568, + "step": 2904 + }, + { + "epoch": 0.22344435043458197, + "learning_rate": 0.0026453562983670895, + "loss": 1.423, + "step": 2905 + }, + { + "epoch": 0.2235212675948004, + "learning_rate": 0.0026451222134847427, + "loss": 0.7131, + "step": 2906 + }, + { + "epoch": 0.22359818475501886, + "learning_rate": 0.0026448880617375168, + "loss": 0.8101, + "step": 2907 + }, + { + "epoch": 0.2236751019152373, + "learning_rate": 0.002644653843139083, + "loss": 0.948, + "step": 2908 + }, + { + "epoch": 0.22375201907545572, + "learning_rate": 0.002644419557703119, + "loss": 1.1125, + "step": 2909 + }, + { + "epoch": 0.22382893623567418, + "learning_rate": 0.0026441852054433046, + "loss": 1.1147, + "step": 2910 + }, + { + "epoch": 0.2239058533958926, + "learning_rate": 0.0026439507863733236, + "loss": 1.4192, + "step": 2911 + }, + { + "epoch": 0.22398277055611107, + "learning_rate": 0.002643716300506863, + "loss": 1.5509, + "step": 2912 + }, + { + "epoch": 0.2240596877163295, + "learning_rate": 0.0026434817478576166, + "loss": 0.9808, + "step": 2913 + }, + { + "epoch": 0.22413660487654796, + "learning_rate": 0.002643247128439279, + "loss": 1.0333, + "step": 2914 + }, + { + "epoch": 0.2242135220367664, + "learning_rate": 0.00264301244226555, + "loss": 1.192, + "step": 2915 + }, + { + "epoch": 0.22429043919698485, + "learning_rate": 0.002642777689350133, + "loss": 1.7771, + "step": 2916 + }, + { + "epoch": 0.22436735635720328, + "learning_rate": 0.002642542869706736, + "loss": 1.197, + "step": 2917 + }, + { + "epoch": 0.22444427351742174, + "learning_rate": 0.00264230798334907, + "loss": 1.1411, + "step": 2918 + }, + { + "epoch": 0.22452119067764018, + "learning_rate": 0.002642073030290851, + "loss": 1.2446, + "step": 2919 + }, + { + "epoch": 0.22459810783785863, + "learning_rate": 0.002641838010545797, + "loss": 1.2408, + "step": 2920 + }, + { + "epoch": 0.22467502499807707, + "learning_rate": 0.0026416029241276315, + "loss": 1.1546, + "step": 2921 + }, + { + "epoch": 0.22475194215829553, + "learning_rate": 0.002641367771050082, + "loss": 1.1168, + "step": 2922 + }, + { + "epoch": 0.22482885931851396, + "learning_rate": 0.0026411325513268783, + "loss": 1.2393, + "step": 2923 + }, + { + "epoch": 0.22490577647873242, + "learning_rate": 0.0026408972649717556, + "loss": 0.9176, + "step": 2924 + }, + { + "epoch": 0.22498269363895085, + "learning_rate": 0.0026406619119984528, + "loss": 1.2219, + "step": 2925 + }, + { + "epoch": 0.2250596107991693, + "learning_rate": 0.0026404264924207127, + "loss": 1.2197, + "step": 2926 + }, + { + "epoch": 0.22513652795938774, + "learning_rate": 0.0026401910062522804, + "loss": 0.7539, + "step": 2927 + }, + { + "epoch": 0.22521344511960617, + "learning_rate": 0.0026399554535069076, + "loss": 1.005, + "step": 2928 + }, + { + "epoch": 0.22529036227982463, + "learning_rate": 0.002639719834198347, + "loss": 0.9408, + "step": 2929 + }, + { + "epoch": 0.22536727944004306, + "learning_rate": 0.0026394841483403583, + "loss": 1.1059, + "step": 2930 + }, + { + "epoch": 0.22544419660026152, + "learning_rate": 0.0026392483959467024, + "loss": 1.0097, + "step": 2931 + }, + { + "epoch": 0.22552111376047995, + "learning_rate": 0.002639012577031145, + "loss": 1.1187, + "step": 2932 + }, + { + "epoch": 0.2255980309206984, + "learning_rate": 0.0026387766916074564, + "loss": 0.6283, + "step": 2933 + }, + { + "epoch": 0.22567494808091684, + "learning_rate": 0.0026385407396894104, + "loss": 1.0521, + "step": 2934 + }, + { + "epoch": 0.2257518652411353, + "learning_rate": 0.0026383047212907838, + "loss": 0.8396, + "step": 2935 + }, + { + "epoch": 0.22582878240135373, + "learning_rate": 0.002638068636425358, + "loss": 1.1625, + "step": 2936 + }, + { + "epoch": 0.2259056995615722, + "learning_rate": 0.0026378324851069183, + "loss": 1.3229, + "step": 2937 + }, + { + "epoch": 0.22598261672179062, + "learning_rate": 0.0026375962673492546, + "loss": 1.1008, + "step": 2938 + }, + { + "epoch": 0.22605953388200908, + "learning_rate": 0.002637359983166159, + "loss": 0.9938, + "step": 2939 + }, + { + "epoch": 0.22613645104222752, + "learning_rate": 0.0026371236325714283, + "loss": 0.9123, + "step": 2940 + }, + { + "epoch": 0.22621336820244597, + "learning_rate": 0.002636887215578864, + "loss": 1.1797, + "step": 2941 + }, + { + "epoch": 0.2262902853626644, + "learning_rate": 0.00263665073220227, + "loss": 1.1467, + "step": 2942 + }, + { + "epoch": 0.22636720252288287, + "learning_rate": 0.0026364141824554554, + "loss": 1.1814, + "step": 2943 + }, + { + "epoch": 0.2264441196831013, + "learning_rate": 0.002636177566352233, + "loss": 1.1443, + "step": 2944 + }, + { + "epoch": 0.22652103684331976, + "learning_rate": 0.002635940883906417, + "loss": 1.1127, + "step": 2945 + }, + { + "epoch": 0.2265979540035382, + "learning_rate": 0.0026357041351318298, + "loss": 1.0529, + "step": 2946 + }, + { + "epoch": 0.22667487116375665, + "learning_rate": 0.002635467320042294, + "loss": 1.28, + "step": 2947 + }, + { + "epoch": 0.22675178832397508, + "learning_rate": 0.002635230438651639, + "loss": 1.2115, + "step": 2948 + }, + { + "epoch": 0.2268287054841935, + "learning_rate": 0.002634993490973695, + "loss": 1.1188, + "step": 2949 + }, + { + "epoch": 0.22690562264441197, + "learning_rate": 0.002634756477022298, + "loss": 1.1421, + "step": 2950 + }, + { + "epoch": 0.2269825398046304, + "learning_rate": 0.0026345193968112884, + "loss": 1.4197, + "step": 2951 + }, + { + "epoch": 0.22705945696484886, + "learning_rate": 0.0026342822503545083, + "loss": 0.92, + "step": 2952 + }, + { + "epoch": 0.2271363741250673, + "learning_rate": 0.002634045037665805, + "loss": 0.9803, + "step": 2953 + }, + { + "epoch": 0.22721329128528575, + "learning_rate": 0.002633807758759031, + "loss": 1.1468, + "step": 2954 + }, + { + "epoch": 0.22729020844550418, + "learning_rate": 0.00263357041364804, + "loss": 1.2968, + "step": 2955 + }, + { + "epoch": 0.22736712560572264, + "learning_rate": 0.0026333330023466913, + "loss": 1.3053, + "step": 2956 + }, + { + "epoch": 0.22744404276594107, + "learning_rate": 0.0026330955248688474, + "loss": 1.1437, + "step": 2957 + }, + { + "epoch": 0.22752095992615953, + "learning_rate": 0.0026328579812283743, + "loss": 0.997, + "step": 2958 + }, + { + "epoch": 0.22759787708637796, + "learning_rate": 0.0026326203714391438, + "loss": 0.9116, + "step": 2959 + }, + { + "epoch": 0.22767479424659642, + "learning_rate": 0.0026323826955150288, + "loss": 0.6847, + "step": 2960 + }, + { + "epoch": 0.22775171140681486, + "learning_rate": 0.002632144953469909, + "loss": 1.314, + "step": 2961 + }, + { + "epoch": 0.22782862856703331, + "learning_rate": 0.0026319071453176648, + "loss": 0.8372, + "step": 2962 + }, + { + "epoch": 0.22790554572725175, + "learning_rate": 0.002631669271072183, + "loss": 1.0382, + "step": 2963 + }, + { + "epoch": 0.2279824628874702, + "learning_rate": 0.0026314313307473527, + "loss": 1.3457, + "step": 2964 + }, + { + "epoch": 0.22805938004768864, + "learning_rate": 0.002631193324357068, + "loss": 0.8498, + "step": 2965 + }, + { + "epoch": 0.2281362972079071, + "learning_rate": 0.0026309552519152265, + "loss": 1.146, + "step": 2966 + }, + { + "epoch": 0.22821321436812553, + "learning_rate": 0.002630717113435729, + "loss": 1.3942, + "step": 2967 + }, + { + "epoch": 0.228290131528344, + "learning_rate": 0.002630478908932481, + "loss": 1.0987, + "step": 2968 + }, + { + "epoch": 0.22836704868856242, + "learning_rate": 0.002630240638419391, + "loss": 1.2825, + "step": 2969 + }, + { + "epoch": 0.22844396584878085, + "learning_rate": 0.0026300023019103723, + "loss": 1.5577, + "step": 2970 + }, + { + "epoch": 0.2285208830089993, + "learning_rate": 0.0026297638994193413, + "loss": 1.0042, + "step": 2971 + }, + { + "epoch": 0.22859780016921774, + "learning_rate": 0.002629525430960219, + "loss": 0.9451, + "step": 2972 + }, + { + "epoch": 0.2286747173294362, + "learning_rate": 0.00262928689654693, + "loss": 1.2184, + "step": 2973 + }, + { + "epoch": 0.22875163448965463, + "learning_rate": 0.002629048296193402, + "loss": 1.1151, + "step": 2974 + }, + { + "epoch": 0.2288285516498731, + "learning_rate": 0.002628809629913567, + "loss": 0.9534, + "step": 2975 + }, + { + "epoch": 0.22890546881009152, + "learning_rate": 0.002628570897721362, + "loss": 1.1324, + "step": 2976 + }, + { + "epoch": 0.22898238597030998, + "learning_rate": 0.0026283320996307253, + "loss": 0.9374, + "step": 2977 + }, + { + "epoch": 0.2290593031305284, + "learning_rate": 0.0026280932356556016, + "loss": 1.3866, + "step": 2978 + }, + { + "epoch": 0.22913622029074687, + "learning_rate": 0.0026278543058099384, + "loss": 0.8686, + "step": 2979 + }, + { + "epoch": 0.2292131374509653, + "learning_rate": 0.0026276153101076867, + "loss": 1.2933, + "step": 2980 + }, + { + "epoch": 0.22929005461118376, + "learning_rate": 0.002627376248562802, + "loss": 1.4172, + "step": 2981 + }, + { + "epoch": 0.2293669717714022, + "learning_rate": 0.002627137121189243, + "loss": 1.0657, + "step": 2982 + }, + { + "epoch": 0.22944388893162065, + "learning_rate": 0.0026268979280009734, + "loss": 1.0249, + "step": 2983 + }, + { + "epoch": 0.22952080609183909, + "learning_rate": 0.0026266586690119587, + "loss": 1.1686, + "step": 2984 + }, + { + "epoch": 0.22959772325205755, + "learning_rate": 0.0026264193442361704, + "loss": 0.9695, + "step": 2985 + }, + { + "epoch": 0.22967464041227598, + "learning_rate": 0.0026261799536875827, + "loss": 1.4658, + "step": 2986 + }, + { + "epoch": 0.22975155757249444, + "learning_rate": 0.0026259404973801737, + "loss": 0.9816, + "step": 2987 + }, + { + "epoch": 0.22982847473271287, + "learning_rate": 0.0026257009753279255, + "loss": 1.3082, + "step": 2988 + }, + { + "epoch": 0.22990539189293133, + "learning_rate": 0.0026254613875448247, + "loss": 1.3418, + "step": 2989 + }, + { + "epoch": 0.22998230905314976, + "learning_rate": 0.0026252217340448606, + "loss": 0.9574, + "step": 2990 + }, + { + "epoch": 0.2300592262133682, + "learning_rate": 0.002624982014842026, + "loss": 1.0907, + "step": 2991 + }, + { + "epoch": 0.23013614337358665, + "learning_rate": 0.0026247422299503194, + "loss": 1.2523, + "step": 2992 + }, + { + "epoch": 0.23021306053380508, + "learning_rate": 0.002624502379383742, + "loss": 1.2739, + "step": 2993 + }, + { + "epoch": 0.23028997769402354, + "learning_rate": 0.002624262463156299, + "loss": 1.037, + "step": 2994 + }, + { + "epoch": 0.23036689485424197, + "learning_rate": 0.002624022481281999, + "loss": 1.1786, + "step": 2995 + }, + { + "epoch": 0.23044381201446043, + "learning_rate": 0.0026237824337748544, + "loss": 0.9719, + "step": 2996 + }, + { + "epoch": 0.23052072917467886, + "learning_rate": 0.0026235423206488824, + "loss": 1.4303, + "step": 2997 + }, + { + "epoch": 0.23059764633489732, + "learning_rate": 0.0026233021419181037, + "loss": 0.7923, + "step": 2998 + }, + { + "epoch": 0.23067456349511575, + "learning_rate": 0.002623061897596542, + "loss": 1.0055, + "step": 2999 + }, + { + "epoch": 0.2307514806553342, + "learning_rate": 0.002622821587698226, + "loss": 1.4625, + "step": 3000 + }, + { + "epoch": 0.23082839781555264, + "learning_rate": 0.002622581212237187, + "loss": 1.0458, + "step": 3001 + }, + { + "epoch": 0.2309053149757711, + "learning_rate": 0.002622340771227461, + "loss": 1.2414, + "step": 3002 + }, + { + "epoch": 0.23098223213598953, + "learning_rate": 0.002622100264683088, + "loss": 1.251, + "step": 3003 + }, + { + "epoch": 0.231059149296208, + "learning_rate": 0.0026218596926181107, + "loss": 0.9071, + "step": 3004 + }, + { + "epoch": 0.23113606645642643, + "learning_rate": 0.002621619055046578, + "loss": 0.8977, + "step": 3005 + }, + { + "epoch": 0.23121298361664488, + "learning_rate": 0.0026213783519825383, + "loss": 1.1652, + "step": 3006 + }, + { + "epoch": 0.23128990077686332, + "learning_rate": 0.0026211375834400484, + "loss": 1.7237, + "step": 3007 + }, + { + "epoch": 0.23136681793708178, + "learning_rate": 0.0026208967494331667, + "loss": 0.8388, + "step": 3008 + }, + { + "epoch": 0.2314437350973002, + "learning_rate": 0.002620655849975956, + "loss": 1.2687, + "step": 3009 + }, + { + "epoch": 0.23152065225751864, + "learning_rate": 0.002620414885082482, + "loss": 1.6059, + "step": 3010 + }, + { + "epoch": 0.2315975694177371, + "learning_rate": 0.002620173854766815, + "loss": 1.1278, + "step": 3011 + }, + { + "epoch": 0.23167448657795553, + "learning_rate": 0.002619932759043029, + "loss": 1.1644, + "step": 3012 + }, + { + "epoch": 0.231751403738174, + "learning_rate": 0.002619691597925203, + "loss": 1.0764, + "step": 3013 + }, + { + "epoch": 0.23182832089839242, + "learning_rate": 0.0026194503714274163, + "loss": 1.094, + "step": 3014 + }, + { + "epoch": 0.23190523805861088, + "learning_rate": 0.0026192090795637566, + "loss": 1.5712, + "step": 3015 + }, + { + "epoch": 0.2319821552188293, + "learning_rate": 0.002618967722348312, + "loss": 1.2202, + "step": 3016 + }, + { + "epoch": 0.23205907237904777, + "learning_rate": 0.0026187262997951766, + "loss": 0.9639, + "step": 3017 + }, + { + "epoch": 0.2321359895392662, + "learning_rate": 0.002618484811918446, + "loss": 1.2411, + "step": 3018 + }, + { + "epoch": 0.23221290669948466, + "learning_rate": 0.002618243258732222, + "loss": 1.0966, + "step": 3019 + }, + { + "epoch": 0.2322898238597031, + "learning_rate": 0.002618001640250608, + "loss": 0.8387, + "step": 3020 + }, + { + "epoch": 0.23236674101992155, + "learning_rate": 0.0026177599564877136, + "loss": 1.5191, + "step": 3021 + }, + { + "epoch": 0.23244365818013998, + "learning_rate": 0.0026175182074576506, + "loss": 1.2134, + "step": 3022 + }, + { + "epoch": 0.23252057534035844, + "learning_rate": 0.0026172763931745346, + "loss": 0.8338, + "step": 3023 + }, + { + "epoch": 0.23259749250057687, + "learning_rate": 0.0026170345136524854, + "loss": 1.2159, + "step": 3024 + }, + { + "epoch": 0.23267440966079533, + "learning_rate": 0.002616792568905627, + "loss": 1.2989, + "step": 3025 + }, + { + "epoch": 0.23275132682101377, + "learning_rate": 0.0026165505589480866, + "loss": 0.9664, + "step": 3026 + }, + { + "epoch": 0.23282824398123222, + "learning_rate": 0.0026163084837939957, + "loss": 1.0549, + "step": 3027 + }, + { + "epoch": 0.23290516114145066, + "learning_rate": 0.002616066343457489, + "loss": 1.5709, + "step": 3028 + }, + { + "epoch": 0.23298207830166912, + "learning_rate": 0.0026158241379527045, + "loss": 1.1979, + "step": 3029 + }, + { + "epoch": 0.23305899546188755, + "learning_rate": 0.0026155818672937868, + "loss": 1.2932, + "step": 3030 + }, + { + "epoch": 0.23313591262210598, + "learning_rate": 0.0026153395314948806, + "loss": 0.9649, + "step": 3031 + }, + { + "epoch": 0.23321282978232444, + "learning_rate": 0.002615097130570137, + "loss": 0.8093, + "step": 3032 + }, + { + "epoch": 0.23328974694254287, + "learning_rate": 0.00261485466453371, + "loss": 1.4998, + "step": 3033 + }, + { + "epoch": 0.23336666410276133, + "learning_rate": 0.0026146121333997566, + "loss": 1.4707, + "step": 3034 + }, + { + "epoch": 0.23344358126297976, + "learning_rate": 0.0026143695371824397, + "loss": 1.1959, + "step": 3035 + }, + { + "epoch": 0.23352049842319822, + "learning_rate": 0.002614126875895924, + "loss": 1.1982, + "step": 3036 + }, + { + "epoch": 0.23359741558341665, + "learning_rate": 0.002613884149554379, + "loss": 1.1764, + "step": 3037 + }, + { + "epoch": 0.2336743327436351, + "learning_rate": 0.0026136413581719776, + "loss": 1.2159, + "step": 3038 + }, + { + "epoch": 0.23375124990385354, + "learning_rate": 0.0026133985017628967, + "loss": 1.1605, + "step": 3039 + }, + { + "epoch": 0.233828167064072, + "learning_rate": 0.002613155580341317, + "loss": 1.3456, + "step": 3040 + }, + { + "epoch": 0.23390508422429043, + "learning_rate": 0.0026129125939214222, + "loss": 1.174, + "step": 3041 + }, + { + "epoch": 0.2339820013845089, + "learning_rate": 0.0026126695425174023, + "loss": 0.9774, + "step": 3042 + }, + { + "epoch": 0.23405891854472732, + "learning_rate": 0.0026124264261434475, + "loss": 1.4991, + "step": 3043 + }, + { + "epoch": 0.23413583570494578, + "learning_rate": 0.0026121832448137545, + "loss": 1.1045, + "step": 3044 + }, + { + "epoch": 0.23421275286516421, + "learning_rate": 0.002611939998542523, + "loss": 0.762, + "step": 3045 + }, + { + "epoch": 0.23428967002538267, + "learning_rate": 0.0026116966873439553, + "loss": 1.2953, + "step": 3046 + }, + { + "epoch": 0.2343665871856011, + "learning_rate": 0.00261145331123226, + "loss": 1.1405, + "step": 3047 + }, + { + "epoch": 0.23444350434581956, + "learning_rate": 0.0026112098702216475, + "loss": 1.0199, + "step": 3048 + }, + { + "epoch": 0.234520421506038, + "learning_rate": 0.002610966364326333, + "loss": 0.9345, + "step": 3049 + }, + { + "epoch": 0.23459733866625646, + "learning_rate": 0.002610722793560534, + "loss": 1.4446, + "step": 3050 + }, + { + "epoch": 0.2346742558264749, + "learning_rate": 0.002610479157938474, + "loss": 0.9241, + "step": 3051 + }, + { + "epoch": 0.23475117298669332, + "learning_rate": 0.0026102354574743785, + "loss": 1.2471, + "step": 3052 + }, + { + "epoch": 0.23482809014691178, + "learning_rate": 0.002609991692182478, + "loss": 1.2094, + "step": 3053 + }, + { + "epoch": 0.2349050073071302, + "learning_rate": 0.002609747862077005, + "loss": 1.0292, + "step": 3054 + }, + { + "epoch": 0.23498192446734867, + "learning_rate": 0.0026095039671721982, + "loss": 1.069, + "step": 3055 + }, + { + "epoch": 0.2350588416275671, + "learning_rate": 0.002609260007482298, + "loss": 0.7545, + "step": 3056 + }, + { + "epoch": 0.23513575878778556, + "learning_rate": 0.0026090159830215506, + "loss": 1.3558, + "step": 3057 + }, + { + "epoch": 0.235212675948004, + "learning_rate": 0.0026087718938042036, + "loss": 1.3914, + "step": 3058 + }, + { + "epoch": 0.23528959310822245, + "learning_rate": 0.0026085277398445105, + "loss": 1.0881, + "step": 3059 + }, + { + "epoch": 0.23536651026844088, + "learning_rate": 0.0026082835211567275, + "loss": 1.2548, + "step": 3060 + }, + { + "epoch": 0.23544342742865934, + "learning_rate": 0.0026080392377551146, + "loss": 1.3836, + "step": 3061 + }, + { + "epoch": 0.23552034458887777, + "learning_rate": 0.0026077948896539357, + "loss": 1.0861, + "step": 3062 + }, + { + "epoch": 0.23559726174909623, + "learning_rate": 0.002607550476867459, + "loss": 1.1892, + "step": 3063 + }, + { + "epoch": 0.23567417890931466, + "learning_rate": 0.0026073059994099557, + "loss": 1.0239, + "step": 3064 + }, + { + "epoch": 0.23575109606953312, + "learning_rate": 0.0026070614572957004, + "loss": 0.9831, + "step": 3065 + }, + { + "epoch": 0.23582801322975155, + "learning_rate": 0.002606816850538973, + "loss": 1.5357, + "step": 3066 + }, + { + "epoch": 0.23590493038997, + "learning_rate": 0.0026065721791540567, + "loss": 1.1955, + "step": 3067 + }, + { + "epoch": 0.23598184755018844, + "learning_rate": 0.002606327443155237, + "loss": 1.1478, + "step": 3068 + }, + { + "epoch": 0.2360587647104069, + "learning_rate": 0.0026060826425568055, + "loss": 1.3183, + "step": 3069 + }, + { + "epoch": 0.23613568187062534, + "learning_rate": 0.002605837777373056, + "loss": 1.1357, + "step": 3070 + }, + { + "epoch": 0.2362125990308438, + "learning_rate": 0.0026055928476182853, + "loss": 0.9016, + "step": 3071 + }, + { + "epoch": 0.23628951619106223, + "learning_rate": 0.002605347853306797, + "loss": 1.3511, + "step": 3072 + }, + { + "epoch": 0.23636643335128066, + "learning_rate": 0.0026051027944528947, + "loss": 1.056, + "step": 3073 + }, + { + "epoch": 0.23644335051149912, + "learning_rate": 0.0026048576710708893, + "loss": 1.1846, + "step": 3074 + }, + { + "epoch": 0.23652026767171755, + "learning_rate": 0.0026046124831750925, + "loss": 1.0967, + "step": 3075 + }, + { + "epoch": 0.236597184831936, + "learning_rate": 0.0026043672307798217, + "loss": 1.1679, + "step": 3076 + }, + { + "epoch": 0.23667410199215444, + "learning_rate": 0.0026041219138993976, + "loss": 1.1025, + "step": 3077 + }, + { + "epoch": 0.2367510191523729, + "learning_rate": 0.002603876532548144, + "loss": 1.0724, + "step": 3078 + }, + { + "epoch": 0.23682793631259133, + "learning_rate": 0.002603631086740389, + "loss": 1.1619, + "step": 3079 + }, + { + "epoch": 0.2369048534728098, + "learning_rate": 0.002603385576490465, + "loss": 1.3809, + "step": 3080 + }, + { + "epoch": 0.23698177063302822, + "learning_rate": 0.002603140001812707, + "loss": 1.0079, + "step": 3081 + }, + { + "epoch": 0.23705868779324668, + "learning_rate": 0.002602894362721455, + "loss": 1.2764, + "step": 3082 + }, + { + "epoch": 0.2371356049534651, + "learning_rate": 0.0026026486592310516, + "loss": 1.2921, + "step": 3083 + }, + { + "epoch": 0.23721252211368357, + "learning_rate": 0.0026024028913558435, + "loss": 1.4317, + "step": 3084 + }, + { + "epoch": 0.237289439273902, + "learning_rate": 0.002602157059110182, + "loss": 1.2007, + "step": 3085 + }, + { + "epoch": 0.23736635643412046, + "learning_rate": 0.0026019111625084214, + "loss": 0.9553, + "step": 3086 + }, + { + "epoch": 0.2374432735943389, + "learning_rate": 0.0026016652015649192, + "loss": 1.2116, + "step": 3087 + }, + { + "epoch": 0.23752019075455735, + "learning_rate": 0.002601419176294038, + "loss": 1.376, + "step": 3088 + }, + { + "epoch": 0.23759710791477578, + "learning_rate": 0.0026011730867101434, + "loss": 1.2945, + "step": 3089 + }, + { + "epoch": 0.23767402507499424, + "learning_rate": 0.002600926932827604, + "loss": 0.9543, + "step": 3090 + }, + { + "epoch": 0.23775094223521268, + "learning_rate": 0.0026006807146607943, + "loss": 1.1269, + "step": 3091 + }, + { + "epoch": 0.2378278593954311, + "learning_rate": 0.0026004344322240903, + "loss": 1.3008, + "step": 3092 + }, + { + "epoch": 0.23790477655564957, + "learning_rate": 0.002600188085531873, + "loss": 0.9291, + "step": 3093 + }, + { + "epoch": 0.237981693715868, + "learning_rate": 0.002599941674598527, + "loss": 0.887, + "step": 3094 + }, + { + "epoch": 0.23805861087608646, + "learning_rate": 0.00259969519943844, + "loss": 0.9833, + "step": 3095 + }, + { + "epoch": 0.2381355280363049, + "learning_rate": 0.0025994486600660043, + "loss": 1.3557, + "step": 3096 + }, + { + "epoch": 0.23821244519652335, + "learning_rate": 0.002599202056495615, + "loss": 1.0028, + "step": 3097 + }, + { + "epoch": 0.23828936235674178, + "learning_rate": 0.002598955388741673, + "loss": 1.2178, + "step": 3098 + }, + { + "epoch": 0.23836627951696024, + "learning_rate": 0.0025987086568185803, + "loss": 0.8754, + "step": 3099 + }, + { + "epoch": 0.23844319667717867, + "learning_rate": 0.0025984618607407437, + "loss": 1.0655, + "step": 3100 + }, + { + "epoch": 0.23852011383739713, + "learning_rate": 0.0025982150005225742, + "loss": 1.1286, + "step": 3101 + }, + { + "epoch": 0.23859703099761556, + "learning_rate": 0.002597968076178487, + "loss": 1.0951, + "step": 3102 + }, + { + "epoch": 0.23867394815783402, + "learning_rate": 0.002597721087722899, + "loss": 1.1224, + "step": 3103 + }, + { + "epoch": 0.23875086531805245, + "learning_rate": 0.002597474035170233, + "loss": 1.2548, + "step": 3104 + }, + { + "epoch": 0.2388277824782709, + "learning_rate": 0.002597226918534914, + "loss": 1.2096, + "step": 3105 + }, + { + "epoch": 0.23890469963848934, + "learning_rate": 0.002596979737831372, + "loss": 1.1102, + "step": 3106 + }, + { + "epoch": 0.2389816167987078, + "learning_rate": 0.0025967324930740395, + "loss": 1.0185, + "step": 3107 + }, + { + "epoch": 0.23905853395892623, + "learning_rate": 0.002596485184277354, + "loss": 1.2696, + "step": 3108 + }, + { + "epoch": 0.2391354511191447, + "learning_rate": 0.002596237811455756, + "loss": 1.011, + "step": 3109 + }, + { + "epoch": 0.23921236827936312, + "learning_rate": 0.002595990374623689, + "loss": 1.3151, + "step": 3110 + }, + { + "epoch": 0.23928928543958158, + "learning_rate": 0.0025957428737956026, + "loss": 1.1528, + "step": 3111 + }, + { + "epoch": 0.23936620259980002, + "learning_rate": 0.0025954953089859477, + "loss": 0.8947, + "step": 3112 + }, + { + "epoch": 0.23944311976001845, + "learning_rate": 0.00259524768020918, + "loss": 1.2746, + "step": 3113 + }, + { + "epoch": 0.2395200369202369, + "learning_rate": 0.0025949999874797587, + "loss": 1.4013, + "step": 3114 + }, + { + "epoch": 0.23959695408045534, + "learning_rate": 0.002594752230812147, + "loss": 1.1671, + "step": 3115 + }, + { + "epoch": 0.2396738712406738, + "learning_rate": 0.0025945044102208114, + "loss": 0.9244, + "step": 3116 + }, + { + "epoch": 0.23975078840089223, + "learning_rate": 0.002594256525720223, + "loss": 1.1692, + "step": 3117 + }, + { + "epoch": 0.2398277055611107, + "learning_rate": 0.002594008577324855, + "loss": 1.4609, + "step": 3118 + }, + { + "epoch": 0.23990462272132912, + "learning_rate": 0.0025937605650491868, + "loss": 1.0407, + "step": 3119 + }, + { + "epoch": 0.23998153988154758, + "learning_rate": 0.0025935124889076987, + "loss": 0.9881, + "step": 3120 + }, + { + "epoch": 0.240058457041766, + "learning_rate": 0.0025932643489148775, + "loss": 1.289, + "step": 3121 + }, + { + "epoch": 0.24013537420198447, + "learning_rate": 0.002593016145085211, + "loss": 0.8986, + "step": 3122 + }, + { + "epoch": 0.2402122913622029, + "learning_rate": 0.0025927678774331927, + "loss": 1.3161, + "step": 3123 + }, + { + "epoch": 0.24028920852242136, + "learning_rate": 0.0025925195459733197, + "loss": 1.1897, + "step": 3124 + }, + { + "epoch": 0.2403661256826398, + "learning_rate": 0.0025922711507200915, + "loss": 1.3164, + "step": 3125 + }, + { + "epoch": 0.24044304284285825, + "learning_rate": 0.002592022691688012, + "loss": 1.0286, + "step": 3126 + }, + { + "epoch": 0.24051996000307668, + "learning_rate": 0.0025917741688915898, + "loss": 0.7483, + "step": 3127 + }, + { + "epoch": 0.24059687716329514, + "learning_rate": 0.002591525582345336, + "loss": 1.1146, + "step": 3128 + }, + { + "epoch": 0.24067379432351357, + "learning_rate": 0.002591276932063766, + "loss": 1.4364, + "step": 3129 + }, + { + "epoch": 0.24075071148373203, + "learning_rate": 0.002591028218061399, + "loss": 0.9923, + "step": 3130 + }, + { + "epoch": 0.24082762864395046, + "learning_rate": 0.0025907794403527565, + "loss": 0.8849, + "step": 3131 + }, + { + "epoch": 0.24090454580416892, + "learning_rate": 0.0025905305989523666, + "loss": 1.4039, + "step": 3132 + }, + { + "epoch": 0.24098146296438736, + "learning_rate": 0.0025902816938747574, + "loss": 1.1842, + "step": 3133 + }, + { + "epoch": 0.2410583801246058, + "learning_rate": 0.002590032725134465, + "loss": 1.0936, + "step": 3134 + }, + { + "epoch": 0.24113529728482425, + "learning_rate": 0.0025897836927460246, + "loss": 1.1739, + "step": 3135 + }, + { + "epoch": 0.24121221444504268, + "learning_rate": 0.00258953459672398, + "loss": 1.2149, + "step": 3136 + }, + { + "epoch": 0.24128913160526114, + "learning_rate": 0.002589285437082874, + "loss": 1.1202, + "step": 3137 + }, + { + "epoch": 0.24136604876547957, + "learning_rate": 0.0025890362138372555, + "loss": 1.5392, + "step": 3138 + }, + { + "epoch": 0.24144296592569803, + "learning_rate": 0.002588786927001678, + "loss": 1.0236, + "step": 3139 + }, + { + "epoch": 0.24151988308591646, + "learning_rate": 0.0025885375765906976, + "loss": 1.0635, + "step": 3140 + }, + { + "epoch": 0.24159680024613492, + "learning_rate": 0.0025882881626188734, + "loss": 0.8874, + "step": 3141 + }, + { + "epoch": 0.24167371740635335, + "learning_rate": 0.002588038685100769, + "loss": 0.9573, + "step": 3142 + }, + { + "epoch": 0.2417506345665718, + "learning_rate": 0.0025877891440509527, + "loss": 1.0899, + "step": 3143 + }, + { + "epoch": 0.24182755172679024, + "learning_rate": 0.0025875395394839933, + "loss": 1.8011, + "step": 3144 + }, + { + "epoch": 0.2419044688870087, + "learning_rate": 0.002587289871414467, + "loss": 1.4425, + "step": 3145 + }, + { + "epoch": 0.24198138604722713, + "learning_rate": 0.002587040139856953, + "loss": 1.1479, + "step": 3146 + }, + { + "epoch": 0.2420583032074456, + "learning_rate": 0.0025867903448260317, + "loss": 1.3082, + "step": 3147 + }, + { + "epoch": 0.24213522036766402, + "learning_rate": 0.0025865404863362894, + "loss": 1.0346, + "step": 3148 + }, + { + "epoch": 0.24221213752788248, + "learning_rate": 0.002586290564402316, + "loss": 1.2068, + "step": 3149 + }, + { + "epoch": 0.2422890546881009, + "learning_rate": 0.002586040579038704, + "loss": 1.2679, + "step": 3150 + }, + { + "epoch": 0.24236597184831937, + "learning_rate": 0.0025857905302600517, + "loss": 1.2732, + "step": 3151 + }, + { + "epoch": 0.2424428890085378, + "learning_rate": 0.0025855404180809578, + "loss": 1.1605, + "step": 3152 + }, + { + "epoch": 0.24251980616875624, + "learning_rate": 0.002585290242516028, + "loss": 1.1359, + "step": 3153 + }, + { + "epoch": 0.2425967233289747, + "learning_rate": 0.0025850400035798697, + "loss": 1.3522, + "step": 3154 + }, + { + "epoch": 0.24267364048919313, + "learning_rate": 0.0025847897012870945, + "loss": 1.2395, + "step": 3155 + }, + { + "epoch": 0.24275055764941159, + "learning_rate": 0.0025845393356523187, + "loss": 1.394, + "step": 3156 + }, + { + "epoch": 0.24282747480963002, + "learning_rate": 0.0025842889066901606, + "loss": 1.3979, + "step": 3157 + }, + { + "epoch": 0.24290439196984848, + "learning_rate": 0.0025840384144152433, + "loss": 1.0034, + "step": 3158 + }, + { + "epoch": 0.2429813091300669, + "learning_rate": 0.002583787858842193, + "loss": 1.0498, + "step": 3159 + }, + { + "epoch": 0.24305822629028537, + "learning_rate": 0.00258353723998564, + "loss": 1.3022, + "step": 3160 + }, + { + "epoch": 0.2431351434505038, + "learning_rate": 0.0025832865578602187, + "loss": 1.4339, + "step": 3161 + }, + { + "epoch": 0.24321206061072226, + "learning_rate": 0.0025830358124805664, + "loss": 1.281, + "step": 3162 + }, + { + "epoch": 0.2432889777709407, + "learning_rate": 0.002582785003861324, + "loss": 1.1118, + "step": 3163 + }, + { + "epoch": 0.24336589493115915, + "learning_rate": 0.002582534132017137, + "loss": 0.7518, + "step": 3164 + }, + { + "epoch": 0.24344281209137758, + "learning_rate": 0.0025822831969626535, + "loss": 1.0684, + "step": 3165 + }, + { + "epoch": 0.24351972925159604, + "learning_rate": 0.002582032198712526, + "loss": 1.2551, + "step": 3166 + }, + { + "epoch": 0.24359664641181447, + "learning_rate": 0.0025817811372814115, + "loss": 0.9243, + "step": 3167 + }, + { + "epoch": 0.24367356357203293, + "learning_rate": 0.0025815300126839683, + "loss": 1.2287, + "step": 3168 + }, + { + "epoch": 0.24375048073225136, + "learning_rate": 0.002581278824934861, + "loss": 1.2283, + "step": 3169 + }, + { + "epoch": 0.24382739789246982, + "learning_rate": 0.002581027574048756, + "loss": 1.0643, + "step": 3170 + }, + { + "epoch": 0.24390431505268825, + "learning_rate": 0.0025807762600403247, + "loss": 1.001, + "step": 3171 + }, + { + "epoch": 0.2439812322129067, + "learning_rate": 0.0025805248829242405, + "loss": 1.1246, + "step": 3172 + }, + { + "epoch": 0.24405814937312514, + "learning_rate": 0.002580273442715183, + "loss": 1.231, + "step": 3173 + }, + { + "epoch": 0.24413506653334358, + "learning_rate": 0.0025800219394278333, + "loss": 1.3881, + "step": 3174 + }, + { + "epoch": 0.24421198369356203, + "learning_rate": 0.0025797703730768766, + "loss": 1.21, + "step": 3175 + }, + { + "epoch": 0.24428890085378047, + "learning_rate": 0.0025795187436770026, + "loss": 1.2494, + "step": 3176 + }, + { + "epoch": 0.24436581801399893, + "learning_rate": 0.002579267051242904, + "loss": 1.3317, + "step": 3177 + }, + { + "epoch": 0.24444273517421736, + "learning_rate": 0.0025790152957892777, + "loss": 1.2651, + "step": 3178 + }, + { + "epoch": 0.24451965233443582, + "learning_rate": 0.0025787634773308237, + "loss": 1.108, + "step": 3179 + }, + { + "epoch": 0.24459656949465425, + "learning_rate": 0.0025785115958822462, + "loss": 0.9391, + "step": 3180 + }, + { + "epoch": 0.2446734866548727, + "learning_rate": 0.002578259651458252, + "loss": 1.1423, + "step": 3181 + }, + { + "epoch": 0.24475040381509114, + "learning_rate": 0.002578007644073554, + "loss": 1.1358, + "step": 3182 + }, + { + "epoch": 0.2448273209753096, + "learning_rate": 0.0025777555737428654, + "loss": 1.2116, + "step": 3183 + }, + { + "epoch": 0.24490423813552803, + "learning_rate": 0.0025775034404809057, + "loss": 1.3328, + "step": 3184 + }, + { + "epoch": 0.2449811552957465, + "learning_rate": 0.0025772512443023973, + "loss": 0.993, + "step": 3185 + }, + { + "epoch": 0.24505807245596492, + "learning_rate": 0.0025769989852220663, + "loss": 0.756, + "step": 3186 + }, + { + "epoch": 0.24513498961618338, + "learning_rate": 0.0025767466632546417, + "loss": 1.2895, + "step": 3187 + }, + { + "epoch": 0.2452119067764018, + "learning_rate": 0.002576494278414858, + "loss": 1.0305, + "step": 3188 + }, + { + "epoch": 0.24528882393662027, + "learning_rate": 0.0025762418307174505, + "loss": 1.4926, + "step": 3189 + }, + { + "epoch": 0.2453657410968387, + "learning_rate": 0.0025759893201771616, + "loss": 1.1996, + "step": 3190 + }, + { + "epoch": 0.24544265825705716, + "learning_rate": 0.0025757367468087346, + "loss": 1.2316, + "step": 3191 + }, + { + "epoch": 0.2455195754172756, + "learning_rate": 0.002575484110626918, + "loss": 0.8187, + "step": 3192 + }, + { + "epoch": 0.24559649257749405, + "learning_rate": 0.0025752314116464633, + "loss": 1.4764, + "step": 3193 + }, + { + "epoch": 0.24567340973771248, + "learning_rate": 0.002574978649882126, + "loss": 1.0287, + "step": 3194 + }, + { + "epoch": 0.24575032689793092, + "learning_rate": 0.0025747258253486643, + "loss": 1.2878, + "step": 3195 + }, + { + "epoch": 0.24582724405814937, + "learning_rate": 0.0025744729380608425, + "loss": 1.1815, + "step": 3196 + }, + { + "epoch": 0.2459041612183678, + "learning_rate": 0.002574219988033426, + "loss": 0.8401, + "step": 3197 + }, + { + "epoch": 0.24598107837858627, + "learning_rate": 0.0025739669752811844, + "loss": 1.1133, + "step": 3198 + }, + { + "epoch": 0.2460579955388047, + "learning_rate": 0.0025737138998188917, + "loss": 1.265, + "step": 3199 + }, + { + "epoch": 0.24613491269902316, + "learning_rate": 0.002573460761661326, + "loss": 1.5385, + "step": 3200 + }, + { + "epoch": 0.2462118298592416, + "learning_rate": 0.0025732075608232677, + "loss": 0.9086, + "step": 3201 + }, + { + "epoch": 0.24628874701946005, + "learning_rate": 0.002572954297319501, + "loss": 0.8488, + "step": 3202 + }, + { + "epoch": 0.24636566417967848, + "learning_rate": 0.002572700971164815, + "loss": 0.7496, + "step": 3203 + }, + { + "epoch": 0.24644258133989694, + "learning_rate": 0.0025724475823740013, + "loss": 1.1925, + "step": 3204 + }, + { + "epoch": 0.24651949850011537, + "learning_rate": 0.0025721941309618554, + "loss": 1.2115, + "step": 3205 + }, + { + "epoch": 0.24659641566033383, + "learning_rate": 0.0025719406169431767, + "loss": 1.2288, + "step": 3206 + }, + { + "epoch": 0.24667333282055226, + "learning_rate": 0.0025716870403327684, + "loss": 1.0751, + "step": 3207 + }, + { + "epoch": 0.24675024998077072, + "learning_rate": 0.0025714334011454373, + "loss": 1.1344, + "step": 3208 + }, + { + "epoch": 0.24682716714098915, + "learning_rate": 0.0025711796993959928, + "loss": 1.2588, + "step": 3209 + }, + { + "epoch": 0.2469040843012076, + "learning_rate": 0.0025709259350992497, + "loss": 1.3606, + "step": 3210 + }, + { + "epoch": 0.24698100146142604, + "learning_rate": 0.0025706721082700244, + "loss": 1.2679, + "step": 3211 + }, + { + "epoch": 0.2470579186216445, + "learning_rate": 0.0025704182189231397, + "loss": 1.1779, + "step": 3212 + }, + { + "epoch": 0.24713483578186293, + "learning_rate": 0.002570164267073419, + "loss": 1.2455, + "step": 3213 + }, + { + "epoch": 0.2472117529420814, + "learning_rate": 0.002569910252735692, + "loss": 1.1148, + "step": 3214 + }, + { + "epoch": 0.24728867010229982, + "learning_rate": 0.00256965617592479, + "loss": 1.2124, + "step": 3215 + }, + { + "epoch": 0.24736558726251826, + "learning_rate": 0.002569402036655549, + "loss": 1.5176, + "step": 3216 + }, + { + "epoch": 0.24744250442273671, + "learning_rate": 0.002569147834942809, + "loss": 1.4819, + "step": 3217 + }, + { + "epoch": 0.24751942158295515, + "learning_rate": 0.0025688935708014123, + "loss": 1.2158, + "step": 3218 + }, + { + "epoch": 0.2475963387431736, + "learning_rate": 0.0025686392442462063, + "loss": 1.4652, + "step": 3219 + }, + { + "epoch": 0.24767325590339204, + "learning_rate": 0.0025683848552920407, + "loss": 1.0987, + "step": 3220 + }, + { + "epoch": 0.2477501730636105, + "learning_rate": 0.0025681304039537702, + "loss": 1.4371, + "step": 3221 + }, + { + "epoch": 0.24782709022382893, + "learning_rate": 0.0025678758902462524, + "loss": 1.0194, + "step": 3222 + }, + { + "epoch": 0.2479040073840474, + "learning_rate": 0.002567621314184348, + "loss": 1.3744, + "step": 3223 + }, + { + "epoch": 0.24798092454426582, + "learning_rate": 0.0025673666757829233, + "loss": 0.9836, + "step": 3224 + }, + { + "epoch": 0.24805784170448428, + "learning_rate": 0.002567111975056845, + "loss": 1.145, + "step": 3225 + }, + { + "epoch": 0.2481347588647027, + "learning_rate": 0.0025668572120209872, + "loss": 1.5748, + "step": 3226 + }, + { + "epoch": 0.24821167602492117, + "learning_rate": 0.0025666023866902244, + "loss": 1.2518, + "step": 3227 + }, + { + "epoch": 0.2482885931851396, + "learning_rate": 0.0025663474990794367, + "loss": 1.1501, + "step": 3228 + }, + { + "epoch": 0.24836551034535806, + "learning_rate": 0.0025660925492035076, + "loss": 1.4456, + "step": 3229 + }, + { + "epoch": 0.2484424275055765, + "learning_rate": 0.0025658375370773232, + "loss": 1.1321, + "step": 3230 + }, + { + "epoch": 0.24851934466579495, + "learning_rate": 0.002565582462715774, + "loss": 1.1396, + "step": 3231 + }, + { + "epoch": 0.24859626182601338, + "learning_rate": 0.0025653273261337546, + "loss": 1.3533, + "step": 3232 + }, + { + "epoch": 0.24867317898623184, + "learning_rate": 0.002565072127346162, + "loss": 0.9253, + "step": 3233 + }, + { + "epoch": 0.24875009614645027, + "learning_rate": 0.002564816866367898, + "loss": 1.325, + "step": 3234 + }, + { + "epoch": 0.2488270133066687, + "learning_rate": 0.002564561543213868, + "loss": 0.815, + "step": 3235 + }, + { + "epoch": 0.24890393046688716, + "learning_rate": 0.0025643061578989796, + "loss": 1.3108, + "step": 3236 + }, + { + "epoch": 0.2489808476271056, + "learning_rate": 0.0025640507104381454, + "loss": 1.0179, + "step": 3237 + }, + { + "epoch": 0.24905776478732405, + "learning_rate": 0.0025637952008462817, + "loss": 1.049, + "step": 3238 + }, + { + "epoch": 0.24913468194754249, + "learning_rate": 0.002563539629138307, + "loss": 1.2534, + "step": 3239 + }, + { + "epoch": 0.24921159910776094, + "learning_rate": 0.0025632839953291455, + "loss": 0.7798, + "step": 3240 + }, + { + "epoch": 0.24928851626797938, + "learning_rate": 0.002563028299433723, + "loss": 1.1917, + "step": 3241 + }, + { + "epoch": 0.24936543342819784, + "learning_rate": 0.0025627725414669703, + "loss": 1.0835, + "step": 3242 + }, + { + "epoch": 0.24944235058841627, + "learning_rate": 0.0025625167214438213, + "loss": 1.2254, + "step": 3243 + }, + { + "epoch": 0.24951926774863473, + "learning_rate": 0.0025622608393792135, + "loss": 1.1073, + "step": 3244 + }, + { + "epoch": 0.24959618490885316, + "learning_rate": 0.0025620048952880886, + "loss": 1.2777, + "step": 3245 + }, + { + "epoch": 0.24967310206907162, + "learning_rate": 0.0025617488891853902, + "loss": 1.0222, + "step": 3246 + }, + { + "epoch": 0.24975001922929005, + "learning_rate": 0.0025614928210860688, + "loss": 1.0946, + "step": 3247 + }, + { + "epoch": 0.2498269363895085, + "learning_rate": 0.0025612366910050745, + "loss": 1.0655, + "step": 3248 + }, + { + "epoch": 0.24990385354972694, + "learning_rate": 0.0025609804989573646, + "loss": 1.0671, + "step": 3249 + }, + { + "epoch": 0.2499807707099454, + "learning_rate": 0.0025607242449578967, + "loss": 1.127, + "step": 3250 + }, + { + "epoch": 0.25005768787016386, + "learning_rate": 0.002560467929021635, + "loss": 1.0283, + "step": 3251 + }, + { + "epoch": 0.25013460503038226, + "learning_rate": 0.0025602115511635464, + "loss": 1.198, + "step": 3252 + }, + { + "epoch": 0.2502115221906007, + "learning_rate": 0.0025599551113985996, + "loss": 1.2752, + "step": 3253 + }, + { + "epoch": 0.2502884393508192, + "learning_rate": 0.0025596986097417692, + "loss": 0.867, + "step": 3254 + }, + { + "epoch": 0.25036535651103764, + "learning_rate": 0.0025594420462080334, + "loss": 1.1996, + "step": 3255 + }, + { + "epoch": 0.25044227367125604, + "learning_rate": 0.0025591854208123713, + "loss": 1.1055, + "step": 3256 + }, + { + "epoch": 0.2505191908314745, + "learning_rate": 0.0025589287335697694, + "loss": 1.3893, + "step": 3257 + }, + { + "epoch": 0.25059610799169296, + "learning_rate": 0.0025586719844952147, + "loss": 1.2048, + "step": 3258 + }, + { + "epoch": 0.25067302515191137, + "learning_rate": 0.0025584151736037, + "loss": 1.031, + "step": 3259 + }, + { + "epoch": 0.2507499423121298, + "learning_rate": 0.00255815830091022, + "loss": 1.6608, + "step": 3260 + }, + { + "epoch": 0.2508268594723483, + "learning_rate": 0.0025579013664297737, + "loss": 1.3659, + "step": 3261 + }, + { + "epoch": 0.25090377663256674, + "learning_rate": 0.0025576443701773644, + "loss": 1.1875, + "step": 3262 + }, + { + "epoch": 0.25098069379278515, + "learning_rate": 0.002557387312167998, + "loss": 1.1799, + "step": 3263 + }, + { + "epoch": 0.2510576109530036, + "learning_rate": 0.0025571301924166845, + "loss": 1.3114, + "step": 3264 + }, + { + "epoch": 0.25113452811322207, + "learning_rate": 0.002556873010938437, + "loss": 1.3133, + "step": 3265 + }, + { + "epoch": 0.2512114452734405, + "learning_rate": 0.002556615767748274, + "loss": 1.0744, + "step": 3266 + }, + { + "epoch": 0.25128836243365893, + "learning_rate": 0.0025563584628612145, + "loss": 1.1149, + "step": 3267 + }, + { + "epoch": 0.2513652795938774, + "learning_rate": 0.0025561010962922835, + "loss": 0.9961, + "step": 3268 + }, + { + "epoch": 0.25144219675409585, + "learning_rate": 0.0025558436680565085, + "loss": 1.5148, + "step": 3269 + }, + { + "epoch": 0.2515191139143143, + "learning_rate": 0.0025555861781689213, + "loss": 1.1389, + "step": 3270 + }, + { + "epoch": 0.2515960310745327, + "learning_rate": 0.002555328626644558, + "loss": 0.9799, + "step": 3271 + }, + { + "epoch": 0.25167294823475117, + "learning_rate": 0.0025550710134984553, + "loss": 1.1286, + "step": 3272 + }, + { + "epoch": 0.25174986539496963, + "learning_rate": 0.002554813338745657, + "loss": 1.282, + "step": 3273 + }, + { + "epoch": 0.2518267825551881, + "learning_rate": 0.0025545556024012084, + "loss": 1.2751, + "step": 3274 + }, + { + "epoch": 0.2519036997154065, + "learning_rate": 0.002554297804480159, + "loss": 1.2496, + "step": 3275 + }, + { + "epoch": 0.25198061687562495, + "learning_rate": 0.0025540399449975625, + "loss": 0.9874, + "step": 3276 + }, + { + "epoch": 0.2520575340358434, + "learning_rate": 0.002553782023968475, + "loss": 1.2749, + "step": 3277 + }, + { + "epoch": 0.2521344511960618, + "learning_rate": 0.002553524041407957, + "loss": 1.1992, + "step": 3278 + }, + { + "epoch": 0.2522113683562803, + "learning_rate": 0.002553265997331072, + "loss": 1.1654, + "step": 3279 + }, + { + "epoch": 0.25228828551649873, + "learning_rate": 0.002553007891752888, + "loss": 1.2483, + "step": 3280 + }, + { + "epoch": 0.2523652026767172, + "learning_rate": 0.002552749724688475, + "loss": 1.2375, + "step": 3281 + }, + { + "epoch": 0.2524421198369356, + "learning_rate": 0.0025524914961529096, + "loss": 1.2751, + "step": 3282 + }, + { + "epoch": 0.25251903699715406, + "learning_rate": 0.0025522332061612683, + "loss": 1.1489, + "step": 3283 + }, + { + "epoch": 0.2525959541573725, + "learning_rate": 0.002551974854728633, + "loss": 1.3221, + "step": 3284 + }, + { + "epoch": 0.252672871317591, + "learning_rate": 0.0025517164418700903, + "loss": 1.2766, + "step": 3285 + }, + { + "epoch": 0.2527497884778094, + "learning_rate": 0.002551457967600728, + "loss": 1.1042, + "step": 3286 + }, + { + "epoch": 0.25282670563802784, + "learning_rate": 0.0025511994319356393, + "loss": 1.1367, + "step": 3287 + }, + { + "epoch": 0.2529036227982463, + "learning_rate": 0.0025509408348899205, + "loss": 1.4032, + "step": 3288 + }, + { + "epoch": 0.25298053995846476, + "learning_rate": 0.002550682176478671, + "loss": 1.0402, + "step": 3289 + }, + { + "epoch": 0.25305745711868316, + "learning_rate": 0.0025504234567169937, + "loss": 1.1797, + "step": 3290 + }, + { + "epoch": 0.2531343742789016, + "learning_rate": 0.0025501646756199967, + "loss": 1.221, + "step": 3291 + }, + { + "epoch": 0.2532112914391201, + "learning_rate": 0.0025499058332027896, + "loss": 1.2086, + "step": 3292 + }, + { + "epoch": 0.25328820859933854, + "learning_rate": 0.0025496469294804865, + "loss": 1.2087, + "step": 3293 + }, + { + "epoch": 0.25336512575955694, + "learning_rate": 0.002549387964468206, + "loss": 0.7126, + "step": 3294 + }, + { + "epoch": 0.2534420429197754, + "learning_rate": 0.002549128938181068, + "loss": 1.2015, + "step": 3295 + }, + { + "epoch": 0.25351896007999386, + "learning_rate": 0.0025488698506341977, + "loss": 1.4902, + "step": 3296 + }, + { + "epoch": 0.25359587724021226, + "learning_rate": 0.0025486107018427243, + "loss": 1.5627, + "step": 3297 + }, + { + "epoch": 0.2536727944004307, + "learning_rate": 0.0025483514918217785, + "loss": 1.1413, + "step": 3298 + }, + { + "epoch": 0.2537497115606492, + "learning_rate": 0.002548092220586497, + "loss": 1.0084, + "step": 3299 + }, + { + "epoch": 0.25382662872086764, + "learning_rate": 0.0025478328881520184, + "loss": 1.2909, + "step": 3300 + }, + { + "epoch": 0.25390354588108605, + "learning_rate": 0.002547573494533486, + "loss": 1.3393, + "step": 3301 + }, + { + "epoch": 0.2539804630413045, + "learning_rate": 0.0025473140397460443, + "loss": 1.1288, + "step": 3302 + }, + { + "epoch": 0.25405738020152296, + "learning_rate": 0.002547054523804845, + "loss": 1.08, + "step": 3303 + }, + { + "epoch": 0.2541342973617414, + "learning_rate": 0.002546794946725041, + "loss": 0.8534, + "step": 3304 + }, + { + "epoch": 0.2542112145219598, + "learning_rate": 0.0025465353085217892, + "loss": 1.3742, + "step": 3305 + }, + { + "epoch": 0.2542881316821783, + "learning_rate": 0.0025462756092102493, + "loss": 0.9702, + "step": 3306 + }, + { + "epoch": 0.25436504884239675, + "learning_rate": 0.0025460158488055866, + "loss": 1.2243, + "step": 3307 + }, + { + "epoch": 0.2544419660026152, + "learning_rate": 0.0025457560273229687, + "loss": 1.3708, + "step": 3308 + }, + { + "epoch": 0.2545188831628336, + "learning_rate": 0.002545496144777566, + "loss": 1.0532, + "step": 3309 + }, + { + "epoch": 0.25459580032305207, + "learning_rate": 0.002545236201184554, + "loss": 1.21, + "step": 3310 + }, + { + "epoch": 0.2546727174832705, + "learning_rate": 0.002544976196559111, + "loss": 1.0845, + "step": 3311 + }, + { + "epoch": 0.254749634643489, + "learning_rate": 0.002544716130916419, + "loss": 0.9006, + "step": 3312 + }, + { + "epoch": 0.2548265518037074, + "learning_rate": 0.002544456004271663, + "loss": 1.0491, + "step": 3313 + }, + { + "epoch": 0.25490346896392585, + "learning_rate": 0.0025441958166400325, + "loss": 1.2703, + "step": 3314 + }, + { + "epoch": 0.2549803861241443, + "learning_rate": 0.00254393556803672, + "loss": 1.1311, + "step": 3315 + }, + { + "epoch": 0.25505730328436277, + "learning_rate": 0.0025436752584769213, + "loss": 1.0844, + "step": 3316 + }, + { + "epoch": 0.2551342204445812, + "learning_rate": 0.0025434148879758373, + "loss": 1.5026, + "step": 3317 + }, + { + "epoch": 0.25521113760479963, + "learning_rate": 0.00254315445654867, + "loss": 1.1845, + "step": 3318 + }, + { + "epoch": 0.2552880547650181, + "learning_rate": 0.0025428939642106276, + "loss": 1.1035, + "step": 3319 + }, + { + "epoch": 0.2553649719252365, + "learning_rate": 0.002542633410976919, + "loss": 1.0475, + "step": 3320 + }, + { + "epoch": 0.25544188908545495, + "learning_rate": 0.002542372796862759, + "loss": 0.9373, + "step": 3321 + }, + { + "epoch": 0.2555188062456734, + "learning_rate": 0.0025421121218833654, + "loss": 0.7934, + "step": 3322 + }, + { + "epoch": 0.2555957234058919, + "learning_rate": 0.0025418513860539587, + "loss": 1.4146, + "step": 3323 + }, + { + "epoch": 0.2556726405661103, + "learning_rate": 0.0025415905893897642, + "loss": 1.0758, + "step": 3324 + }, + { + "epoch": 0.25574955772632874, + "learning_rate": 0.002541329731906009, + "loss": 1.0413, + "step": 3325 + }, + { + "epoch": 0.2558264748865472, + "learning_rate": 0.0025410688136179257, + "loss": 1.2682, + "step": 3326 + }, + { + "epoch": 0.25590339204676565, + "learning_rate": 0.00254080783454075, + "loss": 1.2823, + "step": 3327 + }, + { + "epoch": 0.25598030920698406, + "learning_rate": 0.0025405467946897197, + "loss": 1.1681, + "step": 3328 + }, + { + "epoch": 0.2560572263672025, + "learning_rate": 0.0025402856940800775, + "loss": 1.0219, + "step": 3329 + }, + { + "epoch": 0.256134143527421, + "learning_rate": 0.0025400245327270696, + "loss": 1.0044, + "step": 3330 + }, + { + "epoch": 0.25621106068763944, + "learning_rate": 0.002539763310645945, + "loss": 0.9554, + "step": 3331 + }, + { + "epoch": 0.25628797784785784, + "learning_rate": 0.0025395020278519576, + "loss": 1.429, + "step": 3332 + }, + { + "epoch": 0.2563648950080763, + "learning_rate": 0.0025392406843603634, + "loss": 0.9826, + "step": 3333 + }, + { + "epoch": 0.25644181216829476, + "learning_rate": 0.0025389792801864225, + "loss": 1.1936, + "step": 3334 + }, + { + "epoch": 0.2565187293285132, + "learning_rate": 0.0025387178153453986, + "loss": 0.9941, + "step": 3335 + }, + { + "epoch": 0.2565956464887316, + "learning_rate": 0.0025384562898525588, + "loss": 1.112, + "step": 3336 + }, + { + "epoch": 0.2566725636489501, + "learning_rate": 0.002538194703723174, + "loss": 1.2618, + "step": 3337 + }, + { + "epoch": 0.25674948080916854, + "learning_rate": 0.0025379330569725183, + "loss": 0.7113, + "step": 3338 + }, + { + "epoch": 0.25682639796938694, + "learning_rate": 0.0025376713496158697, + "loss": 1.2442, + "step": 3339 + }, + { + "epoch": 0.2569033151296054, + "learning_rate": 0.0025374095816685103, + "loss": 1.1023, + "step": 3340 + }, + { + "epoch": 0.25698023228982386, + "learning_rate": 0.0025371477531457233, + "loss": 1.3482, + "step": 3341 + }, + { + "epoch": 0.2570571494500423, + "learning_rate": 0.0025368858640627987, + "loss": 1.6154, + "step": 3342 + }, + { + "epoch": 0.2571340666102607, + "learning_rate": 0.002536623914435028, + "loss": 0.895, + "step": 3343 + }, + { + "epoch": 0.2572109837704792, + "learning_rate": 0.0025363619042777065, + "loss": 1.1876, + "step": 3344 + }, + { + "epoch": 0.25728790093069764, + "learning_rate": 0.002536099833606133, + "loss": 1.3241, + "step": 3345 + }, + { + "epoch": 0.2573648180909161, + "learning_rate": 0.0025358377024356105, + "loss": 0.7273, + "step": 3346 + }, + { + "epoch": 0.2574417352511345, + "learning_rate": 0.0025355755107814453, + "loss": 1.1544, + "step": 3347 + }, + { + "epoch": 0.25751865241135297, + "learning_rate": 0.0025353132586589462, + "loss": 1.1422, + "step": 3348 + }, + { + "epoch": 0.2575955695715714, + "learning_rate": 0.0025350509460834274, + "loss": 1.3658, + "step": 3349 + }, + { + "epoch": 0.2576724867317899, + "learning_rate": 0.002534788573070205, + "loss": 1.0071, + "step": 3350 + }, + { + "epoch": 0.2577494038920083, + "learning_rate": 0.0025345261396345995, + "loss": 1.0649, + "step": 3351 + }, + { + "epoch": 0.25782632105222675, + "learning_rate": 0.002534263645791935, + "loss": 1.415, + "step": 3352 + }, + { + "epoch": 0.2579032382124452, + "learning_rate": 0.0025340010915575373, + "loss": 1.492, + "step": 3353 + }, + { + "epoch": 0.25798015537266367, + "learning_rate": 0.002533738476946739, + "loss": 1.0133, + "step": 3354 + }, + { + "epoch": 0.25805707253288207, + "learning_rate": 0.002533475801974874, + "loss": 1.13, + "step": 3355 + }, + { + "epoch": 0.25813398969310053, + "learning_rate": 0.002533213066657279, + "loss": 1.1198, + "step": 3356 + }, + { + "epoch": 0.258210906853319, + "learning_rate": 0.0025329502710092966, + "loss": 0.7984, + "step": 3357 + }, + { + "epoch": 0.2582878240135374, + "learning_rate": 0.0025326874150462713, + "loss": 1.0155, + "step": 3358 + }, + { + "epoch": 0.25836474117375585, + "learning_rate": 0.002532424498783552, + "loss": 1.0591, + "step": 3359 + }, + { + "epoch": 0.2584416583339743, + "learning_rate": 0.0025321615222364895, + "loss": 0.9495, + "step": 3360 + }, + { + "epoch": 0.25851857549419277, + "learning_rate": 0.0025318984854204407, + "loss": 1.349, + "step": 3361 + }, + { + "epoch": 0.2585954926544112, + "learning_rate": 0.0025316353883507634, + "loss": 1.036, + "step": 3362 + }, + { + "epoch": 0.25867240981462963, + "learning_rate": 0.002531372231042821, + "loss": 1.448, + "step": 3363 + }, + { + "epoch": 0.2587493269748481, + "learning_rate": 0.002531109013511979, + "loss": 1.267, + "step": 3364 + }, + { + "epoch": 0.25882624413506655, + "learning_rate": 0.0025308457357736073, + "loss": 1.0893, + "step": 3365 + }, + { + "epoch": 0.25890316129528496, + "learning_rate": 0.0025305823978430788, + "loss": 1.2601, + "step": 3366 + }, + { + "epoch": 0.2589800784555034, + "learning_rate": 0.00253031899973577, + "loss": 1.5132, + "step": 3367 + }, + { + "epoch": 0.2590569956157219, + "learning_rate": 0.0025300555414670607, + "loss": 1.187, + "step": 3368 + }, + { + "epoch": 0.25913391277594033, + "learning_rate": 0.0025297920230523347, + "loss": 1.0362, + "step": 3369 + }, + { + "epoch": 0.25921082993615874, + "learning_rate": 0.00252952844450698, + "loss": 1.0495, + "step": 3370 + }, + { + "epoch": 0.2592877470963772, + "learning_rate": 0.002529264805846386, + "loss": 1.2691, + "step": 3371 + }, + { + "epoch": 0.25936466425659566, + "learning_rate": 0.0025290011070859472, + "loss": 1.1531, + "step": 3372 + }, + { + "epoch": 0.2594415814168141, + "learning_rate": 0.0025287373482410615, + "loss": 1.0173, + "step": 3373 + }, + { + "epoch": 0.2595184985770325, + "learning_rate": 0.00252847352932713, + "loss": 0.9931, + "step": 3374 + }, + { + "epoch": 0.259595415737251, + "learning_rate": 0.002528209650359557, + "loss": 1.2584, + "step": 3375 + }, + { + "epoch": 0.25967233289746944, + "learning_rate": 0.0025279457113537513, + "loss": 1.049, + "step": 3376 + }, + { + "epoch": 0.2597492500576879, + "learning_rate": 0.002527681712325124, + "loss": 1.3439, + "step": 3377 + }, + { + "epoch": 0.2598261672179063, + "learning_rate": 0.0025274176532890905, + "loss": 1.3921, + "step": 3378 + }, + { + "epoch": 0.25990308437812476, + "learning_rate": 0.0025271535342610697, + "loss": 0.5957, + "step": 3379 + }, + { + "epoch": 0.2599800015383432, + "learning_rate": 0.0025268893552564836, + "loss": 0.9592, + "step": 3380 + }, + { + "epoch": 0.2600569186985616, + "learning_rate": 0.0025266251162907576, + "loss": 1.0464, + "step": 3381 + }, + { + "epoch": 0.2601338358587801, + "learning_rate": 0.0025263608173793214, + "loss": 1.1172, + "step": 3382 + }, + { + "epoch": 0.26021075301899854, + "learning_rate": 0.002526096458537607, + "loss": 1.1096, + "step": 3383 + }, + { + "epoch": 0.260287670179217, + "learning_rate": 0.0025258320397810513, + "loss": 0.8709, + "step": 3384 + }, + { + "epoch": 0.2603645873394354, + "learning_rate": 0.0025255675611250933, + "loss": 0.8465, + "step": 3385 + }, + { + "epoch": 0.26044150449965386, + "learning_rate": 0.0025253030225851772, + "loss": 1.331, + "step": 3386 + }, + { + "epoch": 0.2605184216598723, + "learning_rate": 0.0025250384241767487, + "loss": 1.1716, + "step": 3387 + }, + { + "epoch": 0.2605953388200908, + "learning_rate": 0.0025247737659152584, + "loss": 1.1109, + "step": 3388 + }, + { + "epoch": 0.2606722559803092, + "learning_rate": 0.00252450904781616, + "loss": 1.1885, + "step": 3389 + }, + { + "epoch": 0.26074917314052765, + "learning_rate": 0.0025242442698949104, + "loss": 1.655, + "step": 3390 + }, + { + "epoch": 0.2608260903007461, + "learning_rate": 0.0025239794321669706, + "loss": 1.1128, + "step": 3391 + }, + { + "epoch": 0.26090300746096456, + "learning_rate": 0.0025237145346478046, + "loss": 0.9386, + "step": 3392 + }, + { + "epoch": 0.26097992462118297, + "learning_rate": 0.0025234495773528804, + "loss": 1.4328, + "step": 3393 + }, + { + "epoch": 0.2610568417814014, + "learning_rate": 0.002523184560297668, + "loss": 1.1064, + "step": 3394 + }, + { + "epoch": 0.2611337589416199, + "learning_rate": 0.002522919483497643, + "loss": 1.2672, + "step": 3395 + }, + { + "epoch": 0.26121067610183835, + "learning_rate": 0.0025226543469682836, + "loss": 1.6387, + "step": 3396 + }, + { + "epoch": 0.26128759326205675, + "learning_rate": 0.0025223891507250708, + "loss": 1.2837, + "step": 3397 + }, + { + "epoch": 0.2613645104222752, + "learning_rate": 0.00252212389478349, + "loss": 1.0324, + "step": 3398 + }, + { + "epoch": 0.26144142758249367, + "learning_rate": 0.00252185857915903, + "loss": 1.277, + "step": 3399 + }, + { + "epoch": 0.26151834474271207, + "learning_rate": 0.0025215932038671826, + "loss": 1.1001, + "step": 3400 + }, + { + "epoch": 0.26159526190293053, + "learning_rate": 0.002521327768923443, + "loss": 1.6005, + "step": 3401 + }, + { + "epoch": 0.261672179063149, + "learning_rate": 0.002521062274343311, + "loss": 1.1225, + "step": 3402 + }, + { + "epoch": 0.26174909622336745, + "learning_rate": 0.0025207967201422883, + "loss": 0.8252, + "step": 3403 + }, + { + "epoch": 0.26182601338358585, + "learning_rate": 0.0025205311063358814, + "loss": 1.2752, + "step": 3404 + }, + { + "epoch": 0.2619029305438043, + "learning_rate": 0.0025202654329395994, + "loss": 1.1486, + "step": 3405 + }, + { + "epoch": 0.2619798477040228, + "learning_rate": 0.002519999699968956, + "loss": 1.2136, + "step": 3406 + }, + { + "epoch": 0.26205676486424123, + "learning_rate": 0.002519733907439467, + "loss": 1.1214, + "step": 3407 + }, + { + "epoch": 0.26213368202445964, + "learning_rate": 0.0025194680553666526, + "loss": 1.0703, + "step": 3408 + }, + { + "epoch": 0.2622105991846781, + "learning_rate": 0.002519202143766035, + "loss": 1.2793, + "step": 3409 + }, + { + "epoch": 0.26228751634489655, + "learning_rate": 0.0025189361726531425, + "loss": 1.1329, + "step": 3410 + }, + { + "epoch": 0.262364433505115, + "learning_rate": 0.002518670142043505, + "loss": 1.107, + "step": 3411 + }, + { + "epoch": 0.2624413506653334, + "learning_rate": 0.0025184040519526567, + "loss": 0.8191, + "step": 3412 + }, + { + "epoch": 0.2625182678255519, + "learning_rate": 0.002518137902396134, + "loss": 0.9387, + "step": 3413 + }, + { + "epoch": 0.26259518498577034, + "learning_rate": 0.0025178716933894783, + "loss": 0.9596, + "step": 3414 + }, + { + "epoch": 0.2626721021459888, + "learning_rate": 0.0025176054249482337, + "loss": 0.8431, + "step": 3415 + }, + { + "epoch": 0.2627490193062072, + "learning_rate": 0.002517339097087948, + "loss": 1.1453, + "step": 3416 + }, + { + "epoch": 0.26282593646642566, + "learning_rate": 0.0025170727098241716, + "loss": 1.3631, + "step": 3417 + }, + { + "epoch": 0.2629028536266441, + "learning_rate": 0.00251680626317246, + "loss": 1.4974, + "step": 3418 + }, + { + "epoch": 0.2629797707868626, + "learning_rate": 0.002516539757148371, + "loss": 0.9919, + "step": 3419 + }, + { + "epoch": 0.263056687947081, + "learning_rate": 0.0025162731917674664, + "loss": 1.4821, + "step": 3420 + }, + { + "epoch": 0.26313360510729944, + "learning_rate": 0.0025160065670453103, + "loss": 1.1854, + "step": 3421 + }, + { + "epoch": 0.2632105222675179, + "learning_rate": 0.0025157398829974724, + "loss": 1.5294, + "step": 3422 + }, + { + "epoch": 0.2632874394277363, + "learning_rate": 0.002515473139639524, + "loss": 1.1279, + "step": 3423 + }, + { + "epoch": 0.26336435658795476, + "learning_rate": 0.002515206336987041, + "loss": 1.087, + "step": 3424 + }, + { + "epoch": 0.2634412737481732, + "learning_rate": 0.002514939475055602, + "loss": 1.1478, + "step": 3425 + }, + { + "epoch": 0.2635181909083917, + "learning_rate": 0.0025146725538607887, + "loss": 1.114, + "step": 3426 + }, + { + "epoch": 0.2635951080686101, + "learning_rate": 0.002514405573418188, + "loss": 1.2608, + "step": 3427 + }, + { + "epoch": 0.26367202522882854, + "learning_rate": 0.0025141385337433884, + "loss": 1.3321, + "step": 3428 + }, + { + "epoch": 0.263748942389047, + "learning_rate": 0.0025138714348519825, + "loss": 0.8488, + "step": 3429 + }, + { + "epoch": 0.26382585954926546, + "learning_rate": 0.002513604276759568, + "loss": 1.1953, + "step": 3430 + }, + { + "epoch": 0.26390277670948387, + "learning_rate": 0.002513337059481743, + "loss": 1.1408, + "step": 3431 + }, + { + "epoch": 0.2639796938697023, + "learning_rate": 0.002513069783034111, + "loss": 1.1631, + "step": 3432 + }, + { + "epoch": 0.2640566110299208, + "learning_rate": 0.002512802447432278, + "loss": 1.4827, + "step": 3433 + }, + { + "epoch": 0.26413352819013924, + "learning_rate": 0.002512535052691855, + "loss": 1.1841, + "step": 3434 + }, + { + "epoch": 0.26421044535035765, + "learning_rate": 0.0025122675988284555, + "loss": 0.8926, + "step": 3435 + }, + { + "epoch": 0.2642873625105761, + "learning_rate": 0.002512000085857695, + "loss": 1.0023, + "step": 3436 + }, + { + "epoch": 0.26436427967079457, + "learning_rate": 0.0025117325137951952, + "loss": 0.9331, + "step": 3437 + }, + { + "epoch": 0.264441196831013, + "learning_rate": 0.0025114648826565795, + "loss": 1.2008, + "step": 3438 + }, + { + "epoch": 0.26451811399123143, + "learning_rate": 0.0025111971924574755, + "loss": 1.0714, + "step": 3439 + }, + { + "epoch": 0.2645950311514499, + "learning_rate": 0.002510929443213513, + "loss": 1.1681, + "step": 3440 + }, + { + "epoch": 0.26467194831166835, + "learning_rate": 0.0025106616349403277, + "loss": 0.9974, + "step": 3441 + }, + { + "epoch": 0.26474886547188675, + "learning_rate": 0.002510393767653556, + "loss": 1.2625, + "step": 3442 + }, + { + "epoch": 0.2648257826321052, + "learning_rate": 0.0025101258413688385, + "loss": 0.9855, + "step": 3443 + }, + { + "epoch": 0.26490269979232367, + "learning_rate": 0.002509857856101821, + "loss": 1.0247, + "step": 3444 + }, + { + "epoch": 0.26497961695254213, + "learning_rate": 0.0025095898118681504, + "loss": 1.1187, + "step": 3445 + }, + { + "epoch": 0.26505653411276053, + "learning_rate": 0.0025093217086834788, + "loss": 1.1966, + "step": 3446 + }, + { + "epoch": 0.265133451272979, + "learning_rate": 0.0025090535465634605, + "loss": 1.6931, + "step": 3447 + }, + { + "epoch": 0.26521036843319745, + "learning_rate": 0.002508785325523754, + "loss": 1.1623, + "step": 3448 + }, + { + "epoch": 0.2652872855934159, + "learning_rate": 0.002508517045580021, + "loss": 1.3653, + "step": 3449 + }, + { + "epoch": 0.2653642027536343, + "learning_rate": 0.0025082487067479274, + "loss": 1.1628, + "step": 3450 + }, + { + "epoch": 0.2654411199138528, + "learning_rate": 0.0025079803090431404, + "loss": 1.2927, + "step": 3451 + }, + { + "epoch": 0.26551803707407123, + "learning_rate": 0.002507711852481332, + "loss": 0.9913, + "step": 3452 + }, + { + "epoch": 0.2655949542342897, + "learning_rate": 0.0025074433370781794, + "loss": 1.4476, + "step": 3453 + }, + { + "epoch": 0.2656718713945081, + "learning_rate": 0.0025071747628493605, + "loss": 0.8821, + "step": 3454 + }, + { + "epoch": 0.26574878855472656, + "learning_rate": 0.002506906129810557, + "loss": 1.3082, + "step": 3455 + }, + { + "epoch": 0.265825705714945, + "learning_rate": 0.0025066374379774554, + "loss": 1.1008, + "step": 3456 + }, + { + "epoch": 0.2659026228751635, + "learning_rate": 0.0025063686873657443, + "loss": 0.8634, + "step": 3457 + }, + { + "epoch": 0.2659795400353819, + "learning_rate": 0.0025060998779911177, + "loss": 1.2657, + "step": 3458 + }, + { + "epoch": 0.26605645719560034, + "learning_rate": 0.0025058310098692695, + "loss": 1.0036, + "step": 3459 + }, + { + "epoch": 0.2661333743558188, + "learning_rate": 0.002505562083015901, + "loss": 0.9075, + "step": 3460 + }, + { + "epoch": 0.2662102915160372, + "learning_rate": 0.0025052930974467143, + "loss": 1.2736, + "step": 3461 + }, + { + "epoch": 0.26628720867625566, + "learning_rate": 0.0025050240531774163, + "loss": 1.2193, + "step": 3462 + }, + { + "epoch": 0.2663641258364741, + "learning_rate": 0.0025047549502237158, + "loss": 1.2223, + "step": 3463 + }, + { + "epoch": 0.2664410429966926, + "learning_rate": 0.0025044857886013273, + "loss": 1.0444, + "step": 3464 + }, + { + "epoch": 0.266517960156911, + "learning_rate": 0.0025042165683259667, + "loss": 1.1946, + "step": 3465 + }, + { + "epoch": 0.26659487731712944, + "learning_rate": 0.002503947289413354, + "loss": 0.9865, + "step": 3466 + }, + { + "epoch": 0.2666717944773479, + "learning_rate": 0.0025036779518792123, + "loss": 1.3231, + "step": 3467 + }, + { + "epoch": 0.26674871163756636, + "learning_rate": 0.002503408555739269, + "loss": 1.0282, + "step": 3468 + }, + { + "epoch": 0.26682562879778476, + "learning_rate": 0.0025031391010092554, + "loss": 1.5534, + "step": 3469 + }, + { + "epoch": 0.2669025459580032, + "learning_rate": 0.0025028695877049034, + "loss": 1.2378, + "step": 3470 + }, + { + "epoch": 0.2669794631182217, + "learning_rate": 0.0025026000158419512, + "loss": 1.0397, + "step": 3471 + }, + { + "epoch": 0.26705638027844014, + "learning_rate": 0.0025023303854361393, + "loss": 1.2919, + "step": 3472 + }, + { + "epoch": 0.26713329743865855, + "learning_rate": 0.0025020606965032114, + "loss": 1.1562, + "step": 3473 + }, + { + "epoch": 0.267210214598877, + "learning_rate": 0.002501790949058915, + "loss": 1.002, + "step": 3474 + }, + { + "epoch": 0.26728713175909546, + "learning_rate": 0.002501521143119001, + "loss": 1.2503, + "step": 3475 + }, + { + "epoch": 0.2673640489193139, + "learning_rate": 0.0025012512786992246, + "loss": 0.9608, + "step": 3476 + }, + { + "epoch": 0.2674409660795323, + "learning_rate": 0.0025009813558153414, + "loss": 0.7097, + "step": 3477 + }, + { + "epoch": 0.2675178832397508, + "learning_rate": 0.0025007113744831143, + "loss": 0.8568, + "step": 3478 + }, + { + "epoch": 0.26759480039996925, + "learning_rate": 0.0025004413347183073, + "loss": 1.1077, + "step": 3479 + }, + { + "epoch": 0.2676717175601877, + "learning_rate": 0.002500171236536687, + "loss": 0.9338, + "step": 3480 + }, + { + "epoch": 0.2677486347204061, + "learning_rate": 0.0024999010799540275, + "loss": 1.3056, + "step": 3481 + }, + { + "epoch": 0.26782555188062457, + "learning_rate": 0.0024996308649861006, + "loss": 1.1272, + "step": 3482 + }, + { + "epoch": 0.267902469040843, + "learning_rate": 0.002499360591648686, + "loss": 0.8716, + "step": 3483 + }, + { + "epoch": 0.26797938620106143, + "learning_rate": 0.0024990902599575653, + "loss": 0.7946, + "step": 3484 + }, + { + "epoch": 0.2680563033612799, + "learning_rate": 0.002498819869928523, + "loss": 1.2492, + "step": 3485 + }, + { + "epoch": 0.26813322052149835, + "learning_rate": 0.0024985494215773473, + "loss": 1.1077, + "step": 3486 + }, + { + "epoch": 0.2682101376817168, + "learning_rate": 0.00249827891491983, + "loss": 1.0933, + "step": 3487 + }, + { + "epoch": 0.2682870548419352, + "learning_rate": 0.002498008349971767, + "loss": 1.0039, + "step": 3488 + }, + { + "epoch": 0.2683639720021537, + "learning_rate": 0.0024977377267489565, + "loss": 0.8665, + "step": 3489 + }, + { + "epoch": 0.26844088916237213, + "learning_rate": 0.0024974670452671995, + "loss": 1.0237, + "step": 3490 + }, + { + "epoch": 0.2685178063225906, + "learning_rate": 0.0024971963055423025, + "loss": 1.3132, + "step": 3491 + }, + { + "epoch": 0.268594723482809, + "learning_rate": 0.0024969255075900742, + "loss": 0.9821, + "step": 3492 + }, + { + "epoch": 0.26867164064302745, + "learning_rate": 0.0024966546514263266, + "loss": 1.2235, + "step": 3493 + }, + { + "epoch": 0.2687485578032459, + "learning_rate": 0.002496383737066875, + "loss": 0.8197, + "step": 3494 + }, + { + "epoch": 0.2688254749634644, + "learning_rate": 0.0024961127645275385, + "loss": 1.2345, + "step": 3495 + }, + { + "epoch": 0.2689023921236828, + "learning_rate": 0.0024958417338241403, + "loss": 1.2893, + "step": 3496 + }, + { + "epoch": 0.26897930928390124, + "learning_rate": 0.0024955706449725046, + "loss": 0.9128, + "step": 3497 + }, + { + "epoch": 0.2690562264441197, + "learning_rate": 0.0024952994979884615, + "loss": 1.0012, + "step": 3498 + }, + { + "epoch": 0.26913314360433815, + "learning_rate": 0.0024950282928878436, + "loss": 0.927, + "step": 3499 + }, + { + "epoch": 0.26921006076455656, + "learning_rate": 0.0024947570296864866, + "loss": 1.5774, + "step": 3500 + }, + { + "epoch": 0.269286977924775, + "learning_rate": 0.0024944857084002302, + "loss": 1.029, + "step": 3501 + }, + { + "epoch": 0.2693638950849935, + "learning_rate": 0.0024942143290449166, + "loss": 0.9192, + "step": 3502 + }, + { + "epoch": 0.2694408122452119, + "learning_rate": 0.002493942891636392, + "loss": 0.9637, + "step": 3503 + }, + { + "epoch": 0.26951772940543034, + "learning_rate": 0.002493671396190506, + "loss": 1.0466, + "step": 3504 + }, + { + "epoch": 0.2695946465656488, + "learning_rate": 0.002493399842723112, + "loss": 1.235, + "step": 3505 + }, + { + "epoch": 0.26967156372586726, + "learning_rate": 0.002493128231250066, + "loss": 1.0379, + "step": 3506 + }, + { + "epoch": 0.26974848088608566, + "learning_rate": 0.0024928565617872273, + "loss": 0.975, + "step": 3507 + }, + { + "epoch": 0.2698253980463041, + "learning_rate": 0.0024925848343504593, + "loss": 0.7817, + "step": 3508 + }, + { + "epoch": 0.2699023152065226, + "learning_rate": 0.002492313048955628, + "loss": 1.0762, + "step": 3509 + }, + { + "epoch": 0.26997923236674104, + "learning_rate": 0.002492041205618604, + "loss": 1.2626, + "step": 3510 + }, + { + "epoch": 0.27005614952695944, + "learning_rate": 0.00249176930435526, + "loss": 1.6238, + "step": 3511 + }, + { + "epoch": 0.2701330666871779, + "learning_rate": 0.002491497345181473, + "loss": 1.3942, + "step": 3512 + }, + { + "epoch": 0.27020998384739636, + "learning_rate": 0.0024912253281131227, + "loss": 0.9294, + "step": 3513 + }, + { + "epoch": 0.2702869010076148, + "learning_rate": 0.0024909532531660924, + "loss": 1.2417, + "step": 3514 + }, + { + "epoch": 0.2703638181678332, + "learning_rate": 0.0024906811203562686, + "loss": 0.9573, + "step": 3515 + }, + { + "epoch": 0.2704407353280517, + "learning_rate": 0.0024904089296995417, + "loss": 1.312, + "step": 3516 + }, + { + "epoch": 0.27051765248827014, + "learning_rate": 0.0024901366812118057, + "loss": 1.0174, + "step": 3517 + }, + { + "epoch": 0.2705945696484886, + "learning_rate": 0.0024898643749089564, + "loss": 1.3006, + "step": 3518 + }, + { + "epoch": 0.270671486808707, + "learning_rate": 0.0024895920108068954, + "loss": 1.2633, + "step": 3519 + }, + { + "epoch": 0.27074840396892547, + "learning_rate": 0.0024893195889215246, + "loss": 1.3204, + "step": 3520 + }, + { + "epoch": 0.2708253211291439, + "learning_rate": 0.0024890471092687523, + "loss": 0.9997, + "step": 3521 + }, + { + "epoch": 0.27090223828936233, + "learning_rate": 0.0024887745718644894, + "loss": 1.1393, + "step": 3522 + }, + { + "epoch": 0.2709791554495808, + "learning_rate": 0.002488501976724648, + "loss": 1.2127, + "step": 3523 + }, + { + "epoch": 0.27105607260979925, + "learning_rate": 0.0024882293238651454, + "loss": 1.2812, + "step": 3524 + }, + { + "epoch": 0.2711329897700177, + "learning_rate": 0.0024879566133019036, + "loss": 1.0829, + "step": 3525 + }, + { + "epoch": 0.2712099069302361, + "learning_rate": 0.002487683845050845, + "loss": 1.2449, + "step": 3526 + }, + { + "epoch": 0.27128682409045457, + "learning_rate": 0.0024874110191278978, + "loss": 1.3035, + "step": 3527 + }, + { + "epoch": 0.27136374125067303, + "learning_rate": 0.0024871381355489916, + "loss": 0.972, + "step": 3528 + }, + { + "epoch": 0.2714406584108915, + "learning_rate": 0.002486865194330061, + "loss": 0.9873, + "step": 3529 + }, + { + "epoch": 0.2715175755711099, + "learning_rate": 0.0024865921954870442, + "loss": 1.1617, + "step": 3530 + }, + { + "epoch": 0.27159449273132835, + "learning_rate": 0.0024863191390358795, + "loss": 0.7969, + "step": 3531 + }, + { + "epoch": 0.2716714098915468, + "learning_rate": 0.0024860460249925133, + "loss": 1.3353, + "step": 3532 + }, + { + "epoch": 0.27174832705176527, + "learning_rate": 0.0024857728533728917, + "loss": 0.9596, + "step": 3533 + }, + { + "epoch": 0.2718252442119837, + "learning_rate": 0.002485499624192966, + "loss": 1.1102, + "step": 3534 + }, + { + "epoch": 0.27190216137220213, + "learning_rate": 0.0024852263374686902, + "loss": 1.0854, + "step": 3535 + }, + { + "epoch": 0.2719790785324206, + "learning_rate": 0.0024849529932160215, + "loss": 1.0259, + "step": 3536 + }, + { + "epoch": 0.27205599569263905, + "learning_rate": 0.002484679591450921, + "loss": 1.3463, + "step": 3537 + }, + { + "epoch": 0.27213291285285746, + "learning_rate": 0.002484406132189353, + "loss": 1.1481, + "step": 3538 + }, + { + "epoch": 0.2722098300130759, + "learning_rate": 0.002484132615447285, + "loss": 0.9511, + "step": 3539 + }, + { + "epoch": 0.2722867471732944, + "learning_rate": 0.002483859041240688, + "loss": 1.3642, + "step": 3540 + }, + { + "epoch": 0.27236366433351283, + "learning_rate": 0.0024835854095855365, + "loss": 1.2416, + "step": 3541 + }, + { + "epoch": 0.27244058149373124, + "learning_rate": 0.002483311720497808, + "loss": 0.8573, + "step": 3542 + }, + { + "epoch": 0.2725174986539497, + "learning_rate": 0.0024830379739934827, + "loss": 1.0931, + "step": 3543 + }, + { + "epoch": 0.27259441581416816, + "learning_rate": 0.002482764170088546, + "loss": 1.2971, + "step": 3544 + }, + { + "epoch": 0.27267133297438656, + "learning_rate": 0.0024824903087989855, + "loss": 1.4864, + "step": 3545 + }, + { + "epoch": 0.272748250134605, + "learning_rate": 0.0024822163901407914, + "loss": 1.2757, + "step": 3546 + }, + { + "epoch": 0.2728251672948235, + "learning_rate": 0.0024819424141299586, + "loss": 1.1284, + "step": 3547 + }, + { + "epoch": 0.27290208445504194, + "learning_rate": 0.002481668380782485, + "loss": 1.2827, + "step": 3548 + }, + { + "epoch": 0.27297900161526034, + "learning_rate": 0.0024813942901143722, + "loss": 1.1603, + "step": 3549 + }, + { + "epoch": 0.2730559187754788, + "learning_rate": 0.002481120142141623, + "loss": 1.0989, + "step": 3550 + }, + { + "epoch": 0.27313283593569726, + "learning_rate": 0.0024808459368802467, + "loss": 1.1768, + "step": 3551 + }, + { + "epoch": 0.2732097530959157, + "learning_rate": 0.002480571674346254, + "loss": 1.3563, + "step": 3552 + }, + { + "epoch": 0.2732866702561341, + "learning_rate": 0.00248029735455566, + "loss": 1.5579, + "step": 3553 + }, + { + "epoch": 0.2733635874163526, + "learning_rate": 0.0024800229775244803, + "loss": 1.052, + "step": 3554 + }, + { + "epoch": 0.27344050457657104, + "learning_rate": 0.0024797485432687385, + "loss": 0.9302, + "step": 3555 + }, + { + "epoch": 0.2735174217367895, + "learning_rate": 0.002479474051804458, + "loss": 1.2097, + "step": 3556 + }, + { + "epoch": 0.2735943388970079, + "learning_rate": 0.002479199503147667, + "loss": 0.9372, + "step": 3557 + }, + { + "epoch": 0.27367125605722636, + "learning_rate": 0.0024789248973143964, + "loss": 0.8427, + "step": 3558 + }, + { + "epoch": 0.2737481732174448, + "learning_rate": 0.0024786502343206807, + "loss": 1.0251, + "step": 3559 + }, + { + "epoch": 0.2738250903776633, + "learning_rate": 0.0024783755141825578, + "loss": 1.4592, + "step": 3560 + }, + { + "epoch": 0.2739020075378817, + "learning_rate": 0.0024781007369160694, + "loss": 1.468, + "step": 3561 + }, + { + "epoch": 0.27397892469810015, + "learning_rate": 0.0024778259025372592, + "loss": 0.9707, + "step": 3562 + }, + { + "epoch": 0.2740558418583186, + "learning_rate": 0.002477551011062176, + "loss": 1.1006, + "step": 3563 + }, + { + "epoch": 0.274132759018537, + "learning_rate": 0.0024772760625068704, + "loss": 1.2722, + "step": 3564 + }, + { + "epoch": 0.27420967617875547, + "learning_rate": 0.0024770010568873966, + "loss": 1.1913, + "step": 3565 + }, + { + "epoch": 0.2742865933389739, + "learning_rate": 0.002476725994219813, + "loss": 1.4452, + "step": 3566 + }, + { + "epoch": 0.2743635104991924, + "learning_rate": 0.002476450874520181, + "loss": 1.3885, + "step": 3567 + }, + { + "epoch": 0.2744404276594108, + "learning_rate": 0.0024761756978045643, + "loss": 0.9229, + "step": 3568 + }, + { + "epoch": 0.27451734481962925, + "learning_rate": 0.0024759004640890316, + "loss": 1.1716, + "step": 3569 + }, + { + "epoch": 0.2745942619798477, + "learning_rate": 0.0024756251733896537, + "loss": 1.1123, + "step": 3570 + }, + { + "epoch": 0.27467117914006617, + "learning_rate": 0.002475349825722505, + "loss": 1.6297, + "step": 3571 + }, + { + "epoch": 0.27474809630028457, + "learning_rate": 0.002475074421103664, + "loss": 0.9711, + "step": 3572 + }, + { + "epoch": 0.27482501346050303, + "learning_rate": 0.0024747989595492108, + "loss": 1.18, + "step": 3573 + }, + { + "epoch": 0.2749019306207215, + "learning_rate": 0.0024745234410752303, + "loss": 1.1942, + "step": 3574 + }, + { + "epoch": 0.27497884778093995, + "learning_rate": 0.0024742478656978105, + "loss": 1.2194, + "step": 3575 + }, + { + "epoch": 0.27505576494115835, + "learning_rate": 0.0024739722334330434, + "loss": 0.9593, + "step": 3576 + }, + { + "epoch": 0.2751326821013768, + "learning_rate": 0.0024736965442970216, + "loss": 1.352, + "step": 3577 + }, + { + "epoch": 0.2752095992615953, + "learning_rate": 0.0024734207983058442, + "loss": 1.0481, + "step": 3578 + }, + { + "epoch": 0.27528651642181373, + "learning_rate": 0.0024731449954756117, + "loss": 1.1783, + "step": 3579 + }, + { + "epoch": 0.27536343358203214, + "learning_rate": 0.002472869135822429, + "loss": 1.1786, + "step": 3580 + }, + { + "epoch": 0.2754403507422506, + "learning_rate": 0.0024725932193624036, + "loss": 0.9528, + "step": 3581 + }, + { + "epoch": 0.27551726790246905, + "learning_rate": 0.002472317246111646, + "loss": 1.1879, + "step": 3582 + }, + { + "epoch": 0.27559418506268746, + "learning_rate": 0.0024720412160862716, + "loss": 0.9679, + "step": 3583 + }, + { + "epoch": 0.2756711022229059, + "learning_rate": 0.0024717651293023977, + "loss": 1.1578, + "step": 3584 + }, + { + "epoch": 0.2757480193831244, + "learning_rate": 0.0024714889857761443, + "loss": 1.125, + "step": 3585 + }, + { + "epoch": 0.27582493654334284, + "learning_rate": 0.002471212785523637, + "loss": 1.2239, + "step": 3586 + }, + { + "epoch": 0.27590185370356124, + "learning_rate": 0.002470936528561003, + "loss": 1.25, + "step": 3587 + }, + { + "epoch": 0.2759787708637797, + "learning_rate": 0.0024706602149043733, + "loss": 1.1979, + "step": 3588 + }, + { + "epoch": 0.27605568802399816, + "learning_rate": 0.002470383844569882, + "loss": 1.1033, + "step": 3589 + }, + { + "epoch": 0.2761326051842166, + "learning_rate": 0.0024701074175736666, + "loss": 1.0717, + "step": 3590 + }, + { + "epoch": 0.276209522344435, + "learning_rate": 0.0024698309339318686, + "loss": 1.2374, + "step": 3591 + }, + { + "epoch": 0.2762864395046535, + "learning_rate": 0.0024695543936606307, + "loss": 1.2066, + "step": 3592 + }, + { + "epoch": 0.27636335666487194, + "learning_rate": 0.002469277796776102, + "loss": 1.4984, + "step": 3593 + }, + { + "epoch": 0.2764402738250904, + "learning_rate": 0.0024690011432944325, + "loss": 1.2349, + "step": 3594 + }, + { + "epoch": 0.2765171909853088, + "learning_rate": 0.0024687244332317762, + "loss": 1.0037, + "step": 3595 + }, + { + "epoch": 0.27659410814552726, + "learning_rate": 0.0024684476666042908, + "loss": 1.0457, + "step": 3596 + }, + { + "epoch": 0.2766710253057457, + "learning_rate": 0.002468170843428137, + "loss": 0.9947, + "step": 3597 + }, + { + "epoch": 0.2767479424659642, + "learning_rate": 0.002467893963719478, + "loss": 1.3606, + "step": 3598 + }, + { + "epoch": 0.2768248596261826, + "learning_rate": 0.0024676170274944825, + "loss": 1.1684, + "step": 3599 + }, + { + "epoch": 0.27690177678640104, + "learning_rate": 0.00246734003476932, + "loss": 1.1067, + "step": 3600 + }, + { + "epoch": 0.2769786939466195, + "learning_rate": 0.0024670629855601646, + "loss": 1.1872, + "step": 3601 + }, + { + "epoch": 0.27705561110683796, + "learning_rate": 0.0024667858798831936, + "loss": 0.9808, + "step": 3602 + }, + { + "epoch": 0.27713252826705637, + "learning_rate": 0.002466508717754588, + "loss": 1.1883, + "step": 3603 + }, + { + "epoch": 0.2772094454272748, + "learning_rate": 0.0024662314991905306, + "loss": 1.243, + "step": 3604 + }, + { + "epoch": 0.2772863625874933, + "learning_rate": 0.0024659542242072095, + "loss": 1.1596, + "step": 3605 + }, + { + "epoch": 0.2773632797477117, + "learning_rate": 0.0024656768928208142, + "loss": 1.1401, + "step": 3606 + }, + { + "epoch": 0.27744019690793015, + "learning_rate": 0.0024653995050475386, + "loss": 1.2892, + "step": 3607 + }, + { + "epoch": 0.2775171140681486, + "learning_rate": 0.00246512206090358, + "loss": 0.7485, + "step": 3608 + }, + { + "epoch": 0.27759403122836707, + "learning_rate": 0.002464844560405138, + "loss": 0.9778, + "step": 3609 + }, + { + "epoch": 0.27767094838858547, + "learning_rate": 0.0024645670035684172, + "loss": 1.1943, + "step": 3610 + }, + { + "epoch": 0.27774786554880393, + "learning_rate": 0.0024642893904096236, + "loss": 0.7985, + "step": 3611 + }, + { + "epoch": 0.2778247827090224, + "learning_rate": 0.0024640117209449677, + "loss": 0.995, + "step": 3612 + }, + { + "epoch": 0.27790169986924085, + "learning_rate": 0.002463733995190662, + "loss": 1.1795, + "step": 3613 + }, + { + "epoch": 0.27797861702945925, + "learning_rate": 0.002463456213162925, + "loss": 1.114, + "step": 3614 + }, + { + "epoch": 0.2780555341896777, + "learning_rate": 0.002463178374877975, + "loss": 1.2579, + "step": 3615 + }, + { + "epoch": 0.27813245134989617, + "learning_rate": 0.002462900480352036, + "loss": 1.1171, + "step": 3616 + }, + { + "epoch": 0.27820936851011463, + "learning_rate": 0.0024626225296013344, + "loss": 0.9647, + "step": 3617 + }, + { + "epoch": 0.27828628567033303, + "learning_rate": 0.0024623445226421006, + "loss": 1.0202, + "step": 3618 + }, + { + "epoch": 0.2783632028305515, + "learning_rate": 0.002462066459490566, + "loss": 1.4352, + "step": 3619 + }, + { + "epoch": 0.27844011999076995, + "learning_rate": 0.0024617883401629693, + "loss": 1.032, + "step": 3620 + }, + { + "epoch": 0.2785170371509884, + "learning_rate": 0.0024615101646755485, + "loss": 1.2368, + "step": 3621 + }, + { + "epoch": 0.2785939543112068, + "learning_rate": 0.0024612319330445474, + "loss": 0.7482, + "step": 3622 + }, + { + "epoch": 0.2786708714714253, + "learning_rate": 0.002460953645286212, + "loss": 0.9346, + "step": 3623 + }, + { + "epoch": 0.27874778863164373, + "learning_rate": 0.0024606753014167915, + "loss": 0.919, + "step": 3624 + }, + { + "epoch": 0.27882470579186214, + "learning_rate": 0.002460396901452539, + "loss": 1.1329, + "step": 3625 + }, + { + "epoch": 0.2789016229520806, + "learning_rate": 0.0024601184454097107, + "loss": 1.5333, + "step": 3626 + }, + { + "epoch": 0.27897854011229906, + "learning_rate": 0.0024598399333045653, + "loss": 1.2765, + "step": 3627 + }, + { + "epoch": 0.2790554572725175, + "learning_rate": 0.002459561365153366, + "loss": 1.0334, + "step": 3628 + }, + { + "epoch": 0.2791323744327359, + "learning_rate": 0.002459282740972379, + "loss": 1.2858, + "step": 3629 + }, + { + "epoch": 0.2792092915929544, + "learning_rate": 0.002459004060777873, + "loss": 1.2152, + "step": 3630 + }, + { + "epoch": 0.27928620875317284, + "learning_rate": 0.0024587253245861207, + "loss": 1.3838, + "step": 3631 + }, + { + "epoch": 0.2793631259133913, + "learning_rate": 0.0024584465324133968, + "loss": 0.9945, + "step": 3632 + }, + { + "epoch": 0.2794400430736097, + "learning_rate": 0.002458167684275981, + "loss": 1.118, + "step": 3633 + }, + { + "epoch": 0.27951696023382816, + "learning_rate": 0.002457888780190156, + "loss": 1.2525, + "step": 3634 + }, + { + "epoch": 0.2795938773940466, + "learning_rate": 0.0024576098201722065, + "loss": 1.3401, + "step": 3635 + }, + { + "epoch": 0.2796707945542651, + "learning_rate": 0.0024573308042384217, + "loss": 1.0822, + "step": 3636 + }, + { + "epoch": 0.2797477117144835, + "learning_rate": 0.002457051732405094, + "loss": 1.2132, + "step": 3637 + }, + { + "epoch": 0.27982462887470194, + "learning_rate": 0.002456772604688518, + "loss": 1.1069, + "step": 3638 + }, + { + "epoch": 0.2799015460349204, + "learning_rate": 0.0024564934211049917, + "loss": 0.8862, + "step": 3639 + }, + { + "epoch": 0.27997846319513886, + "learning_rate": 0.0024562141816708185, + "loss": 1.2245, + "step": 3640 + }, + { + "epoch": 0.28005538035535726, + "learning_rate": 0.002455934886402303, + "loss": 1.2619, + "step": 3641 + }, + { + "epoch": 0.2801322975155757, + "learning_rate": 0.0024556555353157524, + "loss": 1.1116, + "step": 3642 + }, + { + "epoch": 0.2802092146757942, + "learning_rate": 0.0024553761284274796, + "loss": 1.1542, + "step": 3643 + }, + { + "epoch": 0.28028613183601264, + "learning_rate": 0.002455096665753799, + "loss": 1.2643, + "step": 3644 + }, + { + "epoch": 0.28036304899623105, + "learning_rate": 0.0024548171473110286, + "loss": 1.1555, + "step": 3645 + }, + { + "epoch": 0.2804399661564495, + "learning_rate": 0.0024545375731154897, + "loss": 0.8542, + "step": 3646 + }, + { + "epoch": 0.28051688331666796, + "learning_rate": 0.002454257943183507, + "loss": 2.0818, + "step": 3647 + }, + { + "epoch": 0.28059380047688637, + "learning_rate": 0.0024539782575314093, + "loss": 1.2558, + "step": 3648 + }, + { + "epoch": 0.2806707176371048, + "learning_rate": 0.002453698516175526, + "loss": 0.8071, + "step": 3649 + }, + { + "epoch": 0.2807476347973233, + "learning_rate": 0.0024534187191321928, + "loss": 1.1587, + "step": 3650 + }, + { + "epoch": 0.28082455195754175, + "learning_rate": 0.002453138866417747, + "loss": 1.2945, + "step": 3651 + }, + { + "epoch": 0.28090146911776015, + "learning_rate": 0.00245285895804853, + "loss": 1.2694, + "step": 3652 + }, + { + "epoch": 0.2809783862779786, + "learning_rate": 0.0024525789940408843, + "loss": 0.9074, + "step": 3653 + }, + { + "epoch": 0.28105530343819707, + "learning_rate": 0.0024522989744111593, + "loss": 1.0224, + "step": 3654 + }, + { + "epoch": 0.2811322205984155, + "learning_rate": 0.002452018899175704, + "loss": 1.1235, + "step": 3655 + }, + { + "epoch": 0.28120913775863393, + "learning_rate": 0.0024517387683508736, + "loss": 1.1201, + "step": 3656 + }, + { + "epoch": 0.2812860549188524, + "learning_rate": 0.002451458581953024, + "loss": 1.1645, + "step": 3657 + }, + { + "epoch": 0.28136297207907085, + "learning_rate": 0.002451178339998517, + "loss": 1.1267, + "step": 3658 + }, + { + "epoch": 0.2814398892392893, + "learning_rate": 0.002450898042503715, + "loss": 1.298, + "step": 3659 + }, + { + "epoch": 0.2815168063995077, + "learning_rate": 0.002450617689484985, + "loss": 1.2631, + "step": 3660 + }, + { + "epoch": 0.2815937235597262, + "learning_rate": 0.0024503372809586976, + "loss": 1.2871, + "step": 3661 + }, + { + "epoch": 0.28167064071994463, + "learning_rate": 0.002450056816941226, + "loss": 1.3321, + "step": 3662 + }, + { + "epoch": 0.2817475578801631, + "learning_rate": 0.002449776297448947, + "loss": 0.9775, + "step": 3663 + }, + { + "epoch": 0.2818244750403815, + "learning_rate": 0.00244949572249824, + "loss": 1.445, + "step": 3664 + }, + { + "epoch": 0.28190139220059995, + "learning_rate": 0.002449215092105488, + "loss": 0.9526, + "step": 3665 + }, + { + "epoch": 0.2819783093608184, + "learning_rate": 0.002448934406287078, + "loss": 1.1173, + "step": 3666 + }, + { + "epoch": 0.2820552265210368, + "learning_rate": 0.0024486536650593985, + "loss": 1.3006, + "step": 3667 + }, + { + "epoch": 0.2821321436812553, + "learning_rate": 0.0024483728684388433, + "loss": 0.9413, + "step": 3668 + }, + { + "epoch": 0.28220906084147374, + "learning_rate": 0.002448092016441808, + "loss": 1.1158, + "step": 3669 + }, + { + "epoch": 0.2822859780016922, + "learning_rate": 0.002447811109084691, + "loss": 1.1066, + "step": 3670 + }, + { + "epoch": 0.2823628951619106, + "learning_rate": 0.002447530146383896, + "loss": 1.2595, + "step": 3671 + }, + { + "epoch": 0.28243981232212906, + "learning_rate": 0.0024472491283558284, + "loss": 0.7838, + "step": 3672 + }, + { + "epoch": 0.2825167294823475, + "learning_rate": 0.0024469680550168967, + "loss": 1.1258, + "step": 3673 + }, + { + "epoch": 0.282593646642566, + "learning_rate": 0.002446686926383514, + "loss": 0.934, + "step": 3674 + }, + { + "epoch": 0.2826705638027844, + "learning_rate": 0.002446405742472095, + "loss": 1.1691, + "step": 3675 + }, + { + "epoch": 0.28274748096300284, + "learning_rate": 0.0024461245032990574, + "loss": 1.3394, + "step": 3676 + }, + { + "epoch": 0.2828243981232213, + "learning_rate": 0.002445843208880825, + "loss": 1.3558, + "step": 3677 + }, + { + "epoch": 0.28290131528343976, + "learning_rate": 0.0024455618592338214, + "loss": 1.2262, + "step": 3678 + }, + { + "epoch": 0.28297823244365816, + "learning_rate": 0.0024452804543744757, + "loss": 1.4341, + "step": 3679 + }, + { + "epoch": 0.2830551496038766, + "learning_rate": 0.002444998994319219, + "loss": 1.3195, + "step": 3680 + }, + { + "epoch": 0.2831320667640951, + "learning_rate": 0.002444717479084486, + "loss": 1.0892, + "step": 3681 + }, + { + "epoch": 0.28320898392431354, + "learning_rate": 0.002444435908686715, + "loss": 1.1523, + "step": 3682 + }, + { + "epoch": 0.28328590108453194, + "learning_rate": 0.0024441542831423473, + "loss": 0.8263, + "step": 3683 + }, + { + "epoch": 0.2833628182447504, + "learning_rate": 0.002443872602467826, + "loss": 1.1053, + "step": 3684 + }, + { + "epoch": 0.28343973540496886, + "learning_rate": 0.0024435908666796007, + "loss": 0.9128, + "step": 3685 + }, + { + "epoch": 0.28351665256518727, + "learning_rate": 0.0024433090757941217, + "loss": 1.2603, + "step": 3686 + }, + { + "epoch": 0.2835935697254057, + "learning_rate": 0.002443027229827842, + "loss": 1.3255, + "step": 3687 + }, + { + "epoch": 0.2836704868856242, + "learning_rate": 0.002442745328797219, + "loss": 0.8551, + "step": 3688 + }, + { + "epoch": 0.28374740404584264, + "learning_rate": 0.002442463372718715, + "loss": 1.2226, + "step": 3689 + }, + { + "epoch": 0.28382432120606105, + "learning_rate": 0.002442181361608792, + "loss": 1.0914, + "step": 3690 + }, + { + "epoch": 0.2839012383662795, + "learning_rate": 0.002441899295483917, + "loss": 1.2384, + "step": 3691 + }, + { + "epoch": 0.28397815552649797, + "learning_rate": 0.002441617174360561, + "loss": 1.3121, + "step": 3692 + }, + { + "epoch": 0.2840550726867164, + "learning_rate": 0.002441334998255196, + "loss": 1.0933, + "step": 3693 + }, + { + "epoch": 0.28413198984693483, + "learning_rate": 0.0024410527671843004, + "loss": 0.8185, + "step": 3694 + }, + { + "epoch": 0.2842089070071533, + "learning_rate": 0.0024407704811643524, + "loss": 0.9503, + "step": 3695 + }, + { + "epoch": 0.28428582416737175, + "learning_rate": 0.0024404881402118354, + "loss": 1.2215, + "step": 3696 + }, + { + "epoch": 0.2843627413275902, + "learning_rate": 0.0024402057443432363, + "loss": 0.7783, + "step": 3697 + }, + { + "epoch": 0.2844396584878086, + "learning_rate": 0.002439923293575044, + "loss": 0.9716, + "step": 3698 + }, + { + "epoch": 0.28451657564802707, + "learning_rate": 0.0024396407879237504, + "loss": 1.3816, + "step": 3699 + }, + { + "epoch": 0.28459349280824553, + "learning_rate": 0.002439358227405852, + "loss": 0.9018, + "step": 3700 + }, + { + "epoch": 0.284670409968464, + "learning_rate": 0.002439075612037848, + "loss": 0.8355, + "step": 3701 + }, + { + "epoch": 0.2847473271286824, + "learning_rate": 0.0024387929418362405, + "loss": 1.3953, + "step": 3702 + }, + { + "epoch": 0.28482424428890085, + "learning_rate": 0.0024385102168175345, + "loss": 1.1761, + "step": 3703 + }, + { + "epoch": 0.2849011614491193, + "learning_rate": 0.0024382274369982387, + "loss": 1.2913, + "step": 3704 + }, + { + "epoch": 0.28497807860933777, + "learning_rate": 0.0024379446023948654, + "loss": 0.9577, + "step": 3705 + }, + { + "epoch": 0.2850549957695562, + "learning_rate": 0.0024376617130239292, + "loss": 1.095, + "step": 3706 + }, + { + "epoch": 0.28513191292977463, + "learning_rate": 0.002437378768901948, + "loss": 1.3237, + "step": 3707 + }, + { + "epoch": 0.2852088300899931, + "learning_rate": 0.0024370957700454443, + "loss": 1.1775, + "step": 3708 + }, + { + "epoch": 0.2852857472502115, + "learning_rate": 0.002436812716470941, + "loss": 1.2467, + "step": 3709 + }, + { + "epoch": 0.28536266441042996, + "learning_rate": 0.0024365296081949677, + "loss": 0.9706, + "step": 3710 + }, + { + "epoch": 0.2854395815706484, + "learning_rate": 0.0024362464452340545, + "loss": 1.258, + "step": 3711 + }, + { + "epoch": 0.2855164987308669, + "learning_rate": 0.002435963227604735, + "loss": 1.1666, + "step": 3712 + }, + { + "epoch": 0.2855934158910853, + "learning_rate": 0.0024356799553235478, + "loss": 0.7968, + "step": 3713 + }, + { + "epoch": 0.28567033305130374, + "learning_rate": 0.0024353966284070324, + "loss": 1.0997, + "step": 3714 + }, + { + "epoch": 0.2857472502115222, + "learning_rate": 0.002435113246871733, + "loss": 1.0189, + "step": 3715 + }, + { + "epoch": 0.28582416737174066, + "learning_rate": 0.0024348298107341967, + "loss": 1.1352, + "step": 3716 + }, + { + "epoch": 0.28590108453195906, + "learning_rate": 0.002434546320010974, + "loss": 0.9592, + "step": 3717 + }, + { + "epoch": 0.2859780016921775, + "learning_rate": 0.002434262774718618, + "loss": 1.073, + "step": 3718 + }, + { + "epoch": 0.286054918852396, + "learning_rate": 0.002433979174873684, + "loss": 1.1065, + "step": 3719 + }, + { + "epoch": 0.28613183601261444, + "learning_rate": 0.0024336955204927323, + "loss": 1.1247, + "step": 3720 + }, + { + "epoch": 0.28620875317283284, + "learning_rate": 0.0024334118115923267, + "loss": 1.1813, + "step": 3721 + }, + { + "epoch": 0.2862856703330513, + "learning_rate": 0.0024331280481890327, + "loss": 1.3859, + "step": 3722 + }, + { + "epoch": 0.28636258749326976, + "learning_rate": 0.0024328442302994193, + "loss": 1.3367, + "step": 3723 + }, + { + "epoch": 0.2864395046534882, + "learning_rate": 0.0024325603579400594, + "loss": 1.3159, + "step": 3724 + }, + { + "epoch": 0.2865164218137066, + "learning_rate": 0.002432276431127528, + "loss": 1.1745, + "step": 3725 + }, + { + "epoch": 0.2865933389739251, + "learning_rate": 0.002431992449878404, + "loss": 0.9767, + "step": 3726 + }, + { + "epoch": 0.28667025613414354, + "learning_rate": 0.00243170841420927, + "loss": 1.24, + "step": 3727 + }, + { + "epoch": 0.28674717329436195, + "learning_rate": 0.0024314243241367107, + "loss": 0.8242, + "step": 3728 + }, + { + "epoch": 0.2868240904545804, + "learning_rate": 0.002431140179677314, + "loss": 0.8692, + "step": 3729 + }, + { + "epoch": 0.28690100761479886, + "learning_rate": 0.002430855980847672, + "loss": 1.1353, + "step": 3730 + }, + { + "epoch": 0.2869779247750173, + "learning_rate": 0.002430571727664379, + "loss": 1.2912, + "step": 3731 + }, + { + "epoch": 0.2870548419352357, + "learning_rate": 0.0024302874201440334, + "loss": 1.0579, + "step": 3732 + }, + { + "epoch": 0.2871317590954542, + "learning_rate": 0.0024300030583032355, + "loss": 1.1849, + "step": 3733 + }, + { + "epoch": 0.28720867625567265, + "learning_rate": 0.0024297186421585897, + "loss": 1.1034, + "step": 3734 + }, + { + "epoch": 0.2872855934158911, + "learning_rate": 0.002429434171726704, + "loss": 0.9201, + "step": 3735 + }, + { + "epoch": 0.2873625105761095, + "learning_rate": 0.002429149647024188, + "loss": 1.1395, + "step": 3736 + }, + { + "epoch": 0.28743942773632797, + "learning_rate": 0.0024288650680676558, + "loss": 1.0924, + "step": 3737 + }, + { + "epoch": 0.2875163448965464, + "learning_rate": 0.0024285804348737243, + "loss": 1.451, + "step": 3738 + }, + { + "epoch": 0.2875932620567649, + "learning_rate": 0.002428295747459013, + "loss": 0.8121, + "step": 3739 + }, + { + "epoch": 0.2876701792169833, + "learning_rate": 0.0024280110058401464, + "loss": 0.8656, + "step": 3740 + }, + { + "epoch": 0.28774709637720175, + "learning_rate": 0.0024277262100337493, + "loss": 0.8274, + "step": 3741 + }, + { + "epoch": 0.2878240135374202, + "learning_rate": 0.0024274413600564523, + "loss": 1.1982, + "step": 3742 + }, + { + "epoch": 0.28790093069763867, + "learning_rate": 0.0024271564559248874, + "loss": 1.1877, + "step": 3743 + }, + { + "epoch": 0.28797784785785707, + "learning_rate": 0.0024268714976556915, + "loss": 0.7434, + "step": 3744 + }, + { + "epoch": 0.28805476501807553, + "learning_rate": 0.0024265864852655024, + "loss": 1.0574, + "step": 3745 + }, + { + "epoch": 0.288131682178294, + "learning_rate": 0.002426301418770963, + "loss": 1.2975, + "step": 3746 + }, + { + "epoch": 0.2882085993385124, + "learning_rate": 0.002426016298188718, + "loss": 1.3393, + "step": 3747 + }, + { + "epoch": 0.28828551649873085, + "learning_rate": 0.0024257311235354163, + "loss": 1.0958, + "step": 3748 + }, + { + "epoch": 0.2883624336589493, + "learning_rate": 0.00242544589482771, + "loss": 0.8638, + "step": 3749 + }, + { + "epoch": 0.2884393508191678, + "learning_rate": 0.0024251606120822525, + "loss": 1.2106, + "step": 3750 + }, + { + "epoch": 0.2885162679793862, + "learning_rate": 0.0024248752753157038, + "loss": 1.0024, + "step": 3751 + }, + { + "epoch": 0.28859318513960464, + "learning_rate": 0.0024245898845447235, + "loss": 0.8348, + "step": 3752 + }, + { + "epoch": 0.2886701022998231, + "learning_rate": 0.002424304439785976, + "loss": 1.1999, + "step": 3753 + }, + { + "epoch": 0.28874701946004155, + "learning_rate": 0.0024240189410561293, + "loss": 1.11, + "step": 3754 + }, + { + "epoch": 0.28882393662025996, + "learning_rate": 0.0024237333883718533, + "loss": 1.1857, + "step": 3755 + }, + { + "epoch": 0.2889008537804784, + "learning_rate": 0.002423447781749822, + "loss": 1.0494, + "step": 3756 + }, + { + "epoch": 0.2889777709406969, + "learning_rate": 0.002423162121206712, + "loss": 1.0682, + "step": 3757 + }, + { + "epoch": 0.28905468810091534, + "learning_rate": 0.0024228764067592045, + "loss": 1.0615, + "step": 3758 + }, + { + "epoch": 0.28913160526113374, + "learning_rate": 0.002422590638423982, + "loss": 1.4432, + "step": 3759 + }, + { + "epoch": 0.2892085224213522, + "learning_rate": 0.00242230481621773, + "loss": 1.26, + "step": 3760 + }, + { + "epoch": 0.28928543958157066, + "learning_rate": 0.0024220189401571385, + "loss": 1.0253, + "step": 3761 + }, + { + "epoch": 0.2893623567417891, + "learning_rate": 0.0024217330102589, + "loss": 1.2405, + "step": 3762 + }, + { + "epoch": 0.2894392739020075, + "learning_rate": 0.0024214470265397116, + "loss": 1.1035, + "step": 3763 + }, + { + "epoch": 0.289516191062226, + "learning_rate": 0.00242116098901627, + "loss": 1.0638, + "step": 3764 + }, + { + "epoch": 0.28959310822244444, + "learning_rate": 0.0024208748977052785, + "loss": 1.1296, + "step": 3765 + }, + { + "epoch": 0.2896700253826629, + "learning_rate": 0.002420588752623442, + "loss": 1.0415, + "step": 3766 + }, + { + "epoch": 0.2897469425428813, + "learning_rate": 0.0024203025537874686, + "loss": 1.5781, + "step": 3767 + }, + { + "epoch": 0.28982385970309976, + "learning_rate": 0.00242001630121407, + "loss": 1.0284, + "step": 3768 + }, + { + "epoch": 0.2899007768633182, + "learning_rate": 0.002419729994919961, + "loss": 1.1952, + "step": 3769 + }, + { + "epoch": 0.2899776940235366, + "learning_rate": 0.0024194436349218593, + "loss": 0.9078, + "step": 3770 + }, + { + "epoch": 0.2900546111837551, + "learning_rate": 0.0024191572212364848, + "loss": 0.93, + "step": 3771 + }, + { + "epoch": 0.29013152834397354, + "learning_rate": 0.0024188707538805626, + "loss": 0.9572, + "step": 3772 + }, + { + "epoch": 0.290208445504192, + "learning_rate": 0.0024185842328708195, + "loss": 0.8494, + "step": 3773 + }, + { + "epoch": 0.2902853626644104, + "learning_rate": 0.0024182976582239855, + "loss": 1.2251, + "step": 3774 + }, + { + "epoch": 0.29036227982462887, + "learning_rate": 0.0024180110299567947, + "loss": 1.0756, + "step": 3775 + }, + { + "epoch": 0.2904391969848473, + "learning_rate": 0.0024177243480859828, + "loss": 1.2445, + "step": 3776 + }, + { + "epoch": 0.2905161141450658, + "learning_rate": 0.00241743761262829, + "loss": 1.0082, + "step": 3777 + }, + { + "epoch": 0.2905930313052842, + "learning_rate": 0.002417150823600459, + "loss": 1.6257, + "step": 3778 + }, + { + "epoch": 0.29066994846550265, + "learning_rate": 0.002416863981019235, + "loss": 1.2663, + "step": 3779 + }, + { + "epoch": 0.2907468656257211, + "learning_rate": 0.0024165770849013677, + "loss": 1.21, + "step": 3780 + }, + { + "epoch": 0.29082378278593957, + "learning_rate": 0.002416290135263609, + "loss": 1.0922, + "step": 3781 + }, + { + "epoch": 0.29090069994615797, + "learning_rate": 0.0024160031321227154, + "loss": 1.2834, + "step": 3782 + }, + { + "epoch": 0.29097761710637643, + "learning_rate": 0.0024157160754954437, + "loss": 1.0925, + "step": 3783 + }, + { + "epoch": 0.2910545342665949, + "learning_rate": 0.0024154289653985556, + "loss": 1.0391, + "step": 3784 + }, + { + "epoch": 0.29113145142681335, + "learning_rate": 0.0024151418018488165, + "loss": 1.0189, + "step": 3785 + }, + { + "epoch": 0.29120836858703175, + "learning_rate": 0.002414854584862994, + "loss": 1.3025, + "step": 3786 + }, + { + "epoch": 0.2912852857472502, + "learning_rate": 0.002414567314457859, + "loss": 1.1334, + "step": 3787 + }, + { + "epoch": 0.29136220290746867, + "learning_rate": 0.002414279990650185, + "loss": 1.2363, + "step": 3788 + }, + { + "epoch": 0.2914391200676871, + "learning_rate": 0.0024139926134567496, + "loss": 1.0794, + "step": 3789 + }, + { + "epoch": 0.29151603722790553, + "learning_rate": 0.002413705182894333, + "loss": 1.3255, + "step": 3790 + }, + { + "epoch": 0.291592954388124, + "learning_rate": 0.0024134176989797185, + "loss": 0.7982, + "step": 3791 + }, + { + "epoch": 0.29166987154834245, + "learning_rate": 0.0024131301617296924, + "loss": 1.1689, + "step": 3792 + }, + { + "epoch": 0.29174678870856086, + "learning_rate": 0.0024128425711610452, + "loss": 1.076, + "step": 3793 + }, + { + "epoch": 0.2918237058687793, + "learning_rate": 0.002412554927290568, + "loss": 0.9094, + "step": 3794 + }, + { + "epoch": 0.2919006230289978, + "learning_rate": 0.0024122672301350578, + "loss": 1.0881, + "step": 3795 + }, + { + "epoch": 0.29197754018921623, + "learning_rate": 0.002411979479711314, + "loss": 1.0724, + "step": 3796 + }, + { + "epoch": 0.29205445734943464, + "learning_rate": 0.0024116916760361373, + "loss": 1.0396, + "step": 3797 + }, + { + "epoch": 0.2921313745096531, + "learning_rate": 0.0024114038191263335, + "loss": 1.3407, + "step": 3798 + }, + { + "epoch": 0.29220829166987156, + "learning_rate": 0.0024111159089987106, + "loss": 1.3073, + "step": 3799 + }, + { + "epoch": 0.29228520883009, + "learning_rate": 0.002410827945670081, + "loss": 1.0673, + "step": 3800 + }, + { + "epoch": 0.2923621259903084, + "learning_rate": 0.0024105399291572582, + "loss": 1.3231, + "step": 3801 + }, + { + "epoch": 0.2924390431505269, + "learning_rate": 0.00241025185947706, + "loss": 1.0472, + "step": 3802 + }, + { + "epoch": 0.29251596031074534, + "learning_rate": 0.0024099637366463067, + "loss": 1.1477, + "step": 3803 + }, + { + "epoch": 0.2925928774709638, + "learning_rate": 0.0024096755606818226, + "loss": 1.341, + "step": 3804 + }, + { + "epoch": 0.2926697946311822, + "learning_rate": 0.0024093873316004344, + "loss": 1.3286, + "step": 3805 + }, + { + "epoch": 0.29274671179140066, + "learning_rate": 0.0024090990494189724, + "loss": 1.1217, + "step": 3806 + }, + { + "epoch": 0.2928236289516191, + "learning_rate": 0.0024088107141542695, + "loss": 1.1633, + "step": 3807 + }, + { + "epoch": 0.2929005461118376, + "learning_rate": 0.0024085223258231616, + "loss": 1.3188, + "step": 3808 + }, + { + "epoch": 0.292977463272056, + "learning_rate": 0.0024082338844424885, + "loss": 1.0036, + "step": 3809 + }, + { + "epoch": 0.29305438043227444, + "learning_rate": 0.002407945390029092, + "loss": 1.3416, + "step": 3810 + }, + { + "epoch": 0.2931312975924929, + "learning_rate": 0.0024076568425998182, + "loss": 0.8524, + "step": 3811 + }, + { + "epoch": 0.2932082147527113, + "learning_rate": 0.002407368242171516, + "loss": 1.2356, + "step": 3812 + }, + { + "epoch": 0.29328513191292976, + "learning_rate": 0.002407079588761036, + "loss": 1.1089, + "step": 3813 + }, + { + "epoch": 0.2933620490731482, + "learning_rate": 0.002406790882385233, + "loss": 1.4834, + "step": 3814 + }, + { + "epoch": 0.2934389662333667, + "learning_rate": 0.0024065021230609655, + "loss": 0.8913, + "step": 3815 + }, + { + "epoch": 0.2935158833935851, + "learning_rate": 0.002406213310805095, + "loss": 1.3506, + "step": 3816 + }, + { + "epoch": 0.29359280055380355, + "learning_rate": 0.002405924445634484, + "loss": 0.9484, + "step": 3817 + }, + { + "epoch": 0.293669717714022, + "learning_rate": 0.0024056355275660006, + "loss": 1.6326, + "step": 3818 + }, + { + "epoch": 0.29374663487424046, + "learning_rate": 0.0024053465566165155, + "loss": 1.4045, + "step": 3819 + }, + { + "epoch": 0.29382355203445887, + "learning_rate": 0.002405057532802901, + "loss": 1.3565, + "step": 3820 + }, + { + "epoch": 0.2939004691946773, + "learning_rate": 0.002404768456142034, + "loss": 1.2735, + "step": 3821 + }, + { + "epoch": 0.2939773863548958, + "learning_rate": 0.002404479326650794, + "loss": 1.3249, + "step": 3822 + }, + { + "epoch": 0.29405430351511425, + "learning_rate": 0.0024041901443460634, + "loss": 1.3285, + "step": 3823 + }, + { + "epoch": 0.29413122067533265, + "learning_rate": 0.002403900909244728, + "loss": 1.3567, + "step": 3824 + }, + { + "epoch": 0.2942081378355511, + "learning_rate": 0.0024036116213636766, + "loss": 0.9761, + "step": 3825 + }, + { + "epoch": 0.29428505499576957, + "learning_rate": 0.002403322280719801, + "loss": 0.9587, + "step": 3826 + }, + { + "epoch": 0.294361972155988, + "learning_rate": 0.0024030328873299953, + "loss": 1.0449, + "step": 3827 + }, + { + "epoch": 0.29443888931620643, + "learning_rate": 0.0024027434412111593, + "loss": 1.0609, + "step": 3828 + }, + { + "epoch": 0.2945158064764249, + "learning_rate": 0.002402453942380192, + "loss": 1.0602, + "step": 3829 + }, + { + "epoch": 0.29459272363664335, + "learning_rate": 0.002402164390853999, + "loss": 0.949, + "step": 3830 + }, + { + "epoch": 0.29466964079686175, + "learning_rate": 0.0024018747866494865, + "loss": 1.1173, + "step": 3831 + }, + { + "epoch": 0.2947465579570802, + "learning_rate": 0.0024015851297835658, + "loss": 0.6908, + "step": 3832 + }, + { + "epoch": 0.2948234751172987, + "learning_rate": 0.0024012954202731495, + "loss": 0.8463, + "step": 3833 + }, + { + "epoch": 0.29490039227751713, + "learning_rate": 0.0024010056581351543, + "loss": 1.4411, + "step": 3834 + }, + { + "epoch": 0.29497730943773554, + "learning_rate": 0.0024007158433865003, + "loss": 1.1703, + "step": 3835 + }, + { + "epoch": 0.295054226597954, + "learning_rate": 0.0024004259760441087, + "loss": 1.1127, + "step": 3836 + }, + { + "epoch": 0.29513114375817245, + "learning_rate": 0.0024001360561249066, + "loss": 1.3076, + "step": 3837 + }, + { + "epoch": 0.2952080609183909, + "learning_rate": 0.002399846083645822, + "loss": 0.7568, + "step": 3838 + }, + { + "epoch": 0.2952849780786093, + "learning_rate": 0.002399556058623786, + "loss": 1.1442, + "step": 3839 + }, + { + "epoch": 0.2953618952388278, + "learning_rate": 0.0023992659810757352, + "loss": 0.9839, + "step": 3840 + }, + { + "epoch": 0.29543881239904624, + "learning_rate": 0.0023989758510186057, + "loss": 1.0016, + "step": 3841 + }, + { + "epoch": 0.2955157295592647, + "learning_rate": 0.00239868566846934, + "loss": 1.1556, + "step": 3842 + }, + { + "epoch": 0.2955926467194831, + "learning_rate": 0.0023983954334448813, + "loss": 1.1459, + "step": 3843 + }, + { + "epoch": 0.29566956387970156, + "learning_rate": 0.0023981051459621766, + "loss": 1.6542, + "step": 3844 + }, + { + "epoch": 0.29574648103992, + "learning_rate": 0.0023978148060381775, + "loss": 1.0692, + "step": 3845 + }, + { + "epoch": 0.2958233982001385, + "learning_rate": 0.0023975244136898353, + "loss": 0.9232, + "step": 3846 + }, + { + "epoch": 0.2959003153603569, + "learning_rate": 0.0023972339689341073, + "loss": 1.0787, + "step": 3847 + }, + { + "epoch": 0.29597723252057534, + "learning_rate": 0.0023969434717879532, + "loss": 1.2244, + "step": 3848 + }, + { + "epoch": 0.2960541496807938, + "learning_rate": 0.0023966529222683346, + "loss": 1.1639, + "step": 3849 + }, + { + "epoch": 0.2961310668410122, + "learning_rate": 0.0023963623203922175, + "loss": 1.1043, + "step": 3850 + }, + { + "epoch": 0.29620798400123066, + "learning_rate": 0.0023960716661765706, + "loss": 1.1608, + "step": 3851 + }, + { + "epoch": 0.2962849011614491, + "learning_rate": 0.002395780959638365, + "loss": 1.0279, + "step": 3852 + }, + { + "epoch": 0.2963618183216676, + "learning_rate": 0.0023954902007945755, + "loss": 1.1058, + "step": 3853 + }, + { + "epoch": 0.296438735481886, + "learning_rate": 0.00239519938966218, + "loss": 1.2928, + "step": 3854 + }, + { + "epoch": 0.29651565264210444, + "learning_rate": 0.0023949085262581597, + "loss": 1.0284, + "step": 3855 + }, + { + "epoch": 0.2965925698023229, + "learning_rate": 0.0023946176105994967, + "loss": 1.5161, + "step": 3856 + }, + { + "epoch": 0.29666948696254136, + "learning_rate": 0.00239432664270318, + "loss": 1.0126, + "step": 3857 + }, + { + "epoch": 0.29674640412275977, + "learning_rate": 0.002394035622586199, + "loss": 1.4353, + "step": 3858 + }, + { + "epoch": 0.2968233212829782, + "learning_rate": 0.0023937445502655455, + "loss": 1.2178, + "step": 3859 + }, + { + "epoch": 0.2969002384431967, + "learning_rate": 0.0023934534257582163, + "loss": 1.2485, + "step": 3860 + }, + { + "epoch": 0.29697715560341514, + "learning_rate": 0.0023931622490812104, + "loss": 1.1826, + "step": 3861 + }, + { + "epoch": 0.29705407276363355, + "learning_rate": 0.00239287102025153, + "loss": 1.4365, + "step": 3862 + }, + { + "epoch": 0.297130989923852, + "learning_rate": 0.00239257973928618, + "loss": 1.123, + "step": 3863 + }, + { + "epoch": 0.29720790708407047, + "learning_rate": 0.00239228840620217, + "loss": 1.1501, + "step": 3864 + }, + { + "epoch": 0.2972848242442889, + "learning_rate": 0.0023919970210165087, + "loss": 1.2852, + "step": 3865 + }, + { + "epoch": 0.29736174140450733, + "learning_rate": 0.002391705583746212, + "loss": 1.384, + "step": 3866 + }, + { + "epoch": 0.2974386585647258, + "learning_rate": 0.0023914140944082973, + "loss": 1.5268, + "step": 3867 + }, + { + "epoch": 0.29751557572494425, + "learning_rate": 0.0023911225530197843, + "loss": 1.1828, + "step": 3868 + }, + { + "epoch": 0.2975924928851627, + "learning_rate": 0.0023908309595976965, + "loss": 1.4875, + "step": 3869 + }, + { + "epoch": 0.2976694100453811, + "learning_rate": 0.0023905393141590613, + "loss": 1.1925, + "step": 3870 + }, + { + "epoch": 0.29774632720559957, + "learning_rate": 0.0023902476167209077, + "loss": 0.9692, + "step": 3871 + }, + { + "epoch": 0.29782324436581803, + "learning_rate": 0.002389955867300267, + "loss": 0.8897, + "step": 3872 + }, + { + "epoch": 0.29790016152603643, + "learning_rate": 0.002389664065914176, + "loss": 0.7786, + "step": 3873 + }, + { + "epoch": 0.2979770786862549, + "learning_rate": 0.002389372212579673, + "loss": 1.1258, + "step": 3874 + }, + { + "epoch": 0.29805399584647335, + "learning_rate": 0.0023890803073138, + "loss": 1.0717, + "step": 3875 + }, + { + "epoch": 0.2981309130066918, + "learning_rate": 0.0023887883501336007, + "loss": 0.6882, + "step": 3876 + }, + { + "epoch": 0.2982078301669102, + "learning_rate": 0.0023884963410561234, + "loss": 1.1122, + "step": 3877 + }, + { + "epoch": 0.2982847473271287, + "learning_rate": 0.002388204280098419, + "loss": 1.3263, + "step": 3878 + }, + { + "epoch": 0.29836166448734713, + "learning_rate": 0.0023879121672775415, + "loss": 0.8651, + "step": 3879 + }, + { + "epoch": 0.2984385816475656, + "learning_rate": 0.0023876200026105463, + "loss": 1.3739, + "step": 3880 + }, + { + "epoch": 0.298515498807784, + "learning_rate": 0.002387327786114495, + "loss": 0.985, + "step": 3881 + }, + { + "epoch": 0.29859241596800246, + "learning_rate": 0.002387035517806449, + "loss": 1.0379, + "step": 3882 + }, + { + "epoch": 0.2986693331282209, + "learning_rate": 0.0023867431977034744, + "loss": 1.0018, + "step": 3883 + }, + { + "epoch": 0.2987462502884394, + "learning_rate": 0.0023864508258226408, + "loss": 0.9304, + "step": 3884 + }, + { + "epoch": 0.2988231674486578, + "learning_rate": 0.0023861584021810196, + "loss": 0.8609, + "step": 3885 + }, + { + "epoch": 0.29890008460887624, + "learning_rate": 0.002385865926795686, + "loss": 1.1605, + "step": 3886 + }, + { + "epoch": 0.2989770017690947, + "learning_rate": 0.002385573399683717, + "loss": 1.0836, + "step": 3887 + }, + { + "epoch": 0.29905391892931316, + "learning_rate": 0.0023852808208621947, + "loss": 1.4671, + "step": 3888 + }, + { + "epoch": 0.29913083608953156, + "learning_rate": 0.0023849881903482027, + "loss": 1.2837, + "step": 3889 + }, + { + "epoch": 0.29920775324975, + "learning_rate": 0.0023846955081588276, + "loss": 1.079, + "step": 3890 + }, + { + "epoch": 0.2992846704099685, + "learning_rate": 0.00238440277431116, + "loss": 1.202, + "step": 3891 + }, + { + "epoch": 0.2993615875701869, + "learning_rate": 0.002384109988822293, + "loss": 1.0768, + "step": 3892 + }, + { + "epoch": 0.29943850473040534, + "learning_rate": 0.002383817151709322, + "loss": 1.8944, + "step": 3893 + }, + { + "epoch": 0.2995154218906238, + "learning_rate": 0.002383524262989346, + "loss": 1.1621, + "step": 3894 + }, + { + "epoch": 0.29959233905084226, + "learning_rate": 0.002383231322679468, + "loss": 0.9345, + "step": 3895 + }, + { + "epoch": 0.29966925621106066, + "learning_rate": 0.0023829383307967925, + "loss": 1.178, + "step": 3896 + }, + { + "epoch": 0.2997461733712791, + "learning_rate": 0.002382645287358428, + "loss": 1.193, + "step": 3897 + }, + { + "epoch": 0.2998230905314976, + "learning_rate": 0.002382352192381485, + "loss": 1.0396, + "step": 3898 + }, + { + "epoch": 0.29990000769171604, + "learning_rate": 0.002382059045883078, + "loss": 1.4754, + "step": 3899 + }, + { + "epoch": 0.29997692485193445, + "learning_rate": 0.0023817658478803246, + "loss": 1.4991, + "step": 3900 + }, + { + "epoch": 0.3000538420121529, + "learning_rate": 0.0023814725983903437, + "loss": 0.7582, + "step": 3901 + }, + { + "epoch": 0.30013075917237136, + "learning_rate": 0.002381179297430259, + "loss": 1.2912, + "step": 3902 + }, + { + "epoch": 0.3002076763325898, + "learning_rate": 0.002380885945017197, + "loss": 1.1994, + "step": 3903 + }, + { + "epoch": 0.3002845934928082, + "learning_rate": 0.002380592541168287, + "loss": 1.0211, + "step": 3904 + }, + { + "epoch": 0.3003615106530267, + "learning_rate": 0.00238029908590066, + "loss": 1.4471, + "step": 3905 + }, + { + "epoch": 0.30043842781324515, + "learning_rate": 0.0023800055792314524, + "loss": 1.5325, + "step": 3906 + }, + { + "epoch": 0.3005153449734636, + "learning_rate": 0.002379712021177802, + "loss": 1.1501, + "step": 3907 + }, + { + "epoch": 0.300592262133682, + "learning_rate": 0.00237941841175685, + "loss": 0.637, + "step": 3908 + }, + { + "epoch": 0.30066917929390047, + "learning_rate": 0.00237912475098574, + "loss": 1.2793, + "step": 3909 + }, + { + "epoch": 0.3007460964541189, + "learning_rate": 0.0023788310388816202, + "loss": 1.1716, + "step": 3910 + }, + { + "epoch": 0.30082301361433733, + "learning_rate": 0.00237853727546164, + "loss": 1.6392, + "step": 3911 + }, + { + "epoch": 0.3008999307745558, + "learning_rate": 0.002378243460742952, + "loss": 0.9761, + "step": 3912 + }, + { + "epoch": 0.30097684793477425, + "learning_rate": 0.0023779495947427137, + "loss": 1.1093, + "step": 3913 + }, + { + "epoch": 0.3010537650949927, + "learning_rate": 0.0023776556774780834, + "loss": 1.2699, + "step": 3914 + }, + { + "epoch": 0.3011306822552111, + "learning_rate": 0.002377361708966223, + "loss": 1.3389, + "step": 3915 + }, + { + "epoch": 0.30120759941542957, + "learning_rate": 0.002377067689224299, + "loss": 1.124, + "step": 3916 + }, + { + "epoch": 0.30128451657564803, + "learning_rate": 0.0023767736182694777, + "loss": 1.1271, + "step": 3917 + }, + { + "epoch": 0.3013614337358665, + "learning_rate": 0.002376479496118931, + "loss": 1.1094, + "step": 3918 + }, + { + "epoch": 0.3014383508960849, + "learning_rate": 0.0023761853227898338, + "loss": 1.0094, + "step": 3919 + }, + { + "epoch": 0.30151526805630335, + "learning_rate": 0.0023758910982993623, + "loss": 1.157, + "step": 3920 + }, + { + "epoch": 0.3015921852165218, + "learning_rate": 0.002375596822664696, + "loss": 1.0624, + "step": 3921 + }, + { + "epoch": 0.3016691023767403, + "learning_rate": 0.0023753024959030193, + "loss": 1.0249, + "step": 3922 + }, + { + "epoch": 0.3017460195369587, + "learning_rate": 0.002375008118031518, + "loss": 1.2358, + "step": 3923 + }, + { + "epoch": 0.30182293669717714, + "learning_rate": 0.0023747136890673805, + "loss": 1.0304, + "step": 3924 + }, + { + "epoch": 0.3018998538573956, + "learning_rate": 0.0023744192090277986, + "loss": 1.3563, + "step": 3925 + }, + { + "epoch": 0.30197677101761405, + "learning_rate": 0.0023741246779299687, + "loss": 1.2886, + "step": 3926 + }, + { + "epoch": 0.30205368817783246, + "learning_rate": 0.002373830095791087, + "loss": 1.0715, + "step": 3927 + }, + { + "epoch": 0.3021306053380509, + "learning_rate": 0.0023735354626283557, + "loss": 1.3066, + "step": 3928 + }, + { + "epoch": 0.3022075224982694, + "learning_rate": 0.0023732407784589783, + "loss": 1.2763, + "step": 3929 + }, + { + "epoch": 0.30228443965848784, + "learning_rate": 0.0023729460433001615, + "loss": 1.4041, + "step": 3930 + }, + { + "epoch": 0.30236135681870624, + "learning_rate": 0.0023726512571691162, + "loss": 1.1763, + "step": 3931 + }, + { + "epoch": 0.3024382739789247, + "learning_rate": 0.0023723564200830542, + "loss": 1.3388, + "step": 3932 + }, + { + "epoch": 0.30251519113914316, + "learning_rate": 0.0023720615320591914, + "loss": 1.1631, + "step": 3933 + }, + { + "epoch": 0.30259210829936156, + "learning_rate": 0.0023717665931147474, + "loss": 0.9775, + "step": 3934 + }, + { + "epoch": 0.30266902545958, + "learning_rate": 0.002371471603266943, + "loss": 1.1244, + "step": 3935 + }, + { + "epoch": 0.3027459426197985, + "learning_rate": 0.0023711765625330037, + "loss": 1.2908, + "step": 3936 + }, + { + "epoch": 0.30282285978001694, + "learning_rate": 0.002370881470930157, + "loss": 0.8563, + "step": 3937 + }, + { + "epoch": 0.30289977694023534, + "learning_rate": 0.0023705863284756335, + "loss": 1.1576, + "step": 3938 + }, + { + "epoch": 0.3029766941004538, + "learning_rate": 0.0023702911351866666, + "loss": 1.0741, + "step": 3939 + }, + { + "epoch": 0.30305361126067226, + "learning_rate": 0.0023699958910804943, + "loss": 0.9602, + "step": 3940 + }, + { + "epoch": 0.3031305284208907, + "learning_rate": 0.0023697005961743544, + "loss": 1.0266, + "step": 3941 + }, + { + "epoch": 0.3032074455811091, + "learning_rate": 0.0023694052504854906, + "loss": 1.3567, + "step": 3942 + }, + { + "epoch": 0.3032843627413276, + "learning_rate": 0.0023691098540311478, + "loss": 1.1743, + "step": 3943 + }, + { + "epoch": 0.30336127990154604, + "learning_rate": 0.0023688144068285755, + "loss": 1.0754, + "step": 3944 + }, + { + "epoch": 0.3034381970617645, + "learning_rate": 0.002368518908895024, + "loss": 1.1054, + "step": 3945 + }, + { + "epoch": 0.3035151142219829, + "learning_rate": 0.0023682233602477483, + "loss": 1.0102, + "step": 3946 + }, + { + "epoch": 0.30359203138220137, + "learning_rate": 0.002367927760904006, + "loss": 1.1887, + "step": 3947 + }, + { + "epoch": 0.3036689485424198, + "learning_rate": 0.002367632110881057, + "loss": 0.8444, + "step": 3948 + }, + { + "epoch": 0.3037458657026383, + "learning_rate": 0.002367336410196165, + "loss": 1.2563, + "step": 3949 + }, + { + "epoch": 0.3038227828628567, + "learning_rate": 0.002367040658866596, + "loss": 1.0764, + "step": 3950 + }, + { + "epoch": 0.30389970002307515, + "learning_rate": 0.00236674485690962, + "loss": 1.0052, + "step": 3951 + }, + { + "epoch": 0.3039766171832936, + "learning_rate": 0.0023664490043425078, + "loss": 0.9344, + "step": 3952 + }, + { + "epoch": 0.304053534343512, + "learning_rate": 0.002366153101182535, + "loss": 0.9685, + "step": 3953 + }, + { + "epoch": 0.30413045150373047, + "learning_rate": 0.0023658571474469807, + "loss": 0.9161, + "step": 3954 + }, + { + "epoch": 0.30420736866394893, + "learning_rate": 0.0023655611431531247, + "loss": 1.4992, + "step": 3955 + }, + { + "epoch": 0.3042842858241674, + "learning_rate": 0.0023652650883182512, + "loss": 1.3994, + "step": 3956 + }, + { + "epoch": 0.3043612029843858, + "learning_rate": 0.002364968982959648, + "loss": 1.1665, + "step": 3957 + }, + { + "epoch": 0.30443812014460425, + "learning_rate": 0.002364672827094604, + "loss": 1.2797, + "step": 3958 + }, + { + "epoch": 0.3045150373048227, + "learning_rate": 0.0023643766207404127, + "loss": 1.1326, + "step": 3959 + }, + { + "epoch": 0.30459195446504117, + "learning_rate": 0.00236408036391437, + "loss": 1.1032, + "step": 3960 + }, + { + "epoch": 0.3046688716252596, + "learning_rate": 0.002363784056633774, + "loss": 1.1207, + "step": 3961 + }, + { + "epoch": 0.30474578878547803, + "learning_rate": 0.0023634876989159265, + "loss": 0.9382, + "step": 3962 + }, + { + "epoch": 0.3048227059456965, + "learning_rate": 0.0023631912907781326, + "loss": 0.9859, + "step": 3963 + }, + { + "epoch": 0.30489962310591495, + "learning_rate": 0.0023628948322376997, + "loss": 1.4293, + "step": 3964 + }, + { + "epoch": 0.30497654026613336, + "learning_rate": 0.0023625983233119385, + "loss": 1.1807, + "step": 3965 + }, + { + "epoch": 0.3050534574263518, + "learning_rate": 0.0023623017640181615, + "loss": 1.0156, + "step": 3966 + }, + { + "epoch": 0.3051303745865703, + "learning_rate": 0.002362005154373686, + "loss": 0.7093, + "step": 3967 + }, + { + "epoch": 0.30520729174678873, + "learning_rate": 0.0023617084943958313, + "loss": 0.9847, + "step": 3968 + }, + { + "epoch": 0.30528420890700714, + "learning_rate": 0.00236141178410192, + "loss": 1.0178, + "step": 3969 + }, + { + "epoch": 0.3053611260672256, + "learning_rate": 0.0023611150235092766, + "loss": 0.8512, + "step": 3970 + }, + { + "epoch": 0.30543804322744406, + "learning_rate": 0.0023608182126352295, + "loss": 1.0829, + "step": 3971 + }, + { + "epoch": 0.30551496038766246, + "learning_rate": 0.00236052135149711, + "loss": 1.1115, + "step": 3972 + }, + { + "epoch": 0.3055918775478809, + "learning_rate": 0.0023602244401122515, + "loss": 1.2082, + "step": 3973 + }, + { + "epoch": 0.3056687947080994, + "learning_rate": 0.0023599274784979916, + "loss": 1.5141, + "step": 3974 + }, + { + "epoch": 0.30574571186831784, + "learning_rate": 0.00235963046667167, + "loss": 1.2173, + "step": 3975 + }, + { + "epoch": 0.30582262902853624, + "learning_rate": 0.00235933340465063, + "loss": 1.3931, + "step": 3976 + }, + { + "epoch": 0.3058995461887547, + "learning_rate": 0.0023590362924522166, + "loss": 1.0222, + "step": 3977 + }, + { + "epoch": 0.30597646334897316, + "learning_rate": 0.0023587391300937794, + "loss": 1.5301, + "step": 3978 + }, + { + "epoch": 0.3060533805091916, + "learning_rate": 0.0023584419175926683, + "loss": 1.0222, + "step": 3979 + }, + { + "epoch": 0.30613029766941, + "learning_rate": 0.00235814465496624, + "loss": 0.9644, + "step": 3980 + }, + { + "epoch": 0.3062072148296285, + "learning_rate": 0.0023578473422318507, + "loss": 1.4717, + "step": 3981 + }, + { + "epoch": 0.30628413198984694, + "learning_rate": 0.0023575499794068608, + "loss": 1.1743, + "step": 3982 + }, + { + "epoch": 0.3063610491500654, + "learning_rate": 0.002357252566508634, + "loss": 0.9093, + "step": 3983 + }, + { + "epoch": 0.3064379663102838, + "learning_rate": 0.002356955103554536, + "loss": 1.092, + "step": 3984 + }, + { + "epoch": 0.30651488347050226, + "learning_rate": 0.0023566575905619373, + "loss": 1.0453, + "step": 3985 + }, + { + "epoch": 0.3065918006307207, + "learning_rate": 0.0023563600275482086, + "loss": 0.8073, + "step": 3986 + }, + { + "epoch": 0.3066687177909392, + "learning_rate": 0.0023560624145307256, + "loss": 1.0317, + "step": 3987 + }, + { + "epoch": 0.3067456349511576, + "learning_rate": 0.002355764751526866, + "loss": 0.9002, + "step": 3988 + }, + { + "epoch": 0.30682255211137605, + "learning_rate": 0.0023554670385540107, + "loss": 1.3194, + "step": 3989 + }, + { + "epoch": 0.3068994692715945, + "learning_rate": 0.0023551692756295433, + "loss": 0.9924, + "step": 3990 + }, + { + "epoch": 0.30697638643181296, + "learning_rate": 0.0023548714627708504, + "loss": 1.0192, + "step": 3991 + }, + { + "epoch": 0.30705330359203137, + "learning_rate": 0.0023545735999953225, + "loss": 1.0896, + "step": 3992 + }, + { + "epoch": 0.3071302207522498, + "learning_rate": 0.002354275687320351, + "loss": 1.0743, + "step": 3993 + }, + { + "epoch": 0.3072071379124683, + "learning_rate": 0.002353977724763332, + "loss": 1.3867, + "step": 3994 + }, + { + "epoch": 0.3072840550726867, + "learning_rate": 0.002353679712341664, + "loss": 1.0094, + "step": 3995 + }, + { + "epoch": 0.30736097223290515, + "learning_rate": 0.0023533816500727472, + "loss": 1.2725, + "step": 3996 + }, + { + "epoch": 0.3074378893931236, + "learning_rate": 0.002353083537973987, + "loss": 1.4228, + "step": 3997 + }, + { + "epoch": 0.30751480655334207, + "learning_rate": 0.0023527853760627894, + "loss": 1.0509, + "step": 3998 + }, + { + "epoch": 0.30759172371356047, + "learning_rate": 0.002352487164356566, + "loss": 0.9508, + "step": 3999 + }, + { + "epoch": 0.30766864087377893, + "learning_rate": 0.002352188902872728, + "loss": 1.188, + "step": 4000 + }, + { + "epoch": 0.3077455580339974, + "learning_rate": 0.0023518905916286915, + "loss": 1.5387, + "step": 4001 + }, + { + "epoch": 0.30782247519421585, + "learning_rate": 0.002351592230641876, + "loss": 1.3311, + "step": 4002 + }, + { + "epoch": 0.30789939235443425, + "learning_rate": 0.0023512938199297027, + "loss": 1.3555, + "step": 4003 + }, + { + "epoch": 0.3079763095146527, + "learning_rate": 0.0023509953595095955, + "loss": 0.9321, + "step": 4004 + }, + { + "epoch": 0.3080532266748712, + "learning_rate": 0.0023506968493989828, + "loss": 1.0828, + "step": 4005 + }, + { + "epoch": 0.30813014383508963, + "learning_rate": 0.0023503982896152945, + "loss": 1.1461, + "step": 4006 + }, + { + "epoch": 0.30820706099530804, + "learning_rate": 0.0023500996801759638, + "loss": 1.1811, + "step": 4007 + }, + { + "epoch": 0.3082839781555265, + "learning_rate": 0.0023498010210984267, + "loss": 1.2982, + "step": 4008 + }, + { + "epoch": 0.30836089531574495, + "learning_rate": 0.002349502312400122, + "loss": 1.238, + "step": 4009 + }, + { + "epoch": 0.3084378124759634, + "learning_rate": 0.0023492035540984926, + "loss": 1.1188, + "step": 4010 + }, + { + "epoch": 0.3085147296361818, + "learning_rate": 0.0023489047462109825, + "loss": 1.2138, + "step": 4011 + }, + { + "epoch": 0.3085916467964003, + "learning_rate": 0.0023486058887550393, + "loss": 1.0183, + "step": 4012 + }, + { + "epoch": 0.30866856395661874, + "learning_rate": 0.0023483069817481143, + "loss": 1.0868, + "step": 4013 + }, + { + "epoch": 0.30874548111683714, + "learning_rate": 0.00234800802520766, + "loss": 0.971, + "step": 4014 + }, + { + "epoch": 0.3088223982770556, + "learning_rate": 0.0023477090191511337, + "loss": 1.1807, + "step": 4015 + }, + { + "epoch": 0.30889931543727406, + "learning_rate": 0.002347409963595994, + "loss": 1.1903, + "step": 4016 + }, + { + "epoch": 0.3089762325974925, + "learning_rate": 0.0023471108585597033, + "loss": 1.2067, + "step": 4017 + }, + { + "epoch": 0.3090531497577109, + "learning_rate": 0.0023468117040597275, + "loss": 1.2638, + "step": 4018 + }, + { + "epoch": 0.3091300669179294, + "learning_rate": 0.002346512500113533, + "loss": 1.1632, + "step": 4019 + }, + { + "epoch": 0.30920698407814784, + "learning_rate": 0.002346213246738592, + "loss": 1.1534, + "step": 4020 + }, + { + "epoch": 0.3092839012383663, + "learning_rate": 0.002345913943952377, + "loss": 0.7224, + "step": 4021 + }, + { + "epoch": 0.3093608183985847, + "learning_rate": 0.002345614591772366, + "loss": 0.9565, + "step": 4022 + }, + { + "epoch": 0.30943773555880316, + "learning_rate": 0.002345315190216037, + "loss": 1.1661, + "step": 4023 + }, + { + "epoch": 0.3095146527190216, + "learning_rate": 0.0023450157393008733, + "loss": 1.1602, + "step": 4024 + }, + { + "epoch": 0.3095915698792401, + "learning_rate": 0.00234471623904436, + "loss": 0.9932, + "step": 4025 + }, + { + "epoch": 0.3096684870394585, + "learning_rate": 0.0023444166894639852, + "loss": 1.0469, + "step": 4026 + }, + { + "epoch": 0.30974540419967694, + "learning_rate": 0.00234411709057724, + "loss": 1.4895, + "step": 4027 + }, + { + "epoch": 0.3098223213598954, + "learning_rate": 0.0023438174424016173, + "loss": 1.4801, + "step": 4028 + }, + { + "epoch": 0.30989923852011386, + "learning_rate": 0.0023435177449546156, + "loss": 0.8212, + "step": 4029 + }, + { + "epoch": 0.30997615568033227, + "learning_rate": 0.0023432179982537335, + "loss": 1.0803, + "step": 4030 + }, + { + "epoch": 0.3100530728405507, + "learning_rate": 0.0023429182023164737, + "loss": 1.3106, + "step": 4031 + }, + { + "epoch": 0.3101299900007692, + "learning_rate": 0.0023426183571603415, + "loss": 0.9331, + "step": 4032 + }, + { + "epoch": 0.31020690716098764, + "learning_rate": 0.002342318462802845, + "loss": 1.1308, + "step": 4033 + }, + { + "epoch": 0.31028382432120605, + "learning_rate": 0.002342018519261496, + "loss": 1.585, + "step": 4034 + }, + { + "epoch": 0.3103607414814245, + "learning_rate": 0.0023417185265538084, + "loss": 1.0905, + "step": 4035 + }, + { + "epoch": 0.31043765864164297, + "learning_rate": 0.002341418484697298, + "loss": 1.3169, + "step": 4036 + }, + { + "epoch": 0.31051457580186137, + "learning_rate": 0.002341118393709486, + "loss": 1.0269, + "step": 4037 + }, + { + "epoch": 0.31059149296207983, + "learning_rate": 0.0023408182536078944, + "loss": 1.0275, + "step": 4038 + }, + { + "epoch": 0.3106684101222983, + "learning_rate": 0.0023405180644100487, + "loss": 1.3213, + "step": 4039 + }, + { + "epoch": 0.31074532728251675, + "learning_rate": 0.0023402178261334766, + "loss": 1.1029, + "step": 4040 + }, + { + "epoch": 0.31082224444273515, + "learning_rate": 0.0023399175387957107, + "loss": 1.0443, + "step": 4041 + }, + { + "epoch": 0.3108991616029536, + "learning_rate": 0.0023396172024142846, + "loss": 1.1436, + "step": 4042 + }, + { + "epoch": 0.31097607876317207, + "learning_rate": 0.002339316817006734, + "loss": 1.0799, + "step": 4043 + }, + { + "epoch": 0.31105299592339053, + "learning_rate": 0.0023390163825906004, + "loss": 1.241, + "step": 4044 + }, + { + "epoch": 0.31112991308360893, + "learning_rate": 0.002338715899183426, + "loss": 1.2144, + "step": 4045 + }, + { + "epoch": 0.3112068302438274, + "learning_rate": 0.002338415366802756, + "loss": 1.4211, + "step": 4046 + }, + { + "epoch": 0.31128374740404585, + "learning_rate": 0.0023381147854661384, + "loss": 1.2576, + "step": 4047 + }, + { + "epoch": 0.3113606645642643, + "learning_rate": 0.0023378141551911257, + "loss": 1.1191, + "step": 4048 + }, + { + "epoch": 0.3114375817244827, + "learning_rate": 0.0023375134759952712, + "loss": 1.3006, + "step": 4049 + }, + { + "epoch": 0.3115144988847012, + "learning_rate": 0.0023372127478961317, + "loss": 1.401, + "step": 4050 + }, + { + "epoch": 0.31159141604491963, + "learning_rate": 0.0023369119709112673, + "loss": 0.8778, + "step": 4051 + }, + { + "epoch": 0.3116683332051381, + "learning_rate": 0.0023366111450582407, + "loss": 1.3404, + "step": 4052 + }, + { + "epoch": 0.3117452503653565, + "learning_rate": 0.0023363102703546183, + "loss": 1.3865, + "step": 4053 + }, + { + "epoch": 0.31182216752557496, + "learning_rate": 0.0023360093468179667, + "loss": 1.2789, + "step": 4054 + }, + { + "epoch": 0.3118990846857934, + "learning_rate": 0.0023357083744658586, + "loss": 1.0586, + "step": 4055 + }, + { + "epoch": 0.3119760018460118, + "learning_rate": 0.0023354073533158673, + "loss": 1.1281, + "step": 4056 + }, + { + "epoch": 0.3120529190062303, + "learning_rate": 0.00233510628338557, + "loss": 1.0259, + "step": 4057 + }, + { + "epoch": 0.31212983616644874, + "learning_rate": 0.002334805164692546, + "loss": 1.2627, + "step": 4058 + }, + { + "epoch": 0.3122067533266672, + "learning_rate": 0.0023345039972543797, + "loss": 1.2103, + "step": 4059 + }, + { + "epoch": 0.3122836704868856, + "learning_rate": 0.0023342027810886543, + "loss": 1.1784, + "step": 4060 + }, + { + "epoch": 0.31236058764710406, + "learning_rate": 0.0023339015162129593, + "loss": 1.3992, + "step": 4061 + }, + { + "epoch": 0.3124375048073225, + "learning_rate": 0.0023336002026448856, + "loss": 1.173, + "step": 4062 + }, + { + "epoch": 0.312514421967541, + "learning_rate": 0.0023332988404020276, + "loss": 1.2913, + "step": 4063 + }, + { + "epoch": 0.3125913391277594, + "learning_rate": 0.002332997429501982, + "loss": 0.9053, + "step": 4064 + }, + { + "epoch": 0.31266825628797784, + "learning_rate": 0.002332695969962348, + "loss": 1.0656, + "step": 4065 + }, + { + "epoch": 0.3127451734481963, + "learning_rate": 0.0023323944618007285, + "loss": 1.3543, + "step": 4066 + }, + { + "epoch": 0.31282209060841476, + "learning_rate": 0.002332092905034729, + "loss": 1.1413, + "step": 4067 + }, + { + "epoch": 0.31289900776863316, + "learning_rate": 0.002331791299681958, + "loss": 0.9858, + "step": 4068 + }, + { + "epoch": 0.3129759249288516, + "learning_rate": 0.0023314896457600253, + "loss": 0.8808, + "step": 4069 + }, + { + "epoch": 0.3130528420890701, + "learning_rate": 0.0023311879432865463, + "loss": 1.2249, + "step": 4070 + }, + { + "epoch": 0.31312975924928854, + "learning_rate": 0.0023308861922791365, + "loss": 0.693, + "step": 4071 + }, + { + "epoch": 0.31320667640950695, + "learning_rate": 0.0023305843927554164, + "loss": 1.0655, + "step": 4072 + }, + { + "epoch": 0.3132835935697254, + "learning_rate": 0.0023302825447330078, + "loss": 0.8207, + "step": 4073 + }, + { + "epoch": 0.31336051072994386, + "learning_rate": 0.0023299806482295357, + "loss": 1.6583, + "step": 4074 + }, + { + "epoch": 0.31343742789016227, + "learning_rate": 0.0023296787032626295, + "loss": 1.1852, + "step": 4075 + }, + { + "epoch": 0.3135143450503807, + "learning_rate": 0.0023293767098499186, + "loss": 1.0463, + "step": 4076 + }, + { + "epoch": 0.3135912622105992, + "learning_rate": 0.0023290746680090374, + "loss": 1.3161, + "step": 4077 + }, + { + "epoch": 0.31366817937081765, + "learning_rate": 0.0023287725777576223, + "loss": 1.1284, + "step": 4078 + }, + { + "epoch": 0.31374509653103605, + "learning_rate": 0.002328470439113313, + "loss": 1.3051, + "step": 4079 + }, + { + "epoch": 0.3138220136912545, + "learning_rate": 0.0023281682520937506, + "loss": 1.2625, + "step": 4080 + }, + { + "epoch": 0.31389893085147297, + "learning_rate": 0.0023278660167165813, + "loss": 0.9103, + "step": 4081 + }, + { + "epoch": 0.3139758480116914, + "learning_rate": 0.0023275637329994523, + "loss": 1.3384, + "step": 4082 + }, + { + "epoch": 0.31405276517190983, + "learning_rate": 0.0023272614009600145, + "loss": 1.1082, + "step": 4083 + }, + { + "epoch": 0.3141296823321283, + "learning_rate": 0.0023269590206159213, + "loss": 1.1432, + "step": 4084 + }, + { + "epoch": 0.31420659949234675, + "learning_rate": 0.002326656591984829, + "loss": 1.2934, + "step": 4085 + }, + { + "epoch": 0.3142835166525652, + "learning_rate": 0.0023263541150843974, + "loss": 1.097, + "step": 4086 + }, + { + "epoch": 0.3143604338127836, + "learning_rate": 0.0023260515899322875, + "loss": 0.9245, + "step": 4087 + }, + { + "epoch": 0.31443735097300207, + "learning_rate": 0.0023257490165461644, + "loss": 1.3487, + "step": 4088 + }, + { + "epoch": 0.31451426813322053, + "learning_rate": 0.002325446394943695, + "loss": 1.3961, + "step": 4089 + }, + { + "epoch": 0.314591185293439, + "learning_rate": 0.002325143725142551, + "loss": 0.8543, + "step": 4090 + }, + { + "epoch": 0.3146681024536574, + "learning_rate": 0.002324841007160405, + "loss": 1.5301, + "step": 4091 + }, + { + "epoch": 0.31474501961387585, + "learning_rate": 0.0023245382410149325, + "loss": 1.275, + "step": 4092 + }, + { + "epoch": 0.3148219367740943, + "learning_rate": 0.002324235426723813, + "loss": 1.4179, + "step": 4093 + }, + { + "epoch": 0.3148988539343128, + "learning_rate": 0.002323932564304728, + "loss": 1.6431, + "step": 4094 + }, + { + "epoch": 0.3149757710945312, + "learning_rate": 0.0023236296537753618, + "loss": 1.0404, + "step": 4095 + }, + { + "epoch": 0.31505268825474964, + "learning_rate": 0.002323326695153402, + "loss": 0.6271, + "step": 4096 + }, + { + "epoch": 0.3151296054149681, + "learning_rate": 0.002323023688456538, + "loss": 1.2285, + "step": 4097 + }, + { + "epoch": 0.3152065225751865, + "learning_rate": 0.0023227206337024637, + "loss": 1.1556, + "step": 4098 + }, + { + "epoch": 0.31528343973540496, + "learning_rate": 0.0023224175309088735, + "loss": 0.8867, + "step": 4099 + }, + { + "epoch": 0.3153603568956234, + "learning_rate": 0.002322114380093467, + "loss": 1.1383, + "step": 4100 + }, + { + "epoch": 0.3154372740558419, + "learning_rate": 0.0023218111812739447, + "loss": 1.4028, + "step": 4101 + }, + { + "epoch": 0.3155141912160603, + "learning_rate": 0.0023215079344680116, + "loss": 1.4083, + "step": 4102 + }, + { + "epoch": 0.31559110837627874, + "learning_rate": 0.0023212046396933737, + "loss": 1.3346, + "step": 4103 + }, + { + "epoch": 0.3156680255364972, + "learning_rate": 0.002320901296967741, + "loss": 0.8614, + "step": 4104 + }, + { + "epoch": 0.31574494269671566, + "learning_rate": 0.002320597906308826, + "loss": 1.1192, + "step": 4105 + }, + { + "epoch": 0.31582185985693406, + "learning_rate": 0.0023202944677343444, + "loss": 1.0291, + "step": 4106 + }, + { + "epoch": 0.3158987770171525, + "learning_rate": 0.0023199909812620136, + "loss": 1.2566, + "step": 4107 + }, + { + "epoch": 0.315975694177371, + "learning_rate": 0.002319687446909555, + "loss": 1.0361, + "step": 4108 + }, + { + "epoch": 0.31605261133758944, + "learning_rate": 0.0023193838646946917, + "loss": 1.1147, + "step": 4109 + }, + { + "epoch": 0.31612952849780784, + "learning_rate": 0.0023190802346351514, + "loss": 1.281, + "step": 4110 + }, + { + "epoch": 0.3162064456580263, + "learning_rate": 0.002318776556748662, + "loss": 1.2085, + "step": 4111 + }, + { + "epoch": 0.31628336281824476, + "learning_rate": 0.0023184728310529563, + "loss": 1.219, + "step": 4112 + }, + { + "epoch": 0.3163602799784632, + "learning_rate": 0.002318169057565769, + "loss": 1.1128, + "step": 4113 + }, + { + "epoch": 0.3164371971386816, + "learning_rate": 0.002317865236304838, + "loss": 1.2199, + "step": 4114 + }, + { + "epoch": 0.3165141142989001, + "learning_rate": 0.002317561367287903, + "loss": 0.9051, + "step": 4115 + }, + { + "epoch": 0.31659103145911854, + "learning_rate": 0.0023172574505327076, + "loss": 0.8508, + "step": 4116 + }, + { + "epoch": 0.31666794861933695, + "learning_rate": 0.0023169534860569985, + "loss": 1.1534, + "step": 4117 + }, + { + "epoch": 0.3167448657795554, + "learning_rate": 0.002316649473878524, + "loss": 1.2333, + "step": 4118 + }, + { + "epoch": 0.31682178293977387, + "learning_rate": 0.0023163454140150354, + "loss": 1.0667, + "step": 4119 + }, + { + "epoch": 0.3168987000999923, + "learning_rate": 0.002316041306484287, + "loss": 1.2747, + "step": 4120 + }, + { + "epoch": 0.31697561726021073, + "learning_rate": 0.0023157371513040366, + "loss": 1.0607, + "step": 4121 + }, + { + "epoch": 0.3170525344204292, + "learning_rate": 0.0023154329484920437, + "loss": 1.0513, + "step": 4122 + }, + { + "epoch": 0.31712945158064765, + "learning_rate": 0.002315128698066071, + "loss": 1.1036, + "step": 4123 + }, + { + "epoch": 0.3172063687408661, + "learning_rate": 0.0023148244000438844, + "loss": 1.3047, + "step": 4124 + }, + { + "epoch": 0.3172832859010845, + "learning_rate": 0.0023145200544432516, + "loss": 1.1692, + "step": 4125 + }, + { + "epoch": 0.31736020306130297, + "learning_rate": 0.0023142156612819442, + "loss": 1.297, + "step": 4126 + }, + { + "epoch": 0.31743712022152143, + "learning_rate": 0.0023139112205777354, + "loss": 0.9115, + "step": 4127 + }, + { + "epoch": 0.3175140373817399, + "learning_rate": 0.0023136067323484026, + "loss": 1.3012, + "step": 4128 + }, + { + "epoch": 0.3175909545419583, + "learning_rate": 0.002313302196611725, + "loss": 1.0477, + "step": 4129 + }, + { + "epoch": 0.31766787170217675, + "learning_rate": 0.0023129976133854837, + "loss": 0.9798, + "step": 4130 + }, + { + "epoch": 0.3177447888623952, + "learning_rate": 0.002312692982687465, + "loss": 1.22, + "step": 4131 + }, + { + "epoch": 0.31782170602261367, + "learning_rate": 0.002312388304535456, + "loss": 1.0055, + "step": 4132 + }, + { + "epoch": 0.3178986231828321, + "learning_rate": 0.0023120835789472477, + "loss": 1.3632, + "step": 4133 + }, + { + "epoch": 0.31797554034305053, + "learning_rate": 0.0023117788059406324, + "loss": 1.1977, + "step": 4134 + }, + { + "epoch": 0.318052457503269, + "learning_rate": 0.0023114739855334063, + "loss": 1.1121, + "step": 4135 + }, + { + "epoch": 0.3181293746634874, + "learning_rate": 0.002311169117743369, + "loss": 1.1742, + "step": 4136 + }, + { + "epoch": 0.31820629182370586, + "learning_rate": 0.0023108642025883213, + "loss": 1.1818, + "step": 4137 + }, + { + "epoch": 0.3182832089839243, + "learning_rate": 0.002310559240086068, + "loss": 1.237, + "step": 4138 + }, + { + "epoch": 0.3183601261441428, + "learning_rate": 0.0023102542302544157, + "loss": 1.0121, + "step": 4139 + }, + { + "epoch": 0.3184370433043612, + "learning_rate": 0.002309949173111174, + "loss": 1.2504, + "step": 4140 + }, + { + "epoch": 0.31851396046457964, + "learning_rate": 0.0023096440686741567, + "loss": 0.679, + "step": 4141 + }, + { + "epoch": 0.3185908776247981, + "learning_rate": 0.0023093389169611786, + "loss": 1.0136, + "step": 4142 + }, + { + "epoch": 0.31866779478501656, + "learning_rate": 0.002309033717990057, + "loss": 1.1826, + "step": 4143 + }, + { + "epoch": 0.31874471194523496, + "learning_rate": 0.002308728471778614, + "loss": 1.2552, + "step": 4144 + }, + { + "epoch": 0.3188216291054534, + "learning_rate": 0.0023084231783446723, + "loss": 1.3521, + "step": 4145 + }, + { + "epoch": 0.3188985462656719, + "learning_rate": 0.0023081178377060588, + "loss": 1.0351, + "step": 4146 + }, + { + "epoch": 0.31897546342589034, + "learning_rate": 0.0023078124498806026, + "loss": 1.1174, + "step": 4147 + }, + { + "epoch": 0.31905238058610874, + "learning_rate": 0.0023075070148861356, + "loss": 0.9617, + "step": 4148 + }, + { + "epoch": 0.3191292977463272, + "learning_rate": 0.0023072015327404924, + "loss": 1.0803, + "step": 4149 + }, + { + "epoch": 0.31920621490654566, + "learning_rate": 0.0023068960034615107, + "loss": 1.105, + "step": 4150 + }, + { + "epoch": 0.3192831320667641, + "learning_rate": 0.00230659042706703, + "loss": 0.7961, + "step": 4151 + }, + { + "epoch": 0.3193600492269825, + "learning_rate": 0.0023062848035748937, + "loss": 1.0708, + "step": 4152 + }, + { + "epoch": 0.319436966387201, + "learning_rate": 0.0023059791330029477, + "loss": 1.3105, + "step": 4153 + }, + { + "epoch": 0.31951388354741944, + "learning_rate": 0.0023056734153690398, + "loss": 1.6109, + "step": 4154 + }, + { + "epoch": 0.3195908007076379, + "learning_rate": 0.002305367650691022, + "loss": 1.0377, + "step": 4155 + }, + { + "epoch": 0.3196677178678563, + "learning_rate": 0.0023050618389867474, + "loss": 1.045, + "step": 4156 + }, + { + "epoch": 0.31974463502807476, + "learning_rate": 0.002304755980274073, + "loss": 0.939, + "step": 4157 + }, + { + "epoch": 0.3198215521882932, + "learning_rate": 0.0023044500745708586, + "loss": 0.9277, + "step": 4158 + }, + { + "epoch": 0.3198984693485116, + "learning_rate": 0.0023041441218949656, + "loss": 1.1165, + "step": 4159 + }, + { + "epoch": 0.3199753865087301, + "learning_rate": 0.0023038381222642594, + "loss": 1.1278, + "step": 4160 + }, + { + "epoch": 0.32005230366894855, + "learning_rate": 0.0023035320756966075, + "loss": 1.3572, + "step": 4161 + }, + { + "epoch": 0.320129220829167, + "learning_rate": 0.00230322598220988, + "loss": 1.1209, + "step": 4162 + }, + { + "epoch": 0.3202061379893854, + "learning_rate": 0.0023029198418219507, + "loss": 1.0094, + "step": 4163 + }, + { + "epoch": 0.32028305514960387, + "learning_rate": 0.002302613654550695, + "loss": 1.3123, + "step": 4164 + }, + { + "epoch": 0.3203599723098223, + "learning_rate": 0.0023023074204139913, + "loss": 1.2515, + "step": 4165 + }, + { + "epoch": 0.3204368894700408, + "learning_rate": 0.0023020011394297214, + "loss": 1.5766, + "step": 4166 + }, + { + "epoch": 0.3205138066302592, + "learning_rate": 0.002301694811615769, + "loss": 0.941, + "step": 4167 + }, + { + "epoch": 0.32059072379047765, + "learning_rate": 0.002301388436990021, + "loss": 1.3693, + "step": 4168 + }, + { + "epoch": 0.3206676409506961, + "learning_rate": 0.002301082015570367, + "loss": 1.0084, + "step": 4169 + }, + { + "epoch": 0.32074455811091457, + "learning_rate": 0.002300775547374699, + "loss": 1.3284, + "step": 4170 + }, + { + "epoch": 0.32082147527113297, + "learning_rate": 0.002300469032420913, + "loss": 1.178, + "step": 4171 + }, + { + "epoch": 0.32089839243135143, + "learning_rate": 0.0023001624707269055, + "loss": 1.3273, + "step": 4172 + }, + { + "epoch": 0.3209753095915699, + "learning_rate": 0.002299855862310577, + "loss": 1.3236, + "step": 4173 + }, + { + "epoch": 0.32105222675178835, + "learning_rate": 0.0022995492071898323, + "loss": 1.0124, + "step": 4174 + }, + { + "epoch": 0.32112914391200675, + "learning_rate": 0.002299242505382575, + "loss": 1.075, + "step": 4175 + }, + { + "epoch": 0.3212060610722252, + "learning_rate": 0.002298935756906715, + "loss": 1.2145, + "step": 4176 + }, + { + "epoch": 0.3212829782324437, + "learning_rate": 0.0022986289617801642, + "loss": 1.1242, + "step": 4177 + }, + { + "epoch": 0.3213598953926621, + "learning_rate": 0.002298322120020836, + "loss": 1.0867, + "step": 4178 + }, + { + "epoch": 0.32143681255288054, + "learning_rate": 0.0022980152316466463, + "loss": 0.7148, + "step": 4179 + }, + { + "epoch": 0.321513729713099, + "learning_rate": 0.002297708296675516, + "loss": 1.2394, + "step": 4180 + }, + { + "epoch": 0.32159064687331745, + "learning_rate": 0.002297401315125367, + "loss": 1.2044, + "step": 4181 + }, + { + "epoch": 0.32166756403353586, + "learning_rate": 0.0022970942870141242, + "loss": 0.8243, + "step": 4182 + }, + { + "epoch": 0.3217444811937543, + "learning_rate": 0.002296787212359715, + "loss": 1.3398, + "step": 4183 + }, + { + "epoch": 0.3218213983539728, + "learning_rate": 0.0022964800911800705, + "loss": 1.2309, + "step": 4184 + }, + { + "epoch": 0.32189831551419124, + "learning_rate": 0.002296172923493123, + "loss": 1.1032, + "step": 4185 + }, + { + "epoch": 0.32197523267440964, + "learning_rate": 0.002295865709316809, + "loss": 1.1799, + "step": 4186 + }, + { + "epoch": 0.3220521498346281, + "learning_rate": 0.002295558448669067, + "loss": 1.0306, + "step": 4187 + }, + { + "epoch": 0.32212906699484656, + "learning_rate": 0.0022952511415678376, + "loss": 0.8425, + "step": 4188 + }, + { + "epoch": 0.322205984155065, + "learning_rate": 0.002294943788031065, + "loss": 0.9645, + "step": 4189 + }, + { + "epoch": 0.3222829013152834, + "learning_rate": 0.002294636388076697, + "loss": 1.2359, + "step": 4190 + }, + { + "epoch": 0.3223598184755019, + "learning_rate": 0.0022943289417226816, + "loss": 0.9698, + "step": 4191 + }, + { + "epoch": 0.32243673563572034, + "learning_rate": 0.0022940214489869716, + "loss": 1.5443, + "step": 4192 + }, + { + "epoch": 0.3225136527959388, + "learning_rate": 0.002293713909887522, + "loss": 1.3285, + "step": 4193 + }, + { + "epoch": 0.3225905699561572, + "learning_rate": 0.00229340632444229, + "loss": 1.0976, + "step": 4194 + }, + { + "epoch": 0.32266748711637566, + "learning_rate": 0.002293098692669235, + "loss": 1.0602, + "step": 4195 + }, + { + "epoch": 0.3227444042765941, + "learning_rate": 0.0022927910145863215, + "loss": 0.9043, + "step": 4196 + }, + { + "epoch": 0.3228213214368126, + "learning_rate": 0.002292483290211514, + "loss": 1.242, + "step": 4197 + }, + { + "epoch": 0.322898238597031, + "learning_rate": 0.0022921755195627817, + "loss": 1.4223, + "step": 4198 + }, + { + "epoch": 0.32297515575724944, + "learning_rate": 0.0022918677026580953, + "loss": 1.2745, + "step": 4199 + }, + { + "epoch": 0.3230520729174679, + "learning_rate": 0.002291559839515428, + "loss": 1.0063, + "step": 4200 + }, + { + "epoch": 0.3231289900776863, + "learning_rate": 0.0022912519301527576, + "loss": 1.0889, + "step": 4201 + }, + { + "epoch": 0.32320590723790477, + "learning_rate": 0.002290943974588061, + "loss": 0.9462, + "step": 4202 + }, + { + "epoch": 0.3232828243981232, + "learning_rate": 0.0022906359728393226, + "loss": 1.152, + "step": 4203 + }, + { + "epoch": 0.3233597415583417, + "learning_rate": 0.002290327924924525, + "loss": 1.0191, + "step": 4204 + }, + { + "epoch": 0.3234366587185601, + "learning_rate": 0.0022900198308616567, + "loss": 0.9533, + "step": 4205 + }, + { + "epoch": 0.32351357587877855, + "learning_rate": 0.002289711690668707, + "loss": 1.4973, + "step": 4206 + }, + { + "epoch": 0.323590493038997, + "learning_rate": 0.002289403504363669, + "loss": 1.248, + "step": 4207 + }, + { + "epoch": 0.32366741019921547, + "learning_rate": 0.002289095271964537, + "loss": 1.2784, + "step": 4208 + }, + { + "epoch": 0.32374432735943387, + "learning_rate": 0.00228878699348931, + "loss": 1.163, + "step": 4209 + }, + { + "epoch": 0.32382124451965233, + "learning_rate": 0.0022884786689559887, + "loss": 1.165, + "step": 4210 + }, + { + "epoch": 0.3238981616798708, + "learning_rate": 0.0022881702983825756, + "loss": 1.1404, + "step": 4211 + }, + { + "epoch": 0.32397507884008925, + "learning_rate": 0.002287861881787078, + "loss": 1.1549, + "step": 4212 + }, + { + "epoch": 0.32405199600030765, + "learning_rate": 0.002287553419187503, + "loss": 1.3102, + "step": 4213 + }, + { + "epoch": 0.3241289131605261, + "learning_rate": 0.0022872449106018642, + "loss": 1.3165, + "step": 4214 + }, + { + "epoch": 0.32420583032074457, + "learning_rate": 0.0022869363560481743, + "loss": 0.9043, + "step": 4215 + }, + { + "epoch": 0.32428274748096303, + "learning_rate": 0.00228662775554445, + "loss": 1.0413, + "step": 4216 + }, + { + "epoch": 0.32435966464118143, + "learning_rate": 0.0022863191091087123, + "loss": 1.1427, + "step": 4217 + }, + { + "epoch": 0.3244365818013999, + "learning_rate": 0.0022860104167589813, + "loss": 1.1331, + "step": 4218 + }, + { + "epoch": 0.32451349896161835, + "learning_rate": 0.002285701678513283, + "loss": 1.4227, + "step": 4219 + }, + { + "epoch": 0.32459041612183676, + "learning_rate": 0.0022853928943896455, + "loss": 1.3405, + "step": 4220 + }, + { + "epoch": 0.3246673332820552, + "learning_rate": 0.0022850840644060982, + "loss": 1.2141, + "step": 4221 + }, + { + "epoch": 0.3247442504422737, + "learning_rate": 0.002284775188580674, + "loss": 0.99, + "step": 4222 + }, + { + "epoch": 0.32482116760249213, + "learning_rate": 0.0022844662669314086, + "loss": 1.253, + "step": 4223 + }, + { + "epoch": 0.32489808476271054, + "learning_rate": 0.00228415729947634, + "loss": 1.3627, + "step": 4224 + }, + { + "epoch": 0.324975001922929, + "learning_rate": 0.0022838482862335098, + "loss": 1.1932, + "step": 4225 + }, + { + "epoch": 0.32505191908314746, + "learning_rate": 0.002283539227220961, + "loss": 1.1084, + "step": 4226 + }, + { + "epoch": 0.3251288362433659, + "learning_rate": 0.0022832301224567407, + "loss": 0.7417, + "step": 4227 + }, + { + "epoch": 0.3252057534035843, + "learning_rate": 0.0022829209719588974, + "loss": 1.1925, + "step": 4228 + }, + { + "epoch": 0.3252826705638028, + "learning_rate": 0.0022826117757454817, + "loss": 1.0792, + "step": 4229 + }, + { + "epoch": 0.32535958772402124, + "learning_rate": 0.0022823025338345494, + "loss": 0.8884, + "step": 4230 + }, + { + "epoch": 0.3254365048842397, + "learning_rate": 0.0022819932462441565, + "loss": 1.1749, + "step": 4231 + }, + { + "epoch": 0.3255134220444581, + "learning_rate": 0.002281683912992364, + "loss": 0.7102, + "step": 4232 + }, + { + "epoch": 0.32559033920467656, + "learning_rate": 0.0022813745340972322, + "loss": 1.4558, + "step": 4233 + }, + { + "epoch": 0.325667256364895, + "learning_rate": 0.002281065109576827, + "loss": 0.8879, + "step": 4234 + }, + { + "epoch": 0.3257441735251135, + "learning_rate": 0.002280755639449216, + "loss": 0.9001, + "step": 4235 + }, + { + "epoch": 0.3258210906853319, + "learning_rate": 0.0022804461237324702, + "loss": 1.4333, + "step": 4236 + }, + { + "epoch": 0.32589800784555034, + "learning_rate": 0.0022801365624446615, + "loss": 1.1534, + "step": 4237 + }, + { + "epoch": 0.3259749250057688, + "learning_rate": 0.0022798269556038656, + "loss": 0.814, + "step": 4238 + }, + { + "epoch": 0.3260518421659872, + "learning_rate": 0.0022795173032281618, + "loss": 1.3223, + "step": 4239 + }, + { + "epoch": 0.32612875932620566, + "learning_rate": 0.0022792076053356294, + "loss": 0.7201, + "step": 4240 + }, + { + "epoch": 0.3262056764864241, + "learning_rate": 0.002278897861944353, + "loss": 0.9512, + "step": 4241 + }, + { + "epoch": 0.3262825936466426, + "learning_rate": 0.002278588073072419, + "loss": 1.002, + "step": 4242 + }, + { + "epoch": 0.326359510806861, + "learning_rate": 0.0022782782387379165, + "loss": 1.1497, + "step": 4243 + }, + { + "epoch": 0.32643642796707945, + "learning_rate": 0.002277968358958936, + "loss": 0.7548, + "step": 4244 + }, + { + "epoch": 0.3265133451272979, + "learning_rate": 0.0022776584337535714, + "loss": 0.9711, + "step": 4245 + }, + { + "epoch": 0.32659026228751636, + "learning_rate": 0.0022773484631399216, + "loss": 1.067, + "step": 4246 + }, + { + "epoch": 0.32666717944773477, + "learning_rate": 0.0022770384471360846, + "loss": 0.9141, + "step": 4247 + }, + { + "epoch": 0.3267440966079532, + "learning_rate": 0.0022767283857601625, + "loss": 1.3968, + "step": 4248 + }, + { + "epoch": 0.3268210137681717, + "learning_rate": 0.002276418279030261, + "loss": 0.8375, + "step": 4249 + }, + { + "epoch": 0.32689793092839015, + "learning_rate": 0.0022761081269644864, + "loss": 1.1958, + "step": 4250 + }, + { + "epoch": 0.32697484808860855, + "learning_rate": 0.0022757979295809495, + "loss": 1.0153, + "step": 4251 + }, + { + "epoch": 0.327051765248827, + "learning_rate": 0.0022754876868977636, + "loss": 1.0296, + "step": 4252 + }, + { + "epoch": 0.32712868240904547, + "learning_rate": 0.002275177398933043, + "loss": 1.0223, + "step": 4253 + }, + { + "epoch": 0.3272055995692639, + "learning_rate": 0.0022748670657049062, + "loss": 1.2272, + "step": 4254 + }, + { + "epoch": 0.32728251672948233, + "learning_rate": 0.002274556687231474, + "loss": 1.0495, + "step": 4255 + }, + { + "epoch": 0.3273594338897008, + "learning_rate": 0.00227424626353087, + "loss": 1.1607, + "step": 4256 + }, + { + "epoch": 0.32743635104991925, + "learning_rate": 0.002273935794621219, + "loss": 1.1606, + "step": 4257 + }, + { + "epoch": 0.3275132682101377, + "learning_rate": 0.002273625280520651, + "loss": 1.4155, + "step": 4258 + }, + { + "epoch": 0.3275901853703561, + "learning_rate": 0.0022733147212472965, + "loss": 1.1348, + "step": 4259 + }, + { + "epoch": 0.32766710253057457, + "learning_rate": 0.0022730041168192894, + "loss": 1.1552, + "step": 4260 + }, + { + "epoch": 0.32774401969079303, + "learning_rate": 0.0022726934672547662, + "loss": 1.3827, + "step": 4261 + }, + { + "epoch": 0.32782093685101144, + "learning_rate": 0.002272382772571866, + "loss": 1.7051, + "step": 4262 + }, + { + "epoch": 0.3278978540112299, + "learning_rate": 0.0022720720327887313, + "loss": 1.2682, + "step": 4263 + }, + { + "epoch": 0.32797477117144835, + "learning_rate": 0.0022717612479235053, + "loss": 1.2746, + "step": 4264 + }, + { + "epoch": 0.3280516883316668, + "learning_rate": 0.0022714504179943366, + "loss": 1.46, + "step": 4265 + }, + { + "epoch": 0.3281286054918852, + "learning_rate": 0.002271139543019373, + "loss": 1.0588, + "step": 4266 + }, + { + "epoch": 0.3282055226521037, + "learning_rate": 0.0022708286230167687, + "loss": 1.1343, + "step": 4267 + }, + { + "epoch": 0.32828243981232214, + "learning_rate": 0.0022705176580046773, + "loss": 1.0675, + "step": 4268 + }, + { + "epoch": 0.3283593569725406, + "learning_rate": 0.002270206648001257, + "loss": 1.0478, + "step": 4269 + }, + { + "epoch": 0.328436274132759, + "learning_rate": 0.002269895593024668, + "loss": 1.1152, + "step": 4270 + }, + { + "epoch": 0.32851319129297746, + "learning_rate": 0.0022695844930930727, + "loss": 1.1935, + "step": 4271 + }, + { + "epoch": 0.3285901084531959, + "learning_rate": 0.0022692733482246367, + "loss": 1.316, + "step": 4272 + }, + { + "epoch": 0.3286670256134144, + "learning_rate": 0.0022689621584375285, + "loss": 1.1137, + "step": 4273 + }, + { + "epoch": 0.3287439427736328, + "learning_rate": 0.0022686509237499182, + "loss": 0.7445, + "step": 4274 + }, + { + "epoch": 0.32882085993385124, + "learning_rate": 0.0022683396441799787, + "loss": 1.191, + "step": 4275 + }, + { + "epoch": 0.3288977770940697, + "learning_rate": 0.002268028319745888, + "loss": 1.276, + "step": 4276 + }, + { + "epoch": 0.32897469425428816, + "learning_rate": 0.0022677169504658223, + "loss": 0.9966, + "step": 4277 + }, + { + "epoch": 0.32905161141450656, + "learning_rate": 0.0022674055363579647, + "loss": 1.2092, + "step": 4278 + }, + { + "epoch": 0.329128528574725, + "learning_rate": 0.0022670940774404975, + "loss": 1.024, + "step": 4279 + }, + { + "epoch": 0.3292054457349435, + "learning_rate": 0.002266782573731607, + "loss": 0.9996, + "step": 4280 + }, + { + "epoch": 0.3292823628951619, + "learning_rate": 0.002266471025249484, + "loss": 1.3885, + "step": 4281 + }, + { + "epoch": 0.32935928005538034, + "learning_rate": 0.0022661594320123185, + "loss": 0.7952, + "step": 4282 + }, + { + "epoch": 0.3294361972155988, + "learning_rate": 0.0022658477940383057, + "loss": 1.1963, + "step": 4283 + }, + { + "epoch": 0.32951311437581726, + "learning_rate": 0.0022655361113456418, + "loss": 1.2864, + "step": 4284 + }, + { + "epoch": 0.32959003153603567, + "learning_rate": 0.0022652243839525266, + "loss": 1.033, + "step": 4285 + }, + { + "epoch": 0.3296669486962541, + "learning_rate": 0.002264912611877162, + "loss": 1.2902, + "step": 4286 + }, + { + "epoch": 0.3297438658564726, + "learning_rate": 0.002264600795137753, + "loss": 1.088, + "step": 4287 + }, + { + "epoch": 0.32982078301669104, + "learning_rate": 0.002264288933752506, + "loss": 1.2543, + "step": 4288 + }, + { + "epoch": 0.32989770017690945, + "learning_rate": 0.0022639770277396326, + "loss": 1.0899, + "step": 4289 + }, + { + "epoch": 0.3299746173371279, + "learning_rate": 0.002263665077117344, + "loss": 1.0596, + "step": 4290 + }, + { + "epoch": 0.33005153449734637, + "learning_rate": 0.0022633530819038557, + "loss": 1.5231, + "step": 4291 + }, + { + "epoch": 0.3301284516575648, + "learning_rate": 0.0022630410421173855, + "loss": 0.9046, + "step": 4292 + }, + { + "epoch": 0.33020536881778323, + "learning_rate": 0.002262728957776153, + "loss": 1.118, + "step": 4293 + }, + { + "epoch": 0.3302822859780017, + "learning_rate": 0.0022624168288983824, + "loss": 1.0512, + "step": 4294 + }, + { + "epoch": 0.33035920313822015, + "learning_rate": 0.0022621046555022983, + "loss": 0.9748, + "step": 4295 + }, + { + "epoch": 0.3304361202984386, + "learning_rate": 0.002261792437606129, + "loss": 1.0904, + "step": 4296 + }, + { + "epoch": 0.330513037458657, + "learning_rate": 0.002261480175228106, + "loss": 1.3217, + "step": 4297 + }, + { + "epoch": 0.33058995461887547, + "learning_rate": 0.0022611678683864614, + "loss": 1.1509, + "step": 4298 + }, + { + "epoch": 0.33066687177909393, + "learning_rate": 0.002260855517099432, + "loss": 1.2228, + "step": 4299 + }, + { + "epoch": 0.33074378893931233, + "learning_rate": 0.0022605431213852563, + "loss": 1.2046, + "step": 4300 + }, + { + "epoch": 0.3308207060995308, + "learning_rate": 0.0022602306812621745, + "loss": 1.298, + "step": 4301 + }, + { + "epoch": 0.33089762325974925, + "learning_rate": 0.002259918196748431, + "loss": 1.1577, + "step": 4302 + }, + { + "epoch": 0.3309745404199677, + "learning_rate": 0.002259605667862272, + "loss": 1.0975, + "step": 4303 + }, + { + "epoch": 0.3310514575801861, + "learning_rate": 0.0022592930946219465, + "loss": 1.2118, + "step": 4304 + }, + { + "epoch": 0.3311283747404046, + "learning_rate": 0.0022589804770457057, + "loss": 0.7915, + "step": 4305 + }, + { + "epoch": 0.33120529190062303, + "learning_rate": 0.0022586678151518044, + "loss": 1.1003, + "step": 4306 + }, + { + "epoch": 0.3312822090608415, + "learning_rate": 0.0022583551089584986, + "loss": 1.1668, + "step": 4307 + }, + { + "epoch": 0.3313591262210599, + "learning_rate": 0.0022580423584840476, + "loss": 0.9239, + "step": 4308 + }, + { + "epoch": 0.33143604338127836, + "learning_rate": 0.002257729563746713, + "loss": 1.1236, + "step": 4309 + }, + { + "epoch": 0.3315129605414968, + "learning_rate": 0.0022574167247647594, + "loss": 1.1646, + "step": 4310 + }, + { + "epoch": 0.3315898777017153, + "learning_rate": 0.002257103841556454, + "loss": 1.3413, + "step": 4311 + }, + { + "epoch": 0.3316667948619337, + "learning_rate": 0.0022567909141400664, + "loss": 1.2623, + "step": 4312 + }, + { + "epoch": 0.33174371202215214, + "learning_rate": 0.002256477942533869, + "loss": 1.1116, + "step": 4313 + }, + { + "epoch": 0.3318206291823706, + "learning_rate": 0.0022561649267561355, + "loss": 1.0319, + "step": 4314 + }, + { + "epoch": 0.33189754634258906, + "learning_rate": 0.0022558518668251443, + "loss": 0.8188, + "step": 4315 + }, + { + "epoch": 0.33197446350280746, + "learning_rate": 0.0022555387627591745, + "loss": 1.2136, + "step": 4316 + }, + { + "epoch": 0.3320513806630259, + "learning_rate": 0.0022552256145765086, + "loss": 1.2836, + "step": 4317 + }, + { + "epoch": 0.3321282978232444, + "learning_rate": 0.002254912422295433, + "loss": 1.3871, + "step": 4318 + }, + { + "epoch": 0.33220521498346284, + "learning_rate": 0.002254599185934233, + "loss": 1.4418, + "step": 4319 + }, + { + "epoch": 0.33228213214368124, + "learning_rate": 0.0022542859055112016, + "loss": 1.2481, + "step": 4320 + }, + { + "epoch": 0.3323590493038997, + "learning_rate": 0.0022539725810446294, + "loss": 1.0429, + "step": 4321 + }, + { + "epoch": 0.33243596646411816, + "learning_rate": 0.0022536592125528125, + "loss": 1.0748, + "step": 4322 + }, + { + "epoch": 0.33251288362433656, + "learning_rate": 0.0022533458000540485, + "loss": 1.3121, + "step": 4323 + }, + { + "epoch": 0.332589800784555, + "learning_rate": 0.0022530323435666385, + "loss": 1.3557, + "step": 4324 + }, + { + "epoch": 0.3326667179447735, + "learning_rate": 0.0022527188431088847, + "loss": 1.1272, + "step": 4325 + }, + { + "epoch": 0.33274363510499194, + "learning_rate": 0.002252405298699093, + "loss": 1.029, + "step": 4326 + }, + { + "epoch": 0.33282055226521035, + "learning_rate": 0.0022520917103555724, + "loss": 1.1077, + "step": 4327 + }, + { + "epoch": 0.3328974694254288, + "learning_rate": 0.0022517780780966327, + "loss": 1.3373, + "step": 4328 + }, + { + "epoch": 0.33297438658564726, + "learning_rate": 0.0022514644019405875, + "loss": 1.39, + "step": 4329 + }, + { + "epoch": 0.3330513037458657, + "learning_rate": 0.0022511506819057527, + "loss": 1.1247, + "step": 4330 + }, + { + "epoch": 0.3331282209060841, + "learning_rate": 0.0022508369180104476, + "loss": 1.3734, + "step": 4331 + }, + { + "epoch": 0.3332051380663026, + "learning_rate": 0.002250523110272991, + "loss": 1.5231, + "step": 4332 + }, + { + "epoch": 0.33328205522652105, + "learning_rate": 0.002250209258711709, + "loss": 1.0011, + "step": 4333 + }, + { + "epoch": 0.3333589723867395, + "learning_rate": 0.0022498953633449258, + "loss": 1.0763, + "step": 4334 + }, + { + "epoch": 0.3334358895469579, + "learning_rate": 0.0022495814241909715, + "loss": 1.0338, + "step": 4335 + }, + { + "epoch": 0.33351280670717637, + "learning_rate": 0.0022492674412681758, + "loss": 1.3041, + "step": 4336 + }, + { + "epoch": 0.3335897238673948, + "learning_rate": 0.002248953414594874, + "loss": 1.0705, + "step": 4337 + }, + { + "epoch": 0.3336666410276133, + "learning_rate": 0.002248639344189401, + "loss": 1.1476, + "step": 4338 + }, + { + "epoch": 0.3337435581878317, + "learning_rate": 0.002248325230070097, + "loss": 0.6311, + "step": 4339 + }, + { + "epoch": 0.33382047534805015, + "learning_rate": 0.002248011072255303, + "loss": 1.1914, + "step": 4340 + }, + { + "epoch": 0.3338973925082686, + "learning_rate": 0.002247696870763363, + "loss": 1.354, + "step": 4341 + }, + { + "epoch": 0.333974309668487, + "learning_rate": 0.0022473826256126234, + "loss": 1.2951, + "step": 4342 + }, + { + "epoch": 0.33405122682870547, + "learning_rate": 0.0022470683368214332, + "loss": 0.8793, + "step": 4343 + }, + { + "epoch": 0.33412814398892393, + "learning_rate": 0.0022467540044081445, + "loss": 0.9412, + "step": 4344 + }, + { + "epoch": 0.3342050611491424, + "learning_rate": 0.0022464396283911107, + "loss": 1.0309, + "step": 4345 + }, + { + "epoch": 0.3342819783093608, + "learning_rate": 0.0022461252087886898, + "loss": 1.2234, + "step": 4346 + }, + { + "epoch": 0.33435889546957925, + "learning_rate": 0.00224581074561924, + "loss": 1.4772, + "step": 4347 + }, + { + "epoch": 0.3344358126297977, + "learning_rate": 0.0022454962389011237, + "loss": 1.2051, + "step": 4348 + }, + { + "epoch": 0.3345127297900162, + "learning_rate": 0.0022451816886527045, + "loss": 1.0304, + "step": 4349 + }, + { + "epoch": 0.3345896469502346, + "learning_rate": 0.0022448670948923503, + "loss": 1.6065, + "step": 4350 + }, + { + "epoch": 0.33466656411045304, + "learning_rate": 0.00224455245763843, + "loss": 0.9958, + "step": 4351 + }, + { + "epoch": 0.3347434812706715, + "learning_rate": 0.0022442377769093156, + "loss": 1.2912, + "step": 4352 + }, + { + "epoch": 0.33482039843088995, + "learning_rate": 0.002243923052723382, + "loss": 1.1904, + "step": 4353 + }, + { + "epoch": 0.33489731559110836, + "learning_rate": 0.0022436082850990067, + "loss": 1.496, + "step": 4354 + }, + { + "epoch": 0.3349742327513268, + "learning_rate": 0.002243293474054568, + "loss": 1.3317, + "step": 4355 + }, + { + "epoch": 0.3350511499115453, + "learning_rate": 0.002242978619608449, + "loss": 1.3624, + "step": 4356 + }, + { + "epoch": 0.33512806707176374, + "learning_rate": 0.0022426637217790335, + "loss": 1.3946, + "step": 4357 + }, + { + "epoch": 0.33520498423198214, + "learning_rate": 0.0022423487805847096, + "loss": 0.9476, + "step": 4358 + }, + { + "epoch": 0.3352819013922006, + "learning_rate": 0.0022420337960438667, + "loss": 1.0602, + "step": 4359 + }, + { + "epoch": 0.33535881855241906, + "learning_rate": 0.002241718768174897, + "loss": 1.3431, + "step": 4360 + }, + { + "epoch": 0.33543573571263746, + "learning_rate": 0.0022414036969961963, + "loss": 1.1943, + "step": 4361 + }, + { + "epoch": 0.3355126528728559, + "learning_rate": 0.00224108858252616, + "loss": 0.9365, + "step": 4362 + }, + { + "epoch": 0.3355895700330744, + "learning_rate": 0.0022407734247831897, + "loss": 1.1168, + "step": 4363 + }, + { + "epoch": 0.33566648719329284, + "learning_rate": 0.0022404582237856865, + "loss": 1.6628, + "step": 4364 + }, + { + "epoch": 0.33574340435351124, + "learning_rate": 0.0022401429795520563, + "loss": 1.1019, + "step": 4365 + }, + { + "epoch": 0.3358203215137297, + "learning_rate": 0.0022398276921007064, + "loss": 0.9463, + "step": 4366 + }, + { + "epoch": 0.33589723867394816, + "learning_rate": 0.002239512361450046, + "loss": 1.6078, + "step": 4367 + }, + { + "epoch": 0.3359741558341666, + "learning_rate": 0.002239196987618489, + "loss": 1.3525, + "step": 4368 + }, + { + "epoch": 0.336051072994385, + "learning_rate": 0.0022388815706244486, + "loss": 1.2302, + "step": 4369 + }, + { + "epoch": 0.3361279901546035, + "learning_rate": 0.002238566110486343, + "loss": 1.0367, + "step": 4370 + }, + { + "epoch": 0.33620490731482194, + "learning_rate": 0.0022382506072225936, + "loss": 1.1865, + "step": 4371 + }, + { + "epoch": 0.3362818244750404, + "learning_rate": 0.002237935060851621, + "loss": 1.2263, + "step": 4372 + }, + { + "epoch": 0.3363587416352588, + "learning_rate": 0.0022376194713918515, + "loss": 1.2354, + "step": 4373 + }, + { + "epoch": 0.33643565879547727, + "learning_rate": 0.0022373038388617122, + "loss": 0.9629, + "step": 4374 + }, + { + "epoch": 0.3365125759556957, + "learning_rate": 0.002236988163279634, + "loss": 1.2289, + "step": 4375 + }, + { + "epoch": 0.3365894931159142, + "learning_rate": 0.0022366724446640476, + "loss": 1.3058, + "step": 4376 + }, + { + "epoch": 0.3366664102761326, + "learning_rate": 0.0022363566830333904, + "loss": 1.0201, + "step": 4377 + }, + { + "epoch": 0.33674332743635105, + "learning_rate": 0.0022360408784060986, + "loss": 0.8795, + "step": 4378 + }, + { + "epoch": 0.3368202445965695, + "learning_rate": 0.002235725030800613, + "loss": 1.0216, + "step": 4379 + }, + { + "epoch": 0.33689716175678797, + "learning_rate": 0.002235409140235376, + "loss": 0.9301, + "step": 4380 + }, + { + "epoch": 0.33697407891700637, + "learning_rate": 0.002235093206728833, + "loss": 1.1659, + "step": 4381 + }, + { + "epoch": 0.33705099607722483, + "learning_rate": 0.0022347772302994317, + "loss": 1.1632, + "step": 4382 + }, + { + "epoch": 0.3371279132374433, + "learning_rate": 0.002234461210965622, + "loss": 1.0076, + "step": 4383 + }, + { + "epoch": 0.3372048303976617, + "learning_rate": 0.002234145148745857, + "loss": 1.1557, + "step": 4384 + }, + { + "epoch": 0.33728174755788015, + "learning_rate": 0.002233829043658591, + "loss": 1.2187, + "step": 4385 + }, + { + "epoch": 0.3373586647180986, + "learning_rate": 0.0022335128957222832, + "loss": 1.0021, + "step": 4386 + }, + { + "epoch": 0.33743558187831707, + "learning_rate": 0.002233196704955392, + "loss": 0.8605, + "step": 4387 + }, + { + "epoch": 0.3375124990385355, + "learning_rate": 0.002232880471376382, + "loss": 0.8381, + "step": 4388 + }, + { + "epoch": 0.33758941619875393, + "learning_rate": 0.002232564195003717, + "loss": 1.1981, + "step": 4389 + }, + { + "epoch": 0.3376663333589724, + "learning_rate": 0.0022322478758558654, + "loss": 0.9065, + "step": 4390 + }, + { + "epoch": 0.33774325051919085, + "learning_rate": 0.002231931513951297, + "loss": 0.9244, + "step": 4391 + }, + { + "epoch": 0.33782016767940926, + "learning_rate": 0.0022316151093084847, + "loss": 1.1886, + "step": 4392 + }, + { + "epoch": 0.3378970848396277, + "learning_rate": 0.002231298661945904, + "loss": 1.2158, + "step": 4393 + }, + { + "epoch": 0.3379740019998462, + "learning_rate": 0.002230982171882032, + "loss": 1.2558, + "step": 4394 + }, + { + "epoch": 0.33805091916006463, + "learning_rate": 0.0022306656391353494, + "loss": 1.4155, + "step": 4395 + }, + { + "epoch": 0.33812783632028304, + "learning_rate": 0.0022303490637243383, + "loss": 1.0438, + "step": 4396 + }, + { + "epoch": 0.3382047534805015, + "learning_rate": 0.0022300324456674845, + "loss": 1.3169, + "step": 4397 + }, + { + "epoch": 0.33828167064071996, + "learning_rate": 0.002229715784983275, + "loss": 0.7401, + "step": 4398 + }, + { + "epoch": 0.3383585878009384, + "learning_rate": 0.0022293990816902006, + "loss": 1.0843, + "step": 4399 + }, + { + "epoch": 0.3384355049611568, + "learning_rate": 0.0022290823358067542, + "loss": 1.296, + "step": 4400 + }, + { + "epoch": 0.3385124221213753, + "learning_rate": 0.0022287655473514295, + "loss": 1.1755, + "step": 4401 + }, + { + "epoch": 0.33858933928159374, + "learning_rate": 0.0022284487163427256, + "loss": 1.2204, + "step": 4402 + }, + { + "epoch": 0.33866625644181214, + "learning_rate": 0.0022281318427991416, + "loss": 1.1812, + "step": 4403 + }, + { + "epoch": 0.3387431736020306, + "learning_rate": 0.0022278149267391805, + "loss": 1.5273, + "step": 4404 + }, + { + "epoch": 0.33882009076224906, + "learning_rate": 0.0022274979681813473, + "loss": 1.298, + "step": 4405 + }, + { + "epoch": 0.3388970079224675, + "learning_rate": 0.00222718096714415, + "loss": 1.2752, + "step": 4406 + }, + { + "epoch": 0.3389739250826859, + "learning_rate": 0.0022268639236460975, + "loss": 1.1704, + "step": 4407 + }, + { + "epoch": 0.3390508422429044, + "learning_rate": 0.0022265468377057037, + "loss": 0.9175, + "step": 4408 + }, + { + "epoch": 0.33912775940312284, + "learning_rate": 0.0022262297093414824, + "loss": 1.1228, + "step": 4409 + }, + { + "epoch": 0.3392046765633413, + "learning_rate": 0.0022259125385719514, + "loss": 1.1993, + "step": 4410 + }, + { + "epoch": 0.3392815937235597, + "learning_rate": 0.002225595325415631, + "loss": 0.747, + "step": 4411 + }, + { + "epoch": 0.33935851088377816, + "learning_rate": 0.0022252780698910433, + "loss": 1.2871, + "step": 4412 + }, + { + "epoch": 0.3394354280439966, + "learning_rate": 0.002224960772016713, + "loss": 1.3381, + "step": 4413 + }, + { + "epoch": 0.3395123452042151, + "learning_rate": 0.0022246434318111685, + "loss": 1.0126, + "step": 4414 + }, + { + "epoch": 0.3395892623644335, + "learning_rate": 0.0022243260492929378, + "loss": 1.4034, + "step": 4415 + }, + { + "epoch": 0.33966617952465195, + "learning_rate": 0.0022240086244805545, + "loss": 1.2425, + "step": 4416 + }, + { + "epoch": 0.3397430966848704, + "learning_rate": 0.0022236911573925533, + "loss": 1.7546, + "step": 4417 + }, + { + "epoch": 0.33982001384508886, + "learning_rate": 0.002223373648047471, + "loss": 1.0383, + "step": 4418 + }, + { + "epoch": 0.33989693100530727, + "learning_rate": 0.002223056096463848, + "loss": 1.3741, + "step": 4419 + }, + { + "epoch": 0.3399738481655257, + "learning_rate": 0.0022227385026602252, + "loss": 0.7326, + "step": 4420 + }, + { + "epoch": 0.3400507653257442, + "learning_rate": 0.0022224208666551484, + "loss": 1.4531, + "step": 4421 + }, + { + "epoch": 0.34012768248596265, + "learning_rate": 0.0022221031884671646, + "loss": 1.0761, + "step": 4422 + }, + { + "epoch": 0.34020459964618105, + "learning_rate": 0.002221785468114823, + "loss": 1.086, + "step": 4423 + }, + { + "epoch": 0.3402815168063995, + "learning_rate": 0.002221467705616675, + "loss": 1.0769, + "step": 4424 + }, + { + "epoch": 0.34035843396661797, + "learning_rate": 0.0022211499009912763, + "loss": 1.2885, + "step": 4425 + }, + { + "epoch": 0.34043535112683637, + "learning_rate": 0.0022208320542571837, + "loss": 1.1621, + "step": 4426 + }, + { + "epoch": 0.34051226828705483, + "learning_rate": 0.0022205141654329554, + "loss": 1.306, + "step": 4427 + }, + { + "epoch": 0.3405891854472733, + "learning_rate": 0.0022201962345371543, + "loss": 1.0948, + "step": 4428 + }, + { + "epoch": 0.34066610260749175, + "learning_rate": 0.0022198782615883445, + "loss": 1.179, + "step": 4429 + }, + { + "epoch": 0.34074301976771015, + "learning_rate": 0.0022195602466050933, + "loss": 1.0467, + "step": 4430 + }, + { + "epoch": 0.3408199369279286, + "learning_rate": 0.002219242189605969, + "loss": 1.0063, + "step": 4431 + }, + { + "epoch": 0.34089685408814707, + "learning_rate": 0.002218924090609544, + "loss": 0.9891, + "step": 4432 + }, + { + "epoch": 0.34097377124836553, + "learning_rate": 0.0022186059496343913, + "loss": 1.3963, + "step": 4433 + }, + { + "epoch": 0.34105068840858394, + "learning_rate": 0.0022182877666990885, + "loss": 1.4076, + "step": 4434 + }, + { + "epoch": 0.3411276055688024, + "learning_rate": 0.002217969541822215, + "loss": 1.2832, + "step": 4435 + }, + { + "epoch": 0.34120452272902085, + "learning_rate": 0.002217651275022352, + "loss": 1.2185, + "step": 4436 + }, + { + "epoch": 0.3412814398892393, + "learning_rate": 0.002217332966318082, + "loss": 1.4446, + "step": 4437 + }, + { + "epoch": 0.3413583570494577, + "learning_rate": 0.0022170146157279927, + "loss": 1.1896, + "step": 4438 + }, + { + "epoch": 0.3414352742096762, + "learning_rate": 0.0022166962232706734, + "loss": 1.1249, + "step": 4439 + }, + { + "epoch": 0.34151219136989464, + "learning_rate": 0.002216377788964714, + "loss": 1.2731, + "step": 4440 + }, + { + "epoch": 0.3415891085301131, + "learning_rate": 0.00221605931282871, + "loss": 1.2442, + "step": 4441 + }, + { + "epoch": 0.3416660256903315, + "learning_rate": 0.0022157407948812558, + "loss": 1.1938, + "step": 4442 + }, + { + "epoch": 0.34174294285054996, + "learning_rate": 0.0022154222351409504, + "loss": 0.8131, + "step": 4443 + }, + { + "epoch": 0.3418198600107684, + "learning_rate": 0.002215103633626395, + "loss": 1.3119, + "step": 4444 + }, + { + "epoch": 0.3418967771709868, + "learning_rate": 0.0022147849903561933, + "loss": 1.3002, + "step": 4445 + }, + { + "epoch": 0.3419736943312053, + "learning_rate": 0.002214466305348952, + "loss": 1.6118, + "step": 4446 + }, + { + "epoch": 0.34205061149142374, + "learning_rate": 0.0022141475786232774, + "loss": 1.2568, + "step": 4447 + }, + { + "epoch": 0.3421275286516422, + "learning_rate": 0.0022138288101977814, + "loss": 1.3951, + "step": 4448 + }, + { + "epoch": 0.3422044458118606, + "learning_rate": 0.002213510000091078, + "loss": 1.1839, + "step": 4449 + }, + { + "epoch": 0.34228136297207906, + "learning_rate": 0.0022131911483217817, + "loss": 0.766, + "step": 4450 + }, + { + "epoch": 0.3423582801322975, + "learning_rate": 0.002212872254908511, + "loss": 1.0955, + "step": 4451 + }, + { + "epoch": 0.342435197292516, + "learning_rate": 0.002212553319869886, + "loss": 1.0926, + "step": 4452 + }, + { + "epoch": 0.3425121144527344, + "learning_rate": 0.002212234343224531, + "loss": 1.3951, + "step": 4453 + }, + { + "epoch": 0.34258903161295284, + "learning_rate": 0.0022119153249910697, + "loss": 1.2506, + "step": 4454 + }, + { + "epoch": 0.3426659487731713, + "learning_rate": 0.0022115962651881307, + "loss": 1.0962, + "step": 4455 + }, + { + "epoch": 0.34274286593338976, + "learning_rate": 0.0022112771638343443, + "loss": 0.963, + "step": 4456 + }, + { + "epoch": 0.34281978309360817, + "learning_rate": 0.002210958020948343, + "loss": 1.0749, + "step": 4457 + }, + { + "epoch": 0.3428967002538266, + "learning_rate": 0.002210638836548762, + "loss": 0.9101, + "step": 4458 + }, + { + "epoch": 0.3429736174140451, + "learning_rate": 0.0022103196106542386, + "loss": 0.7482, + "step": 4459 + }, + { + "epoch": 0.34305053457426354, + "learning_rate": 0.0022100003432834127, + "loss": 1.1059, + "step": 4460 + }, + { + "epoch": 0.34312745173448195, + "learning_rate": 0.0022096810344549274, + "loss": 1.025, + "step": 4461 + }, + { + "epoch": 0.3432043688947004, + "learning_rate": 0.0022093616841874263, + "loss": 1.4481, + "step": 4462 + }, + { + "epoch": 0.34328128605491887, + "learning_rate": 0.0022090422924995576, + "loss": 1.3447, + "step": 4463 + }, + { + "epoch": 0.34335820321513727, + "learning_rate": 0.00220872285940997, + "loss": 0.6945, + "step": 4464 + }, + { + "epoch": 0.34343512037535573, + "learning_rate": 0.0022084033849373167, + "loss": 1.2389, + "step": 4465 + }, + { + "epoch": 0.3435120375355742, + "learning_rate": 0.0022080838691002505, + "loss": 0.7735, + "step": 4466 + }, + { + "epoch": 0.34358895469579265, + "learning_rate": 0.0022077643119174303, + "loss": 1.4623, + "step": 4467 + }, + { + "epoch": 0.34366587185601105, + "learning_rate": 0.002207444713407514, + "loss": 1.0244, + "step": 4468 + }, + { + "epoch": 0.3437427890162295, + "learning_rate": 0.002207125073589163, + "loss": 0.9686, + "step": 4469 + }, + { + "epoch": 0.34381970617644797, + "learning_rate": 0.0022068053924810425, + "loss": 1.1557, + "step": 4470 + }, + { + "epoch": 0.34389662333666643, + "learning_rate": 0.0022064856701018184, + "loss": 1.211, + "step": 4471 + }, + { + "epoch": 0.34397354049688483, + "learning_rate": 0.00220616590647016, + "loss": 1.4653, + "step": 4472 + }, + { + "epoch": 0.3440504576571033, + "learning_rate": 0.002205846101604738, + "loss": 0.9713, + "step": 4473 + }, + { + "epoch": 0.34412737481732175, + "learning_rate": 0.002205526255524227, + "loss": 1.053, + "step": 4474 + }, + { + "epoch": 0.3442042919775402, + "learning_rate": 0.002205206368247302, + "loss": 0.9828, + "step": 4475 + }, + { + "epoch": 0.3442812091377586, + "learning_rate": 0.002204886439792643, + "loss": 1.0335, + "step": 4476 + }, + { + "epoch": 0.3443581262979771, + "learning_rate": 0.0022045664701789295, + "loss": 1.0159, + "step": 4477 + }, + { + "epoch": 0.34443504345819553, + "learning_rate": 0.002204246459424846, + "loss": 0.9004, + "step": 4478 + }, + { + "epoch": 0.344511960618414, + "learning_rate": 0.002203926407549077, + "loss": 1.3003, + "step": 4479 + }, + { + "epoch": 0.3445888777786324, + "learning_rate": 0.002203606314570312, + "loss": 1.1376, + "step": 4480 + }, + { + "epoch": 0.34466579493885086, + "learning_rate": 0.002203286180507241, + "loss": 1.2093, + "step": 4481 + }, + { + "epoch": 0.3447427120990693, + "learning_rate": 0.0022029660053785568, + "loss": 0.9017, + "step": 4482 + }, + { + "epoch": 0.3448196292592878, + "learning_rate": 0.002202645789202955, + "loss": 1.3401, + "step": 4483 + }, + { + "epoch": 0.3448965464195062, + "learning_rate": 0.0022023255319991338, + "loss": 0.975, + "step": 4484 + }, + { + "epoch": 0.34497346357972464, + "learning_rate": 0.0022020052337857923, + "loss": 0.9787, + "step": 4485 + }, + { + "epoch": 0.3450503807399431, + "learning_rate": 0.0022016848945816334, + "loss": 0.9551, + "step": 4486 + }, + { + "epoch": 0.3451272979001615, + "learning_rate": 0.0022013645144053633, + "loss": 0.9308, + "step": 4487 + }, + { + "epoch": 0.34520421506037996, + "learning_rate": 0.002201044093275687, + "loss": 1.2331, + "step": 4488 + }, + { + "epoch": 0.3452811322205984, + "learning_rate": 0.002200723631211316, + "loss": 0.8028, + "step": 4489 + }, + { + "epoch": 0.3453580493808169, + "learning_rate": 0.002200403128230962, + "loss": 1.0846, + "step": 4490 + }, + { + "epoch": 0.3454349665410353, + "learning_rate": 0.0022000825843533395, + "loss": 1.3261, + "step": 4491 + }, + { + "epoch": 0.34551188370125374, + "learning_rate": 0.0021997619995971648, + "loss": 0.9967, + "step": 4492 + }, + { + "epoch": 0.3455888008614722, + "learning_rate": 0.0021994413739811576, + "loss": 1.2422, + "step": 4493 + }, + { + "epoch": 0.34566571802169066, + "learning_rate": 0.00219912070752404, + "loss": 1.0371, + "step": 4494 + }, + { + "epoch": 0.34574263518190906, + "learning_rate": 0.002198800000244536, + "loss": 1.0514, + "step": 4495 + }, + { + "epoch": 0.3458195523421275, + "learning_rate": 0.0021984792521613714, + "loss": 1.101, + "step": 4496 + }, + { + "epoch": 0.345896469502346, + "learning_rate": 0.002198158463293275, + "loss": 0.8815, + "step": 4497 + }, + { + "epoch": 0.34597338666256444, + "learning_rate": 0.002197837633658979, + "loss": 1.0826, + "step": 4498 + }, + { + "epoch": 0.34605030382278285, + "learning_rate": 0.002197516763277216, + "loss": 1.1511, + "step": 4499 + }, + { + "epoch": 0.3461272209830013, + "learning_rate": 0.002197195852166722, + "loss": 0.7556, + "step": 4500 + }, + { + "epoch": 0.34620413814321976, + "learning_rate": 0.0021968749003462356, + "loss": 1.0577, + "step": 4501 + }, + { + "epoch": 0.3462810553034382, + "learning_rate": 0.0021965539078344976, + "loss": 1.1914, + "step": 4502 + }, + { + "epoch": 0.3463579724636566, + "learning_rate": 0.0021962328746502513, + "loss": 0.974, + "step": 4503 + }, + { + "epoch": 0.3464348896238751, + "learning_rate": 0.002195911800812241, + "loss": 1.6516, + "step": 4504 + }, + { + "epoch": 0.34651180678409355, + "learning_rate": 0.0021955906863392162, + "loss": 1.2723, + "step": 4505 + }, + { + "epoch": 0.34658872394431195, + "learning_rate": 0.0021952695312499266, + "loss": 1.1857, + "step": 4506 + }, + { + "epoch": 0.3466656411045304, + "learning_rate": 0.002194948335563124, + "loss": 1.0173, + "step": 4507 + }, + { + "epoch": 0.34674255826474887, + "learning_rate": 0.0021946270992975634, + "loss": 1.3723, + "step": 4508 + }, + { + "epoch": 0.3468194754249673, + "learning_rate": 0.002194305822472003, + "loss": 1.0406, + "step": 4509 + }, + { + "epoch": 0.34689639258518573, + "learning_rate": 0.0021939845051052024, + "loss": 0.9578, + "step": 4510 + }, + { + "epoch": 0.3469733097454042, + "learning_rate": 0.0021936631472159224, + "loss": 1.8563, + "step": 4511 + }, + { + "epoch": 0.34705022690562265, + "learning_rate": 0.002193341748822929, + "loss": 1.5305, + "step": 4512 + }, + { + "epoch": 0.3471271440658411, + "learning_rate": 0.002193020309944988, + "loss": 1.223, + "step": 4513 + }, + { + "epoch": 0.3472040612260595, + "learning_rate": 0.002192698830600869, + "loss": 1.1849, + "step": 4514 + }, + { + "epoch": 0.34728097838627797, + "learning_rate": 0.0021923773108093432, + "loss": 1.4335, + "step": 4515 + }, + { + "epoch": 0.34735789554649643, + "learning_rate": 0.002192055750589185, + "loss": 0.8563, + "step": 4516 + }, + { + "epoch": 0.3474348127067149, + "learning_rate": 0.0021917341499591695, + "loss": 0.8439, + "step": 4517 + }, + { + "epoch": 0.3475117298669333, + "learning_rate": 0.0021914125089380772, + "loss": 0.7949, + "step": 4518 + }, + { + "epoch": 0.34758864702715175, + "learning_rate": 0.0021910908275446874, + "loss": 1.1507, + "step": 4519 + }, + { + "epoch": 0.3476655641873702, + "learning_rate": 0.0021907691057977837, + "loss": 1.2038, + "step": 4520 + }, + { + "epoch": 0.34774248134758867, + "learning_rate": 0.002190447343716152, + "loss": 1.1221, + "step": 4521 + }, + { + "epoch": 0.3478193985078071, + "learning_rate": 0.002190125541318581, + "loss": 1.1422, + "step": 4522 + }, + { + "epoch": 0.34789631566802554, + "learning_rate": 0.00218980369862386, + "loss": 1.2733, + "step": 4523 + }, + { + "epoch": 0.347973232828244, + "learning_rate": 0.002189481815650782, + "loss": 1.0066, + "step": 4524 + }, + { + "epoch": 0.3480501499884624, + "learning_rate": 0.0021891598924181424, + "loss": 1.1536, + "step": 4525 + }, + { + "epoch": 0.34812706714868086, + "learning_rate": 0.0021888379289447383, + "loss": 1.15, + "step": 4526 + }, + { + "epoch": 0.3482039843088993, + "learning_rate": 0.0021885159252493696, + "loss": 1.2677, + "step": 4527 + }, + { + "epoch": 0.3482809014691178, + "learning_rate": 0.0021881938813508384, + "loss": 1.1407, + "step": 4528 + }, + { + "epoch": 0.3483578186293362, + "learning_rate": 0.0021878717972679497, + "loss": 0.9848, + "step": 4529 + }, + { + "epoch": 0.34843473578955464, + "learning_rate": 0.0021875496730195095, + "loss": 1.2132, + "step": 4530 + }, + { + "epoch": 0.3485116529497731, + "learning_rate": 0.0021872275086243275, + "loss": 1.2289, + "step": 4531 + }, + { + "epoch": 0.34858857010999156, + "learning_rate": 0.002186905304101215, + "loss": 1.279, + "step": 4532 + }, + { + "epoch": 0.34866548727020996, + "learning_rate": 0.002186583059468986, + "loss": 1.3782, + "step": 4533 + }, + { + "epoch": 0.3487424044304284, + "learning_rate": 0.0021862607747464564, + "loss": 1.2075, + "step": 4534 + }, + { + "epoch": 0.3488193215906469, + "learning_rate": 0.002185938449952445, + "loss": 1.2174, + "step": 4535 + }, + { + "epoch": 0.34889623875086534, + "learning_rate": 0.0021856160851057727, + "loss": 1.2171, + "step": 4536 + }, + { + "epoch": 0.34897315591108374, + "learning_rate": 0.002185293680225263, + "loss": 0.6313, + "step": 4537 + }, + { + "epoch": 0.3490500730713022, + "learning_rate": 0.002184971235329741, + "loss": 0.7088, + "step": 4538 + }, + { + "epoch": 0.34912699023152066, + "learning_rate": 0.002184648750438034, + "loss": 1.198, + "step": 4539 + }, + { + "epoch": 0.3492039073917391, + "learning_rate": 0.0021843262255689735, + "loss": 0.9856, + "step": 4540 + }, + { + "epoch": 0.3492808245519575, + "learning_rate": 0.0021840036607413916, + "loss": 1.2084, + "step": 4541 + }, + { + "epoch": 0.349357741712176, + "learning_rate": 0.002183681055974123, + "loss": 1.125, + "step": 4542 + }, + { + "epoch": 0.34943465887239444, + "learning_rate": 0.002183358411286005, + "loss": 1.0228, + "step": 4543 + }, + { + "epoch": 0.3495115760326129, + "learning_rate": 0.002183035726695877, + "loss": 1.4435, + "step": 4544 + }, + { + "epoch": 0.3495884931928313, + "learning_rate": 0.0021827130022225812, + "loss": 1.1466, + "step": 4545 + }, + { + "epoch": 0.34966541035304977, + "learning_rate": 0.002182390237884962, + "loss": 0.8455, + "step": 4546 + }, + { + "epoch": 0.3497423275132682, + "learning_rate": 0.0021820674337018654, + "loss": 1.5281, + "step": 4547 + }, + { + "epoch": 0.34981924467348663, + "learning_rate": 0.0021817445896921404, + "loss": 0.7684, + "step": 4548 + }, + { + "epoch": 0.3498961618337051, + "learning_rate": 0.0021814217058746385, + "loss": 1.2232, + "step": 4549 + }, + { + "epoch": 0.34997307899392355, + "learning_rate": 0.0021810987822682127, + "loss": 1.0272, + "step": 4550 + }, + { + "epoch": 0.350049996154142, + "learning_rate": 0.0021807758188917194, + "loss": 0.9636, + "step": 4551 + }, + { + "epoch": 0.3501269133143604, + "learning_rate": 0.002180452815764017, + "loss": 1.3657, + "step": 4552 + }, + { + "epoch": 0.35020383047457887, + "learning_rate": 0.0021801297729039655, + "loss": 1.1887, + "step": 4553 + }, + { + "epoch": 0.35028074763479733, + "learning_rate": 0.0021798066903304268, + "loss": 1.1966, + "step": 4554 + }, + { + "epoch": 0.3503576647950158, + "learning_rate": 0.0021794835680622677, + "loss": 1.034, + "step": 4555 + }, + { + "epoch": 0.3504345819552342, + "learning_rate": 0.0021791604061183545, + "loss": 1.2712, + "step": 4556 + }, + { + "epoch": 0.35051149911545265, + "learning_rate": 0.0021788372045175577, + "loss": 0.8929, + "step": 4557 + }, + { + "epoch": 0.3505884162756711, + "learning_rate": 0.0021785139632787493, + "loss": 1.1922, + "step": 4558 + }, + { + "epoch": 0.35066533343588957, + "learning_rate": 0.0021781906824208026, + "loss": 0.8638, + "step": 4559 + }, + { + "epoch": 0.350742250596108, + "learning_rate": 0.002177867361962596, + "loss": 1.1992, + "step": 4560 + }, + { + "epoch": 0.35081916775632643, + "learning_rate": 0.0021775440019230073, + "loss": 1.2379, + "step": 4561 + }, + { + "epoch": 0.3508960849165449, + "learning_rate": 0.0021772206023209185, + "loss": 1.1376, + "step": 4562 + }, + { + "epoch": 0.35097300207676335, + "learning_rate": 0.002176897163175212, + "loss": 1.2832, + "step": 4563 + }, + { + "epoch": 0.35104991923698176, + "learning_rate": 0.0021765736845047764, + "loss": 1.3489, + "step": 4564 + }, + { + "epoch": 0.3511268363972002, + "learning_rate": 0.0021762501663284968, + "loss": 0.9143, + "step": 4565 + }, + { + "epoch": 0.3512037535574187, + "learning_rate": 0.002175926608665266, + "loss": 1.1331, + "step": 4566 + }, + { + "epoch": 0.3512806707176371, + "learning_rate": 0.002175603011533976, + "loss": 1.1869, + "step": 4567 + }, + { + "epoch": 0.35135758787785554, + "learning_rate": 0.0021752793749535214, + "loss": 0.9946, + "step": 4568 + }, + { + "epoch": 0.351434505038074, + "learning_rate": 0.002174955698942801, + "loss": 1.334, + "step": 4569 + }, + { + "epoch": 0.35151142219829246, + "learning_rate": 0.0021746319835207134, + "loss": 0.83, + "step": 4570 + }, + { + "epoch": 0.35158833935851086, + "learning_rate": 0.002174308228706162, + "loss": 1.3723, + "step": 4571 + }, + { + "epoch": 0.3516652565187293, + "learning_rate": 0.00217398443451805, + "loss": 1.5225, + "step": 4572 + }, + { + "epoch": 0.3517421736789478, + "learning_rate": 0.002173660600975284, + "loss": 1.3333, + "step": 4573 + }, + { + "epoch": 0.35181909083916624, + "learning_rate": 0.002173336728096774, + "loss": 0.8326, + "step": 4574 + }, + { + "epoch": 0.35189600799938464, + "learning_rate": 0.002173012815901431, + "loss": 1.3905, + "step": 4575 + }, + { + "epoch": 0.3519729251596031, + "learning_rate": 0.002172688864408167, + "loss": 0.7822, + "step": 4576 + }, + { + "epoch": 0.35204984231982156, + "learning_rate": 0.0021723648736358997, + "loss": 1.1313, + "step": 4577 + }, + { + "epoch": 0.35212675948004, + "learning_rate": 0.002172040843603547, + "loss": 1.4056, + "step": 4578 + }, + { + "epoch": 0.3522036766402584, + "learning_rate": 0.002171716774330029, + "loss": 1.1409, + "step": 4579 + }, + { + "epoch": 0.3522805938004769, + "learning_rate": 0.0021713926658342687, + "loss": 1.4108, + "step": 4580 + }, + { + "epoch": 0.35235751096069534, + "learning_rate": 0.0021710685181351905, + "loss": 1.1529, + "step": 4581 + }, + { + "epoch": 0.3524344281209138, + "learning_rate": 0.0021707443312517223, + "loss": 1.3354, + "step": 4582 + }, + { + "epoch": 0.3525113452811322, + "learning_rate": 0.0021704201052027937, + "loss": 1.0511, + "step": 4583 + }, + { + "epoch": 0.35258826244135066, + "learning_rate": 0.0021700958400073366, + "loss": 1.4036, + "step": 4584 + }, + { + "epoch": 0.3526651796015691, + "learning_rate": 0.0021697715356842847, + "loss": 1.0696, + "step": 4585 + }, + { + "epoch": 0.3527420967617876, + "learning_rate": 0.002169447192252575, + "loss": 1.4884, + "step": 4586 + }, + { + "epoch": 0.352819013922006, + "learning_rate": 0.002169122809731146, + "loss": 1.3654, + "step": 4587 + }, + { + "epoch": 0.35289593108222445, + "learning_rate": 0.002168798388138939, + "loss": 0.849, + "step": 4588 + }, + { + "epoch": 0.3529728482424429, + "learning_rate": 0.0021684739274948966, + "loss": 1.448, + "step": 4589 + }, + { + "epoch": 0.3530497654026613, + "learning_rate": 0.002168149427817966, + "loss": 0.9463, + "step": 4590 + }, + { + "epoch": 0.35312668256287977, + "learning_rate": 0.0021678248891270928, + "loss": 1.4599, + "step": 4591 + }, + { + "epoch": 0.3532035997230982, + "learning_rate": 0.0021675003114412284, + "loss": 1.377, + "step": 4592 + }, + { + "epoch": 0.3532805168833167, + "learning_rate": 0.0021671756947793256, + "loss": 1.0482, + "step": 4593 + }, + { + "epoch": 0.3533574340435351, + "learning_rate": 0.0021668510391603386, + "loss": 1.351, + "step": 4594 + }, + { + "epoch": 0.35343435120375355, + "learning_rate": 0.0021665263446032246, + "loss": 0.9677, + "step": 4595 + }, + { + "epoch": 0.353511268363972, + "learning_rate": 0.0021662016111269424, + "loss": 1.4628, + "step": 4596 + }, + { + "epoch": 0.35358818552419047, + "learning_rate": 0.0021658768387504542, + "loss": 1.2718, + "step": 4597 + }, + { + "epoch": 0.35366510268440887, + "learning_rate": 0.0021655520274927232, + "loss": 1.4732, + "step": 4598 + }, + { + "epoch": 0.35374201984462733, + "learning_rate": 0.0021652271773727155, + "loss": 0.6943, + "step": 4599 + }, + { + "epoch": 0.3538189370048458, + "learning_rate": 0.0021649022884094, + "loss": 1.192, + "step": 4600 + }, + { + "epoch": 0.35389585416506425, + "learning_rate": 0.0021645773606217464, + "loss": 1.2953, + "step": 4601 + }, + { + "epoch": 0.35397277132528265, + "learning_rate": 0.0021642523940287287, + "loss": 1.2194, + "step": 4602 + }, + { + "epoch": 0.3540496884855011, + "learning_rate": 0.002163927388649321, + "loss": 1.2613, + "step": 4603 + }, + { + "epoch": 0.35412660564571957, + "learning_rate": 0.0021636023445025013, + "loss": 0.9221, + "step": 4604 + }, + { + "epoch": 0.35420352280593803, + "learning_rate": 0.002163277261607249, + "loss": 1.1065, + "step": 4605 + }, + { + "epoch": 0.35428043996615644, + "learning_rate": 0.002162952139982547, + "loss": 1.0068, + "step": 4606 + }, + { + "epoch": 0.3543573571263749, + "learning_rate": 0.0021626269796473775, + "loss": 1.3094, + "step": 4607 + }, + { + "epoch": 0.35443427428659335, + "learning_rate": 0.0021623017806207284, + "loss": 1.2667, + "step": 4608 + }, + { + "epoch": 0.35451119144681176, + "learning_rate": 0.0021619765429215882, + "loss": 1.2654, + "step": 4609 + }, + { + "epoch": 0.3545881086070302, + "learning_rate": 0.002161651266568948, + "loss": 0.8474, + "step": 4610 + }, + { + "epoch": 0.3546650257672487, + "learning_rate": 0.0021613259515818007, + "loss": 1.1862, + "step": 4611 + }, + { + "epoch": 0.35474194292746714, + "learning_rate": 0.002161000597979141, + "loss": 1.1444, + "step": 4612 + }, + { + "epoch": 0.35481886008768554, + "learning_rate": 0.002160675205779969, + "loss": 1.1831, + "step": 4613 + }, + { + "epoch": 0.354895777247904, + "learning_rate": 0.0021603497750032822, + "loss": 1.1047, + "step": 4614 + }, + { + "epoch": 0.35497269440812246, + "learning_rate": 0.002160024305668084, + "loss": 1.4083, + "step": 4615 + }, + { + "epoch": 0.3550496115683409, + "learning_rate": 0.002159698797793379, + "loss": 0.9487, + "step": 4616 + }, + { + "epoch": 0.3551265287285593, + "learning_rate": 0.0021593732513981736, + "loss": 1.0903, + "step": 4617 + }, + { + "epoch": 0.3552034458887778, + "learning_rate": 0.0021590476665014767, + "loss": 1.258, + "step": 4618 + }, + { + "epoch": 0.35528036304899624, + "learning_rate": 0.0021587220431222996, + "loss": 1.1773, + "step": 4619 + }, + { + "epoch": 0.3553572802092147, + "learning_rate": 0.002158396381279656, + "loss": 1.1955, + "step": 4620 + }, + { + "epoch": 0.3554341973694331, + "learning_rate": 0.0021580706809925616, + "loss": 0.899, + "step": 4621 + }, + { + "epoch": 0.35551111452965156, + "learning_rate": 0.0021577449422800344, + "loss": 0.905, + "step": 4622 + }, + { + "epoch": 0.35558803168987, + "learning_rate": 0.0021574191651610943, + "loss": 0.9677, + "step": 4623 + }, + { + "epoch": 0.3556649488500885, + "learning_rate": 0.002157093349654764, + "loss": 1.1556, + "step": 4624 + }, + { + "epoch": 0.3557418660103069, + "learning_rate": 0.002156767495780069, + "loss": 1.2715, + "step": 4625 + }, + { + "epoch": 0.35581878317052534, + "learning_rate": 0.0021564416035560344, + "loss": 1.2876, + "step": 4626 + }, + { + "epoch": 0.3558957003307438, + "learning_rate": 0.002156115673001691, + "loss": 1.1753, + "step": 4627 + }, + { + "epoch": 0.3559726174909622, + "learning_rate": 0.0021557897041360697, + "loss": 1.2324, + "step": 4628 + }, + { + "epoch": 0.35604953465118067, + "learning_rate": 0.0021554636969782037, + "loss": 0.9744, + "step": 4629 + }, + { + "epoch": 0.3561264518113991, + "learning_rate": 0.0021551376515471297, + "loss": 1.1218, + "step": 4630 + }, + { + "epoch": 0.3562033689716176, + "learning_rate": 0.002154811567861885, + "loss": 1.1404, + "step": 4631 + }, + { + "epoch": 0.356280286131836, + "learning_rate": 0.002154485445941511, + "loss": 1.2168, + "step": 4632 + }, + { + "epoch": 0.35635720329205445, + "learning_rate": 0.0021541592858050494, + "loss": 0.9979, + "step": 4633 + }, + { + "epoch": 0.3564341204522729, + "learning_rate": 0.002153833087471545, + "loss": 0.9219, + "step": 4634 + }, + { + "epoch": 0.35651103761249137, + "learning_rate": 0.002153506850960046, + "loss": 1.2117, + "step": 4635 + }, + { + "epoch": 0.35658795477270977, + "learning_rate": 0.0021531805762896004, + "loss": 1.0654, + "step": 4636 + }, + { + "epoch": 0.35666487193292823, + "learning_rate": 0.0021528542634792606, + "loss": 1.3072, + "step": 4637 + }, + { + "epoch": 0.3567417890931467, + "learning_rate": 0.0021525279125480796, + "loss": 1.0846, + "step": 4638 + }, + { + "epoch": 0.35681870625336515, + "learning_rate": 0.002152201523515114, + "loss": 1.2969, + "step": 4639 + }, + { + "epoch": 0.35689562341358355, + "learning_rate": 0.0021518750963994216, + "loss": 1.1594, + "step": 4640 + }, + { + "epoch": 0.356972540573802, + "learning_rate": 0.0021515486312200633, + "loss": 1.2355, + "step": 4641 + }, + { + "epoch": 0.35704945773402047, + "learning_rate": 0.0021512221279961, + "loss": 0.7651, + "step": 4642 + }, + { + "epoch": 0.35712637489423893, + "learning_rate": 0.0021508955867465995, + "loss": 1.0978, + "step": 4643 + }, + { + "epoch": 0.35720329205445733, + "learning_rate": 0.0021505690074906272, + "loss": 1.047, + "step": 4644 + }, + { + "epoch": 0.3572802092146758, + "learning_rate": 0.0021502423902472517, + "loss": 0.964, + "step": 4645 + }, + { + "epoch": 0.35735712637489425, + "learning_rate": 0.0021499157350355464, + "loss": 1.0864, + "step": 4646 + }, + { + "epoch": 0.3574340435351127, + "learning_rate": 0.002149589041874583, + "loss": 1.4225, + "step": 4647 + }, + { + "epoch": 0.3575109606953311, + "learning_rate": 0.0021492623107834397, + "loss": 1.3418, + "step": 4648 + }, + { + "epoch": 0.3575878778555496, + "learning_rate": 0.0021489355417811918, + "loss": 1.0769, + "step": 4649 + }, + { + "epoch": 0.35766479501576803, + "learning_rate": 0.0021486087348869224, + "loss": 1.0327, + "step": 4650 + }, + { + "epoch": 0.35774171217598644, + "learning_rate": 0.002148281890119713, + "loss": 1.295, + "step": 4651 + }, + { + "epoch": 0.3578186293362049, + "learning_rate": 0.002147955007498648, + "loss": 1.1204, + "step": 4652 + }, + { + "epoch": 0.35789554649642336, + "learning_rate": 0.0021476280870428147, + "loss": 1.094, + "step": 4653 + }, + { + "epoch": 0.3579724636566418, + "learning_rate": 0.002147301128771303, + "loss": 1.2706, + "step": 4654 + }, + { + "epoch": 0.3580493808168602, + "learning_rate": 0.0021469741327032037, + "loss": 1.2103, + "step": 4655 + }, + { + "epoch": 0.3581262979770787, + "learning_rate": 0.0021466470988576103, + "loss": 1.4753, + "step": 4656 + }, + { + "epoch": 0.35820321513729714, + "learning_rate": 0.002146320027253619, + "loss": 1.0931, + "step": 4657 + }, + { + "epoch": 0.3582801322975156, + "learning_rate": 0.0021459929179103274, + "loss": 1.0186, + "step": 4658 + }, + { + "epoch": 0.358357049457734, + "learning_rate": 0.002145665770846837, + "loss": 1.0179, + "step": 4659 + }, + { + "epoch": 0.35843396661795246, + "learning_rate": 0.0021453385860822484, + "loss": 1.1274, + "step": 4660 + }, + { + "epoch": 0.3585108837781709, + "learning_rate": 0.0021450113636356676, + "loss": 0.8537, + "step": 4661 + }, + { + "epoch": 0.3585878009383894, + "learning_rate": 0.002144684103526201, + "loss": 1.2804, + "step": 4662 + }, + { + "epoch": 0.3586647180986078, + "learning_rate": 0.0021443568057729575, + "loss": 0.9907, + "step": 4663 + }, + { + "epoch": 0.35874163525882624, + "learning_rate": 0.002144029470395049, + "loss": 0.9742, + "step": 4664 + }, + { + "epoch": 0.3588185524190447, + "learning_rate": 0.0021437020974115887, + "loss": 1.071, + "step": 4665 + }, + { + "epoch": 0.35889546957926316, + "learning_rate": 0.002143374686841691, + "loss": 1.1727, + "step": 4666 + }, + { + "epoch": 0.35897238673948156, + "learning_rate": 0.002143047238704476, + "loss": 0.862, + "step": 4667 + }, + { + "epoch": 0.3590493038997, + "learning_rate": 0.002142719753019062, + "loss": 1.2982, + "step": 4668 + }, + { + "epoch": 0.3591262210599185, + "learning_rate": 0.002142392229804572, + "loss": 0.9353, + "step": 4669 + }, + { + "epoch": 0.3592031382201369, + "learning_rate": 0.0021420646690801302, + "loss": 1.0221, + "step": 4670 + }, + { + "epoch": 0.35928005538035535, + "learning_rate": 0.0021417370708648634, + "loss": 1.4891, + "step": 4671 + }, + { + "epoch": 0.3593569725405738, + "learning_rate": 0.0021414094351778997, + "loss": 0.89, + "step": 4672 + }, + { + "epoch": 0.35943388970079226, + "learning_rate": 0.0021410817620383703, + "loss": 0.9063, + "step": 4673 + }, + { + "epoch": 0.35951080686101067, + "learning_rate": 0.0021407540514654095, + "loss": 0.9539, + "step": 4674 + }, + { + "epoch": 0.3595877240212291, + "learning_rate": 0.002140426303478152, + "loss": 0.9472, + "step": 4675 + }, + { + "epoch": 0.3596646411814476, + "learning_rate": 0.0021400985180957347, + "loss": 1.0589, + "step": 4676 + }, + { + "epoch": 0.35974155834166605, + "learning_rate": 0.0021397706953372978, + "loss": 1.1529, + "step": 4677 + }, + { + "epoch": 0.35981847550188445, + "learning_rate": 0.0021394428352219837, + "loss": 1.4477, + "step": 4678 + }, + { + "epoch": 0.3598953926621029, + "learning_rate": 0.0021391149377689355, + "loss": 1.0915, + "step": 4679 + }, + { + "epoch": 0.35997230982232137, + "learning_rate": 0.0021387870029973004, + "loss": 1.4642, + "step": 4680 + }, + { + "epoch": 0.3600492269825398, + "learning_rate": 0.002138459030926226, + "loss": 0.9607, + "step": 4681 + }, + { + "epoch": 0.36012614414275823, + "learning_rate": 0.0021381310215748644, + "loss": 1.2426, + "step": 4682 + }, + { + "epoch": 0.3602030613029767, + "learning_rate": 0.0021378029749623663, + "loss": 1.3444, + "step": 4683 + }, + { + "epoch": 0.36027997846319515, + "learning_rate": 0.0021374748911078883, + "loss": 1.2366, + "step": 4684 + }, + { + "epoch": 0.3603568956234136, + "learning_rate": 0.0021371467700305874, + "loss": 1.1309, + "step": 4685 + }, + { + "epoch": 0.360433812783632, + "learning_rate": 0.002136818611749622, + "loss": 0.8353, + "step": 4686 + }, + { + "epoch": 0.36051072994385047, + "learning_rate": 0.0021364904162841543, + "loss": 1.1354, + "step": 4687 + }, + { + "epoch": 0.36058764710406893, + "learning_rate": 0.0021361621836533478, + "loss": 1.1264, + "step": 4688 + }, + { + "epoch": 0.36066456426428734, + "learning_rate": 0.0021358339138763684, + "loss": 0.9663, + "step": 4689 + }, + { + "epoch": 0.3607414814245058, + "learning_rate": 0.0021355056069723845, + "loss": 1.0954, + "step": 4690 + }, + { + "epoch": 0.36081839858472425, + "learning_rate": 0.0021351772629605655, + "loss": 1.057, + "step": 4691 + }, + { + "epoch": 0.3608953157449427, + "learning_rate": 0.0021348488818600845, + "loss": 1.1182, + "step": 4692 + }, + { + "epoch": 0.3609722329051611, + "learning_rate": 0.0021345204636901154, + "loss": 1.0405, + "step": 4693 + }, + { + "epoch": 0.3610491500653796, + "learning_rate": 0.002134192008469835, + "loss": 1.2886, + "step": 4694 + }, + { + "epoch": 0.36112606722559804, + "learning_rate": 0.0021338635162184223, + "loss": 1.0584, + "step": 4695 + }, + { + "epoch": 0.3612029843858165, + "learning_rate": 0.0021335349869550588, + "loss": 1.389, + "step": 4696 + }, + { + "epoch": 0.3612799015460349, + "learning_rate": 0.002133206420698927, + "loss": 1.0482, + "step": 4697 + }, + { + "epoch": 0.36135681870625336, + "learning_rate": 0.002132877817469212, + "loss": 0.8862, + "step": 4698 + }, + { + "epoch": 0.3614337358664718, + "learning_rate": 0.0021325491772851016, + "loss": 1.2062, + "step": 4699 + }, + { + "epoch": 0.3615106530266903, + "learning_rate": 0.0021322205001657857, + "loss": 1.4097, + "step": 4700 + }, + { + "epoch": 0.3615875701869087, + "learning_rate": 0.0021318917861304566, + "loss": 1.3983, + "step": 4701 + }, + { + "epoch": 0.36166448734712714, + "learning_rate": 0.002131563035198307, + "loss": 1.1813, + "step": 4702 + }, + { + "epoch": 0.3617414045073456, + "learning_rate": 0.0021312342473885333, + "loss": 1.4315, + "step": 4703 + }, + { + "epoch": 0.36181832166756406, + "learning_rate": 0.0021309054227203338, + "loss": 1.0769, + "step": 4704 + }, + { + "epoch": 0.36189523882778246, + "learning_rate": 0.0021305765612129103, + "loss": 1.2647, + "step": 4705 + }, + { + "epoch": 0.3619721559880009, + "learning_rate": 0.002130247662885463, + "loss": 1.2099, + "step": 4706 + }, + { + "epoch": 0.3620490731482194, + "learning_rate": 0.002129918727757199, + "loss": 1.0531, + "step": 4707 + }, + { + "epoch": 0.36212599030843784, + "learning_rate": 0.002129589755847323, + "loss": 1.0395, + "step": 4708 + }, + { + "epoch": 0.36220290746865624, + "learning_rate": 0.002129260747175046, + "loss": 1.4338, + "step": 4709 + }, + { + "epoch": 0.3622798246288747, + "learning_rate": 0.0021289317017595772, + "loss": 1.1014, + "step": 4710 + }, + { + "epoch": 0.36235674178909316, + "learning_rate": 0.002128602619620131, + "loss": 1.3059, + "step": 4711 + }, + { + "epoch": 0.36243365894931157, + "learning_rate": 0.0021282735007759233, + "loss": 1.2196, + "step": 4712 + }, + { + "epoch": 0.36251057610953, + "learning_rate": 0.0021279443452461707, + "loss": 1.1374, + "step": 4713 + }, + { + "epoch": 0.3625874932697485, + "learning_rate": 0.0021276151530500937, + "loss": 1.1438, + "step": 4714 + }, + { + "epoch": 0.36266441042996694, + "learning_rate": 0.0021272859242069133, + "loss": 1.2911, + "step": 4715 + }, + { + "epoch": 0.36274132759018535, + "learning_rate": 0.0021269566587358547, + "loss": 1.1752, + "step": 4716 + }, + { + "epoch": 0.3628182447504038, + "learning_rate": 0.002126627356656143, + "loss": 1.3372, + "step": 4717 + }, + { + "epoch": 0.36289516191062227, + "learning_rate": 0.0021262980179870064, + "loss": 1.0532, + "step": 4718 + }, + { + "epoch": 0.3629720790708407, + "learning_rate": 0.0021259686427476765, + "loss": 1.0698, + "step": 4719 + }, + { + "epoch": 0.36304899623105913, + "learning_rate": 0.002125639230957385, + "loss": 0.8697, + "step": 4720 + }, + { + "epoch": 0.3631259133912776, + "learning_rate": 0.002125309782635367, + "loss": 1.317, + "step": 4721 + }, + { + "epoch": 0.36320283055149605, + "learning_rate": 0.0021249802978008585, + "loss": 1.3131, + "step": 4722 + }, + { + "epoch": 0.3632797477117145, + "learning_rate": 0.0021246507764731, + "loss": 1.5172, + "step": 4723 + }, + { + "epoch": 0.3633566648719329, + "learning_rate": 0.002124321218671331, + "loss": 1.3545, + "step": 4724 + }, + { + "epoch": 0.36343358203215137, + "learning_rate": 0.0021239916244147955, + "loss": 1.3479, + "step": 4725 + }, + { + "epoch": 0.36351049919236983, + "learning_rate": 0.0021236619937227384, + "loss": 1.0363, + "step": 4726 + }, + { + "epoch": 0.3635874163525883, + "learning_rate": 0.002123332326614408, + "loss": 0.9226, + "step": 4727 + }, + { + "epoch": 0.3636643335128067, + "learning_rate": 0.002123002623109053, + "loss": 1.0268, + "step": 4728 + }, + { + "epoch": 0.36374125067302515, + "learning_rate": 0.0021226728832259264, + "loss": 1.3165, + "step": 4729 + }, + { + "epoch": 0.3638181678332436, + "learning_rate": 0.0021223431069842804, + "loss": 1.081, + "step": 4730 + }, + { + "epoch": 0.363895084993462, + "learning_rate": 0.002122013294403372, + "loss": 1.0528, + "step": 4731 + }, + { + "epoch": 0.3639720021536805, + "learning_rate": 0.0021216834455024592, + "loss": 1.3662, + "step": 4732 + }, + { + "epoch": 0.36404891931389893, + "learning_rate": 0.002121353560300802, + "loss": 1.4345, + "step": 4733 + }, + { + "epoch": 0.3641258364741174, + "learning_rate": 0.0021210236388176634, + "loss": 1.5348, + "step": 4734 + }, + { + "epoch": 0.3642027536343358, + "learning_rate": 0.002120693681072307, + "loss": 1.0566, + "step": 4735 + }, + { + "epoch": 0.36427967079455426, + "learning_rate": 0.0021203636870839996, + "loss": 1.2474, + "step": 4736 + }, + { + "epoch": 0.3643565879547727, + "learning_rate": 0.00212003365687201, + "loss": 1.013, + "step": 4737 + }, + { + "epoch": 0.3644335051149912, + "learning_rate": 0.0021197035904556086, + "loss": 0.9083, + "step": 4738 + }, + { + "epoch": 0.3645104222752096, + "learning_rate": 0.002119373487854069, + "loss": 1.1171, + "step": 4739 + }, + { + "epoch": 0.36458733943542804, + "learning_rate": 0.002119043349086666, + "loss": 1.1389, + "step": 4740 + }, + { + "epoch": 0.3646642565956465, + "learning_rate": 0.0021187131741726766, + "loss": 1.0876, + "step": 4741 + }, + { + "epoch": 0.36474117375586496, + "learning_rate": 0.0021183829631313805, + "loss": 0.8338, + "step": 4742 + }, + { + "epoch": 0.36481809091608336, + "learning_rate": 0.0021180527159820584, + "loss": 1.3758, + "step": 4743 + }, + { + "epoch": 0.3648950080763018, + "learning_rate": 0.002117722432743994, + "loss": 1.0835, + "step": 4744 + }, + { + "epoch": 0.3649719252365203, + "learning_rate": 0.002117392113436473, + "loss": 0.727, + "step": 4745 + }, + { + "epoch": 0.36504884239673874, + "learning_rate": 0.002117061758078783, + "loss": 1.5522, + "step": 4746 + }, + { + "epoch": 0.36512575955695714, + "learning_rate": 0.0021167313666902144, + "loss": 1.0084, + "step": 4747 + }, + { + "epoch": 0.3652026767171756, + "learning_rate": 0.002116400939290058, + "loss": 0.9914, + "step": 4748 + }, + { + "epoch": 0.36527959387739406, + "learning_rate": 0.002116070475897609, + "loss": 1.3308, + "step": 4749 + }, + { + "epoch": 0.36535651103761246, + "learning_rate": 0.0021157399765321626, + "loss": 1.1763, + "step": 4750 + }, + { + "epoch": 0.3654334281978309, + "learning_rate": 0.0021154094412130166, + "loss": 1.0322, + "step": 4751 + }, + { + "epoch": 0.3655103453580494, + "learning_rate": 0.002115078869959473, + "loss": 1.1166, + "step": 4752 + }, + { + "epoch": 0.36558726251826784, + "learning_rate": 0.002114748262790832, + "loss": 1.0376, + "step": 4753 + }, + { + "epoch": 0.36566417967848625, + "learning_rate": 0.0021144176197264004, + "loss": 0.983, + "step": 4754 + }, + { + "epoch": 0.3657410968387047, + "learning_rate": 0.0021140869407854828, + "loss": 1.2964, + "step": 4755 + }, + { + "epoch": 0.36581801399892316, + "learning_rate": 0.002113756225987389, + "loss": 1.2203, + "step": 4756 + }, + { + "epoch": 0.3658949311591416, + "learning_rate": 0.0021134254753514296, + "loss": 1.4343, + "step": 4757 + }, + { + "epoch": 0.36597184831936, + "learning_rate": 0.0021130946888969176, + "loss": 1.0308, + "step": 4758 + }, + { + "epoch": 0.3660487654795785, + "learning_rate": 0.002112763866643167, + "loss": 1.3455, + "step": 4759 + }, + { + "epoch": 0.36612568263979695, + "learning_rate": 0.0021124330086094967, + "loss": 1.2272, + "step": 4760 + }, + { + "epoch": 0.3662025998000154, + "learning_rate": 0.002112102114815224, + "loss": 1.3347, + "step": 4761 + }, + { + "epoch": 0.3662795169602338, + "learning_rate": 0.002111771185279671, + "loss": 1.0937, + "step": 4762 + }, + { + "epoch": 0.36635643412045227, + "learning_rate": 0.0021114402200221615, + "loss": 1.0203, + "step": 4763 + }, + { + "epoch": 0.3664333512806707, + "learning_rate": 0.0021111092190620193, + "loss": 1.4053, + "step": 4764 + }, + { + "epoch": 0.3665102684408892, + "learning_rate": 0.002110778182418574, + "loss": 1.3736, + "step": 4765 + }, + { + "epoch": 0.3665871856011076, + "learning_rate": 0.0021104471101111534, + "loss": 0.801, + "step": 4766 + }, + { + "epoch": 0.36666410276132605, + "learning_rate": 0.00211011600215909, + "loss": 1.0116, + "step": 4767 + }, + { + "epoch": 0.3667410199215445, + "learning_rate": 0.0021097848585817174, + "loss": 0.9977, + "step": 4768 + }, + { + "epoch": 0.36681793708176297, + "learning_rate": 0.0021094536793983716, + "loss": 1.1082, + "step": 4769 + }, + { + "epoch": 0.36689485424198137, + "learning_rate": 0.00210912246462839, + "loss": 1.3029, + "step": 4770 + }, + { + "epoch": 0.36697177140219983, + "learning_rate": 0.002108791214291113, + "loss": 1.1388, + "step": 4771 + }, + { + "epoch": 0.3670486885624183, + "learning_rate": 0.0021084599284058826, + "loss": 1.2999, + "step": 4772 + }, + { + "epoch": 0.3671256057226367, + "learning_rate": 0.002108128606992043, + "loss": 1.3385, + "step": 4773 + }, + { + "epoch": 0.36720252288285515, + "learning_rate": 0.00210779725006894, + "loss": 1.3433, + "step": 4774 + }, + { + "epoch": 0.3672794400430736, + "learning_rate": 0.0021074658576559217, + "loss": 0.8713, + "step": 4775 + }, + { + "epoch": 0.36735635720329207, + "learning_rate": 0.0021071344297723398, + "loss": 1.1533, + "step": 4776 + }, + { + "epoch": 0.3674332743635105, + "learning_rate": 0.0021068029664375455, + "loss": 1.1066, + "step": 4777 + }, + { + "epoch": 0.36751019152372894, + "learning_rate": 0.0021064714676708933, + "loss": 0.8034, + "step": 4778 + }, + { + "epoch": 0.3675871086839474, + "learning_rate": 0.0021061399334917404, + "loss": 1.0343, + "step": 4779 + }, + { + "epoch": 0.36766402584416585, + "learning_rate": 0.002105808363919445, + "loss": 1.1599, + "step": 4780 + }, + { + "epoch": 0.36774094300438426, + "learning_rate": 0.002105476758973368, + "loss": 1.3468, + "step": 4781 + }, + { + "epoch": 0.3678178601646027, + "learning_rate": 0.0021051451186728715, + "loss": 1.0666, + "step": 4782 + }, + { + "epoch": 0.3678947773248212, + "learning_rate": 0.002104813443037321, + "loss": 1.1095, + "step": 4783 + }, + { + "epoch": 0.36797169448503964, + "learning_rate": 0.0021044817320860837, + "loss": 1.1295, + "step": 4784 + }, + { + "epoch": 0.36804861164525804, + "learning_rate": 0.0021041499858385276, + "loss": 1.027, + "step": 4785 + }, + { + "epoch": 0.3681255288054765, + "learning_rate": 0.0021038182043140244, + "loss": 0.867, + "step": 4786 + }, + { + "epoch": 0.36820244596569496, + "learning_rate": 0.002103486387531947, + "loss": 1.0818, + "step": 4787 + }, + { + "epoch": 0.3682793631259134, + "learning_rate": 0.00210315453551167, + "loss": 1.0079, + "step": 4788 + }, + { + "epoch": 0.3683562802861318, + "learning_rate": 0.002102822648272572, + "loss": 1.2499, + "step": 4789 + }, + { + "epoch": 0.3684331974463503, + "learning_rate": 0.0021024907258340307, + "loss": 1.4547, + "step": 4790 + }, + { + "epoch": 0.36851011460656874, + "learning_rate": 0.0021021587682154286, + "loss": 0.6778, + "step": 4791 + }, + { + "epoch": 0.36858703176678714, + "learning_rate": 0.002101826775436148, + "loss": 0.8478, + "step": 4792 + }, + { + "epoch": 0.3686639489270056, + "learning_rate": 0.002101494747515575, + "loss": 1.2081, + "step": 4793 + }, + { + "epoch": 0.36874086608722406, + "learning_rate": 0.0021011626844730964, + "loss": 1.1856, + "step": 4794 + }, + { + "epoch": 0.3688177832474425, + "learning_rate": 0.002100830586328102, + "loss": 1.2232, + "step": 4795 + }, + { + "epoch": 0.3688947004076609, + "learning_rate": 0.0021004984530999843, + "loss": 1.1324, + "step": 4796 + }, + { + "epoch": 0.3689716175678794, + "learning_rate": 0.0021001662848081355, + "loss": 1.1762, + "step": 4797 + }, + { + "epoch": 0.36904853472809784, + "learning_rate": 0.002099834081471952, + "loss": 1.0931, + "step": 4798 + }, + { + "epoch": 0.3691254518883163, + "learning_rate": 0.0020995018431108313, + "loss": 1.0817, + "step": 4799 + }, + { + "epoch": 0.3692023690485347, + "learning_rate": 0.0020991695697441734, + "loss": 1.2891, + "step": 4800 + }, + { + "epoch": 0.36927928620875317, + "learning_rate": 0.0020988372613913796, + "loss": 1.1474, + "step": 4801 + }, + { + "epoch": 0.3693562033689716, + "learning_rate": 0.0020985049180718544, + "loss": 1.2557, + "step": 4802 + }, + { + "epoch": 0.3694331205291901, + "learning_rate": 0.002098172539805003, + "loss": 1.2381, + "step": 4803 + }, + { + "epoch": 0.3695100376894085, + "learning_rate": 0.0020978401266102337, + "loss": 1.2017, + "step": 4804 + }, + { + "epoch": 0.36958695484962695, + "learning_rate": 0.002097507678506956, + "loss": 1.2147, + "step": 4805 + }, + { + "epoch": 0.3696638720098454, + "learning_rate": 0.002097175195514583, + "loss": 1.2062, + "step": 4806 + }, + { + "epoch": 0.36974078917006387, + "learning_rate": 0.002096842677652527, + "loss": 1.3095, + "step": 4807 + }, + { + "epoch": 0.36981770633028227, + "learning_rate": 0.002096510124940206, + "loss": 1.2723, + "step": 4808 + }, + { + "epoch": 0.36989462349050073, + "learning_rate": 0.0020961775373970364, + "loss": 0.704, + "step": 4809 + }, + { + "epoch": 0.3699715406507192, + "learning_rate": 0.0020958449150424393, + "loss": 1.3423, + "step": 4810 + }, + { + "epoch": 0.37004845781093765, + "learning_rate": 0.002095512257895837, + "loss": 0.9298, + "step": 4811 + }, + { + "epoch": 0.37012537497115605, + "learning_rate": 0.002095179565976653, + "loss": 1.1521, + "step": 4812 + }, + { + "epoch": 0.3702022921313745, + "learning_rate": 0.002094846839304314, + "loss": 1.0047, + "step": 4813 + }, + { + "epoch": 0.37027920929159297, + "learning_rate": 0.0020945140778982478, + "loss": 0.7626, + "step": 4814 + }, + { + "epoch": 0.3703561264518114, + "learning_rate": 0.0020941812817778854, + "loss": 1.0414, + "step": 4815 + }, + { + "epoch": 0.37043304361202983, + "learning_rate": 0.0020938484509626584, + "loss": 1.0078, + "step": 4816 + }, + { + "epoch": 0.3705099607722483, + "learning_rate": 0.0020935155854720014, + "loss": 1.2156, + "step": 4817 + }, + { + "epoch": 0.37058687793246675, + "learning_rate": 0.0020931826853253514, + "loss": 1.2674, + "step": 4818 + }, + { + "epoch": 0.37066379509268516, + "learning_rate": 0.0020928497505421457, + "loss": 1.275, + "step": 4819 + }, + { + "epoch": 0.3707407122529036, + "learning_rate": 0.0020925167811418257, + "loss": 1.058, + "step": 4820 + }, + { + "epoch": 0.3708176294131221, + "learning_rate": 0.0020921837771438324, + "loss": 1.2275, + "step": 4821 + }, + { + "epoch": 0.37089454657334053, + "learning_rate": 0.0020918507385676117, + "loss": 1.0988, + "step": 4822 + }, + { + "epoch": 0.37097146373355894, + "learning_rate": 0.00209151766543261, + "loss": 1.2692, + "step": 4823 + }, + { + "epoch": 0.3710483808937774, + "learning_rate": 0.0020911845577582747, + "loss": 1.3823, + "step": 4824 + }, + { + "epoch": 0.37112529805399586, + "learning_rate": 0.002090851415564057, + "loss": 1.0632, + "step": 4825 + }, + { + "epoch": 0.3712022152142143, + "learning_rate": 0.00209051823886941, + "loss": 0.8388, + "step": 4826 + }, + { + "epoch": 0.3712791323744327, + "learning_rate": 0.0020901850276937874, + "loss": 0.7805, + "step": 4827 + }, + { + "epoch": 0.3713560495346512, + "learning_rate": 0.0020898517820566458, + "loss": 1.2459, + "step": 4828 + }, + { + "epoch": 0.37143296669486964, + "learning_rate": 0.002089518501977444, + "loss": 0.8033, + "step": 4829 + }, + { + "epoch": 0.3715098838550881, + "learning_rate": 0.002089185187475642, + "loss": 1.2512, + "step": 4830 + }, + { + "epoch": 0.3715868010153065, + "learning_rate": 0.0020888518385707034, + "loss": 1.0718, + "step": 4831 + }, + { + "epoch": 0.37166371817552496, + "learning_rate": 0.002088518455282092, + "loss": 1.4192, + "step": 4832 + }, + { + "epoch": 0.3717406353357434, + "learning_rate": 0.002088185037629275, + "loss": 0.953, + "step": 4833 + }, + { + "epoch": 0.3718175524959618, + "learning_rate": 0.00208785158563172, + "loss": 0.9038, + "step": 4834 + }, + { + "epoch": 0.3718944696561803, + "learning_rate": 0.002087518099308899, + "loss": 0.8231, + "step": 4835 + }, + { + "epoch": 0.37197138681639874, + "learning_rate": 0.0020871845786802836, + "loss": 1.0362, + "step": 4836 + }, + { + "epoch": 0.3720483039766172, + "learning_rate": 0.0020868510237653484, + "loss": 0.929, + "step": 4837 + }, + { + "epoch": 0.3721252211368356, + "learning_rate": 0.002086517434583571, + "loss": 1.0692, + "step": 4838 + }, + { + "epoch": 0.37220213829705406, + "learning_rate": 0.0020861838111544285, + "loss": 1.5116, + "step": 4839 + }, + { + "epoch": 0.3722790554572725, + "learning_rate": 0.002085850153497403, + "loss": 1.0373, + "step": 4840 + }, + { + "epoch": 0.372355972617491, + "learning_rate": 0.002085516461631976, + "loss": 1.1705, + "step": 4841 + }, + { + "epoch": 0.3724328897777094, + "learning_rate": 0.0020851827355776336, + "loss": 1.1292, + "step": 4842 + }, + { + "epoch": 0.37250980693792785, + "learning_rate": 0.0020848489753538604, + "loss": 0.9182, + "step": 4843 + }, + { + "epoch": 0.3725867240981463, + "learning_rate": 0.002084515180980146, + "loss": 1.3056, + "step": 4844 + }, + { + "epoch": 0.37266364125836476, + "learning_rate": 0.002084181352475981, + "loss": 1.1036, + "step": 4845 + }, + { + "epoch": 0.37274055841858317, + "learning_rate": 0.002083847489860859, + "loss": 1.0138, + "step": 4846 + }, + { + "epoch": 0.3728174755788016, + "learning_rate": 0.0020835135931542726, + "loss": 1.2371, + "step": 4847 + }, + { + "epoch": 0.3728943927390201, + "learning_rate": 0.00208317966237572, + "loss": 1.3178, + "step": 4848 + }, + { + "epoch": 0.37297130989923855, + "learning_rate": 0.0020828456975446985, + "loss": 1.6001, + "step": 4849 + }, + { + "epoch": 0.37304822705945695, + "learning_rate": 0.0020825116986807096, + "loss": 0.8634, + "step": 4850 + }, + { + "epoch": 0.3731251442196754, + "learning_rate": 0.002082177665803256, + "loss": 0.7771, + "step": 4851 + }, + { + "epoch": 0.37320206137989387, + "learning_rate": 0.0020818435989318406, + "loss": 1.1008, + "step": 4852 + }, + { + "epoch": 0.37327897854011227, + "learning_rate": 0.0020815094980859723, + "loss": 1.4486, + "step": 4853 + }, + { + "epoch": 0.37335589570033073, + "learning_rate": 0.002081175363285158, + "loss": 1.0082, + "step": 4854 + }, + { + "epoch": 0.3734328128605492, + "learning_rate": 0.0020808411945489086, + "loss": 1.3112, + "step": 4855 + }, + { + "epoch": 0.37350973002076765, + "learning_rate": 0.002080506991896736, + "loss": 0.7854, + "step": 4856 + }, + { + "epoch": 0.37358664718098605, + "learning_rate": 0.002080172755348156, + "loss": 1.2396, + "step": 4857 + }, + { + "epoch": 0.3736635643412045, + "learning_rate": 0.0020798384849226833, + "loss": 0.8316, + "step": 4858 + }, + { + "epoch": 0.37374048150142297, + "learning_rate": 0.0020795041806398375, + "loss": 0.9986, + "step": 4859 + }, + { + "epoch": 0.37381739866164143, + "learning_rate": 0.0020791698425191383, + "loss": 1.3407, + "step": 4860 + }, + { + "epoch": 0.37389431582185984, + "learning_rate": 0.002078835470580109, + "loss": 1.4374, + "step": 4861 + }, + { + "epoch": 0.3739712329820783, + "learning_rate": 0.0020785010648422734, + "loss": 1.0916, + "step": 4862 + }, + { + "epoch": 0.37404815014229675, + "learning_rate": 0.0020781666253251573, + "loss": 1.3573, + "step": 4863 + }, + { + "epoch": 0.3741250673025152, + "learning_rate": 0.0020778321520482893, + "loss": 1.072, + "step": 4864 + }, + { + "epoch": 0.3742019844627336, + "learning_rate": 0.0020774976450312, + "loss": 1.1448, + "step": 4865 + }, + { + "epoch": 0.3742789016229521, + "learning_rate": 0.0020771631042934215, + "loss": 1.396, + "step": 4866 + }, + { + "epoch": 0.37435581878317054, + "learning_rate": 0.0020768285298544873, + "loss": 1.3147, + "step": 4867 + }, + { + "epoch": 0.374432735943389, + "learning_rate": 0.002076493921733935, + "loss": 1.1093, + "step": 4868 + }, + { + "epoch": 0.3745096531036074, + "learning_rate": 0.002076159279951301, + "loss": 1.1783, + "step": 4869 + }, + { + "epoch": 0.37458657026382586, + "learning_rate": 0.0020758246045261266, + "loss": 0.8997, + "step": 4870 + }, + { + "epoch": 0.3746634874240443, + "learning_rate": 0.002075489895477953, + "loss": 1.4723, + "step": 4871 + }, + { + "epoch": 0.3747404045842628, + "learning_rate": 0.002075155152826325, + "loss": 1.376, + "step": 4872 + }, + { + "epoch": 0.3748173217444812, + "learning_rate": 0.002074820376590788, + "loss": 1.0331, + "step": 4873 + }, + { + "epoch": 0.37489423890469964, + "learning_rate": 0.0020744855667908903, + "loss": 0.9642, + "step": 4874 + }, + { + "epoch": 0.3749711560649181, + "learning_rate": 0.0020741507234461813, + "loss": 0.8382, + "step": 4875 + }, + { + "epoch": 0.3750480732251365, + "learning_rate": 0.002073815846576214, + "loss": 1.3105, + "step": 4876 + }, + { + "epoch": 0.37512499038535496, + "learning_rate": 0.0020734809362005406, + "loss": 1.418, + "step": 4877 + }, + { + "epoch": 0.3752019075455734, + "learning_rate": 0.002073145992338718, + "loss": 1.347, + "step": 4878 + }, + { + "epoch": 0.3752788247057919, + "learning_rate": 0.0020728110150103033, + "loss": 1.0774, + "step": 4879 + }, + { + "epoch": 0.3753557418660103, + "learning_rate": 0.002072476004234857, + "loss": 1.0817, + "step": 4880 + }, + { + "epoch": 0.37543265902622874, + "learning_rate": 0.0020721409600319403, + "loss": 0.6908, + "step": 4881 + }, + { + "epoch": 0.3755095761864472, + "learning_rate": 0.0020718058824211157, + "loss": 1.1856, + "step": 4882 + }, + { + "epoch": 0.37558649334666566, + "learning_rate": 0.0020714707714219504, + "loss": 1.329, + "step": 4883 + }, + { + "epoch": 0.37566341050688407, + "learning_rate": 0.0020711356270540116, + "loss": 1.0417, + "step": 4884 + }, + { + "epoch": 0.3757403276671025, + "learning_rate": 0.002070800449336867, + "loss": 1.2722, + "step": 4885 + }, + { + "epoch": 0.375817244827321, + "learning_rate": 0.0020704652382900904, + "loss": 1.1854, + "step": 4886 + }, + { + "epoch": 0.37589416198753944, + "learning_rate": 0.002070129993933254, + "loss": 1.0445, + "step": 4887 + }, + { + "epoch": 0.37597107914775785, + "learning_rate": 0.0020697947162859327, + "loss": 1.4291, + "step": 4888 + }, + { + "epoch": 0.3760479963079763, + "learning_rate": 0.002069459405367704, + "loss": 1.4924, + "step": 4889 + }, + { + "epoch": 0.37612491346819477, + "learning_rate": 0.002069124061198147, + "loss": 1.0078, + "step": 4890 + }, + { + "epoch": 0.3762018306284132, + "learning_rate": 0.0020687886837968436, + "loss": 1.2222, + "step": 4891 + }, + { + "epoch": 0.37627874778863163, + "learning_rate": 0.0020684532731833753, + "loss": 1.2222, + "step": 4892 + }, + { + "epoch": 0.3763556649488501, + "learning_rate": 0.002068117829377329, + "loss": 1.1769, + "step": 4893 + }, + { + "epoch": 0.37643258210906855, + "learning_rate": 0.00206778235239829, + "loss": 1.2531, + "step": 4894 + }, + { + "epoch": 0.37650949926928695, + "learning_rate": 0.0020674468422658475, + "loss": 1.5151, + "step": 4895 + }, + { + "epoch": 0.3765864164295054, + "learning_rate": 0.002067111298999593, + "loss": 1.4172, + "step": 4896 + }, + { + "epoch": 0.37666333358972387, + "learning_rate": 0.0020667757226191185, + "loss": 1.1472, + "step": 4897 + }, + { + "epoch": 0.37674025074994233, + "learning_rate": 0.0020664401131440186, + "loss": 1.3908, + "step": 4898 + }, + { + "epoch": 0.37681716791016073, + "learning_rate": 0.0020661044705938907, + "loss": 0.8493, + "step": 4899 + }, + { + "epoch": 0.3768940850703792, + "learning_rate": 0.0020657687949883325, + "loss": 1.1832, + "step": 4900 + }, + { + "epoch": 0.37697100223059765, + "learning_rate": 0.0020654330863469446, + "loss": 1.0361, + "step": 4901 + }, + { + "epoch": 0.3770479193908161, + "learning_rate": 0.0020650973446893293, + "loss": 1.26, + "step": 4902 + }, + { + "epoch": 0.3771248365510345, + "learning_rate": 0.0020647615700350916, + "loss": 0.8732, + "step": 4903 + }, + { + "epoch": 0.377201753711253, + "learning_rate": 0.002064425762403837, + "loss": 0.8881, + "step": 4904 + }, + { + "epoch": 0.37727867087147143, + "learning_rate": 0.002064089921815174, + "loss": 1.1377, + "step": 4905 + }, + { + "epoch": 0.3773555880316899, + "learning_rate": 0.0020637540482887125, + "loss": 1.2488, + "step": 4906 + }, + { + "epoch": 0.3774325051919083, + "learning_rate": 0.002063418141844065, + "loss": 1.2511, + "step": 4907 + }, + { + "epoch": 0.37750942235212676, + "learning_rate": 0.0020630822025008446, + "loss": 1.0231, + "step": 4908 + }, + { + "epoch": 0.3775863395123452, + "learning_rate": 0.002062746230278667, + "loss": 0.8397, + "step": 4909 + }, + { + "epoch": 0.3776632566725637, + "learning_rate": 0.002062410225197152, + "loss": 1.0901, + "step": 4910 + }, + { + "epoch": 0.3777401738327821, + "learning_rate": 0.0020620741872759167, + "loss": 1.2004, + "step": 4911 + }, + { + "epoch": 0.37781709099300054, + "learning_rate": 0.0020617381165345845, + "loss": 1.1119, + "step": 4912 + }, + { + "epoch": 0.377894008153219, + "learning_rate": 0.0020614020129927774, + "loss": 1.1782, + "step": 4913 + }, + { + "epoch": 0.3779709253134374, + "learning_rate": 0.0020610658766701223, + "loss": 1.2043, + "step": 4914 + }, + { + "epoch": 0.37804784247365586, + "learning_rate": 0.002060729707586246, + "loss": 1.001, + "step": 4915 + }, + { + "epoch": 0.3781247596338743, + "learning_rate": 0.002060393505760777, + "loss": 1.0096, + "step": 4916 + }, + { + "epoch": 0.3782016767940928, + "learning_rate": 0.002060057271213348, + "loss": 1.0906, + "step": 4917 + }, + { + "epoch": 0.3782785939543112, + "learning_rate": 0.0020597210039635908, + "loss": 1.4346, + "step": 4918 + }, + { + "epoch": 0.37835551111452964, + "learning_rate": 0.002059384704031141, + "loss": 1.2484, + "step": 4919 + }, + { + "epoch": 0.3784324282747481, + "learning_rate": 0.0020590483714356352, + "loss": 0.989, + "step": 4920 + }, + { + "epoch": 0.37850934543496656, + "learning_rate": 0.0020587120061967127, + "loss": 1.3396, + "step": 4921 + }, + { + "epoch": 0.37858626259518496, + "learning_rate": 0.0020583756083340137, + "loss": 1.2484, + "step": 4922 + }, + { + "epoch": 0.3786631797554034, + "learning_rate": 0.002058039177867181, + "loss": 1.3944, + "step": 4923 + }, + { + "epoch": 0.3787400969156219, + "learning_rate": 0.002057702714815859, + "loss": 1.0276, + "step": 4924 + }, + { + "epoch": 0.37881701407584034, + "learning_rate": 0.002057366219199694, + "loss": 1.53, + "step": 4925 + }, + { + "epoch": 0.37889393123605875, + "learning_rate": 0.002057029691038335, + "loss": 0.8265, + "step": 4926 + }, + { + "epoch": 0.3789708483962772, + "learning_rate": 0.002056693130351432, + "loss": 1.056, + "step": 4927 + }, + { + "epoch": 0.37904776555649566, + "learning_rate": 0.0020563565371586367, + "loss": 1.2299, + "step": 4928 + }, + { + "epoch": 0.3791246827167141, + "learning_rate": 0.0020560199114796032, + "loss": 0.9996, + "step": 4929 + }, + { + "epoch": 0.3792015998769325, + "learning_rate": 0.0020556832533339883, + "loss": 1.2088, + "step": 4930 + }, + { + "epoch": 0.379278517037151, + "learning_rate": 0.002055346562741448, + "loss": 0.9918, + "step": 4931 + }, + { + "epoch": 0.37935543419736945, + "learning_rate": 0.0020550098397216442, + "loss": 1.0546, + "step": 4932 + }, + { + "epoch": 0.3794323513575879, + "learning_rate": 0.002054673084294237, + "loss": 1.1932, + "step": 4933 + }, + { + "epoch": 0.3795092685178063, + "learning_rate": 0.0020543362964788904, + "loss": 1.0308, + "step": 4934 + }, + { + "epoch": 0.37958618567802477, + "learning_rate": 0.002053999476295269, + "loss": 1.1372, + "step": 4935 + }, + { + "epoch": 0.3796631028382432, + "learning_rate": 0.002053662623763042, + "loss": 1.0867, + "step": 4936 + }, + { + "epoch": 0.37974001999846163, + "learning_rate": 0.0020533257389018767, + "loss": 0.9925, + "step": 4937 + }, + { + "epoch": 0.3798169371586801, + "learning_rate": 0.002052988821731445, + "loss": 1.2389, + "step": 4938 + }, + { + "epoch": 0.37989385431889855, + "learning_rate": 0.0020526518722714195, + "loss": 1.3574, + "step": 4939 + }, + { + "epoch": 0.379970771479117, + "learning_rate": 0.002052314890541475, + "loss": 1.0758, + "step": 4940 + }, + { + "epoch": 0.3800476886393354, + "learning_rate": 0.0020519778765612897, + "loss": 1.0655, + "step": 4941 + }, + { + "epoch": 0.38012460579955387, + "learning_rate": 0.00205164083035054, + "loss": 1.2907, + "step": 4942 + }, + { + "epoch": 0.38020152295977233, + "learning_rate": 0.002051303751928907, + "loss": 1.4228, + "step": 4943 + }, + { + "epoch": 0.3802784401199908, + "learning_rate": 0.0020509666413160736, + "loss": 1.1147, + "step": 4944 + }, + { + "epoch": 0.3803553572802092, + "learning_rate": 0.0020506294985317246, + "loss": 1.4504, + "step": 4945 + }, + { + "epoch": 0.38043227444042765, + "learning_rate": 0.0020502923235955445, + "loss": 1.242, + "step": 4946 + }, + { + "epoch": 0.3805091916006461, + "learning_rate": 0.0020499551165272223, + "loss": 0.7567, + "step": 4947 + }, + { + "epoch": 0.38058610876086457, + "learning_rate": 0.0020496178773464473, + "loss": 0.9129, + "step": 4948 + }, + { + "epoch": 0.380663025921083, + "learning_rate": 0.002049280606072912, + "loss": 1.1387, + "step": 4949 + }, + { + "epoch": 0.38073994308130144, + "learning_rate": 0.00204894330272631, + "loss": 1.184, + "step": 4950 + }, + { + "epoch": 0.3808168602415199, + "learning_rate": 0.002048605967326336, + "loss": 1.222, + "step": 4951 + }, + { + "epoch": 0.38089377740173835, + "learning_rate": 0.002048268599892688, + "loss": 0.8492, + "step": 4952 + }, + { + "epoch": 0.38097069456195676, + "learning_rate": 0.0020479312004450653, + "loss": 1.3435, + "step": 4953 + }, + { + "epoch": 0.3810476117221752, + "learning_rate": 0.0020475937690031683, + "loss": 1.1365, + "step": 4954 + }, + { + "epoch": 0.3811245288823937, + "learning_rate": 0.002047256305586701, + "loss": 1.2877, + "step": 4955 + }, + { + "epoch": 0.3812014460426121, + "learning_rate": 0.0020469188102153675, + "loss": 0.9223, + "step": 4956 + }, + { + "epoch": 0.38127836320283054, + "learning_rate": 0.0020465812829088743, + "loss": 1.0167, + "step": 4957 + }, + { + "epoch": 0.381355280363049, + "learning_rate": 0.0020462437236869304, + "loss": 1.1759, + "step": 4958 + }, + { + "epoch": 0.38143219752326746, + "learning_rate": 0.0020459061325692465, + "loss": 1.2252, + "step": 4959 + }, + { + "epoch": 0.38150911468348586, + "learning_rate": 0.0020455685095755343, + "loss": 0.9734, + "step": 4960 + }, + { + "epoch": 0.3815860318437043, + "learning_rate": 0.0020452308547255086, + "loss": 1.2894, + "step": 4961 + }, + { + "epoch": 0.3816629490039228, + "learning_rate": 0.0020448931680388843, + "loss": 1.0756, + "step": 4962 + }, + { + "epoch": 0.38173986616414124, + "learning_rate": 0.002044555449535381, + "loss": 0.8025, + "step": 4963 + }, + { + "epoch": 0.38181678332435964, + "learning_rate": 0.002044217699234717, + "loss": 1.1754, + "step": 4964 + }, + { + "epoch": 0.3818937004845781, + "learning_rate": 0.0020438799171566147, + "loss": 1.1561, + "step": 4965 + }, + { + "epoch": 0.38197061764479656, + "learning_rate": 0.002043542103320797, + "loss": 1.0023, + "step": 4966 + }, + { + "epoch": 0.382047534805015, + "learning_rate": 0.002043204257746989, + "loss": 1.3553, + "step": 4967 + }, + { + "epoch": 0.3821244519652334, + "learning_rate": 0.002042866380454919, + "loss": 1.31, + "step": 4968 + }, + { + "epoch": 0.3822013691254519, + "learning_rate": 0.002042528471464315, + "loss": 1.6954, + "step": 4969 + }, + { + "epoch": 0.38227828628567034, + "learning_rate": 0.002042190530794908, + "loss": 1.036, + "step": 4970 + }, + { + "epoch": 0.3823552034458888, + "learning_rate": 0.002041852558466431, + "loss": 1.1154, + "step": 4971 + }, + { + "epoch": 0.3824321206061072, + "learning_rate": 0.002041514554498619, + "loss": 1.462, + "step": 4972 + }, + { + "epoch": 0.38250903776632567, + "learning_rate": 0.0020411765189112066, + "loss": 1.0796, + "step": 4973 + }, + { + "epoch": 0.3825859549265441, + "learning_rate": 0.002040838451723934, + "loss": 1.0786, + "step": 4974 + }, + { + "epoch": 0.3826628720867626, + "learning_rate": 0.00204050035295654, + "loss": 1.1372, + "step": 4975 + }, + { + "epoch": 0.382739789246981, + "learning_rate": 0.0020401622226287677, + "loss": 0.897, + "step": 4976 + }, + { + "epoch": 0.38281670640719945, + "learning_rate": 0.00203982406076036, + "loss": 1.1475, + "step": 4977 + }, + { + "epoch": 0.3828936235674179, + "learning_rate": 0.002039485867371063, + "loss": 0.601, + "step": 4978 + }, + { + "epoch": 0.3829705407276363, + "learning_rate": 0.002039147642480624, + "loss": 0.9801, + "step": 4979 + }, + { + "epoch": 0.38304745788785477, + "learning_rate": 0.002038809386108792, + "loss": 1.0897, + "step": 4980 + }, + { + "epoch": 0.38312437504807323, + "learning_rate": 0.0020384710982753185, + "loss": 0.9211, + "step": 4981 + }, + { + "epoch": 0.3832012922082917, + "learning_rate": 0.002038132778999956, + "loss": 1.1034, + "step": 4982 + }, + { + "epoch": 0.3832782093685101, + "learning_rate": 0.0020377944283024607, + "loss": 0.9202, + "step": 4983 + }, + { + "epoch": 0.38335512652872855, + "learning_rate": 0.0020374560462025873, + "loss": 1.1294, + "step": 4984 + }, + { + "epoch": 0.383432043688947, + "learning_rate": 0.0020371176327200957, + "loss": 1.0268, + "step": 4985 + }, + { + "epoch": 0.38350896084916547, + "learning_rate": 0.0020367791878747456, + "loss": 1.0809, + "step": 4986 + }, + { + "epoch": 0.3835858780093839, + "learning_rate": 0.0020364407116862998, + "loss": 1.0792, + "step": 4987 + }, + { + "epoch": 0.38366279516960233, + "learning_rate": 0.002036102204174521, + "loss": 1.3723, + "step": 4988 + }, + { + "epoch": 0.3837397123298208, + "learning_rate": 0.002035763665359176, + "loss": 1.2951, + "step": 4989 + }, + { + "epoch": 0.38381662949003925, + "learning_rate": 0.0020354250952600325, + "loss": 1.113, + "step": 4990 + }, + { + "epoch": 0.38389354665025766, + "learning_rate": 0.0020350864938968594, + "loss": 1.2646, + "step": 4991 + }, + { + "epoch": 0.3839704638104761, + "learning_rate": 0.0020347478612894286, + "loss": 1.4198, + "step": 4992 + }, + { + "epoch": 0.3840473809706946, + "learning_rate": 0.002034409197457512, + "loss": 1.0156, + "step": 4993 + }, + { + "epoch": 0.38412429813091303, + "learning_rate": 0.0020340705024208864, + "loss": 0.9134, + "step": 4994 + }, + { + "epoch": 0.38420121529113144, + "learning_rate": 0.002033731776199328, + "loss": 1.1205, + "step": 4995 + }, + { + "epoch": 0.3842781324513499, + "learning_rate": 0.002033393018812614, + "loss": 1.2747, + "step": 4996 + }, + { + "epoch": 0.38435504961156836, + "learning_rate": 0.0020330542302805254, + "loss": 1.4218, + "step": 4997 + }, + { + "epoch": 0.38443196677178676, + "learning_rate": 0.0020327154106228457, + "loss": 1.263, + "step": 4998 + }, + { + "epoch": 0.3845088839320052, + "learning_rate": 0.0020323765598593575, + "loss": 1.347, + "step": 4999 + }, + { + "epoch": 0.3845858010922237, + "learning_rate": 0.0020320376780098475, + "loss": 1.3136, + "step": 5000 + }, + { + "epoch": 0.38466271825244214, + "learning_rate": 0.0020316987650941024, + "loss": 1.0214, + "step": 5001 + }, + { + "epoch": 0.38473963541266054, + "learning_rate": 0.002031359821131913, + "loss": 1.0561, + "step": 5002 + }, + { + "epoch": 0.384816552572879, + "learning_rate": 0.0020310208461430694, + "loss": 1.1358, + "step": 5003 + }, + { + "epoch": 0.38489346973309746, + "learning_rate": 0.0020306818401473648, + "loss": 0.9896, + "step": 5004 + }, + { + "epoch": 0.3849703868933159, + "learning_rate": 0.0020303428031645952, + "loss": 0.966, + "step": 5005 + }, + { + "epoch": 0.3850473040535343, + "learning_rate": 0.0020300037352145566, + "loss": 1.275, + "step": 5006 + }, + { + "epoch": 0.3851242212137528, + "learning_rate": 0.0020296646363170473, + "loss": 1.2206, + "step": 5007 + }, + { + "epoch": 0.38520113837397124, + "learning_rate": 0.0020293255064918678, + "loss": 1.3341, + "step": 5008 + }, + { + "epoch": 0.3852780555341897, + "learning_rate": 0.0020289863457588204, + "loss": 1.0249, + "step": 5009 + }, + { + "epoch": 0.3853549726944081, + "learning_rate": 0.002028647154137709, + "loss": 0.8251, + "step": 5010 + }, + { + "epoch": 0.38543188985462656, + "learning_rate": 0.0020283079316483392, + "loss": 1.453, + "step": 5011 + }, + { + "epoch": 0.385508807014845, + "learning_rate": 0.0020279686783105186, + "loss": 1.4294, + "step": 5012 + }, + { + "epoch": 0.3855857241750635, + "learning_rate": 0.002027629394144057, + "loss": 1.3009, + "step": 5013 + }, + { + "epoch": 0.3856626413352819, + "learning_rate": 0.0020272900791687655, + "loss": 1.0109, + "step": 5014 + }, + { + "epoch": 0.38573955849550035, + "learning_rate": 0.0020269507334044557, + "loss": 1.0238, + "step": 5015 + }, + { + "epoch": 0.3858164756557188, + "learning_rate": 0.0020266113568709445, + "loss": 1.1444, + "step": 5016 + }, + { + "epoch": 0.3858933928159372, + "learning_rate": 0.0020262719495880467, + "loss": 1.1523, + "step": 5017 + }, + { + "epoch": 0.38597030997615567, + "learning_rate": 0.0020259325115755816, + "loss": 1.1241, + "step": 5018 + }, + { + "epoch": 0.3860472271363741, + "learning_rate": 0.0020255930428533684, + "loss": 1.1907, + "step": 5019 + }, + { + "epoch": 0.3861241442965926, + "learning_rate": 0.0020252535434412302, + "loss": 1.1154, + "step": 5020 + }, + { + "epoch": 0.386201061456811, + "learning_rate": 0.00202491401335899, + "loss": 1.5667, + "step": 5021 + }, + { + "epoch": 0.38627797861702945, + "learning_rate": 0.0020245744526264744, + "loss": 1.1757, + "step": 5022 + }, + { + "epoch": 0.3863548957772479, + "learning_rate": 0.0020242348612635086, + "loss": 1.249, + "step": 5023 + }, + { + "epoch": 0.38643181293746637, + "learning_rate": 0.0020238952392899233, + "loss": 1.3065, + "step": 5024 + }, + { + "epoch": 0.38650873009768477, + "learning_rate": 0.0020235555867255495, + "loss": 1.2946, + "step": 5025 + }, + { + "epoch": 0.38658564725790323, + "learning_rate": 0.0020232159035902184, + "loss": 1.0711, + "step": 5026 + }, + { + "epoch": 0.3866625644181217, + "learning_rate": 0.002022876189903766, + "loss": 1.0078, + "step": 5027 + }, + { + "epoch": 0.38673948157834015, + "learning_rate": 0.0020225364456860277, + "loss": 1.0949, + "step": 5028 + }, + { + "epoch": 0.38681639873855855, + "learning_rate": 0.002022196670956842, + "loss": 1.0695, + "step": 5029 + }, + { + "epoch": 0.386893315898777, + "learning_rate": 0.0020218568657360486, + "loss": 1.1325, + "step": 5030 + }, + { + "epoch": 0.38697023305899547, + "learning_rate": 0.0020215170300434885, + "loss": 1.1046, + "step": 5031 + }, + { + "epoch": 0.38704715021921393, + "learning_rate": 0.0020211771638990058, + "loss": 1.2303, + "step": 5032 + }, + { + "epoch": 0.38712406737943234, + "learning_rate": 0.0020208372673224456, + "loss": 0.9923, + "step": 5033 + }, + { + "epoch": 0.3872009845396508, + "learning_rate": 0.002020497340333654, + "loss": 0.8864, + "step": 5034 + }, + { + "epoch": 0.38727790169986925, + "learning_rate": 0.0020201573829524803, + "loss": 1.0268, + "step": 5035 + }, + { + "epoch": 0.3873548188600877, + "learning_rate": 0.0020198173951987753, + "loss": 1.0813, + "step": 5036 + }, + { + "epoch": 0.3874317360203061, + "learning_rate": 0.0020194773770923907, + "loss": 1.0885, + "step": 5037 + }, + { + "epoch": 0.3875086531805246, + "learning_rate": 0.0020191373286531805, + "loss": 1.3219, + "step": 5038 + }, + { + "epoch": 0.38758557034074304, + "learning_rate": 0.0020187972499010006, + "loss": 0.8751, + "step": 5039 + }, + { + "epoch": 0.38766248750096144, + "learning_rate": 0.0020184571408557095, + "loss": 1.1717, + "step": 5040 + }, + { + "epoch": 0.3877394046611799, + "learning_rate": 0.0020181170015371643, + "loss": 1.1097, + "step": 5041 + }, + { + "epoch": 0.38781632182139836, + "learning_rate": 0.0020177768319652283, + "loss": 1.1087, + "step": 5042 + }, + { + "epoch": 0.3878932389816168, + "learning_rate": 0.0020174366321597625, + "loss": 0.9766, + "step": 5043 + }, + { + "epoch": 0.3879701561418352, + "learning_rate": 0.0020170964021406336, + "loss": 1.1057, + "step": 5044 + }, + { + "epoch": 0.3880470733020537, + "learning_rate": 0.0020167561419277063, + "loss": 1.1334, + "step": 5045 + }, + { + "epoch": 0.38812399046227214, + "learning_rate": 0.0020164158515408497, + "loss": 0.9277, + "step": 5046 + }, + { + "epoch": 0.3882009076224906, + "learning_rate": 0.002016075530999933, + "loss": 1.266, + "step": 5047 + }, + { + "epoch": 0.388277824782709, + "learning_rate": 0.0020157351803248283, + "loss": 0.7329, + "step": 5048 + }, + { + "epoch": 0.38835474194292746, + "learning_rate": 0.002015394799535409, + "loss": 1.0342, + "step": 5049 + }, + { + "epoch": 0.3884316591031459, + "learning_rate": 0.00201505438865155, + "loss": 1.0407, + "step": 5050 + }, + { + "epoch": 0.3885085762633644, + "learning_rate": 0.0020147139476931284, + "loss": 0.8504, + "step": 5051 + }, + { + "epoch": 0.3885854934235828, + "learning_rate": 0.0020143734766800234, + "loss": 1.3714, + "step": 5052 + }, + { + "epoch": 0.38866241058380124, + "learning_rate": 0.002014032975632115, + "loss": 1.4464, + "step": 5053 + }, + { + "epoch": 0.3887393277440197, + "learning_rate": 0.002013692444569285, + "loss": 1.2104, + "step": 5054 + }, + { + "epoch": 0.38881624490423816, + "learning_rate": 0.002013351883511418, + "loss": 1.2731, + "step": 5055 + }, + { + "epoch": 0.38889316206445657, + "learning_rate": 0.0020130112924784, + "loss": 0.982, + "step": 5056 + }, + { + "epoch": 0.388970079224675, + "learning_rate": 0.002012670671490117, + "loss": 1.4885, + "step": 5057 + }, + { + "epoch": 0.3890469963848935, + "learning_rate": 0.00201233002056646, + "loss": 1.1467, + "step": 5058 + }, + { + "epoch": 0.3891239135451119, + "learning_rate": 0.002011989339727319, + "loss": 0.7181, + "step": 5059 + }, + { + "epoch": 0.38920083070533035, + "learning_rate": 0.0020116486289925867, + "loss": 1.0285, + "step": 5060 + }, + { + "epoch": 0.3892777478655488, + "learning_rate": 0.0020113078883821577, + "loss": 1.1083, + "step": 5061 + }, + { + "epoch": 0.38935466502576727, + "learning_rate": 0.0020109671179159285, + "loss": 1.218, + "step": 5062 + }, + { + "epoch": 0.38943158218598567, + "learning_rate": 0.0020106263176137965, + "loss": 1.3801, + "step": 5063 + }, + { + "epoch": 0.38950849934620413, + "learning_rate": 0.002010285487495662, + "loss": 0.8797, + "step": 5064 + }, + { + "epoch": 0.3895854165064226, + "learning_rate": 0.0020099446275814257, + "loss": 1.1166, + "step": 5065 + }, + { + "epoch": 0.38966233366664105, + "learning_rate": 0.0020096037378909914, + "loss": 1.5569, + "step": 5066 + }, + { + "epoch": 0.38973925082685945, + "learning_rate": 0.0020092628184442637, + "loss": 1.3894, + "step": 5067 + }, + { + "epoch": 0.3898161679870779, + "learning_rate": 0.002008921869261149, + "loss": 1.2409, + "step": 5068 + }, + { + "epoch": 0.38989308514729637, + "learning_rate": 0.0020085808903615564, + "loss": 1.0281, + "step": 5069 + }, + { + "epoch": 0.38997000230751483, + "learning_rate": 0.002008239881765395, + "loss": 1.0496, + "step": 5070 + }, + { + "epoch": 0.39004691946773323, + "learning_rate": 0.002007898843492578, + "loss": 1.1025, + "step": 5071 + }, + { + "epoch": 0.3901238366279517, + "learning_rate": 0.0020075577755630175, + "loss": 1.0, + "step": 5072 + }, + { + "epoch": 0.39020075378817015, + "learning_rate": 0.00200721667799663, + "loss": 1.55, + "step": 5073 + }, + { + "epoch": 0.3902776709483886, + "learning_rate": 0.0020068755508133316, + "loss": 1.0485, + "step": 5074 + }, + { + "epoch": 0.390354588108607, + "learning_rate": 0.0020065343940330422, + "loss": 1.0761, + "step": 5075 + }, + { + "epoch": 0.3904315052688255, + "learning_rate": 0.002006193207675681, + "loss": 1.1752, + "step": 5076 + }, + { + "epoch": 0.39050842242904393, + "learning_rate": 0.0020058519917611714, + "loss": 1.0596, + "step": 5077 + }, + { + "epoch": 0.39058533958926234, + "learning_rate": 0.002005510746309437, + "loss": 1.1141, + "step": 5078 + }, + { + "epoch": 0.3906622567494808, + "learning_rate": 0.0020051694713404023, + "loss": 0.8887, + "step": 5079 + }, + { + "epoch": 0.39073917390969926, + "learning_rate": 0.002004828166873996, + "loss": 1.158, + "step": 5080 + }, + { + "epoch": 0.3908160910699177, + "learning_rate": 0.002004486832930147, + "loss": 1.6662, + "step": 5081 + }, + { + "epoch": 0.3908930082301361, + "learning_rate": 0.0020041454695287867, + "loss": 1.2052, + "step": 5082 + }, + { + "epoch": 0.3909699253903546, + "learning_rate": 0.0020038040766898464, + "loss": 1.0947, + "step": 5083 + }, + { + "epoch": 0.39104684255057304, + "learning_rate": 0.002003462654433261, + "loss": 1.421, + "step": 5084 + }, + { + "epoch": 0.3911237597107915, + "learning_rate": 0.0020031212027789664, + "loss": 1.3049, + "step": 5085 + }, + { + "epoch": 0.3912006768710099, + "learning_rate": 0.002002779721746901, + "loss": 0.9061, + "step": 5086 + }, + { + "epoch": 0.39127759403122836, + "learning_rate": 0.002002438211357003, + "loss": 0.6874, + "step": 5087 + }, + { + "epoch": 0.3913545111914468, + "learning_rate": 0.002002096671629215, + "loss": 1.3376, + "step": 5088 + }, + { + "epoch": 0.3914314283516653, + "learning_rate": 0.0020017551025834786, + "loss": 1.5181, + "step": 5089 + }, + { + "epoch": 0.3915083455118837, + "learning_rate": 0.0020014135042397386, + "loss": 1.1969, + "step": 5090 + }, + { + "epoch": 0.39158526267210214, + "learning_rate": 0.002001071876617942, + "loss": 1.165, + "step": 5091 + }, + { + "epoch": 0.3916621798323206, + "learning_rate": 0.002000730219738036, + "loss": 0.9387, + "step": 5092 + }, + { + "epoch": 0.39173909699253906, + "learning_rate": 0.0020003885336199708, + "loss": 1.0339, + "step": 5093 + }, + { + "epoch": 0.39181601415275746, + "learning_rate": 0.0020000468182836974, + "loss": 0.985, + "step": 5094 + }, + { + "epoch": 0.3918929313129759, + "learning_rate": 0.0019997050737491696, + "loss": 1.1823, + "step": 5095 + }, + { + "epoch": 0.3919698484731944, + "learning_rate": 0.0019993633000363414, + "loss": 1.2862, + "step": 5096 + }, + { + "epoch": 0.39204676563341284, + "learning_rate": 0.00199902149716517, + "loss": 0.7466, + "step": 5097 + }, + { + "epoch": 0.39212368279363125, + "learning_rate": 0.0019986796651556126, + "loss": 1.0475, + "step": 5098 + }, + { + "epoch": 0.3922005999538497, + "learning_rate": 0.0019983378040276307, + "loss": 1.203, + "step": 5099 + }, + { + "epoch": 0.39227751711406816, + "learning_rate": 0.0019979959138011847, + "loss": 1.3668, + "step": 5100 + }, + { + "epoch": 0.39235443427428657, + "learning_rate": 0.001997653994496238, + "loss": 1.1902, + "step": 5101 + }, + { + "epoch": 0.392431351434505, + "learning_rate": 0.0019973120461327567, + "loss": 1.1621, + "step": 5102 + }, + { + "epoch": 0.3925082685947235, + "learning_rate": 0.0019969700687307054, + "loss": 1.4672, + "step": 5103 + }, + { + "epoch": 0.39258518575494195, + "learning_rate": 0.001996628062310055, + "loss": 1.1216, + "step": 5104 + }, + { + "epoch": 0.39266210291516035, + "learning_rate": 0.0019962860268907742, + "loss": 1.1515, + "step": 5105 + }, + { + "epoch": 0.3927390200753788, + "learning_rate": 0.001995943962492835, + "loss": 1.3502, + "step": 5106 + }, + { + "epoch": 0.39281593723559727, + "learning_rate": 0.0019956018691362114, + "loss": 1.1658, + "step": 5107 + }, + { + "epoch": 0.3928928543958157, + "learning_rate": 0.0019952597468408774, + "loss": 1.0952, + "step": 5108 + }, + { + "epoch": 0.39296977155603413, + "learning_rate": 0.001994917595626811, + "loss": 1.0106, + "step": 5109 + }, + { + "epoch": 0.3930466887162526, + "learning_rate": 0.0019945754155139906, + "loss": 1.2692, + "step": 5110 + }, + { + "epoch": 0.39312360587647105, + "learning_rate": 0.001994233206522396, + "loss": 0.8599, + "step": 5111 + }, + { + "epoch": 0.3932005230366895, + "learning_rate": 0.0019938909686720097, + "loss": 1.0223, + "step": 5112 + }, + { + "epoch": 0.3932774401969079, + "learning_rate": 0.0019935487019828147, + "loss": 0.9486, + "step": 5113 + }, + { + "epoch": 0.39335435735712637, + "learning_rate": 0.0019932064064747965, + "loss": 1.1028, + "step": 5114 + }, + { + "epoch": 0.39343127451734483, + "learning_rate": 0.0019928640821679426, + "loss": 1.1744, + "step": 5115 + }, + { + "epoch": 0.3935081916775633, + "learning_rate": 0.0019925217290822405, + "loss": 1.1037, + "step": 5116 + }, + { + "epoch": 0.3935851088377817, + "learning_rate": 0.0019921793472376824, + "loss": 0.9918, + "step": 5117 + }, + { + "epoch": 0.39366202599800015, + "learning_rate": 0.001991836936654259, + "loss": 1.2017, + "step": 5118 + }, + { + "epoch": 0.3937389431582186, + "learning_rate": 0.0019914944973519638, + "loss": 1.2374, + "step": 5119 + }, + { + "epoch": 0.393815860318437, + "learning_rate": 0.001991152029350793, + "loss": 1.2245, + "step": 5120 + }, + { + "epoch": 0.3938927774786555, + "learning_rate": 0.0019908095326707435, + "loss": 0.9647, + "step": 5121 + }, + { + "epoch": 0.39396969463887394, + "learning_rate": 0.0019904670073318136, + "loss": 1.2698, + "step": 5122 + }, + { + "epoch": 0.3940466117990924, + "learning_rate": 0.001990124453354004, + "loss": 0.8463, + "step": 5123 + }, + { + "epoch": 0.3941235289593108, + "learning_rate": 0.001989781870757317, + "loss": 1.1296, + "step": 5124 + }, + { + "epoch": 0.39420044611952926, + "learning_rate": 0.001989439259561756, + "loss": 1.0608, + "step": 5125 + }, + { + "epoch": 0.3942773632797477, + "learning_rate": 0.001989096619787326, + "loss": 1.0291, + "step": 5126 + }, + { + "epoch": 0.3943542804399662, + "learning_rate": 0.0019887539514540356, + "loss": 0.788, + "step": 5127 + }, + { + "epoch": 0.3944311976001846, + "learning_rate": 0.0019884112545818927, + "loss": 0.974, + "step": 5128 + }, + { + "epoch": 0.39450811476040304, + "learning_rate": 0.0019880685291909068, + "loss": 1.3288, + "step": 5129 + }, + { + "epoch": 0.3945850319206215, + "learning_rate": 0.001987725775301091, + "loss": 1.181, + "step": 5130 + }, + { + "epoch": 0.39466194908083996, + "learning_rate": 0.001987382992932459, + "loss": 1.2275, + "step": 5131 + }, + { + "epoch": 0.39473886624105836, + "learning_rate": 0.001987040182105026, + "loss": 1.5413, + "step": 5132 + }, + { + "epoch": 0.3948157834012768, + "learning_rate": 0.00198669734283881, + "loss": 1.2808, + "step": 5133 + }, + { + "epoch": 0.3948927005614953, + "learning_rate": 0.001986354475153828, + "loss": 0.9555, + "step": 5134 + }, + { + "epoch": 0.39496961772171374, + "learning_rate": 0.001986011579070101, + "loss": 0.8047, + "step": 5135 + }, + { + "epoch": 0.39504653488193214, + "learning_rate": 0.0019856686546076523, + "loss": 1.4716, + "step": 5136 + }, + { + "epoch": 0.3951234520421506, + "learning_rate": 0.001985325701786505, + "loss": 0.8409, + "step": 5137 + }, + { + "epoch": 0.39520036920236906, + "learning_rate": 0.001984982720626683, + "loss": 0.8102, + "step": 5138 + }, + { + "epoch": 0.39527728636258747, + "learning_rate": 0.0019846397111482152, + "loss": 1.1077, + "step": 5139 + }, + { + "epoch": 0.3953542035228059, + "learning_rate": 0.0019842966733711297, + "loss": 1.0789, + "step": 5140 + }, + { + "epoch": 0.3954311206830244, + "learning_rate": 0.0019839536073154563, + "loss": 0.9913, + "step": 5141 + }, + { + "epoch": 0.39550803784324284, + "learning_rate": 0.0019836105130012277, + "loss": 1.097, + "step": 5142 + }, + { + "epoch": 0.39558495500346125, + "learning_rate": 0.001983267390448477, + "loss": 1.2007, + "step": 5143 + }, + { + "epoch": 0.3956618721636797, + "learning_rate": 0.00198292423967724, + "loss": 0.8776, + "step": 5144 + }, + { + "epoch": 0.39573878932389817, + "learning_rate": 0.001982581060707553, + "loss": 1.526, + "step": 5145 + }, + { + "epoch": 0.3958157064841166, + "learning_rate": 0.0019822378535594553, + "loss": 0.8039, + "step": 5146 + }, + { + "epoch": 0.39589262364433503, + "learning_rate": 0.0019818946182529867, + "loss": 0.8459, + "step": 5147 + }, + { + "epoch": 0.3959695408045535, + "learning_rate": 0.001981551354808189, + "loss": 1.148, + "step": 5148 + }, + { + "epoch": 0.39604645796477195, + "learning_rate": 0.0019812080632451053, + "loss": 1.0183, + "step": 5149 + }, + { + "epoch": 0.3961233751249904, + "learning_rate": 0.0019808647435837824, + "loss": 1.1046, + "step": 5150 + }, + { + "epoch": 0.3962002922852088, + "learning_rate": 0.001980521395844265, + "loss": 0.8268, + "step": 5151 + }, + { + "epoch": 0.39627720944542727, + "learning_rate": 0.0019801780200466032, + "loss": 0.7725, + "step": 5152 + }, + { + "epoch": 0.39635412660564573, + "learning_rate": 0.001979834616210846, + "loss": 1.1379, + "step": 5153 + }, + { + "epoch": 0.3964310437658642, + "learning_rate": 0.0019794911843570456, + "loss": 1.2747, + "step": 5154 + }, + { + "epoch": 0.3965079609260826, + "learning_rate": 0.0019791477245052554, + "loss": 1.1457, + "step": 5155 + }, + { + "epoch": 0.39658487808630105, + "learning_rate": 0.0019788042366755297, + "loss": 1.4812, + "step": 5156 + }, + { + "epoch": 0.3966617952465195, + "learning_rate": 0.001978460720887926, + "loss": 1.0438, + "step": 5157 + }, + { + "epoch": 0.39673871240673797, + "learning_rate": 0.001978117177162502, + "loss": 1.4135, + "step": 5158 + }, + { + "epoch": 0.3968156295669564, + "learning_rate": 0.0019777736055193182, + "loss": 1.7645, + "step": 5159 + }, + { + "epoch": 0.39689254672717483, + "learning_rate": 0.001977430005978435, + "loss": 1.4837, + "step": 5160 + }, + { + "epoch": 0.3969694638873933, + "learning_rate": 0.0019770863785599168, + "loss": 1.1554, + "step": 5161 + }, + { + "epoch": 0.3970463810476117, + "learning_rate": 0.001976742723283827, + "loss": 1.0866, + "step": 5162 + }, + { + "epoch": 0.39712329820783016, + "learning_rate": 0.0019763990401702337, + "loss": 1.3217, + "step": 5163 + }, + { + "epoch": 0.3972002153680486, + "learning_rate": 0.001976055329239203, + "loss": 1.1724, + "step": 5164 + }, + { + "epoch": 0.3972771325282671, + "learning_rate": 0.0019757115905108063, + "loss": 0.9498, + "step": 5165 + }, + { + "epoch": 0.3973540496884855, + "learning_rate": 0.0019753678240051137, + "loss": 1.0203, + "step": 5166 + }, + { + "epoch": 0.39743096684870394, + "learning_rate": 0.001975024029742198, + "loss": 1.16, + "step": 5167 + }, + { + "epoch": 0.3975078840089224, + "learning_rate": 0.0019746802077421346, + "loss": 1.2596, + "step": 5168 + }, + { + "epoch": 0.39758480116914086, + "learning_rate": 0.001974336358024999, + "loss": 1.1585, + "step": 5169 + }, + { + "epoch": 0.39766171832935926, + "learning_rate": 0.0019739924806108696, + "loss": 1.3912, + "step": 5170 + }, + { + "epoch": 0.3977386354895777, + "learning_rate": 0.0019736485755198247, + "loss": 1.162, + "step": 5171 + }, + { + "epoch": 0.3978155526497962, + "learning_rate": 0.0019733046427719463, + "loss": 0.9619, + "step": 5172 + }, + { + "epoch": 0.39789246981001464, + "learning_rate": 0.001972960682387316, + "loss": 1.134, + "step": 5173 + }, + { + "epoch": 0.39796938697023304, + "learning_rate": 0.001972616694386019, + "loss": 1.0416, + "step": 5174 + }, + { + "epoch": 0.3980463041304515, + "learning_rate": 0.0019722726787881405, + "loss": 1.3365, + "step": 5175 + }, + { + "epoch": 0.39812322129066996, + "learning_rate": 0.001971928635613768, + "loss": 1.395, + "step": 5176 + }, + { + "epoch": 0.3982001384508884, + "learning_rate": 0.001971584564882991, + "loss": 1.0601, + "step": 5177 + }, + { + "epoch": 0.3982770556111068, + "learning_rate": 0.0019712404666158995, + "loss": 0.9681, + "step": 5178 + }, + { + "epoch": 0.3983539727713253, + "learning_rate": 0.001970896340832586, + "loss": 1.0862, + "step": 5179 + }, + { + "epoch": 0.39843088993154374, + "learning_rate": 0.0019705521875531444, + "loss": 1.1627, + "step": 5180 + }, + { + "epoch": 0.39850780709176215, + "learning_rate": 0.0019702080067976703, + "loss": 1.2503, + "step": 5181 + }, + { + "epoch": 0.3985847242519806, + "learning_rate": 0.001969863798586261, + "loss": 1.2246, + "step": 5182 + }, + { + "epoch": 0.39866164141219906, + "learning_rate": 0.0019695195629390143, + "loss": 1.2009, + "step": 5183 + }, + { + "epoch": 0.3987385585724175, + "learning_rate": 0.0019691752998760308, + "loss": 1.3125, + "step": 5184 + }, + { + "epoch": 0.3988154757326359, + "learning_rate": 0.0019688310094174133, + "loss": 0.9761, + "step": 5185 + }, + { + "epoch": 0.3988923928928544, + "learning_rate": 0.001968486691583264, + "loss": 0.8742, + "step": 5186 + }, + { + "epoch": 0.39896931005307285, + "learning_rate": 0.001968142346393689, + "loss": 1.2871, + "step": 5187 + }, + { + "epoch": 0.3990462272132913, + "learning_rate": 0.0019677979738687936, + "loss": 1.4258, + "step": 5188 + }, + { + "epoch": 0.3991231443735097, + "learning_rate": 0.001967453574028688, + "loss": 1.2961, + "step": 5189 + }, + { + "epoch": 0.39920006153372817, + "learning_rate": 0.0019671091468934803, + "loss": 0.9601, + "step": 5190 + }, + { + "epoch": 0.3992769786939466, + "learning_rate": 0.001966764692483283, + "loss": 1.4248, + "step": 5191 + }, + { + "epoch": 0.3993538958541651, + "learning_rate": 0.001966420210818209, + "loss": 1.3774, + "step": 5192 + }, + { + "epoch": 0.3994308130143835, + "learning_rate": 0.0019660757019183727, + "loss": 1.2007, + "step": 5193 + }, + { + "epoch": 0.39950773017460195, + "learning_rate": 0.0019657311658038902, + "loss": 1.1962, + "step": 5194 + }, + { + "epoch": 0.3995846473348204, + "learning_rate": 0.00196538660249488, + "loss": 0.8821, + "step": 5195 + }, + { + "epoch": 0.39966156449503887, + "learning_rate": 0.0019650420120114606, + "loss": 1.1384, + "step": 5196 + }, + { + "epoch": 0.39973848165525727, + "learning_rate": 0.0019646973943737538, + "loss": 0.992, + "step": 5197 + }, + { + "epoch": 0.39981539881547573, + "learning_rate": 0.0019643527496018822, + "loss": 1.0595, + "step": 5198 + }, + { + "epoch": 0.3998923159756942, + "learning_rate": 0.001964008077715969, + "loss": 1.2635, + "step": 5199 + }, + { + "epoch": 0.39996923313591265, + "learning_rate": 0.00196366337873614, + "loss": 1.101, + "step": 5200 + }, + { + "epoch": 0.40004615029613105, + "learning_rate": 0.0019633186526825243, + "loss": 0.9079, + "step": 5201 + }, + { + "epoch": 0.4001230674563495, + "learning_rate": 0.0019629738995752487, + "loss": 1.1907, + "step": 5202 + }, + { + "epoch": 0.40019998461656797, + "learning_rate": 0.0019626291194344454, + "loss": 1.1628, + "step": 5203 + }, + { + "epoch": 0.4002769017767864, + "learning_rate": 0.001962284312280245, + "loss": 1.7953, + "step": 5204 + }, + { + "epoch": 0.40035381893700484, + "learning_rate": 0.0019619394781327827, + "loss": 1.1781, + "step": 5205 + }, + { + "epoch": 0.4004307360972233, + "learning_rate": 0.001961594617012192, + "loss": 1.2836, + "step": 5206 + }, + { + "epoch": 0.40050765325744175, + "learning_rate": 0.001961249728938611, + "loss": 1.0594, + "step": 5207 + }, + { + "epoch": 0.40058457041766016, + "learning_rate": 0.0019609048139321776, + "loss": 0.2582, + "step": 5208 + }, + { + "epoch": 0.4006614875778786, + "learning_rate": 0.001960559872013031, + "loss": 1.0238, + "step": 5209 + }, + { + "epoch": 0.4007384047380971, + "learning_rate": 0.001960214903201314, + "loss": 1.1124, + "step": 5210 + }, + { + "epoch": 0.40081532189831554, + "learning_rate": 0.0019598699075171694, + "loss": 1.0666, + "step": 5211 + }, + { + "epoch": 0.40089223905853394, + "learning_rate": 0.0019595248849807417, + "loss": 1.2561, + "step": 5212 + }, + { + "epoch": 0.4009691562187524, + "learning_rate": 0.001959179835612177, + "loss": 1.0237, + "step": 5213 + }, + { + "epoch": 0.40104607337897086, + "learning_rate": 0.001958834759431623, + "loss": 1.1575, + "step": 5214 + }, + { + "epoch": 0.4011229905391893, + "learning_rate": 0.001958489656459229, + "loss": 1.2128, + "step": 5215 + }, + { + "epoch": 0.4011999076994077, + "learning_rate": 0.001958144526715147, + "loss": 1.1695, + "step": 5216 + }, + { + "epoch": 0.4012768248596262, + "learning_rate": 0.0019577993702195274, + "loss": 1.2952, + "step": 5217 + }, + { + "epoch": 0.40135374201984464, + "learning_rate": 0.0019574541869925265, + "loss": 1.0493, + "step": 5218 + }, + { + "epoch": 0.4014306591800631, + "learning_rate": 0.001957108977054298, + "loss": 1.0179, + "step": 5219 + }, + { + "epoch": 0.4015075763402815, + "learning_rate": 0.0019567637404250006, + "loss": 1.3156, + "step": 5220 + }, + { + "epoch": 0.40158449350049996, + "learning_rate": 0.001956418477124792, + "loss": 1.2617, + "step": 5221 + }, + { + "epoch": 0.4016614106607184, + "learning_rate": 0.0019560731871738333, + "loss": 1.7767, + "step": 5222 + }, + { + "epoch": 0.4017383278209368, + "learning_rate": 0.0019557278705922853, + "loss": 1.0942, + "step": 5223 + }, + { + "epoch": 0.4018152449811553, + "learning_rate": 0.0019553825274003125, + "loss": 1.2732, + "step": 5224 + }, + { + "epoch": 0.40189216214137374, + "learning_rate": 0.001955037157618079, + "loss": 1.2963, + "step": 5225 + }, + { + "epoch": 0.4019690793015922, + "learning_rate": 0.001954691761265752, + "loss": 1.0306, + "step": 5226 + }, + { + "epoch": 0.4020459964618106, + "learning_rate": 0.001954346338363499, + "loss": 1.1381, + "step": 5227 + }, + { + "epoch": 0.40212291362202907, + "learning_rate": 0.0019540008889314896, + "loss": 1.1768, + "step": 5228 + }, + { + "epoch": 0.4021998307822475, + "learning_rate": 0.001953655412989895, + "loss": 0.9864, + "step": 5229 + }, + { + "epoch": 0.402276747942466, + "learning_rate": 0.001953309910558888, + "loss": 1.0862, + "step": 5230 + }, + { + "epoch": 0.4023536651026844, + "learning_rate": 0.0019529643816586437, + "loss": 1.0612, + "step": 5231 + }, + { + "epoch": 0.40243058226290285, + "learning_rate": 0.0019526188263093366, + "loss": 1.0074, + "step": 5232 + }, + { + "epoch": 0.4025074994231213, + "learning_rate": 0.0019522732445311445, + "loss": 1.1565, + "step": 5233 + }, + { + "epoch": 0.40258441658333977, + "learning_rate": 0.0019519276363442464, + "loss": 0.9626, + "step": 5234 + }, + { + "epoch": 0.40266133374355817, + "learning_rate": 0.0019515820017688226, + "loss": 1.3013, + "step": 5235 + }, + { + "epoch": 0.40273825090377663, + "learning_rate": 0.001951236340825055, + "loss": 1.0675, + "step": 5236 + }, + { + "epoch": 0.4028151680639951, + "learning_rate": 0.0019508906535331272, + "loss": 0.8322, + "step": 5237 + }, + { + "epoch": 0.40289208522421355, + "learning_rate": 0.0019505449399132245, + "loss": 1.354, + "step": 5238 + }, + { + "epoch": 0.40296900238443195, + "learning_rate": 0.001950199199985533, + "loss": 0.9497, + "step": 5239 + }, + { + "epoch": 0.4030459195446504, + "learning_rate": 0.001949853433770241, + "loss": 1.0087, + "step": 5240 + }, + { + "epoch": 0.40312283670486887, + "learning_rate": 0.001949507641287538, + "loss": 1.4448, + "step": 5241 + }, + { + "epoch": 0.4031997538650873, + "learning_rate": 0.0019491618225576158, + "loss": 1.2243, + "step": 5242 + }, + { + "epoch": 0.40327667102530573, + "learning_rate": 0.0019488159776006664, + "loss": 1.516, + "step": 5243 + }, + { + "epoch": 0.4033535881855242, + "learning_rate": 0.0019484701064368842, + "loss": 1.4581, + "step": 5244 + }, + { + "epoch": 0.40343050534574265, + "learning_rate": 0.0019481242090864655, + "loss": 1.1713, + "step": 5245 + }, + { + "epoch": 0.40350742250596106, + "learning_rate": 0.0019477782855696069, + "loss": 1.663, + "step": 5246 + }, + { + "epoch": 0.4035843396661795, + "learning_rate": 0.0019474323359065077, + "loss": 1.2683, + "step": 5247 + }, + { + "epoch": 0.403661256826398, + "learning_rate": 0.0019470863601173678, + "loss": 1.7624, + "step": 5248 + }, + { + "epoch": 0.40373817398661643, + "learning_rate": 0.0019467403582223894, + "loss": 0.7888, + "step": 5249 + }, + { + "epoch": 0.40381509114683484, + "learning_rate": 0.0019463943302417766, + "loss": 1.2357, + "step": 5250 + }, + { + "epoch": 0.4038920083070533, + "learning_rate": 0.0019460482761957333, + "loss": 1.1859, + "step": 5251 + }, + { + "epoch": 0.40396892546727176, + "learning_rate": 0.0019457021961044659, + "loss": 0.9087, + "step": 5252 + }, + { + "epoch": 0.4040458426274902, + "learning_rate": 0.001945356089988183, + "loss": 0.9273, + "step": 5253 + }, + { + "epoch": 0.4041227597877086, + "learning_rate": 0.001945009957867094, + "loss": 1.2712, + "step": 5254 + }, + { + "epoch": 0.4041996769479271, + "learning_rate": 0.0019446637997614093, + "loss": 1.2907, + "step": 5255 + }, + { + "epoch": 0.40427659410814554, + "learning_rate": 0.0019443176156913424, + "loss": 1.1121, + "step": 5256 + }, + { + "epoch": 0.404353511268364, + "learning_rate": 0.0019439714056771063, + "loss": 1.2598, + "step": 5257 + }, + { + "epoch": 0.4044304284285824, + "learning_rate": 0.0019436251697389183, + "loss": 1.3612, + "step": 5258 + }, + { + "epoch": 0.40450734558880086, + "learning_rate": 0.001943278907896993, + "loss": 1.1818, + "step": 5259 + }, + { + "epoch": 0.4045842627490193, + "learning_rate": 0.0019429326201715511, + "loss": 0.9665, + "step": 5260 + }, + { + "epoch": 0.4046611799092378, + "learning_rate": 0.0019425863065828115, + "loss": 1.2597, + "step": 5261 + }, + { + "epoch": 0.4047380970694562, + "learning_rate": 0.0019422399671509968, + "loss": 1.4157, + "step": 5262 + }, + { + "epoch": 0.40481501422967464, + "learning_rate": 0.0019418936018963289, + "loss": 1.0955, + "step": 5263 + }, + { + "epoch": 0.4048919313898931, + "learning_rate": 0.0019415472108390331, + "loss": 1.1623, + "step": 5264 + }, + { + "epoch": 0.4049688485501115, + "learning_rate": 0.0019412007939993356, + "loss": 1.0951, + "step": 5265 + }, + { + "epoch": 0.40504576571032996, + "learning_rate": 0.0019408543513974637, + "loss": 0.9792, + "step": 5266 + }, + { + "epoch": 0.4051226828705484, + "learning_rate": 0.0019405078830536473, + "loss": 1.039, + "step": 5267 + }, + { + "epoch": 0.4051996000307669, + "learning_rate": 0.0019401613889881162, + "loss": 1.0961, + "step": 5268 + }, + { + "epoch": 0.4052765171909853, + "learning_rate": 0.0019398148692211033, + "loss": 1.0801, + "step": 5269 + }, + { + "epoch": 0.40535343435120375, + "learning_rate": 0.001939468323772841, + "loss": 0.7251, + "step": 5270 + }, + { + "epoch": 0.4054303515114222, + "learning_rate": 0.001939121752663566, + "loss": 1.0864, + "step": 5271 + }, + { + "epoch": 0.40550726867164066, + "learning_rate": 0.0019387751559135137, + "loss": 1.3339, + "step": 5272 + }, + { + "epoch": 0.40558418583185907, + "learning_rate": 0.0019384285335429228, + "loss": 0.8261, + "step": 5273 + }, + { + "epoch": 0.4056611029920775, + "learning_rate": 0.0019380818855720334, + "loss": 0.9781, + "step": 5274 + }, + { + "epoch": 0.405738020152296, + "learning_rate": 0.0019377352120210861, + "loss": 1.5219, + "step": 5275 + }, + { + "epoch": 0.40581493731251445, + "learning_rate": 0.0019373885129103231, + "loss": 1.1983, + "step": 5276 + }, + { + "epoch": 0.40589185447273285, + "learning_rate": 0.0019370417882599892, + "loss": 1.3282, + "step": 5277 + }, + { + "epoch": 0.4059687716329513, + "learning_rate": 0.0019366950380903297, + "loss": 1.3461, + "step": 5278 + }, + { + "epoch": 0.40604568879316977, + "learning_rate": 0.0019363482624215917, + "loss": 1.0908, + "step": 5279 + }, + { + "epoch": 0.4061226059533882, + "learning_rate": 0.0019360014612740245, + "loss": 1.3082, + "step": 5280 + }, + { + "epoch": 0.40619952311360663, + "learning_rate": 0.0019356546346678773, + "loss": 1.2681, + "step": 5281 + }, + { + "epoch": 0.4062764402738251, + "learning_rate": 0.001935307782623402, + "loss": 0.9953, + "step": 5282 + }, + { + "epoch": 0.40635335743404355, + "learning_rate": 0.0019349609051608511, + "loss": 1.1353, + "step": 5283 + }, + { + "epoch": 0.40643027459426195, + "learning_rate": 0.0019346140023004802, + "loss": 1.6298, + "step": 5284 + }, + { + "epoch": 0.4065071917544804, + "learning_rate": 0.0019342670740625447, + "loss": 1.1867, + "step": 5285 + }, + { + "epoch": 0.40658410891469887, + "learning_rate": 0.0019339201204673017, + "loss": 0.8309, + "step": 5286 + }, + { + "epoch": 0.40666102607491733, + "learning_rate": 0.0019335731415350112, + "loss": 1.4024, + "step": 5287 + }, + { + "epoch": 0.40673794323513573, + "learning_rate": 0.0019332261372859332, + "loss": 1.2121, + "step": 5288 + }, + { + "epoch": 0.4068148603953542, + "learning_rate": 0.0019328791077403294, + "loss": 1.2537, + "step": 5289 + }, + { + "epoch": 0.40689177755557265, + "learning_rate": 0.0019325320529184628, + "loss": 1.2006, + "step": 5290 + }, + { + "epoch": 0.4069686947157911, + "learning_rate": 0.0019321849728405995, + "loss": 1.4749, + "step": 5291 + }, + { + "epoch": 0.4070456118760095, + "learning_rate": 0.0019318378675270051, + "loss": 1.3887, + "step": 5292 + }, + { + "epoch": 0.407122529036228, + "learning_rate": 0.0019314907369979474, + "loss": 0.9718, + "step": 5293 + }, + { + "epoch": 0.40719944619644644, + "learning_rate": 0.0019311435812736958, + "loss": 1.0322, + "step": 5294 + }, + { + "epoch": 0.4072763633566649, + "learning_rate": 0.0019307964003745215, + "loss": 1.1825, + "step": 5295 + }, + { + "epoch": 0.4073532805168833, + "learning_rate": 0.0019304491943206963, + "loss": 0.9931, + "step": 5296 + }, + { + "epoch": 0.40743019767710176, + "learning_rate": 0.001930101963132494, + "loss": 1.2207, + "step": 5297 + }, + { + "epoch": 0.4075071148373202, + "learning_rate": 0.0019297547068301895, + "loss": 1.8569, + "step": 5298 + }, + { + "epoch": 0.4075840319975387, + "learning_rate": 0.00192940742543406, + "loss": 1.2515, + "step": 5299 + }, + { + "epoch": 0.4076609491577571, + "learning_rate": 0.0019290601189643837, + "loss": 1.0231, + "step": 5300 + }, + { + "epoch": 0.40773786631797554, + "learning_rate": 0.0019287127874414395, + "loss": 1.1391, + "step": 5301 + }, + { + "epoch": 0.407814783478194, + "learning_rate": 0.0019283654308855092, + "loss": 1.1344, + "step": 5302 + }, + { + "epoch": 0.4078917006384124, + "learning_rate": 0.001928018049316875, + "loss": 1.3173, + "step": 5303 + }, + { + "epoch": 0.40796861779863086, + "learning_rate": 0.0019276706427558206, + "loss": 1.2478, + "step": 5304 + }, + { + "epoch": 0.4080455349588493, + "learning_rate": 0.0019273232112226315, + "loss": 0.9447, + "step": 5305 + }, + { + "epoch": 0.4081224521190678, + "learning_rate": 0.0019269757547375952, + "loss": 1.4481, + "step": 5306 + }, + { + "epoch": 0.4081993692792862, + "learning_rate": 0.0019266282733209994, + "loss": 1.2329, + "step": 5307 + }, + { + "epoch": 0.40827628643950464, + "learning_rate": 0.001926280766993134, + "loss": 1.2065, + "step": 5308 + }, + { + "epoch": 0.4083532035997231, + "learning_rate": 0.0019259332357742902, + "loss": 1.0585, + "step": 5309 + }, + { + "epoch": 0.40843012075994156, + "learning_rate": 0.0019255856796847608, + "loss": 0.7993, + "step": 5310 + }, + { + "epoch": 0.40850703792015997, + "learning_rate": 0.0019252380987448408, + "loss": 1.4017, + "step": 5311 + }, + { + "epoch": 0.4085839550803784, + "learning_rate": 0.001924890492974824, + "loss": 1.3804, + "step": 5312 + }, + { + "epoch": 0.4086608722405969, + "learning_rate": 0.0019245428623950091, + "loss": 0.8605, + "step": 5313 + }, + { + "epoch": 0.40873778940081534, + "learning_rate": 0.0019241952070256938, + "loss": 1.347, + "step": 5314 + }, + { + "epoch": 0.40881470656103375, + "learning_rate": 0.001923847526887179, + "loss": 1.0736, + "step": 5315 + }, + { + "epoch": 0.4088916237212522, + "learning_rate": 0.0019234998219997649, + "loss": 1.2824, + "step": 5316 + }, + { + "epoch": 0.40896854088147067, + "learning_rate": 0.0019231520923837547, + "loss": 0.9933, + "step": 5317 + }, + { + "epoch": 0.4090454580416891, + "learning_rate": 0.0019228043380594524, + "loss": 1.2015, + "step": 5318 + }, + { + "epoch": 0.40912237520190753, + "learning_rate": 0.0019224565590471648, + "loss": 0.7587, + "step": 5319 + }, + { + "epoch": 0.409199292362126, + "learning_rate": 0.0019221087553671983, + "loss": 1.2312, + "step": 5320 + }, + { + "epoch": 0.40927620952234445, + "learning_rate": 0.0019217609270398615, + "loss": 1.2008, + "step": 5321 + }, + { + "epoch": 0.4093531266825629, + "learning_rate": 0.0019214130740854648, + "loss": 1.3925, + "step": 5322 + }, + { + "epoch": 0.4094300438427813, + "learning_rate": 0.0019210651965243196, + "loss": 1.0074, + "step": 5323 + }, + { + "epoch": 0.40950696100299977, + "learning_rate": 0.001920717294376739, + "loss": 1.5791, + "step": 5324 + }, + { + "epoch": 0.40958387816321823, + "learning_rate": 0.0019203693676630364, + "loss": 1.162, + "step": 5325 + }, + { + "epoch": 0.40966079532343663, + "learning_rate": 0.0019200214164035291, + "loss": 1.3833, + "step": 5326 + }, + { + "epoch": 0.4097377124836551, + "learning_rate": 0.001919673440618533, + "loss": 1.287, + "step": 5327 + }, + { + "epoch": 0.40981462964387355, + "learning_rate": 0.0019193254403283675, + "loss": 1.3652, + "step": 5328 + }, + { + "epoch": 0.409891546804092, + "learning_rate": 0.0019189774155533525, + "loss": 0.8447, + "step": 5329 + }, + { + "epoch": 0.4099684639643104, + "learning_rate": 0.0019186293663138092, + "loss": 1.4372, + "step": 5330 + }, + { + "epoch": 0.4100453811245289, + "learning_rate": 0.0019182812926300614, + "loss": 1.3139, + "step": 5331 + }, + { + "epoch": 0.41012229828474733, + "learning_rate": 0.0019179331945224328, + "loss": 0.9102, + "step": 5332 + }, + { + "epoch": 0.4101992154449658, + "learning_rate": 0.0019175850720112495, + "loss": 1.0867, + "step": 5333 + }, + { + "epoch": 0.4102761326051842, + "learning_rate": 0.001917236925116839, + "loss": 1.0796, + "step": 5334 + }, + { + "epoch": 0.41035304976540266, + "learning_rate": 0.0019168887538595291, + "loss": 0.8754, + "step": 5335 + }, + { + "epoch": 0.4104299669256211, + "learning_rate": 0.0019165405582596503, + "loss": 1.4631, + "step": 5336 + }, + { + "epoch": 0.4105068840858396, + "learning_rate": 0.0019161923383375344, + "loss": 1.0335, + "step": 5337 + }, + { + "epoch": 0.410583801246058, + "learning_rate": 0.0019158440941135144, + "loss": 1.2179, + "step": 5338 + }, + { + "epoch": 0.41066071840627644, + "learning_rate": 0.0019154958256079242, + "loss": 1.4298, + "step": 5339 + }, + { + "epoch": 0.4107376355664949, + "learning_rate": 0.0019151475328410994, + "loss": 1.6193, + "step": 5340 + }, + { + "epoch": 0.41081455272671336, + "learning_rate": 0.001914799215833378, + "loss": 1.0342, + "step": 5341 + }, + { + "epoch": 0.41089146988693176, + "learning_rate": 0.0019144508746050976, + "loss": 1.6808, + "step": 5342 + }, + { + "epoch": 0.4109683870471502, + "learning_rate": 0.0019141025091765989, + "loss": 1.2041, + "step": 5343 + }, + { + "epoch": 0.4110453042073687, + "learning_rate": 0.0019137541195682233, + "loss": 0.9751, + "step": 5344 + }, + { + "epoch": 0.4111222213675871, + "learning_rate": 0.0019134057058003133, + "loss": 1.1871, + "step": 5345 + }, + { + "epoch": 0.41119913852780554, + "learning_rate": 0.0019130572678932134, + "loss": 0.9748, + "step": 5346 + }, + { + "epoch": 0.411276055688024, + "learning_rate": 0.001912708805867269, + "loss": 1.3121, + "step": 5347 + }, + { + "epoch": 0.41135297284824246, + "learning_rate": 0.0019123603197428272, + "loss": 1.0555, + "step": 5348 + }, + { + "epoch": 0.41142989000846086, + "learning_rate": 0.001912011809540237, + "loss": 1.4181, + "step": 5349 + }, + { + "epoch": 0.4115068071686793, + "learning_rate": 0.0019116632752798474, + "loss": 0.8591, + "step": 5350 + }, + { + "epoch": 0.4115837243288978, + "learning_rate": 0.0019113147169820103, + "loss": 1.5652, + "step": 5351 + }, + { + "epoch": 0.41166064148911624, + "learning_rate": 0.0019109661346670785, + "loss": 1.2278, + "step": 5352 + }, + { + "epoch": 0.41173755864933465, + "learning_rate": 0.0019106175283554055, + "loss": 1.597, + "step": 5353 + }, + { + "epoch": 0.4118144758095531, + "learning_rate": 0.0019102688980673467, + "loss": 1.3299, + "step": 5354 + }, + { + "epoch": 0.41189139296977156, + "learning_rate": 0.0019099202438232598, + "loss": 1.1454, + "step": 5355 + }, + { + "epoch": 0.41196831012999, + "learning_rate": 0.0019095715656435028, + "loss": 1.018, + "step": 5356 + }, + { + "epoch": 0.4120452272902084, + "learning_rate": 0.0019092228635484354, + "loss": 1.1463, + "step": 5357 + }, + { + "epoch": 0.4121221444504269, + "learning_rate": 0.001908874137558418, + "loss": 1.1272, + "step": 5358 + }, + { + "epoch": 0.41219906161064535, + "learning_rate": 0.0019085253876938138, + "loss": 1.3921, + "step": 5359 + }, + { + "epoch": 0.4122759787708638, + "learning_rate": 0.0019081766139749862, + "loss": 1.11, + "step": 5360 + }, + { + "epoch": 0.4123528959310822, + "learning_rate": 0.0019078278164223013, + "loss": 1.4963, + "step": 5361 + }, + { + "epoch": 0.41242981309130067, + "learning_rate": 0.001907478995056125, + "loss": 1.184, + "step": 5362 + }, + { + "epoch": 0.4125067302515191, + "learning_rate": 0.001907130149896825, + "loss": 1.1806, + "step": 5363 + }, + { + "epoch": 0.4125836474117376, + "learning_rate": 0.0019067812809647716, + "loss": 1.0272, + "step": 5364 + }, + { + "epoch": 0.412660564571956, + "learning_rate": 0.0019064323882803352, + "loss": 1.2647, + "step": 5365 + }, + { + "epoch": 0.41273748173217445, + "learning_rate": 0.0019060834718638883, + "loss": 1.3375, + "step": 5366 + }, + { + "epoch": 0.4128143988923929, + "learning_rate": 0.001905734531735804, + "loss": 1.114, + "step": 5367 + }, + { + "epoch": 0.4128913160526113, + "learning_rate": 0.001905385567916458, + "loss": 1.443, + "step": 5368 + }, + { + "epoch": 0.41296823321282977, + "learning_rate": 0.0019050365804262255, + "loss": 1.3102, + "step": 5369 + }, + { + "epoch": 0.41304515037304823, + "learning_rate": 0.0019046875692854857, + "loss": 1.1855, + "step": 5370 + }, + { + "epoch": 0.4131220675332667, + "learning_rate": 0.001904338534514616, + "loss": 0.974, + "step": 5371 + }, + { + "epoch": 0.4131989846934851, + "learning_rate": 0.0019039894761339988, + "loss": 0.8566, + "step": 5372 + }, + { + "epoch": 0.41327590185370355, + "learning_rate": 0.0019036403941640146, + "loss": 0.8169, + "step": 5373 + }, + { + "epoch": 0.413352819013922, + "learning_rate": 0.0019032912886250471, + "loss": 1.17, + "step": 5374 + }, + { + "epoch": 0.41342973617414047, + "learning_rate": 0.0019029421595374814, + "loss": 0.9946, + "step": 5375 + }, + { + "epoch": 0.4135066533343589, + "learning_rate": 0.0019025930069217025, + "loss": 1.0334, + "step": 5376 + }, + { + "epoch": 0.41358357049457734, + "learning_rate": 0.001902243830798099, + "loss": 1.1297, + "step": 5377 + }, + { + "epoch": 0.4136604876547958, + "learning_rate": 0.0019018946311870582, + "loss": 0.894, + "step": 5378 + }, + { + "epoch": 0.41373740481501425, + "learning_rate": 0.0019015454081089717, + "loss": 0.9276, + "step": 5379 + }, + { + "epoch": 0.41381432197523266, + "learning_rate": 0.0019011961615842303, + "loss": 1.2543, + "step": 5380 + }, + { + "epoch": 0.4138912391354511, + "learning_rate": 0.001900846891633227, + "loss": 0.8844, + "step": 5381 + }, + { + "epoch": 0.4139681562956696, + "learning_rate": 0.0019004975982763552, + "loss": 1.4388, + "step": 5382 + }, + { + "epoch": 0.41404507345588804, + "learning_rate": 0.0019001482815340118, + "loss": 1.2338, + "step": 5383 + }, + { + "epoch": 0.41412199061610644, + "learning_rate": 0.001899798941426593, + "loss": 1.1151, + "step": 5384 + }, + { + "epoch": 0.4141989077763249, + "learning_rate": 0.0018994495779744976, + "loss": 1.1242, + "step": 5385 + }, + { + "epoch": 0.41427582493654336, + "learning_rate": 0.0018991001911981247, + "loss": 1.0195, + "step": 5386 + }, + { + "epoch": 0.41435274209676176, + "learning_rate": 0.0018987507811178762, + "loss": 1.0197, + "step": 5387 + }, + { + "epoch": 0.4144296592569802, + "learning_rate": 0.0018984013477541537, + "loss": 1.2167, + "step": 5388 + }, + { + "epoch": 0.4145065764171987, + "learning_rate": 0.0018980518911273611, + "loss": 0.9649, + "step": 5389 + }, + { + "epoch": 0.41458349357741714, + "learning_rate": 0.0018977024112579043, + "loss": 0.9731, + "step": 5390 + }, + { + "epoch": 0.41466041073763554, + "learning_rate": 0.001897352908166189, + "loss": 1.101, + "step": 5391 + }, + { + "epoch": 0.414737327897854, + "learning_rate": 0.0018970033818726231, + "loss": 1.3306, + "step": 5392 + }, + { + "epoch": 0.41481424505807246, + "learning_rate": 0.001896653832397616, + "loss": 1.1095, + "step": 5393 + }, + { + "epoch": 0.4148911622182909, + "learning_rate": 0.0018963042597615785, + "loss": 1.1044, + "step": 5394 + }, + { + "epoch": 0.4149680793785093, + "learning_rate": 0.0018959546639849224, + "loss": 0.8383, + "step": 5395 + }, + { + "epoch": 0.4150449965387278, + "learning_rate": 0.00189560504508806, + "loss": 1.2859, + "step": 5396 + }, + { + "epoch": 0.41512191369894624, + "learning_rate": 0.0018952554030914075, + "loss": 1.0922, + "step": 5397 + }, + { + "epoch": 0.4151988308591647, + "learning_rate": 0.0018949057380153794, + "loss": 1.3724, + "step": 5398 + }, + { + "epoch": 0.4152757480193831, + "learning_rate": 0.0018945560498803945, + "loss": 0.913, + "step": 5399 + }, + { + "epoch": 0.41535266517960157, + "learning_rate": 0.00189420633870687, + "loss": 1.3185, + "step": 5400 + }, + { + "epoch": 0.41542958233982, + "learning_rate": 0.001893856604515227, + "loss": 1.3982, + "step": 5401 + }, + { + "epoch": 0.4155064995000385, + "learning_rate": 0.001893506847325886, + "loss": 1.3029, + "step": 5402 + }, + { + "epoch": 0.4155834166602569, + "learning_rate": 0.001893157067159271, + "loss": 1.2488, + "step": 5403 + }, + { + "epoch": 0.41566033382047535, + "learning_rate": 0.001892807264035804, + "loss": 0.968, + "step": 5404 + }, + { + "epoch": 0.4157372509806938, + "learning_rate": 0.0018924574379759116, + "loss": 0.9755, + "step": 5405 + }, + { + "epoch": 0.4158141681409122, + "learning_rate": 0.001892107589000021, + "loss": 1.2083, + "step": 5406 + }, + { + "epoch": 0.41589108530113067, + "learning_rate": 0.001891757717128559, + "loss": 1.126, + "step": 5407 + }, + { + "epoch": 0.41596800246134913, + "learning_rate": 0.001891407822381956, + "loss": 1.2708, + "step": 5408 + }, + { + "epoch": 0.4160449196215676, + "learning_rate": 0.001891057904780642, + "loss": 1.1473, + "step": 5409 + }, + { + "epoch": 0.416121836781786, + "learning_rate": 0.0018907079643450501, + "loss": 1.0215, + "step": 5410 + }, + { + "epoch": 0.41619875394200445, + "learning_rate": 0.0018903580010956125, + "loss": 1.1171, + "step": 5411 + }, + { + "epoch": 0.4162756711022229, + "learning_rate": 0.0018900080150527645, + "loss": 1.1427, + "step": 5412 + }, + { + "epoch": 0.41635258826244137, + "learning_rate": 0.0018896580062369415, + "loss": 1.4094, + "step": 5413 + }, + { + "epoch": 0.4164295054226598, + "learning_rate": 0.0018893079746685822, + "loss": 1.1212, + "step": 5414 + }, + { + "epoch": 0.41650642258287823, + "learning_rate": 0.0018889579203681243, + "loss": 1.337, + "step": 5415 + }, + { + "epoch": 0.4165833397430967, + "learning_rate": 0.001888607843356008, + "loss": 1.1354, + "step": 5416 + }, + { + "epoch": 0.41666025690331515, + "learning_rate": 0.0018882577436526743, + "loss": 1.4084, + "step": 5417 + }, + { + "epoch": 0.41673717406353356, + "learning_rate": 0.0018879076212785669, + "loss": 0.8814, + "step": 5418 + }, + { + "epoch": 0.416814091223752, + "learning_rate": 0.001887557476254129, + "loss": 0.5657, + "step": 5419 + }, + { + "epoch": 0.4168910083839705, + "learning_rate": 0.001887207308599806, + "loss": 1.0559, + "step": 5420 + }, + { + "epoch": 0.41696792554418893, + "learning_rate": 0.001886857118336045, + "loss": 0.9819, + "step": 5421 + }, + { + "epoch": 0.41704484270440734, + "learning_rate": 0.001886506905483294, + "loss": 1.298, + "step": 5422 + }, + { + "epoch": 0.4171217598646258, + "learning_rate": 0.0018861566700620015, + "loss": 1.2829, + "step": 5423 + }, + { + "epoch": 0.41719867702484426, + "learning_rate": 0.0018858064120926183, + "loss": 0.927, + "step": 5424 + }, + { + "epoch": 0.4172755941850627, + "learning_rate": 0.0018854561315955975, + "loss": 0.9657, + "step": 5425 + }, + { + "epoch": 0.4173525113452811, + "learning_rate": 0.0018851058285913912, + "loss": 0.9917, + "step": 5426 + }, + { + "epoch": 0.4174294285054996, + "learning_rate": 0.001884755503100454, + "loss": 1.196, + "step": 5427 + }, + { + "epoch": 0.41750634566571804, + "learning_rate": 0.0018844051551432416, + "loss": 1.2799, + "step": 5428 + }, + { + "epoch": 0.41758326282593644, + "learning_rate": 0.001884054784740212, + "loss": 0.8154, + "step": 5429 + }, + { + "epoch": 0.4176601799861549, + "learning_rate": 0.0018837043919118232, + "loss": 1.239, + "step": 5430 + }, + { + "epoch": 0.41773709714637336, + "learning_rate": 0.0018833539766785349, + "loss": 0.971, + "step": 5431 + }, + { + "epoch": 0.4178140143065918, + "learning_rate": 0.0018830035390608089, + "loss": 0.7035, + "step": 5432 + }, + { + "epoch": 0.4178909314668102, + "learning_rate": 0.0018826530790791065, + "loss": 1.2166, + "step": 5433 + }, + { + "epoch": 0.4179678486270287, + "learning_rate": 0.0018823025967538926, + "loss": 1.1455, + "step": 5434 + }, + { + "epoch": 0.41804476578724714, + "learning_rate": 0.001881952092105631, + "loss": 1.0731, + "step": 5435 + }, + { + "epoch": 0.4181216829474656, + "learning_rate": 0.0018816015651547892, + "loss": 0.9191, + "step": 5436 + }, + { + "epoch": 0.418198600107684, + "learning_rate": 0.0018812510159218344, + "loss": 1.2045, + "step": 5437 + }, + { + "epoch": 0.41827551726790246, + "learning_rate": 0.0018809004444272348, + "loss": 1.2937, + "step": 5438 + }, + { + "epoch": 0.4183524344281209, + "learning_rate": 0.0018805498506914615, + "loss": 1.1727, + "step": 5439 + }, + { + "epoch": 0.4184293515883394, + "learning_rate": 0.0018801992347349862, + "loss": 0.9741, + "step": 5440 + }, + { + "epoch": 0.4185062687485578, + "learning_rate": 0.0018798485965782813, + "loss": 1.2351, + "step": 5441 + }, + { + "epoch": 0.41858318590877625, + "learning_rate": 0.0018794979362418206, + "loss": 0.9889, + "step": 5442 + }, + { + "epoch": 0.4186601030689947, + "learning_rate": 0.0018791472537460802, + "loss": 1.1544, + "step": 5443 + }, + { + "epoch": 0.41873702022921316, + "learning_rate": 0.0018787965491115363, + "loss": 1.1413, + "step": 5444 + }, + { + "epoch": 0.41881393738943157, + "learning_rate": 0.0018784458223586678, + "loss": 1.21, + "step": 5445 + }, + { + "epoch": 0.41889085454965, + "learning_rate": 0.0018780950735079523, + "loss": 0.8953, + "step": 5446 + }, + { + "epoch": 0.4189677717098685, + "learning_rate": 0.001877744302579872, + "loss": 0.7567, + "step": 5447 + }, + { + "epoch": 0.4190446888700869, + "learning_rate": 0.0018773935095949084, + "loss": 1.1003, + "step": 5448 + }, + { + "epoch": 0.41912160603030535, + "learning_rate": 0.0018770426945735436, + "loss": 0.9116, + "step": 5449 + }, + { + "epoch": 0.4191985231905238, + "learning_rate": 0.001876691857536264, + "loss": 1.0287, + "step": 5450 + }, + { + "epoch": 0.41927544035074227, + "learning_rate": 0.0018763409985035533, + "loss": 1.4729, + "step": 5451 + }, + { + "epoch": 0.41935235751096067, + "learning_rate": 0.0018759901174959003, + "loss": 1.2076, + "step": 5452 + }, + { + "epoch": 0.41942927467117913, + "learning_rate": 0.001875639214533792, + "loss": 1.1054, + "step": 5453 + }, + { + "epoch": 0.4195061918313976, + "learning_rate": 0.0018752882896377189, + "loss": 1.0048, + "step": 5454 + }, + { + "epoch": 0.41958310899161605, + "learning_rate": 0.0018749373428281707, + "loss": 1.3677, + "step": 5455 + }, + { + "epoch": 0.41966002615183445, + "learning_rate": 0.0018745863741256412, + "loss": 1.2008, + "step": 5456 + }, + { + "epoch": 0.4197369433120529, + "learning_rate": 0.0018742353835506224, + "loss": 0.9679, + "step": 5457 + }, + { + "epoch": 0.41981386047227137, + "learning_rate": 0.0018738843711236094, + "loss": 1.0629, + "step": 5458 + }, + { + "epoch": 0.41989077763248983, + "learning_rate": 0.0018735333368650982, + "loss": 1.288, + "step": 5459 + }, + { + "epoch": 0.41996769479270823, + "learning_rate": 0.0018731822807955868, + "loss": 1.1632, + "step": 5460 + }, + { + "epoch": 0.4200446119529267, + "learning_rate": 0.001872831202935573, + "loss": 1.4204, + "step": 5461 + }, + { + "epoch": 0.42012152911314515, + "learning_rate": 0.0018724801033055563, + "loss": 1.1945, + "step": 5462 + }, + { + "epoch": 0.4201984462733636, + "learning_rate": 0.0018721289819260383, + "loss": 1.0537, + "step": 5463 + }, + { + "epoch": 0.420275363433582, + "learning_rate": 0.0018717778388175213, + "loss": 0.9874, + "step": 5464 + }, + { + "epoch": 0.4203522805938005, + "learning_rate": 0.001871426674000509, + "loss": 0.6534, + "step": 5465 + }, + { + "epoch": 0.42042919775401894, + "learning_rate": 0.001871075487495506, + "loss": 1.3472, + "step": 5466 + }, + { + "epoch": 0.42050611491423734, + "learning_rate": 0.0018707242793230185, + "loss": 0.9003, + "step": 5467 + }, + { + "epoch": 0.4205830320744558, + "learning_rate": 0.001870373049503554, + "loss": 1.5472, + "step": 5468 + }, + { + "epoch": 0.42065994923467426, + "learning_rate": 0.0018700217980576212, + "loss": 1.2338, + "step": 5469 + }, + { + "epoch": 0.4207368663948927, + "learning_rate": 0.0018696705250057296, + "loss": 1.6406, + "step": 5470 + }, + { + "epoch": 0.4208137835551111, + "learning_rate": 0.0018693192303683914, + "loss": 0.986, + "step": 5471 + }, + { + "epoch": 0.4208907007153296, + "learning_rate": 0.0018689679141661183, + "loss": 1.5575, + "step": 5472 + }, + { + "epoch": 0.42096761787554804, + "learning_rate": 0.001868616576419424, + "loss": 1.0817, + "step": 5473 + }, + { + "epoch": 0.4210445350357665, + "learning_rate": 0.0018682652171488235, + "loss": 1.5008, + "step": 5474 + }, + { + "epoch": 0.4211214521959849, + "learning_rate": 0.0018679138363748334, + "loss": 1.1259, + "step": 5475 + }, + { + "epoch": 0.42119836935620336, + "learning_rate": 0.001867562434117971, + "loss": 1.2873, + "step": 5476 + }, + { + "epoch": 0.4212752865164218, + "learning_rate": 0.0018672110103987547, + "loss": 1.1144, + "step": 5477 + }, + { + "epoch": 0.4213522036766403, + "learning_rate": 0.0018668595652377052, + "loss": 1.278, + "step": 5478 + }, + { + "epoch": 0.4214291208368587, + "learning_rate": 0.001866508098655343, + "loss": 0.724, + "step": 5479 + }, + { + "epoch": 0.42150603799707714, + "learning_rate": 0.0018661566106721913, + "loss": 1.1591, + "step": 5480 + }, + { + "epoch": 0.4215829551572956, + "learning_rate": 0.0018658051013087731, + "loss": 1.2845, + "step": 5481 + }, + { + "epoch": 0.42165987231751406, + "learning_rate": 0.001865453570585614, + "loss": 0.8868, + "step": 5482 + }, + { + "epoch": 0.42173678947773247, + "learning_rate": 0.0018651020185232404, + "loss": 1.2903, + "step": 5483 + }, + { + "epoch": 0.4218137066379509, + "learning_rate": 0.0018647504451421786, + "loss": 0.896, + "step": 5484 + }, + { + "epoch": 0.4218906237981694, + "learning_rate": 0.001864398850462959, + "loss": 0.9226, + "step": 5485 + }, + { + "epoch": 0.42196754095838784, + "learning_rate": 0.00186404723450611, + "loss": 1.099, + "step": 5486 + }, + { + "epoch": 0.42204445811860625, + "learning_rate": 0.0018636955972921644, + "loss": 1.0727, + "step": 5487 + }, + { + "epoch": 0.4221213752788247, + "learning_rate": 0.001863343938841653, + "loss": 1.1478, + "step": 5488 + }, + { + "epoch": 0.42219829243904317, + "learning_rate": 0.0018629922591751112, + "loss": 1.372, + "step": 5489 + }, + { + "epoch": 0.42227520959926157, + "learning_rate": 0.0018626405583130727, + "loss": 1.057, + "step": 5490 + }, + { + "epoch": 0.42235212675948003, + "learning_rate": 0.0018622888362760741, + "loss": 1.0178, + "step": 5491 + }, + { + "epoch": 0.4224290439196985, + "learning_rate": 0.0018619370930846525, + "loss": 1.108, + "step": 5492 + }, + { + "epoch": 0.42250596107991695, + "learning_rate": 0.0018615853287593472, + "loss": 1.1004, + "step": 5493 + }, + { + "epoch": 0.42258287824013535, + "learning_rate": 0.0018612335433206976, + "loss": 1.305, + "step": 5494 + }, + { + "epoch": 0.4226597954003538, + "learning_rate": 0.0018608817367892446, + "loss": 1.0859, + "step": 5495 + }, + { + "epoch": 0.42273671256057227, + "learning_rate": 0.0018605299091855316, + "loss": 1.0684, + "step": 5496 + }, + { + "epoch": 0.42281362972079073, + "learning_rate": 0.0018601780605301006, + "loss": 1.174, + "step": 5497 + }, + { + "epoch": 0.42289054688100913, + "learning_rate": 0.0018598261908434983, + "loss": 0.8731, + "step": 5498 + }, + { + "epoch": 0.4229674640412276, + "learning_rate": 0.001859474300146269, + "loss": 1.2816, + "step": 5499 + }, + { + "epoch": 0.42304438120144605, + "learning_rate": 0.001859122388458961, + "loss": 1.2088, + "step": 5500 + }, + { + "epoch": 0.4231212983616645, + "learning_rate": 0.0018587704558021218, + "loss": 1.0698, + "step": 5501 + }, + { + "epoch": 0.4231982155218829, + "learning_rate": 0.0018584185021963023, + "loss": 0.8157, + "step": 5502 + }, + { + "epoch": 0.4232751326821014, + "learning_rate": 0.0018580665276620532, + "loss": 0.8452, + "step": 5503 + }, + { + "epoch": 0.42335204984231983, + "learning_rate": 0.001857714532219926, + "loss": 1.0249, + "step": 5504 + }, + { + "epoch": 0.4234289670025383, + "learning_rate": 0.0018573625158904744, + "loss": 1.104, + "step": 5505 + }, + { + "epoch": 0.4235058841627567, + "learning_rate": 0.0018570104786942534, + "loss": 1.1015, + "step": 5506 + }, + { + "epoch": 0.42358280132297516, + "learning_rate": 0.0018566584206518182, + "loss": 1.105, + "step": 5507 + }, + { + "epoch": 0.4236597184831936, + "learning_rate": 0.0018563063417837262, + "loss": 1.0311, + "step": 5508 + }, + { + "epoch": 0.423736635643412, + "learning_rate": 0.001855954242110536, + "loss": 1.395, + "step": 5509 + }, + { + "epoch": 0.4238135528036305, + "learning_rate": 0.001855602121652806, + "loss": 1.0516, + "step": 5510 + }, + { + "epoch": 0.42389046996384894, + "learning_rate": 0.0018552499804310976, + "loss": 1.8601, + "step": 5511 + }, + { + "epoch": 0.4239673871240674, + "learning_rate": 0.0018548978184659728, + "loss": 1.1041, + "step": 5512 + }, + { + "epoch": 0.4240443042842858, + "learning_rate": 0.0018545456357779945, + "loss": 1.058, + "step": 5513 + }, + { + "epoch": 0.42412122144450426, + "learning_rate": 0.001854193432387727, + "loss": 0.8967, + "step": 5514 + }, + { + "epoch": 0.4241981386047227, + "learning_rate": 0.0018538412083157355, + "loss": 1.004, + "step": 5515 + }, + { + "epoch": 0.4242750557649412, + "learning_rate": 0.0018534889635825878, + "loss": 1.0781, + "step": 5516 + }, + { + "epoch": 0.4243519729251596, + "learning_rate": 0.0018531366982088506, + "loss": 0.9593, + "step": 5517 + }, + { + "epoch": 0.42442889008537804, + "learning_rate": 0.001852784412215094, + "loss": 1.1843, + "step": 5518 + }, + { + "epoch": 0.4245058072455965, + "learning_rate": 0.0018524321056218873, + "loss": 1.1723, + "step": 5519 + }, + { + "epoch": 0.42458272440581496, + "learning_rate": 0.0018520797784498028, + "loss": 1.0247, + "step": 5520 + }, + { + "epoch": 0.42465964156603336, + "learning_rate": 0.0018517274307194135, + "loss": 1.1903, + "step": 5521 + }, + { + "epoch": 0.4247365587262518, + "learning_rate": 0.001851375062451293, + "loss": 1.2003, + "step": 5522 + }, + { + "epoch": 0.4248134758864703, + "learning_rate": 0.0018510226736660157, + "loss": 1.0411, + "step": 5523 + }, + { + "epoch": 0.42489039304668874, + "learning_rate": 0.0018506702643841593, + "loss": 1.2441, + "step": 5524 + }, + { + "epoch": 0.42496731020690715, + "learning_rate": 0.0018503178346263006, + "loss": 1.2971, + "step": 5525 + }, + { + "epoch": 0.4250442273671256, + "learning_rate": 0.001849965384413018, + "loss": 0.9883, + "step": 5526 + }, + { + "epoch": 0.42512114452734406, + "learning_rate": 0.0018496129137648926, + "loss": 0.9447, + "step": 5527 + }, + { + "epoch": 0.42519806168756247, + "learning_rate": 0.0018492604227025042, + "loss": 1.2593, + "step": 5528 + }, + { + "epoch": 0.4252749788477809, + "learning_rate": 0.0018489079112464363, + "loss": 1.2315, + "step": 5529 + }, + { + "epoch": 0.4253518960079994, + "learning_rate": 0.0018485553794172712, + "loss": 1.2355, + "step": 5530 + }, + { + "epoch": 0.42542881316821785, + "learning_rate": 0.0018482028272355947, + "loss": 1.0028, + "step": 5531 + }, + { + "epoch": 0.42550573032843625, + "learning_rate": 0.0018478502547219923, + "loss": 1.4217, + "step": 5532 + }, + { + "epoch": 0.4255826474886547, + "learning_rate": 0.0018474976618970509, + "loss": 1.092, + "step": 5533 + }, + { + "epoch": 0.42565956464887317, + "learning_rate": 0.0018471450487813584, + "loss": 1.2022, + "step": 5534 + }, + { + "epoch": 0.4257364818090916, + "learning_rate": 0.001846792415395505, + "loss": 1.1645, + "step": 5535 + }, + { + "epoch": 0.42581339896931003, + "learning_rate": 0.0018464397617600813, + "loss": 1.3899, + "step": 5536 + }, + { + "epoch": 0.4258903161295285, + "learning_rate": 0.0018460870878956785, + "loss": 1.3774, + "step": 5537 + }, + { + "epoch": 0.42596723328974695, + "learning_rate": 0.0018457343938228902, + "loss": 1.3572, + "step": 5538 + }, + { + "epoch": 0.4260441504499654, + "learning_rate": 0.0018453816795623102, + "loss": 1.3027, + "step": 5539 + }, + { + "epoch": 0.4261210676101838, + "learning_rate": 0.0018450289451345343, + "loss": 1.1324, + "step": 5540 + }, + { + "epoch": 0.42619798477040227, + "learning_rate": 0.001844676190560158, + "loss": 1.3033, + "step": 5541 + }, + { + "epoch": 0.42627490193062073, + "learning_rate": 0.0018443234158597806, + "loss": 1.0375, + "step": 5542 + }, + { + "epoch": 0.4263518190908392, + "learning_rate": 0.0018439706210539991, + "loss": 1.0948, + "step": 5543 + }, + { + "epoch": 0.4264287362510576, + "learning_rate": 0.0018436178061634156, + "loss": 1.4625, + "step": 5544 + }, + { + "epoch": 0.42650565341127605, + "learning_rate": 0.0018432649712086296, + "loss": 0.951, + "step": 5545 + }, + { + "epoch": 0.4265825705714945, + "learning_rate": 0.0018429121162102447, + "loss": 1.0098, + "step": 5546 + }, + { + "epoch": 0.42665948773171297, + "learning_rate": 0.0018425592411888635, + "loss": 0.9701, + "step": 5547 + }, + { + "epoch": 0.4267364048919314, + "learning_rate": 0.0018422063461650915, + "loss": 1.4429, + "step": 5548 + }, + { + "epoch": 0.42681332205214983, + "learning_rate": 0.0018418534311595341, + "loss": 1.1941, + "step": 5549 + }, + { + "epoch": 0.4268902392123683, + "learning_rate": 0.0018415004961927983, + "loss": 1.0569, + "step": 5550 + }, + { + "epoch": 0.4269671563725867, + "learning_rate": 0.0018411475412854934, + "loss": 1.1631, + "step": 5551 + }, + { + "epoch": 0.42704407353280516, + "learning_rate": 0.0018407945664582273, + "loss": 1.0821, + "step": 5552 + }, + { + "epoch": 0.4271209906930236, + "learning_rate": 0.0018404415717316119, + "loss": 1.229, + "step": 5553 + }, + { + "epoch": 0.4271979078532421, + "learning_rate": 0.0018400885571262576, + "loss": 1.1604, + "step": 5554 + }, + { + "epoch": 0.4272748250134605, + "learning_rate": 0.0018397355226627793, + "loss": 1.2475, + "step": 5555 + }, + { + "epoch": 0.42735174217367894, + "learning_rate": 0.0018393824683617885, + "loss": 0.8418, + "step": 5556 + }, + { + "epoch": 0.4274286593338974, + "learning_rate": 0.0018390293942439023, + "loss": 0.9409, + "step": 5557 + }, + { + "epoch": 0.42750557649411586, + "learning_rate": 0.0018386763003297357, + "loss": 0.9576, + "step": 5558 + }, + { + "epoch": 0.42758249365433426, + "learning_rate": 0.0018383231866399077, + "loss": 1.0819, + "step": 5559 + }, + { + "epoch": 0.4276594108145527, + "learning_rate": 0.0018379700531950362, + "loss": 1.0788, + "step": 5560 + }, + { + "epoch": 0.4277363279747712, + "learning_rate": 0.0018376169000157406, + "loss": 1.0406, + "step": 5561 + }, + { + "epoch": 0.42781324513498964, + "learning_rate": 0.0018372637271226427, + "loss": 1.2434, + "step": 5562 + }, + { + "epoch": 0.42789016229520804, + "learning_rate": 0.0018369105345363646, + "loss": 1.2091, + "step": 5563 + }, + { + "epoch": 0.4279670794554265, + "learning_rate": 0.0018365573222775288, + "loss": 1.5499, + "step": 5564 + }, + { + "epoch": 0.42804399661564496, + "learning_rate": 0.0018362040903667602, + "loss": 1.2836, + "step": 5565 + }, + { + "epoch": 0.4281209137758634, + "learning_rate": 0.0018358508388246844, + "loss": 1.3395, + "step": 5566 + }, + { + "epoch": 0.4281978309360818, + "learning_rate": 0.0018354975676719285, + "loss": 1.2794, + "step": 5567 + }, + { + "epoch": 0.4282747480963003, + "learning_rate": 0.0018351442769291195, + "loss": 0.9135, + "step": 5568 + }, + { + "epoch": 0.42835166525651874, + "learning_rate": 0.001834790966616887, + "loss": 0.9336, + "step": 5569 + }, + { + "epoch": 0.42842858241673715, + "learning_rate": 0.0018344376367558613, + "loss": 1.334, + "step": 5570 + }, + { + "epoch": 0.4285054995769556, + "learning_rate": 0.0018340842873666734, + "loss": 0.8759, + "step": 5571 + }, + { + "epoch": 0.42858241673717407, + "learning_rate": 0.0018337309184699555, + "loss": 1.2565, + "step": 5572 + }, + { + "epoch": 0.4286593338973925, + "learning_rate": 0.0018333775300863415, + "loss": 1.0318, + "step": 5573 + }, + { + "epoch": 0.42873625105761093, + "learning_rate": 0.0018330241222364663, + "loss": 1.0225, + "step": 5574 + }, + { + "epoch": 0.4288131682178294, + "learning_rate": 0.0018326706949409657, + "loss": 1.1291, + "step": 5575 + }, + { + "epoch": 0.42889008537804785, + "learning_rate": 0.001832317248220476, + "loss": 1.0597, + "step": 5576 + }, + { + "epoch": 0.4289670025382663, + "learning_rate": 0.0018319637820956361, + "loss": 1.4029, + "step": 5577 + }, + { + "epoch": 0.4290439196984847, + "learning_rate": 0.0018316102965870852, + "loss": 1.3606, + "step": 5578 + }, + { + "epoch": 0.42912083685870317, + "learning_rate": 0.0018312567917154632, + "loss": 0.8228, + "step": 5579 + }, + { + "epoch": 0.42919775401892163, + "learning_rate": 0.001830903267501412, + "loss": 0.9363, + "step": 5580 + }, + { + "epoch": 0.4292746711791401, + "learning_rate": 0.001830549723965574, + "loss": 1.0098, + "step": 5581 + }, + { + "epoch": 0.4293515883393585, + "learning_rate": 0.0018301961611285936, + "loss": 1.3514, + "step": 5582 + }, + { + "epoch": 0.42942850549957695, + "learning_rate": 0.0018298425790111146, + "loss": 1.3572, + "step": 5583 + }, + { + "epoch": 0.4295054226597954, + "learning_rate": 0.001829488977633784, + "loss": 0.7179, + "step": 5584 + }, + { + "epoch": 0.42958233982001387, + "learning_rate": 0.0018291353570172484, + "loss": 1.2272, + "step": 5585 + }, + { + "epoch": 0.4296592569802323, + "learning_rate": 0.0018287817171821568, + "loss": 1.2496, + "step": 5586 + }, + { + "epoch": 0.42973617414045073, + "learning_rate": 0.0018284280581491576, + "loss": 1.0349, + "step": 5587 + }, + { + "epoch": 0.4298130913006692, + "learning_rate": 0.001828074379938902, + "loss": 1.124, + "step": 5588 + }, + { + "epoch": 0.42989000846088765, + "learning_rate": 0.001827720682572041, + "loss": 0.9621, + "step": 5589 + }, + { + "epoch": 0.42996692562110606, + "learning_rate": 0.0018273669660692285, + "loss": 0.7076, + "step": 5590 + }, + { + "epoch": 0.4300438427813245, + "learning_rate": 0.0018270132304511173, + "loss": 1.2028, + "step": 5591 + }, + { + "epoch": 0.430120759941543, + "learning_rate": 0.0018266594757383626, + "loss": 1.197, + "step": 5592 + }, + { + "epoch": 0.4301976771017614, + "learning_rate": 0.0018263057019516213, + "loss": 1.3546, + "step": 5593 + }, + { + "epoch": 0.43027459426197984, + "learning_rate": 0.0018259519091115492, + "loss": 1.001, + "step": 5594 + }, + { + "epoch": 0.4303515114221983, + "learning_rate": 0.0018255980972388063, + "loss": 0.8959, + "step": 5595 + }, + { + "epoch": 0.43042842858241676, + "learning_rate": 0.0018252442663540505, + "loss": 1.3395, + "step": 5596 + }, + { + "epoch": 0.43050534574263516, + "learning_rate": 0.0018248904164779437, + "loss": 1.051, + "step": 5597 + }, + { + "epoch": 0.4305822629028536, + "learning_rate": 0.001824536547631146, + "loss": 1.0275, + "step": 5598 + }, + { + "epoch": 0.4306591800630721, + "learning_rate": 0.0018241826598343217, + "loss": 1.1403, + "step": 5599 + }, + { + "epoch": 0.43073609722329054, + "learning_rate": 0.0018238287531081334, + "loss": 1.1042, + "step": 5600 + }, + { + "epoch": 0.43081301438350894, + "learning_rate": 0.0018234748274732473, + "loss": 1.3162, + "step": 5601 + }, + { + "epoch": 0.4308899315437274, + "learning_rate": 0.0018231208829503286, + "loss": 1.0446, + "step": 5602 + }, + { + "epoch": 0.43096684870394586, + "learning_rate": 0.0018227669195600448, + "loss": 1.2103, + "step": 5603 + }, + { + "epoch": 0.4310437658641643, + "learning_rate": 0.0018224129373230647, + "loss": 1.2208, + "step": 5604 + }, + { + "epoch": 0.4311206830243827, + "learning_rate": 0.0018220589362600568, + "loss": 1.1432, + "step": 5605 + }, + { + "epoch": 0.4311976001846012, + "learning_rate": 0.001821704916391692, + "loss": 1.3191, + "step": 5606 + }, + { + "epoch": 0.43127451734481964, + "learning_rate": 0.0018213508777386418, + "loss": 1.0648, + "step": 5607 + }, + { + "epoch": 0.4313514345050381, + "learning_rate": 0.0018209968203215788, + "loss": 1.197, + "step": 5608 + }, + { + "epoch": 0.4314283516652565, + "learning_rate": 0.0018206427441611773, + "loss": 1.1027, + "step": 5609 + }, + { + "epoch": 0.43150526882547496, + "learning_rate": 0.0018202886492781117, + "loss": 1.0762, + "step": 5610 + }, + { + "epoch": 0.4315821859856934, + "learning_rate": 0.0018199345356930579, + "loss": 1.2543, + "step": 5611 + }, + { + "epoch": 0.4316591031459118, + "learning_rate": 0.001819580403426693, + "loss": 1.2031, + "step": 5612 + }, + { + "epoch": 0.4317360203061303, + "learning_rate": 0.0018192262524996956, + "loss": 1.2362, + "step": 5613 + }, + { + "epoch": 0.43181293746634875, + "learning_rate": 0.0018188720829327447, + "loss": 1.3068, + "step": 5614 + }, + { + "epoch": 0.4318898546265672, + "learning_rate": 0.0018185178947465201, + "loss": 1.1597, + "step": 5615 + }, + { + "epoch": 0.4319667717867856, + "learning_rate": 0.0018181636879617041, + "loss": 1.1482, + "step": 5616 + }, + { + "epoch": 0.43204368894700407, + "learning_rate": 0.0018178094625989787, + "loss": 1.3016, + "step": 5617 + }, + { + "epoch": 0.4321206061072225, + "learning_rate": 0.0018174552186790273, + "loss": 1.0175, + "step": 5618 + }, + { + "epoch": 0.432197523267441, + "learning_rate": 0.0018171009562225353, + "loss": 1.2541, + "step": 5619 + }, + { + "epoch": 0.4322744404276594, + "learning_rate": 0.0018167466752501877, + "loss": 1.147, + "step": 5620 + }, + { + "epoch": 0.43235135758787785, + "learning_rate": 0.0018163923757826717, + "loss": 0.9687, + "step": 5621 + }, + { + "epoch": 0.4324282747480963, + "learning_rate": 0.001816038057840675, + "loss": 1.0988, + "step": 5622 + }, + { + "epoch": 0.43250519190831477, + "learning_rate": 0.0018156837214448871, + "loss": 1.2579, + "step": 5623 + }, + { + "epoch": 0.43258210906853317, + "learning_rate": 0.0018153293666159976, + "loss": 1.4583, + "step": 5624 + }, + { + "epoch": 0.43265902622875163, + "learning_rate": 0.0018149749933746972, + "loss": 1.2231, + "step": 5625 + }, + { + "epoch": 0.4327359433889701, + "learning_rate": 0.0018146206017416793, + "loss": 1.2069, + "step": 5626 + }, + { + "epoch": 0.43281286054918855, + "learning_rate": 0.0018142661917376361, + "loss": 1.0469, + "step": 5627 + }, + { + "epoch": 0.43288977770940695, + "learning_rate": 0.001813911763383263, + "loss": 1.0712, + "step": 5628 + }, + { + "epoch": 0.4329666948696254, + "learning_rate": 0.0018135573166992542, + "loss": 0.8022, + "step": 5629 + }, + { + "epoch": 0.43304361202984387, + "learning_rate": 0.0018132028517063076, + "loss": 1.0543, + "step": 5630 + }, + { + "epoch": 0.4331205291900623, + "learning_rate": 0.0018128483684251198, + "loss": 1.0692, + "step": 5631 + }, + { + "epoch": 0.43319744635028073, + "learning_rate": 0.0018124938668763896, + "loss": 1.1745, + "step": 5632 + }, + { + "epoch": 0.4332743635104992, + "learning_rate": 0.001812139347080817, + "loss": 1.2819, + "step": 5633 + }, + { + "epoch": 0.43335128067071765, + "learning_rate": 0.0018117848090591026, + "loss": 1.1315, + "step": 5634 + }, + { + "epoch": 0.43342819783093606, + "learning_rate": 0.001811430252831948, + "loss": 0.8297, + "step": 5635 + }, + { + "epoch": 0.4335051149911545, + "learning_rate": 0.0018110756784200563, + "loss": 1.1009, + "step": 5636 + }, + { + "epoch": 0.433582032151373, + "learning_rate": 0.001810721085844132, + "loss": 1.7179, + "step": 5637 + }, + { + "epoch": 0.43365894931159144, + "learning_rate": 0.0018103664751248792, + "loss": 1.2815, + "step": 5638 + }, + { + "epoch": 0.43373586647180984, + "learning_rate": 0.0018100118462830055, + "loss": 0.7268, + "step": 5639 + }, + { + "epoch": 0.4338127836320283, + "learning_rate": 0.0018096571993392156, + "loss": 1.2658, + "step": 5640 + }, + { + "epoch": 0.43388970079224676, + "learning_rate": 0.00180930253431422, + "loss": 1.2305, + "step": 5641 + }, + { + "epoch": 0.4339666179524652, + "learning_rate": 0.0018089478512287268, + "loss": 1.2178, + "step": 5642 + }, + { + "epoch": 0.4340435351126836, + "learning_rate": 0.001808593150103447, + "loss": 1.0977, + "step": 5643 + }, + { + "epoch": 0.4341204522729021, + "learning_rate": 0.0018082384309590909, + "loss": 0.9142, + "step": 5644 + }, + { + "epoch": 0.43419736943312054, + "learning_rate": 0.0018078836938163721, + "loss": 1.1176, + "step": 5645 + }, + { + "epoch": 0.434274286593339, + "learning_rate": 0.001807528938696003, + "loss": 1.1686, + "step": 5646 + }, + { + "epoch": 0.4343512037535574, + "learning_rate": 0.001807174165618699, + "loss": 1.4266, + "step": 5647 + }, + { + "epoch": 0.43442812091377586, + "learning_rate": 0.0018068193746051755, + "loss": 0.9128, + "step": 5648 + }, + { + "epoch": 0.4345050380739943, + "learning_rate": 0.0018064645656761485, + "loss": 1.3591, + "step": 5649 + }, + { + "epoch": 0.4345819552342128, + "learning_rate": 0.0018061097388523365, + "loss": 0.9289, + "step": 5650 + }, + { + "epoch": 0.4346588723944312, + "learning_rate": 0.0018057548941544576, + "loss": 1.333, + "step": 5651 + }, + { + "epoch": 0.43473578955464964, + "learning_rate": 0.0018054000316032324, + "loss": 1.2954, + "step": 5652 + }, + { + "epoch": 0.4348127067148681, + "learning_rate": 0.0018050451512193805, + "loss": 1.2906, + "step": 5653 + }, + { + "epoch": 0.4348896238750865, + "learning_rate": 0.0018046902530236245, + "loss": 1.2782, + "step": 5654 + }, + { + "epoch": 0.43496654103530497, + "learning_rate": 0.001804335337036687, + "loss": 1.3585, + "step": 5655 + }, + { + "epoch": 0.4350434581955234, + "learning_rate": 0.001803980403279292, + "loss": 1.176, + "step": 5656 + }, + { + "epoch": 0.4351203753557419, + "learning_rate": 0.001803625451772164, + "loss": 1.3991, + "step": 5657 + }, + { + "epoch": 0.4351972925159603, + "learning_rate": 0.0018032704825360303, + "loss": 1.5181, + "step": 5658 + }, + { + "epoch": 0.43527420967617875, + "learning_rate": 0.0018029154955916166, + "loss": 1.3791, + "step": 5659 + }, + { + "epoch": 0.4353511268363972, + "learning_rate": 0.0018025604909596513, + "loss": 1.3091, + "step": 5660 + }, + { + "epoch": 0.43542804399661567, + "learning_rate": 0.0018022054686608639, + "loss": 0.866, + "step": 5661 + }, + { + "epoch": 0.43550496115683407, + "learning_rate": 0.0018018504287159842, + "loss": 1.2028, + "step": 5662 + }, + { + "epoch": 0.43558187831705253, + "learning_rate": 0.0018014953711457432, + "loss": 1.0731, + "step": 5663 + }, + { + "epoch": 0.435658795477271, + "learning_rate": 0.0018011402959708734, + "loss": 1.0611, + "step": 5664 + }, + { + "epoch": 0.43573571263748945, + "learning_rate": 0.0018007852032121076, + "loss": 1.2305, + "step": 5665 + }, + { + "epoch": 0.43581262979770785, + "learning_rate": 0.0018004300928901806, + "loss": 0.7995, + "step": 5666 + }, + { + "epoch": 0.4358895469579263, + "learning_rate": 0.0018000749650258273, + "loss": 1.0112, + "step": 5667 + }, + { + "epoch": 0.43596646411814477, + "learning_rate": 0.001799719819639784, + "loss": 1.1825, + "step": 5668 + }, + { + "epoch": 0.43604338127836323, + "learning_rate": 0.0017993646567527886, + "loss": 1.0716, + "step": 5669 + }, + { + "epoch": 0.43612029843858163, + "learning_rate": 0.0017990094763855784, + "loss": 1.0991, + "step": 5670 + }, + { + "epoch": 0.4361972155988001, + "learning_rate": 0.001798654278558893, + "loss": 1.1066, + "step": 5671 + }, + { + "epoch": 0.43627413275901855, + "learning_rate": 0.0017982990632934735, + "loss": 1.0699, + "step": 5672 + }, + { + "epoch": 0.43635104991923696, + "learning_rate": 0.0017979438306100605, + "loss": 1.3438, + "step": 5673 + }, + { + "epoch": 0.4364279670794554, + "learning_rate": 0.001797588580529397, + "loss": 1.0997, + "step": 5674 + }, + { + "epoch": 0.4365048842396739, + "learning_rate": 0.001797233313072226, + "loss": 1.1237, + "step": 5675 + }, + { + "epoch": 0.43658180139989233, + "learning_rate": 0.001796878028259292, + "loss": 1.2894, + "step": 5676 + }, + { + "epoch": 0.43665871856011074, + "learning_rate": 0.0017965227261113408, + "loss": 1.348, + "step": 5677 + }, + { + "epoch": 0.4367356357203292, + "learning_rate": 0.0017961674066491184, + "loss": 1.0409, + "step": 5678 + }, + { + "epoch": 0.43681255288054766, + "learning_rate": 0.0017958120698933725, + "loss": 1.4251, + "step": 5679 + }, + { + "epoch": 0.4368894700407661, + "learning_rate": 0.0017954567158648513, + "loss": 1.2653, + "step": 5680 + }, + { + "epoch": 0.4369663872009845, + "learning_rate": 0.0017951013445843054, + "loss": 0.9678, + "step": 5681 + }, + { + "epoch": 0.437043304361203, + "learning_rate": 0.0017947459560724833, + "loss": 1.2467, + "step": 5682 + }, + { + "epoch": 0.43712022152142144, + "learning_rate": 0.0017943905503501384, + "loss": 1.444, + "step": 5683 + }, + { + "epoch": 0.4371971386816399, + "learning_rate": 0.001794035127438022, + "loss": 1.026, + "step": 5684 + }, + { + "epoch": 0.4372740558418583, + "learning_rate": 0.0017936796873568887, + "loss": 1.3701, + "step": 5685 + }, + { + "epoch": 0.43735097300207676, + "learning_rate": 0.0017933242301274917, + "loss": 1.5726, + "step": 5686 + }, + { + "epoch": 0.4374278901622952, + "learning_rate": 0.0017929687557705876, + "loss": 1.0995, + "step": 5687 + }, + { + "epoch": 0.4375048073225137, + "learning_rate": 0.0017926132643069322, + "loss": 1.2614, + "step": 5688 + }, + { + "epoch": 0.4375817244827321, + "learning_rate": 0.0017922577557572838, + "loss": 1.1429, + "step": 5689 + }, + { + "epoch": 0.43765864164295054, + "learning_rate": 0.0017919022301424003, + "loss": 0.8692, + "step": 5690 + }, + { + "epoch": 0.437735558803169, + "learning_rate": 0.0017915466874830412, + "loss": 1.3014, + "step": 5691 + }, + { + "epoch": 0.4378124759633874, + "learning_rate": 0.0017911911277999676, + "loss": 1.3442, + "step": 5692 + }, + { + "epoch": 0.43788939312360586, + "learning_rate": 0.00179083555111394, + "loss": 1.2153, + "step": 5693 + }, + { + "epoch": 0.4379663102838243, + "learning_rate": 0.0017904799574457219, + "loss": 0.8997, + "step": 5694 + }, + { + "epoch": 0.4380432274440428, + "learning_rate": 0.001790124346816076, + "loss": 1.1885, + "step": 5695 + }, + { + "epoch": 0.4381201446042612, + "learning_rate": 0.0017897687192457675, + "loss": 1.3203, + "step": 5696 + }, + { + "epoch": 0.43819706176447965, + "learning_rate": 0.001789413074755561, + "loss": 1.3783, + "step": 5697 + }, + { + "epoch": 0.4382739789246981, + "learning_rate": 0.001789057413366224, + "loss": 1.1416, + "step": 5698 + }, + { + "epoch": 0.43835089608491656, + "learning_rate": 0.001788701735098523, + "loss": 1.0234, + "step": 5699 + }, + { + "epoch": 0.43842781324513497, + "learning_rate": 0.0017883460399732267, + "loss": 1.4129, + "step": 5700 + }, + { + "epoch": 0.4385047304053534, + "learning_rate": 0.0017879903280111052, + "loss": 0.9845, + "step": 5701 + }, + { + "epoch": 0.4385816475655719, + "learning_rate": 0.0017876345992329278, + "loss": 1.2713, + "step": 5702 + }, + { + "epoch": 0.43865856472579035, + "learning_rate": 0.0017872788536594665, + "loss": 1.2226, + "step": 5703 + }, + { + "epoch": 0.43873548188600875, + "learning_rate": 0.0017869230913114937, + "loss": 1.0973, + "step": 5704 + }, + { + "epoch": 0.4388123990462272, + "learning_rate": 0.0017865673122097825, + "loss": 0.9595, + "step": 5705 + }, + { + "epoch": 0.43888931620644567, + "learning_rate": 0.001786211516375107, + "loss": 1.3162, + "step": 5706 + }, + { + "epoch": 0.4389662333666641, + "learning_rate": 0.0017858557038282433, + "loss": 1.8682, + "step": 5707 + }, + { + "epoch": 0.43904315052688253, + "learning_rate": 0.001785499874589967, + "loss": 1.0754, + "step": 5708 + }, + { + "epoch": 0.439120067687101, + "learning_rate": 0.0017851440286810554, + "loss": 0.9753, + "step": 5709 + }, + { + "epoch": 0.43919698484731945, + "learning_rate": 0.0017847881661222865, + "loss": 0.9731, + "step": 5710 + }, + { + "epoch": 0.4392739020075379, + "learning_rate": 0.00178443228693444, + "loss": 1.3326, + "step": 5711 + }, + { + "epoch": 0.4393508191677563, + "learning_rate": 0.0017840763911382961, + "loss": 1.1624, + "step": 5712 + }, + { + "epoch": 0.43942773632797477, + "learning_rate": 0.0017837204787546353, + "loss": 1.0467, + "step": 5713 + }, + { + "epoch": 0.43950465348819323, + "learning_rate": 0.0017833645498042403, + "loss": 1.3491, + "step": 5714 + }, + { + "epoch": 0.43958157064841163, + "learning_rate": 0.0017830086043078936, + "loss": 1.1612, + "step": 5715 + }, + { + "epoch": 0.4396584878086301, + "learning_rate": 0.0017826526422863801, + "loss": 1.4372, + "step": 5716 + }, + { + "epoch": 0.43973540496884855, + "learning_rate": 0.0017822966637604837, + "loss": 1.0515, + "step": 5717 + }, + { + "epoch": 0.439812322129067, + "learning_rate": 0.001781940668750991, + "loss": 1.3669, + "step": 5718 + }, + { + "epoch": 0.4398892392892854, + "learning_rate": 0.0017815846572786893, + "loss": 0.7621, + "step": 5719 + }, + { + "epoch": 0.4399661564495039, + "learning_rate": 0.0017812286293643657, + "loss": 1.3997, + "step": 5720 + }, + { + "epoch": 0.44004307360972233, + "learning_rate": 0.0017808725850288088, + "loss": 1.1387, + "step": 5721 + }, + { + "epoch": 0.4401199907699408, + "learning_rate": 0.0017805165242928095, + "loss": 0.9814, + "step": 5722 + }, + { + "epoch": 0.4401969079301592, + "learning_rate": 0.0017801604471771582, + "loss": 1.1478, + "step": 5723 + }, + { + "epoch": 0.44027382509037766, + "learning_rate": 0.001779804353702646, + "loss": 1.3558, + "step": 5724 + }, + { + "epoch": 0.4403507422505961, + "learning_rate": 0.001779448243890066, + "loss": 1.2087, + "step": 5725 + }, + { + "epoch": 0.4404276594108146, + "learning_rate": 0.0017790921177602117, + "loss": 0.8501, + "step": 5726 + }, + { + "epoch": 0.440504576571033, + "learning_rate": 0.0017787359753338786, + "loss": 1.232, + "step": 5727 + }, + { + "epoch": 0.44058149373125144, + "learning_rate": 0.0017783798166318605, + "loss": 0.9937, + "step": 5728 + }, + { + "epoch": 0.4406584108914699, + "learning_rate": 0.001778023641674955, + "loss": 1.2035, + "step": 5729 + }, + { + "epoch": 0.44073532805168836, + "learning_rate": 0.0017776674504839593, + "loss": 0.9493, + "step": 5730 + }, + { + "epoch": 0.44081224521190676, + "learning_rate": 0.001777311243079672, + "loss": 1.2342, + "step": 5731 + }, + { + "epoch": 0.4408891623721252, + "learning_rate": 0.0017769550194828923, + "loss": 1.0352, + "step": 5732 + }, + { + "epoch": 0.4409660795323437, + "learning_rate": 0.0017765987797144203, + "loss": 0.9839, + "step": 5733 + }, + { + "epoch": 0.4410429966925621, + "learning_rate": 0.0017762425237950572, + "loss": 1.2379, + "step": 5734 + }, + { + "epoch": 0.44111991385278054, + "learning_rate": 0.0017758862517456048, + "loss": 1.3134, + "step": 5735 + }, + { + "epoch": 0.441196831012999, + "learning_rate": 0.0017755299635868673, + "loss": 1.2512, + "step": 5736 + }, + { + "epoch": 0.44127374817321746, + "learning_rate": 0.0017751736593396474, + "loss": 1.0343, + "step": 5737 + }, + { + "epoch": 0.44135066533343587, + "learning_rate": 0.0017748173390247518, + "loss": 1.2389, + "step": 5738 + }, + { + "epoch": 0.4414275824936543, + "learning_rate": 0.0017744610026629846, + "loss": 1.0723, + "step": 5739 + }, + { + "epoch": 0.4415044996538728, + "learning_rate": 0.0017741046502751535, + "loss": 0.9816, + "step": 5740 + }, + { + "epoch": 0.44158141681409124, + "learning_rate": 0.001773748281882066, + "loss": 1.0717, + "step": 5741 + }, + { + "epoch": 0.44165833397430965, + "learning_rate": 0.0017733918975045314, + "loss": 1.069, + "step": 5742 + }, + { + "epoch": 0.4417352511345281, + "learning_rate": 0.001773035497163359, + "loss": 0.9084, + "step": 5743 + }, + { + "epoch": 0.44181216829474657, + "learning_rate": 0.0017726790808793589, + "loss": 1.2703, + "step": 5744 + }, + { + "epoch": 0.441889085454965, + "learning_rate": 0.0017723226486733436, + "loss": 1.0079, + "step": 5745 + }, + { + "epoch": 0.44196600261518343, + "learning_rate": 0.001771966200566125, + "loss": 1.2455, + "step": 5746 + }, + { + "epoch": 0.4420429197754019, + "learning_rate": 0.0017716097365785164, + "loss": 1.0852, + "step": 5747 + }, + { + "epoch": 0.44211983693562035, + "learning_rate": 0.0017712532567313321, + "loss": 1.359, + "step": 5748 + }, + { + "epoch": 0.4421967540958388, + "learning_rate": 0.0017708967610453876, + "loss": 0.7557, + "step": 5749 + }, + { + "epoch": 0.4422736712560572, + "learning_rate": 0.001770540249541499, + "loss": 1.3303, + "step": 5750 + }, + { + "epoch": 0.44235058841627567, + "learning_rate": 0.0017701837222404834, + "loss": 1.401, + "step": 5751 + }, + { + "epoch": 0.44242750557649413, + "learning_rate": 0.0017698271791631577, + "loss": 1.2947, + "step": 5752 + }, + { + "epoch": 0.4425044227367126, + "learning_rate": 0.0017694706203303428, + "loss": 0.7679, + "step": 5753 + }, + { + "epoch": 0.442581339896931, + "learning_rate": 0.0017691140457628573, + "loss": 1.3605, + "step": 5754 + }, + { + "epoch": 0.44265825705714945, + "learning_rate": 0.001768757455481522, + "loss": 1.3312, + "step": 5755 + }, + { + "epoch": 0.4427351742173679, + "learning_rate": 0.0017684008495071591, + "loss": 1.252, + "step": 5756 + }, + { + "epoch": 0.4428120913775863, + "learning_rate": 0.0017680442278605907, + "loss": 1.215, + "step": 5757 + }, + { + "epoch": 0.4428890085378048, + "learning_rate": 0.0017676875905626404, + "loss": 0.9961, + "step": 5758 + }, + { + "epoch": 0.44296592569802323, + "learning_rate": 0.0017673309376341328, + "loss": 0.9695, + "step": 5759 + }, + { + "epoch": 0.4430428428582417, + "learning_rate": 0.0017669742690958933, + "loss": 1.3749, + "step": 5760 + }, + { + "epoch": 0.4431197600184601, + "learning_rate": 0.0017666175849687478, + "loss": 1.0788, + "step": 5761 + }, + { + "epoch": 0.44319667717867856, + "learning_rate": 0.001766260885273524, + "loss": 1.4633, + "step": 5762 + }, + { + "epoch": 0.443273594338897, + "learning_rate": 0.001765904170031049, + "loss": 1.0799, + "step": 5763 + }, + { + "epoch": 0.4433505114991155, + "learning_rate": 0.0017655474392621526, + "loss": 1.1455, + "step": 5764 + }, + { + "epoch": 0.4434274286593339, + "learning_rate": 0.0017651906929876647, + "loss": 1.2592, + "step": 5765 + }, + { + "epoch": 0.44350434581955234, + "learning_rate": 0.0017648339312284155, + "loss": 1.2114, + "step": 5766 + }, + { + "epoch": 0.4435812629797708, + "learning_rate": 0.0017644771540052376, + "loss": 1.0834, + "step": 5767 + }, + { + "epoch": 0.44365818013998926, + "learning_rate": 0.0017641203613389625, + "loss": 1.2354, + "step": 5768 + }, + { + "epoch": 0.44373509730020766, + "learning_rate": 0.0017637635532504249, + "loss": 1.2626, + "step": 5769 + }, + { + "epoch": 0.4438120144604261, + "learning_rate": 0.001763406729760458, + "loss": 1.2834, + "step": 5770 + }, + { + "epoch": 0.4438889316206446, + "learning_rate": 0.0017630498908898977, + "loss": 1.4358, + "step": 5771 + }, + { + "epoch": 0.44396584878086304, + "learning_rate": 0.0017626930366595797, + "loss": 1.4712, + "step": 5772 + }, + { + "epoch": 0.44404276594108144, + "learning_rate": 0.0017623361670903426, + "loss": 1.3903, + "step": 5773 + }, + { + "epoch": 0.4441196831012999, + "learning_rate": 0.0017619792822030226, + "loss": 1.1489, + "step": 5774 + }, + { + "epoch": 0.44419660026151836, + "learning_rate": 0.0017616223820184594, + "loss": 1.3875, + "step": 5775 + }, + { + "epoch": 0.44427351742173676, + "learning_rate": 0.0017612654665574926, + "loss": 0.8662, + "step": 5776 + }, + { + "epoch": 0.4443504345819552, + "learning_rate": 0.001760908535840963, + "loss": 1.3263, + "step": 5777 + }, + { + "epoch": 0.4444273517421737, + "learning_rate": 0.0017605515898897121, + "loss": 1.2111, + "step": 5778 + }, + { + "epoch": 0.44450426890239214, + "learning_rate": 0.0017601946287245818, + "loss": 1.7218, + "step": 5779 + }, + { + "epoch": 0.44458118606261054, + "learning_rate": 0.0017598376523664172, + "loss": 1.4569, + "step": 5780 + }, + { + "epoch": 0.444658103222829, + "learning_rate": 0.0017594806608360603, + "loss": 1.1493, + "step": 5781 + }, + { + "epoch": 0.44473502038304746, + "learning_rate": 0.0017591236541543574, + "loss": 1.0143, + "step": 5782 + }, + { + "epoch": 0.4448119375432659, + "learning_rate": 0.001758766632342154, + "loss": 1.2816, + "step": 5783 + }, + { + "epoch": 0.4448888547034843, + "learning_rate": 0.001758409595420298, + "loss": 1.386, + "step": 5784 + }, + { + "epoch": 0.4449657718637028, + "learning_rate": 0.0017580525434096354, + "loss": 1.3543, + "step": 5785 + }, + { + "epoch": 0.44504268902392125, + "learning_rate": 0.0017576954763310168, + "loss": 1.3586, + "step": 5786 + }, + { + "epoch": 0.4451196061841397, + "learning_rate": 0.0017573383942052902, + "loss": 1.0857, + "step": 5787 + }, + { + "epoch": 0.4451965233443581, + "learning_rate": 0.0017569812970533068, + "loss": 1.2279, + "step": 5788 + }, + { + "epoch": 0.44527344050457657, + "learning_rate": 0.0017566241848959176, + "loss": 1.2447, + "step": 5789 + }, + { + "epoch": 0.445350357664795, + "learning_rate": 0.0017562670577539747, + "loss": 0.8943, + "step": 5790 + }, + { + "epoch": 0.4454272748250135, + "learning_rate": 0.0017559099156483312, + "loss": 1.0126, + "step": 5791 + }, + { + "epoch": 0.4455041919852319, + "learning_rate": 0.0017555527585998414, + "loss": 1.0517, + "step": 5792 + }, + { + "epoch": 0.44558110914545035, + "learning_rate": 0.0017551955866293596, + "loss": 1.0309, + "step": 5793 + }, + { + "epoch": 0.4456580263056688, + "learning_rate": 0.0017548383997577411, + "loss": 0.8542, + "step": 5794 + }, + { + "epoch": 0.4457349434658872, + "learning_rate": 0.0017544811980058437, + "loss": 1.2182, + "step": 5795 + }, + { + "epoch": 0.44581186062610567, + "learning_rate": 0.001754123981394523, + "loss": 1.0964, + "step": 5796 + }, + { + "epoch": 0.44588877778632413, + "learning_rate": 0.0017537667499446386, + "loss": 1.03, + "step": 5797 + }, + { + "epoch": 0.4459656949465426, + "learning_rate": 0.0017534095036770491, + "loss": 1.2094, + "step": 5798 + }, + { + "epoch": 0.446042612106761, + "learning_rate": 0.0017530522426126148, + "loss": 1.1494, + "step": 5799 + }, + { + "epoch": 0.44611952926697945, + "learning_rate": 0.0017526949667721959, + "loss": 1.0827, + "step": 5800 + }, + { + "epoch": 0.4461964464271979, + "learning_rate": 0.0017523376761766547, + "loss": 1.0839, + "step": 5801 + }, + { + "epoch": 0.44627336358741637, + "learning_rate": 0.0017519803708468536, + "loss": 1.3063, + "step": 5802 + }, + { + "epoch": 0.4463502807476348, + "learning_rate": 0.0017516230508036565, + "loss": 1.3156, + "step": 5803 + }, + { + "epoch": 0.44642719790785323, + "learning_rate": 0.001751265716067927, + "loss": 1.3592, + "step": 5804 + }, + { + "epoch": 0.4465041150680717, + "learning_rate": 0.00175090836666053, + "loss": 0.8599, + "step": 5805 + }, + { + "epoch": 0.44658103222829015, + "learning_rate": 0.0017505510026023326, + "loss": 0.965, + "step": 5806 + }, + { + "epoch": 0.44665794938850856, + "learning_rate": 0.0017501936239142008, + "loss": 1.2277, + "step": 5807 + }, + { + "epoch": 0.446734866548727, + "learning_rate": 0.0017498362306170028, + "loss": 0.9069, + "step": 5808 + }, + { + "epoch": 0.4468117837089455, + "learning_rate": 0.0017494788227316062, + "loss": 1.3905, + "step": 5809 + }, + { + "epoch": 0.44688870086916394, + "learning_rate": 0.0017491214002788816, + "loss": 1.2333, + "step": 5810 + }, + { + "epoch": 0.44696561802938234, + "learning_rate": 0.001748763963279699, + "loss": 1.1398, + "step": 5811 + }, + { + "epoch": 0.4470425351896008, + "learning_rate": 0.0017484065117549289, + "loss": 1.2049, + "step": 5812 + }, + { + "epoch": 0.44711945234981926, + "learning_rate": 0.001748049045725444, + "loss": 1.0972, + "step": 5813 + }, + { + "epoch": 0.4471963695100377, + "learning_rate": 0.0017476915652121165, + "loss": 0.8411, + "step": 5814 + }, + { + "epoch": 0.4472732866702561, + "learning_rate": 0.001747334070235821, + "loss": 1.3048, + "step": 5815 + }, + { + "epoch": 0.4473502038304746, + "learning_rate": 0.0017469765608174306, + "loss": 0.9927, + "step": 5816 + }, + { + "epoch": 0.44742712099069304, + "learning_rate": 0.0017466190369778219, + "loss": 1.1908, + "step": 5817 + }, + { + "epoch": 0.44750403815091144, + "learning_rate": 0.0017462614987378707, + "loss": 1.0945, + "step": 5818 + }, + { + "epoch": 0.4475809553111299, + "learning_rate": 0.0017459039461184536, + "loss": 1.183, + "step": 5819 + }, + { + "epoch": 0.44765787247134836, + "learning_rate": 0.001745546379140449, + "loss": 0.9948, + "step": 5820 + }, + { + "epoch": 0.4477347896315668, + "learning_rate": 0.0017451887978247353, + "loss": 1.2216, + "step": 5821 + }, + { + "epoch": 0.4478117067917852, + "learning_rate": 0.0017448312021921923, + "loss": 1.1367, + "step": 5822 + }, + { + "epoch": 0.4478886239520037, + "learning_rate": 0.0017444735922637, + "loss": 0.8306, + "step": 5823 + }, + { + "epoch": 0.44796554111222214, + "learning_rate": 0.0017441159680601404, + "loss": 1.0624, + "step": 5824 + }, + { + "epoch": 0.4480424582724406, + "learning_rate": 0.0017437583296023944, + "loss": 1.1724, + "step": 5825 + }, + { + "epoch": 0.448119375432659, + "learning_rate": 0.0017434006769113462, + "loss": 1.2228, + "step": 5826 + }, + { + "epoch": 0.44819629259287747, + "learning_rate": 0.001743043010007878, + "loss": 1.3137, + "step": 5827 + }, + { + "epoch": 0.4482732097530959, + "learning_rate": 0.0017426853289128758, + "loss": 1.1691, + "step": 5828 + }, + { + "epoch": 0.4483501269133144, + "learning_rate": 0.0017423276336472238, + "loss": 1.2528, + "step": 5829 + }, + { + "epoch": 0.4484270440735328, + "learning_rate": 0.001741969924231809, + "loss": 1.4442, + "step": 5830 + }, + { + "epoch": 0.44850396123375125, + "learning_rate": 0.0017416122006875183, + "loss": 1.3941, + "step": 5831 + }, + { + "epoch": 0.4485808783939697, + "learning_rate": 0.0017412544630352389, + "loss": 1.0074, + "step": 5832 + }, + { + "epoch": 0.44865779555418817, + "learning_rate": 0.0017408967112958606, + "loss": 0.8635, + "step": 5833 + }, + { + "epoch": 0.44873471271440657, + "learning_rate": 0.0017405389454902718, + "loss": 1.213, + "step": 5834 + }, + { + "epoch": 0.44881162987462503, + "learning_rate": 0.0017401811656393637, + "loss": 1.4124, + "step": 5835 + }, + { + "epoch": 0.4488885470348435, + "learning_rate": 0.0017398233717640264, + "loss": 1.3143, + "step": 5836 + }, + { + "epoch": 0.4489654641950619, + "learning_rate": 0.0017394655638851533, + "loss": 0.8716, + "step": 5837 + }, + { + "epoch": 0.44904238135528035, + "learning_rate": 0.0017391077420236358, + "loss": 1.4332, + "step": 5838 + }, + { + "epoch": 0.4491192985154988, + "learning_rate": 0.0017387499062003684, + "loss": 1.2571, + "step": 5839 + }, + { + "epoch": 0.44919621567571727, + "learning_rate": 0.0017383920564362446, + "loss": 1.1221, + "step": 5840 + }, + { + "epoch": 0.4492731328359357, + "learning_rate": 0.0017380341927521604, + "loss": 0.9311, + "step": 5841 + }, + { + "epoch": 0.44935004999615413, + "learning_rate": 0.001737676315169012, + "loss": 1.265, + "step": 5842 + }, + { + "epoch": 0.4494269671563726, + "learning_rate": 0.0017373184237076953, + "loss": 1.0609, + "step": 5843 + }, + { + "epoch": 0.44950388431659105, + "learning_rate": 0.001736960518389109, + "loss": 1.4313, + "step": 5844 + }, + { + "epoch": 0.44958080147680946, + "learning_rate": 0.0017366025992341509, + "loss": 1.187, + "step": 5845 + }, + { + "epoch": 0.4496577186370279, + "learning_rate": 0.0017362446662637203, + "loss": 0.9655, + "step": 5846 + }, + { + "epoch": 0.4497346357972464, + "learning_rate": 0.0017358867194987176, + "loss": 1.354, + "step": 5847 + }, + { + "epoch": 0.44981155295746483, + "learning_rate": 0.0017355287589600435, + "loss": 1.2368, + "step": 5848 + }, + { + "epoch": 0.44988847011768324, + "learning_rate": 0.0017351707846685997, + "loss": 1.2649, + "step": 5849 + }, + { + "epoch": 0.4499653872779017, + "learning_rate": 0.0017348127966452889, + "loss": 0.961, + "step": 5850 + }, + { + "epoch": 0.45004230443812016, + "learning_rate": 0.0017344547949110138, + "loss": 1.3079, + "step": 5851 + }, + { + "epoch": 0.4501192215983386, + "learning_rate": 0.0017340967794866788, + "loss": 0.9241, + "step": 5852 + }, + { + "epoch": 0.450196138758557, + "learning_rate": 0.0017337387503931895, + "loss": 1.0215, + "step": 5853 + }, + { + "epoch": 0.4502730559187755, + "learning_rate": 0.0017333807076514503, + "loss": 1.0799, + "step": 5854 + }, + { + "epoch": 0.45034997307899394, + "learning_rate": 0.0017330226512823688, + "loss": 1.0383, + "step": 5855 + }, + { + "epoch": 0.45042689023921234, + "learning_rate": 0.0017326645813068517, + "loss": 1.2903, + "step": 5856 + }, + { + "epoch": 0.4505038073994308, + "learning_rate": 0.001732306497745807, + "loss": 1.0836, + "step": 5857 + }, + { + "epoch": 0.45058072455964926, + "learning_rate": 0.001731948400620144, + "loss": 0.9747, + "step": 5858 + }, + { + "epoch": 0.4506576417198677, + "learning_rate": 0.0017315902899507722, + "loss": 1.3402, + "step": 5859 + }, + { + "epoch": 0.4507345588800861, + "learning_rate": 0.0017312321657586018, + "loss": 1.0289, + "step": 5860 + }, + { + "epoch": 0.4508114760403046, + "learning_rate": 0.0017308740280645447, + "loss": 1.0714, + "step": 5861 + }, + { + "epoch": 0.45088839320052304, + "learning_rate": 0.0017305158768895121, + "loss": 1.3094, + "step": 5862 + }, + { + "epoch": 0.4509653103607415, + "learning_rate": 0.0017301577122544175, + "loss": 1.3609, + "step": 5863 + }, + { + "epoch": 0.4510422275209599, + "learning_rate": 0.001729799534180174, + "loss": 1.097, + "step": 5864 + }, + { + "epoch": 0.45111914468117836, + "learning_rate": 0.0017294413426876961, + "loss": 1.1695, + "step": 5865 + }, + { + "epoch": 0.4511960618413968, + "learning_rate": 0.0017290831377978991, + "loss": 1.1512, + "step": 5866 + }, + { + "epoch": 0.4512729790016153, + "learning_rate": 0.0017287249195316989, + "loss": 1.2946, + "step": 5867 + }, + { + "epoch": 0.4513498961618337, + "learning_rate": 0.0017283666879100127, + "loss": 1.1683, + "step": 5868 + }, + { + "epoch": 0.45142681332205215, + "learning_rate": 0.0017280084429537573, + "loss": 1.0327, + "step": 5869 + }, + { + "epoch": 0.4515037304822706, + "learning_rate": 0.0017276501846838514, + "loss": 1.0762, + "step": 5870 + }, + { + "epoch": 0.45158064764248906, + "learning_rate": 0.0017272919131212138, + "loss": 1.0974, + "step": 5871 + }, + { + "epoch": 0.45165756480270747, + "learning_rate": 0.001726933628286765, + "loss": 1.1356, + "step": 5872 + }, + { + "epoch": 0.4517344819629259, + "learning_rate": 0.0017265753302014245, + "loss": 0.9996, + "step": 5873 + }, + { + "epoch": 0.4518113991231444, + "learning_rate": 0.0017262170188861144, + "loss": 1.367, + "step": 5874 + }, + { + "epoch": 0.45188831628336285, + "learning_rate": 0.001725858694361757, + "loss": 1.3565, + "step": 5875 + }, + { + "epoch": 0.45196523344358125, + "learning_rate": 0.001725500356649275, + "loss": 1.0324, + "step": 5876 + }, + { + "epoch": 0.4520421506037997, + "learning_rate": 0.0017251420057695923, + "loss": 1.1557, + "step": 5877 + }, + { + "epoch": 0.45211906776401817, + "learning_rate": 0.0017247836417436332, + "loss": 1.046, + "step": 5878 + }, + { + "epoch": 0.45219598492423657, + "learning_rate": 0.0017244252645923232, + "loss": 1.2456, + "step": 5879 + }, + { + "epoch": 0.45227290208445503, + "learning_rate": 0.0017240668743365877, + "loss": 0.9816, + "step": 5880 + }, + { + "epoch": 0.4523498192446735, + "learning_rate": 0.001723708470997354, + "loss": 0.8587, + "step": 5881 + }, + { + "epoch": 0.45242673640489195, + "learning_rate": 0.0017233500545955492, + "loss": 1.2587, + "step": 5882 + }, + { + "epoch": 0.45250365356511035, + "learning_rate": 0.0017229916251521027, + "loss": 1.4925, + "step": 5883 + }, + { + "epoch": 0.4525805707253288, + "learning_rate": 0.0017226331826879421, + "loss": 0.8827, + "step": 5884 + }, + { + "epoch": 0.45265748788554727, + "learning_rate": 0.0017222747272239987, + "loss": 1.0074, + "step": 5885 + }, + { + "epoch": 0.45273440504576573, + "learning_rate": 0.0017219162587812015, + "loss": 0.9539, + "step": 5886 + }, + { + "epoch": 0.45281132220598413, + "learning_rate": 0.0017215577773804832, + "loss": 1.323, + "step": 5887 + }, + { + "epoch": 0.4528882393662026, + "learning_rate": 0.0017211992830427755, + "loss": 1.3439, + "step": 5888 + }, + { + "epoch": 0.45296515652642105, + "learning_rate": 0.0017208407757890107, + "loss": 1.5041, + "step": 5889 + }, + { + "epoch": 0.4530420736866395, + "learning_rate": 0.001720482255640123, + "loss": 0.9133, + "step": 5890 + }, + { + "epoch": 0.4531189908468579, + "learning_rate": 0.0017201237226170468, + "loss": 1.3905, + "step": 5891 + }, + { + "epoch": 0.4531959080070764, + "learning_rate": 0.001719765176740717, + "loss": 1.0082, + "step": 5892 + }, + { + "epoch": 0.45327282516729483, + "learning_rate": 0.0017194066180320692, + "loss": 0.8672, + "step": 5893 + }, + { + "epoch": 0.4533497423275133, + "learning_rate": 0.0017190480465120405, + "loss": 1.1731, + "step": 5894 + }, + { + "epoch": 0.4534266594877317, + "learning_rate": 0.001718689462201568, + "loss": 1.1501, + "step": 5895 + }, + { + "epoch": 0.45350357664795016, + "learning_rate": 0.00171833086512159, + "loss": 1.1021, + "step": 5896 + }, + { + "epoch": 0.4535804938081686, + "learning_rate": 0.001717972255293045, + "loss": 1.3561, + "step": 5897 + }, + { + "epoch": 0.453657410968387, + "learning_rate": 0.0017176136327368735, + "loss": 1.106, + "step": 5898 + }, + { + "epoch": 0.4537343281286055, + "learning_rate": 0.0017172549974740147, + "loss": 1.1652, + "step": 5899 + }, + { + "epoch": 0.45381124528882394, + "learning_rate": 0.0017168963495254098, + "loss": 0.9884, + "step": 5900 + }, + { + "epoch": 0.4538881624490424, + "learning_rate": 0.0017165376889120018, + "loss": 1.6408, + "step": 5901 + }, + { + "epoch": 0.4539650796092608, + "learning_rate": 0.001716179015654732, + "loss": 1.4123, + "step": 5902 + }, + { + "epoch": 0.45404199676947926, + "learning_rate": 0.0017158203297745443, + "loss": 1.0873, + "step": 5903 + }, + { + "epoch": 0.4541189139296977, + "learning_rate": 0.0017154616312923825, + "loss": 0.9099, + "step": 5904 + }, + { + "epoch": 0.4541958310899162, + "learning_rate": 0.0017151029202291917, + "loss": 0.9838, + "step": 5905 + }, + { + "epoch": 0.4542727482501346, + "learning_rate": 0.0017147441966059173, + "loss": 0.8756, + "step": 5906 + }, + { + "epoch": 0.45434966541035304, + "learning_rate": 0.001714385460443505, + "loss": 0.9054, + "step": 5907 + }, + { + "epoch": 0.4544265825705715, + "learning_rate": 0.0017140267117629027, + "loss": 1.375, + "step": 5908 + }, + { + "epoch": 0.45450349973078996, + "learning_rate": 0.0017136679505850572, + "loss": 1.2602, + "step": 5909 + }, + { + "epoch": 0.45458041689100837, + "learning_rate": 0.001713309176930918, + "loss": 1.0451, + "step": 5910 + }, + { + "epoch": 0.4546573340512268, + "learning_rate": 0.0017129503908214331, + "loss": 1.2153, + "step": 5911 + }, + { + "epoch": 0.4547342512114453, + "learning_rate": 0.001712591592277553, + "loss": 1.3271, + "step": 5912 + }, + { + "epoch": 0.45481116837166374, + "learning_rate": 0.0017122327813202283, + "loss": 1.3862, + "step": 5913 + }, + { + "epoch": 0.45488808553188215, + "learning_rate": 0.001711873957970411, + "loss": 1.17, + "step": 5914 + }, + { + "epoch": 0.4549650026921006, + "learning_rate": 0.0017115151222490516, + "loss": 1.4342, + "step": 5915 + }, + { + "epoch": 0.45504191985231907, + "learning_rate": 0.0017111562741771043, + "loss": 1.3069, + "step": 5916 + }, + { + "epoch": 0.45511883701253747, + "learning_rate": 0.0017107974137755219, + "loss": 1.0684, + "step": 5917 + }, + { + "epoch": 0.45519575417275593, + "learning_rate": 0.0017104385410652585, + "loss": 1.1121, + "step": 5918 + }, + { + "epoch": 0.4552726713329744, + "learning_rate": 0.0017100796560672697, + "loss": 1.2236, + "step": 5919 + }, + { + "epoch": 0.45534958849319285, + "learning_rate": 0.0017097207588025105, + "loss": 1.1302, + "step": 5920 + }, + { + "epoch": 0.45542650565341125, + "learning_rate": 0.0017093618492919386, + "loss": 1.1827, + "step": 5921 + }, + { + "epoch": 0.4555034228136297, + "learning_rate": 0.0017090029275565093, + "loss": 1.0719, + "step": 5922 + }, + { + "epoch": 0.45558033997384817, + "learning_rate": 0.0017086439936171815, + "loss": 1.2071, + "step": 5923 + }, + { + "epoch": 0.45565725713406663, + "learning_rate": 0.0017082850474949129, + "loss": 1.2975, + "step": 5924 + }, + { + "epoch": 0.45573417429428503, + "learning_rate": 0.0017079260892106643, + "loss": 1.4163, + "step": 5925 + }, + { + "epoch": 0.4558110914545035, + "learning_rate": 0.001707567118785394, + "loss": 1.0864, + "step": 5926 + }, + { + "epoch": 0.45588800861472195, + "learning_rate": 0.0017072081362400635, + "loss": 1.3652, + "step": 5927 + }, + { + "epoch": 0.4559649257749404, + "learning_rate": 0.001706849141595634, + "loss": 1.2553, + "step": 5928 + }, + { + "epoch": 0.4560418429351588, + "learning_rate": 0.0017064901348730673, + "loss": 1.3937, + "step": 5929 + }, + { + "epoch": 0.4561187600953773, + "learning_rate": 0.0017061311160933268, + "loss": 0.9029, + "step": 5930 + }, + { + "epoch": 0.45619567725559573, + "learning_rate": 0.0017057720852773754, + "loss": 1.3697, + "step": 5931 + }, + { + "epoch": 0.4562725944158142, + "learning_rate": 0.0017054130424461775, + "loss": 0.9505, + "step": 5932 + }, + { + "epoch": 0.4563495115760326, + "learning_rate": 0.001705053987620698, + "loss": 0.9906, + "step": 5933 + }, + { + "epoch": 0.45642642873625106, + "learning_rate": 0.0017046949208219028, + "loss": 1.4199, + "step": 5934 + }, + { + "epoch": 0.4565033458964695, + "learning_rate": 0.0017043358420707568, + "loss": 1.0986, + "step": 5935 + }, + { + "epoch": 0.456580263056688, + "learning_rate": 0.0017039767513882293, + "loss": 1.0399, + "step": 5936 + }, + { + "epoch": 0.4566571802169064, + "learning_rate": 0.0017036176487952859, + "loss": 1.2513, + "step": 5937 + }, + { + "epoch": 0.45673409737712484, + "learning_rate": 0.001703258534312896, + "loss": 0.9261, + "step": 5938 + }, + { + "epoch": 0.4568110145373433, + "learning_rate": 0.0017028994079620284, + "loss": 1.0229, + "step": 5939 + }, + { + "epoch": 0.4568879316975617, + "learning_rate": 0.001702540269763653, + "loss": 0.8337, + "step": 5940 + }, + { + "epoch": 0.45696484885778016, + "learning_rate": 0.0017021811197387403, + "loss": 1.044, + "step": 5941 + }, + { + "epoch": 0.4570417660179986, + "learning_rate": 0.0017018219579082608, + "loss": 1.3212, + "step": 5942 + }, + { + "epoch": 0.4571186831782171, + "learning_rate": 0.0017014627842931872, + "loss": 1.3879, + "step": 5943 + }, + { + "epoch": 0.4571956003384355, + "learning_rate": 0.001701103598914492, + "loss": 0.8131, + "step": 5944 + }, + { + "epoch": 0.45727251749865394, + "learning_rate": 0.001700744401793148, + "loss": 1.0003, + "step": 5945 + }, + { + "epoch": 0.4573494346588724, + "learning_rate": 0.0017003851929501283, + "loss": 0.9824, + "step": 5946 + }, + { + "epoch": 0.45742635181909086, + "learning_rate": 0.0017000259724064093, + "loss": 1.1699, + "step": 5947 + }, + { + "epoch": 0.45750326897930926, + "learning_rate": 0.001699666740182965, + "loss": 1.2166, + "step": 5948 + }, + { + "epoch": 0.4575801861395277, + "learning_rate": 0.0016993074963007715, + "loss": 1.1913, + "step": 5949 + }, + { + "epoch": 0.4576571032997462, + "learning_rate": 0.0016989482407808055, + "loss": 1.3081, + "step": 5950 + }, + { + "epoch": 0.45773402045996464, + "learning_rate": 0.001698588973644045, + "loss": 1.4225, + "step": 5951 + }, + { + "epoch": 0.45781093762018304, + "learning_rate": 0.0016982296949114666, + "loss": 0.9996, + "step": 5952 + }, + { + "epoch": 0.4578878547804015, + "learning_rate": 0.00169787040460405, + "loss": 1.4416, + "step": 5953 + }, + { + "epoch": 0.45796477194061996, + "learning_rate": 0.0016975111027427744, + "loss": 1.1461, + "step": 5954 + }, + { + "epoch": 0.4580416891008384, + "learning_rate": 0.0016971517893486193, + "loss": 1.0895, + "step": 5955 + }, + { + "epoch": 0.4581186062610568, + "learning_rate": 0.0016967924644425666, + "loss": 1.5107, + "step": 5956 + }, + { + "epoch": 0.4581955234212753, + "learning_rate": 0.001696433128045596, + "loss": 1.0817, + "step": 5957 + }, + { + "epoch": 0.45827244058149375, + "learning_rate": 0.0016960737801786907, + "loss": 1.1311, + "step": 5958 + }, + { + "epoch": 0.45834935774171215, + "learning_rate": 0.001695714420862833, + "loss": 1.1138, + "step": 5959 + }, + { + "epoch": 0.4584262749019306, + "learning_rate": 0.0016953550501190066, + "loss": 1.1194, + "step": 5960 + }, + { + "epoch": 0.45850319206214907, + "learning_rate": 0.0016949956679681944, + "loss": 1.5385, + "step": 5961 + }, + { + "epoch": 0.4585801092223675, + "learning_rate": 0.0016946362744313821, + "loss": 1.1169, + "step": 5962 + }, + { + "epoch": 0.45865702638258593, + "learning_rate": 0.0016942768695295555, + "loss": 0.9635, + "step": 5963 + }, + { + "epoch": 0.4587339435428044, + "learning_rate": 0.0016939174532836997, + "loss": 0.9801, + "step": 5964 + }, + { + "epoch": 0.45881086070302285, + "learning_rate": 0.0016935580257148013, + "loss": 1.3425, + "step": 5965 + }, + { + "epoch": 0.4588877778632413, + "learning_rate": 0.0016931985868438486, + "loss": 1.2531, + "step": 5966 + }, + { + "epoch": 0.4589646950234597, + "learning_rate": 0.001692839136691829, + "loss": 0.8637, + "step": 5967 + }, + { + "epoch": 0.45904161218367817, + "learning_rate": 0.001692479675279731, + "loss": 1.1803, + "step": 5968 + }, + { + "epoch": 0.45911852934389663, + "learning_rate": 0.0016921202026285447, + "loss": 1.106, + "step": 5969 + }, + { + "epoch": 0.4591954465041151, + "learning_rate": 0.0016917607187592588, + "loss": 1.7642, + "step": 5970 + }, + { + "epoch": 0.4592723636643335, + "learning_rate": 0.001691401223692865, + "loss": 0.9494, + "step": 5971 + }, + { + "epoch": 0.45934928082455195, + "learning_rate": 0.0016910417174503544, + "loss": 1.2692, + "step": 5972 + }, + { + "epoch": 0.4594261979847704, + "learning_rate": 0.0016906822000527185, + "loss": 1.1, + "step": 5973 + }, + { + "epoch": 0.45950311514498887, + "learning_rate": 0.0016903226715209506, + "loss": 1.1767, + "step": 5974 + }, + { + "epoch": 0.4595800323052073, + "learning_rate": 0.0016899631318760436, + "loss": 1.1886, + "step": 5975 + }, + { + "epoch": 0.45965694946542573, + "learning_rate": 0.0016896035811389911, + "loss": 1.0692, + "step": 5976 + }, + { + "epoch": 0.4597338666256442, + "learning_rate": 0.001689244019330788, + "loss": 1.1346, + "step": 5977 + }, + { + "epoch": 0.45981078378586265, + "learning_rate": 0.0016888844464724294, + "loss": 1.1183, + "step": 5978 + }, + { + "epoch": 0.45988770094608106, + "learning_rate": 0.0016885248625849111, + "loss": 1.3074, + "step": 5979 + }, + { + "epoch": 0.4599646181062995, + "learning_rate": 0.0016881652676892294, + "loss": 1.518, + "step": 5980 + }, + { + "epoch": 0.460041535266518, + "learning_rate": 0.0016878056618063817, + "loss": 1.3334, + "step": 5981 + }, + { + "epoch": 0.4601184524267364, + "learning_rate": 0.001687446044957366, + "loss": 1.162, + "step": 5982 + }, + { + "epoch": 0.46019536958695484, + "learning_rate": 0.00168708641716318, + "loss": 1.3031, + "step": 5983 + }, + { + "epoch": 0.4602722867471733, + "learning_rate": 0.0016867267784448233, + "loss": 1.6319, + "step": 5984 + }, + { + "epoch": 0.46034920390739176, + "learning_rate": 0.0016863671288232956, + "loss": 1.1516, + "step": 5985 + }, + { + "epoch": 0.46042612106761016, + "learning_rate": 0.0016860074683195966, + "loss": 1.4387, + "step": 5986 + }, + { + "epoch": 0.4605030382278286, + "learning_rate": 0.001685647796954728, + "loss": 0.7825, + "step": 5987 + }, + { + "epoch": 0.4605799553880471, + "learning_rate": 0.0016852881147496907, + "loss": 1.2067, + "step": 5988 + }, + { + "epoch": 0.46065687254826554, + "learning_rate": 0.0016849284217254877, + "loss": 1.3889, + "step": 5989 + }, + { + "epoch": 0.46073378970848394, + "learning_rate": 0.0016845687179031215, + "loss": 1.0775, + "step": 5990 + }, + { + "epoch": 0.4608107068687024, + "learning_rate": 0.0016842090033035951, + "loss": 1.0139, + "step": 5991 + }, + { + "epoch": 0.46088762402892086, + "learning_rate": 0.001683849277947913, + "loss": 1.4182, + "step": 5992 + }, + { + "epoch": 0.4609645411891393, + "learning_rate": 0.00168348954185708, + "loss": 1.1822, + "step": 5993 + }, + { + "epoch": 0.4610414583493577, + "learning_rate": 0.0016831297950521016, + "loss": 1.0535, + "step": 5994 + }, + { + "epoch": 0.4611183755095762, + "learning_rate": 0.0016827700375539837, + "loss": 1.0001, + "step": 5995 + }, + { + "epoch": 0.46119529266979464, + "learning_rate": 0.0016824102693837326, + "loss": 1.5293, + "step": 5996 + }, + { + "epoch": 0.4612722098300131, + "learning_rate": 0.0016820504905623557, + "loss": 1.057, + "step": 5997 + }, + { + "epoch": 0.4613491269902315, + "learning_rate": 0.0016816907011108612, + "loss": 0.9821, + "step": 5998 + }, + { + "epoch": 0.46142604415044997, + "learning_rate": 0.0016813309010502574, + "loss": 1.3119, + "step": 5999 + }, + { + "epoch": 0.4615029613106684, + "learning_rate": 0.0016809710904015529, + "loss": 1.0554, + "step": 6000 + }, + { + "epoch": 0.46157987847088683, + "learning_rate": 0.0016806112691857584, + "loss": 1.0643, + "step": 6001 + }, + { + "epoch": 0.4616567956311053, + "learning_rate": 0.0016802514374238831, + "loss": 0.9915, + "step": 6002 + }, + { + "epoch": 0.46173371279132375, + "learning_rate": 0.0016798915951369386, + "loss": 1.2561, + "step": 6003 + }, + { + "epoch": 0.4618106299515422, + "learning_rate": 0.0016795317423459366, + "loss": 1.2189, + "step": 6004 + }, + { + "epoch": 0.4618875471117606, + "learning_rate": 0.0016791718790718892, + "loss": 1.2343, + "step": 6005 + }, + { + "epoch": 0.46196446427197907, + "learning_rate": 0.0016788120053358084, + "loss": 1.0122, + "step": 6006 + }, + { + "epoch": 0.46204138143219753, + "learning_rate": 0.0016784521211587087, + "loss": 1.3605, + "step": 6007 + }, + { + "epoch": 0.462118298592416, + "learning_rate": 0.0016780922265616034, + "loss": 1.2304, + "step": 6008 + }, + { + "epoch": 0.4621952157526344, + "learning_rate": 0.001677732321565508, + "loss": 0.8988, + "step": 6009 + }, + { + "epoch": 0.46227213291285285, + "learning_rate": 0.0016773724061914367, + "loss": 1.105, + "step": 6010 + }, + { + "epoch": 0.4623490500730713, + "learning_rate": 0.0016770124804604057, + "loss": 1.3922, + "step": 6011 + }, + { + "epoch": 0.46242596723328977, + "learning_rate": 0.0016766525443934314, + "loss": 1.0368, + "step": 6012 + }, + { + "epoch": 0.4625028843935082, + "learning_rate": 0.0016762925980115314, + "loss": 0.9141, + "step": 6013 + }, + { + "epoch": 0.46257980155372663, + "learning_rate": 0.0016759326413357224, + "loss": 1.1835, + "step": 6014 + }, + { + "epoch": 0.4626567187139451, + "learning_rate": 0.0016755726743870234, + "loss": 1.0215, + "step": 6015 + }, + { + "epoch": 0.46273363587416355, + "learning_rate": 0.0016752126971864526, + "loss": 1.1779, + "step": 6016 + }, + { + "epoch": 0.46281055303438196, + "learning_rate": 0.00167485270975503, + "loss": 1.2531, + "step": 6017 + }, + { + "epoch": 0.4628874701946004, + "learning_rate": 0.0016744927121137756, + "loss": 1.2628, + "step": 6018 + }, + { + "epoch": 0.4629643873548189, + "learning_rate": 0.0016741327042837097, + "loss": 1.3413, + "step": 6019 + }, + { + "epoch": 0.4630413045150373, + "learning_rate": 0.0016737726862858543, + "loss": 1.256, + "step": 6020 + }, + { + "epoch": 0.46311822167525574, + "learning_rate": 0.00167341265814123, + "loss": 1.0495, + "step": 6021 + }, + { + "epoch": 0.4631951388354742, + "learning_rate": 0.0016730526198708601, + "loss": 1.2592, + "step": 6022 + }, + { + "epoch": 0.46327205599569266, + "learning_rate": 0.0016726925714957669, + "loss": 0.8827, + "step": 6023 + }, + { + "epoch": 0.46334897315591106, + "learning_rate": 0.0016723325130369754, + "loss": 0.9721, + "step": 6024 + }, + { + "epoch": 0.4634258903161295, + "learning_rate": 0.0016719724445155083, + "loss": 1.4127, + "step": 6025 + }, + { + "epoch": 0.463502807476348, + "learning_rate": 0.0016716123659523908, + "loss": 0.8854, + "step": 6026 + }, + { + "epoch": 0.46357972463656644, + "learning_rate": 0.0016712522773686485, + "loss": 1.5001, + "step": 6027 + }, + { + "epoch": 0.46365664179678484, + "learning_rate": 0.0016708921787853073, + "loss": 0.7913, + "step": 6028 + }, + { + "epoch": 0.4637335589570033, + "learning_rate": 0.001670532070223394, + "loss": 1.0606, + "step": 6029 + }, + { + "epoch": 0.46381047611722176, + "learning_rate": 0.0016701719517039349, + "loss": 1.1232, + "step": 6030 + }, + { + "epoch": 0.4638873932774402, + "learning_rate": 0.0016698118232479583, + "loss": 1.4346, + "step": 6031 + }, + { + "epoch": 0.4639643104376586, + "learning_rate": 0.0016694516848764927, + "loss": 1.0104, + "step": 6032 + }, + { + "epoch": 0.4640412275978771, + "learning_rate": 0.0016690915366105665, + "loss": 1.0202, + "step": 6033 + }, + { + "epoch": 0.46411814475809554, + "learning_rate": 0.001668731378471209, + "loss": 0.9859, + "step": 6034 + }, + { + "epoch": 0.464195061918314, + "learning_rate": 0.0016683712104794506, + "loss": 1.0598, + "step": 6035 + }, + { + "epoch": 0.4642719790785324, + "learning_rate": 0.0016680110326563217, + "loss": 1.3897, + "step": 6036 + }, + { + "epoch": 0.46434889623875086, + "learning_rate": 0.001667650845022854, + "loss": 1.1436, + "step": 6037 + }, + { + "epoch": 0.4644258133989693, + "learning_rate": 0.0016672906476000778, + "loss": 1.3955, + "step": 6038 + }, + { + "epoch": 0.4645027305591878, + "learning_rate": 0.001666930440409027, + "loss": 1.0686, + "step": 6039 + }, + { + "epoch": 0.4645796477194062, + "learning_rate": 0.0016665702234707337, + "loss": 0.8183, + "step": 6040 + }, + { + "epoch": 0.46465656487962465, + "learning_rate": 0.0016662099968062312, + "loss": 1.3284, + "step": 6041 + }, + { + "epoch": 0.4647334820398431, + "learning_rate": 0.001665849760436554, + "loss": 1.5238, + "step": 6042 + }, + { + "epoch": 0.4648103992000615, + "learning_rate": 0.0016654895143827367, + "loss": 1.0193, + "step": 6043 + }, + { + "epoch": 0.46488731636027997, + "learning_rate": 0.001665129258665814, + "loss": 1.0677, + "step": 6044 + }, + { + "epoch": 0.4649642335204984, + "learning_rate": 0.0016647689933068216, + "loss": 1.2368, + "step": 6045 + }, + { + "epoch": 0.4650411506807169, + "learning_rate": 0.001664408718326796, + "loss": 0.9672, + "step": 6046 + }, + { + "epoch": 0.4651180678409353, + "learning_rate": 0.0016640484337467742, + "loss": 1.1939, + "step": 6047 + }, + { + "epoch": 0.46519498500115375, + "learning_rate": 0.001663688139587793, + "loss": 0.9955, + "step": 6048 + }, + { + "epoch": 0.4652719021613722, + "learning_rate": 0.001663327835870891, + "loss": 0.8805, + "step": 6049 + }, + { + "epoch": 0.46534881932159067, + "learning_rate": 0.001662967522617106, + "loss": 1.0091, + "step": 6050 + }, + { + "epoch": 0.46542573648180907, + "learning_rate": 0.0016626071998474786, + "loss": 0.996, + "step": 6051 + }, + { + "epoch": 0.46550265364202753, + "learning_rate": 0.0016622468675830462, + "loss": 1.1809, + "step": 6052 + }, + { + "epoch": 0.465579570802246, + "learning_rate": 0.001661886525844851, + "loss": 1.4364, + "step": 6053 + }, + { + "epoch": 0.46565648796246445, + "learning_rate": 0.0016615261746539321, + "loss": 1.4154, + "step": 6054 + }, + { + "epoch": 0.46573340512268285, + "learning_rate": 0.0016611658140313324, + "loss": 1.246, + "step": 6055 + }, + { + "epoch": 0.4658103222829013, + "learning_rate": 0.001660805443998092, + "loss": 0.8958, + "step": 6056 + }, + { + "epoch": 0.46588723944311977, + "learning_rate": 0.0016604450645752544, + "loss": 1.1856, + "step": 6057 + }, + { + "epoch": 0.46596415660333823, + "learning_rate": 0.001660084675783862, + "loss": 1.2212, + "step": 6058 + }, + { + "epoch": 0.46604107376355663, + "learning_rate": 0.0016597242776449591, + "loss": 1.1431, + "step": 6059 + }, + { + "epoch": 0.4661179909237751, + "learning_rate": 0.001659363870179589, + "loss": 1.0644, + "step": 6060 + }, + { + "epoch": 0.46619490808399355, + "learning_rate": 0.001659003453408796, + "loss": 1.1554, + "step": 6061 + }, + { + "epoch": 0.46627182524421196, + "learning_rate": 0.0016586430273536265, + "loss": 0.9181, + "step": 6062 + }, + { + "epoch": 0.4663487424044304, + "learning_rate": 0.0016582825920351246, + "loss": 0.976, + "step": 6063 + }, + { + "epoch": 0.4664256595646489, + "learning_rate": 0.0016579221474743375, + "loss": 1.2245, + "step": 6064 + }, + { + "epoch": 0.46650257672486733, + "learning_rate": 0.0016575616936923114, + "loss": 1.2207, + "step": 6065 + }, + { + "epoch": 0.46657949388508574, + "learning_rate": 0.0016572012307100941, + "loss": 1.3792, + "step": 6066 + }, + { + "epoch": 0.4666564110453042, + "learning_rate": 0.0016568407585487329, + "loss": 1.3489, + "step": 6067 + }, + { + "epoch": 0.46673332820552266, + "learning_rate": 0.0016564802772292765, + "loss": 1.539, + "step": 6068 + }, + { + "epoch": 0.4668102453657411, + "learning_rate": 0.0016561197867727734, + "loss": 1.4299, + "step": 6069 + }, + { + "epoch": 0.4668871625259595, + "learning_rate": 0.0016557592872002732, + "loss": 0.9731, + "step": 6070 + }, + { + "epoch": 0.466964079686178, + "learning_rate": 0.001655398778532826, + "loss": 1.2043, + "step": 6071 + }, + { + "epoch": 0.46704099684639644, + "learning_rate": 0.0016550382607914823, + "loss": 0.9383, + "step": 6072 + }, + { + "epoch": 0.4671179140066149, + "learning_rate": 0.001654677733997293, + "loss": 1.1071, + "step": 6073 + }, + { + "epoch": 0.4671948311668333, + "learning_rate": 0.0016543171981713094, + "loss": 0.7593, + "step": 6074 + }, + { + "epoch": 0.46727174832705176, + "learning_rate": 0.001653956653334584, + "loss": 1.056, + "step": 6075 + }, + { + "epoch": 0.4673486654872702, + "learning_rate": 0.0016535960995081689, + "loss": 0.9245, + "step": 6076 + }, + { + "epoch": 0.4674255826474887, + "learning_rate": 0.0016532355367131176, + "loss": 1.3049, + "step": 6077 + }, + { + "epoch": 0.4675024998077071, + "learning_rate": 0.001652874964970484, + "loss": 1.1381, + "step": 6078 + }, + { + "epoch": 0.46757941696792554, + "learning_rate": 0.0016525143843013216, + "loss": 0.9822, + "step": 6079 + }, + { + "epoch": 0.467656334128144, + "learning_rate": 0.0016521537947266852, + "loss": 1.0136, + "step": 6080 + }, + { + "epoch": 0.4677332512883624, + "learning_rate": 0.0016517931962676301, + "loss": 1.2798, + "step": 6081 + }, + { + "epoch": 0.46781016844858087, + "learning_rate": 0.0016514325889452125, + "loss": 1.2401, + "step": 6082 + }, + { + "epoch": 0.4678870856087993, + "learning_rate": 0.0016510719727804875, + "loss": 1.2683, + "step": 6083 + }, + { + "epoch": 0.4679640027690178, + "learning_rate": 0.0016507113477945133, + "loss": 1.091, + "step": 6084 + }, + { + "epoch": 0.4680409199292362, + "learning_rate": 0.001650350714008346, + "loss": 0.9794, + "step": 6085 + }, + { + "epoch": 0.46811783708945465, + "learning_rate": 0.001649990071443044, + "loss": 1.0378, + "step": 6086 + }, + { + "epoch": 0.4681947542496731, + "learning_rate": 0.001649629420119665, + "loss": 1.0775, + "step": 6087 + }, + { + "epoch": 0.46827167140989157, + "learning_rate": 0.0016492687600592685, + "loss": 1.3136, + "step": 6088 + }, + { + "epoch": 0.46834858857010997, + "learning_rate": 0.0016489080912829139, + "loss": 1.2109, + "step": 6089 + }, + { + "epoch": 0.46842550573032843, + "learning_rate": 0.0016485474138116605, + "loss": 1.06, + "step": 6090 + }, + { + "epoch": 0.4685024228905469, + "learning_rate": 0.0016481867276665683, + "loss": 1.2542, + "step": 6091 + }, + { + "epoch": 0.46857934005076535, + "learning_rate": 0.001647826032868699, + "loss": 1.1768, + "step": 6092 + }, + { + "epoch": 0.46865625721098375, + "learning_rate": 0.0016474653294391139, + "loss": 1.1091, + "step": 6093 + }, + { + "epoch": 0.4687331743712022, + "learning_rate": 0.0016471046173988737, + "loss": 1.4344, + "step": 6094 + }, + { + "epoch": 0.46881009153142067, + "learning_rate": 0.0016467438967690423, + "loss": 0.8547, + "step": 6095 + }, + { + "epoch": 0.46888700869163913, + "learning_rate": 0.0016463831675706815, + "loss": 1.3919, + "step": 6096 + }, + { + "epoch": 0.46896392585185753, + "learning_rate": 0.0016460224298248554, + "loss": 1.4495, + "step": 6097 + }, + { + "epoch": 0.469040843012076, + "learning_rate": 0.001645661683552627, + "loss": 1.2432, + "step": 6098 + }, + { + "epoch": 0.46911776017229445, + "learning_rate": 0.0016453009287750611, + "loss": 0.7972, + "step": 6099 + }, + { + "epoch": 0.4691946773325129, + "learning_rate": 0.0016449401655132224, + "loss": 1.3865, + "step": 6100 + }, + { + "epoch": 0.4692715944927313, + "learning_rate": 0.0016445793937881772, + "loss": 1.1304, + "step": 6101 + }, + { + "epoch": 0.4693485116529498, + "learning_rate": 0.0016442186136209898, + "loss": 0.9214, + "step": 6102 + }, + { + "epoch": 0.46942542881316823, + "learning_rate": 0.0016438578250327276, + "loss": 1.3439, + "step": 6103 + }, + { + "epoch": 0.46950234597338664, + "learning_rate": 0.0016434970280444568, + "loss": 1.0002, + "step": 6104 + }, + { + "epoch": 0.4695792631336051, + "learning_rate": 0.001643136222677245, + "loss": 1.0149, + "step": 6105 + }, + { + "epoch": 0.46965618029382356, + "learning_rate": 0.0016427754089521603, + "loss": 1.0766, + "step": 6106 + }, + { + "epoch": 0.469733097454042, + "learning_rate": 0.00164241458689027, + "loss": 0.9845, + "step": 6107 + }, + { + "epoch": 0.4698100146142604, + "learning_rate": 0.0016420537565126443, + "loss": 1.3217, + "step": 6108 + }, + { + "epoch": 0.4698869317744789, + "learning_rate": 0.0016416929178403514, + "loss": 1.1593, + "step": 6109 + }, + { + "epoch": 0.46996384893469734, + "learning_rate": 0.0016413320708944612, + "loss": 1.4989, + "step": 6110 + }, + { + "epoch": 0.4700407660949158, + "learning_rate": 0.001640971215696044, + "loss": 0.9828, + "step": 6111 + }, + { + "epoch": 0.4701176832551342, + "learning_rate": 0.0016406103522661711, + "loss": 1.1914, + "step": 6112 + }, + { + "epoch": 0.47019460041535266, + "learning_rate": 0.001640249480625913, + "loss": 0.7579, + "step": 6113 + }, + { + "epoch": 0.4702715175755711, + "learning_rate": 0.0016398886007963416, + "loss": 0.8694, + "step": 6114 + }, + { + "epoch": 0.4703484347357896, + "learning_rate": 0.0016395277127985285, + "loss": 0.8679, + "step": 6115 + }, + { + "epoch": 0.470425351896008, + "learning_rate": 0.0016391668166535474, + "loss": 0.9447, + "step": 6116 + }, + { + "epoch": 0.47050226905622644, + "learning_rate": 0.001638805912382471, + "loss": 1.3569, + "step": 6117 + }, + { + "epoch": 0.4705791862164449, + "learning_rate": 0.001638445000006372, + "loss": 1.097, + "step": 6118 + }, + { + "epoch": 0.47065610337666336, + "learning_rate": 0.0016380840795463258, + "loss": 1.1325, + "step": 6119 + }, + { + "epoch": 0.47073302053688176, + "learning_rate": 0.0016377231510234062, + "loss": 1.335, + "step": 6120 + }, + { + "epoch": 0.4708099376971002, + "learning_rate": 0.001637362214458688, + "loss": 1.3611, + "step": 6121 + }, + { + "epoch": 0.4708868548573187, + "learning_rate": 0.001637001269873247, + "loss": 1.2211, + "step": 6122 + }, + { + "epoch": 0.4709637720175371, + "learning_rate": 0.0016366403172881593, + "loss": 1.0244, + "step": 6123 + }, + { + "epoch": 0.47104068917775554, + "learning_rate": 0.001636279356724501, + "loss": 1.1608, + "step": 6124 + }, + { + "epoch": 0.471117606337974, + "learning_rate": 0.0016359183882033487, + "loss": 1.3197, + "step": 6125 + }, + { + "epoch": 0.47119452349819246, + "learning_rate": 0.0016355574117457802, + "loss": 1.0408, + "step": 6126 + }, + { + "epoch": 0.47127144065841087, + "learning_rate": 0.0016351964273728736, + "loss": 0.9261, + "step": 6127 + }, + { + "epoch": 0.4713483578186293, + "learning_rate": 0.0016348354351057065, + "loss": 1.2108, + "step": 6128 + }, + { + "epoch": 0.4714252749788478, + "learning_rate": 0.0016344744349653573, + "loss": 1.2655, + "step": 6129 + }, + { + "epoch": 0.47150219213906625, + "learning_rate": 0.001634113426972906, + "loss": 0.8518, + "step": 6130 + }, + { + "epoch": 0.47157910929928465, + "learning_rate": 0.0016337524111494325, + "loss": 1.2985, + "step": 6131 + }, + { + "epoch": 0.4716560264595031, + "learning_rate": 0.0016333913875160157, + "loss": 1.1613, + "step": 6132 + }, + { + "epoch": 0.47173294361972157, + "learning_rate": 0.0016330303560937367, + "loss": 0.7984, + "step": 6133 + }, + { + "epoch": 0.47180986077994, + "learning_rate": 0.0016326693169036771, + "loss": 1.2623, + "step": 6134 + }, + { + "epoch": 0.47188677794015843, + "learning_rate": 0.0016323082699669174, + "loss": 0.7766, + "step": 6135 + }, + { + "epoch": 0.4719636951003769, + "learning_rate": 0.00163194721530454, + "loss": 1.1435, + "step": 6136 + }, + { + "epoch": 0.47204061226059535, + "learning_rate": 0.0016315861529376272, + "loss": 1.1985, + "step": 6137 + }, + { + "epoch": 0.4721175294208138, + "learning_rate": 0.0016312250828872615, + "loss": 1.299, + "step": 6138 + }, + { + "epoch": 0.4721944465810322, + "learning_rate": 0.0016308640051745274, + "loss": 1.5076, + "step": 6139 + }, + { + "epoch": 0.47227136374125067, + "learning_rate": 0.0016305029198205067, + "loss": 1.1281, + "step": 6140 + }, + { + "epoch": 0.47234828090146913, + "learning_rate": 0.001630141826846285, + "loss": 1.2295, + "step": 6141 + }, + { + "epoch": 0.4724251980616876, + "learning_rate": 0.0016297807262729458, + "loss": 1.2271, + "step": 6142 + }, + { + "epoch": 0.472502115221906, + "learning_rate": 0.0016294196181215756, + "loss": 1.1294, + "step": 6143 + }, + { + "epoch": 0.47257903238212445, + "learning_rate": 0.0016290585024132582, + "loss": 1.0136, + "step": 6144 + }, + { + "epoch": 0.4726559495423429, + "learning_rate": 0.0016286973791690808, + "loss": 1.2991, + "step": 6145 + }, + { + "epoch": 0.4727328667025613, + "learning_rate": 0.0016283362484101295, + "loss": 1.1902, + "step": 6146 + }, + { + "epoch": 0.4728097838627798, + "learning_rate": 0.0016279751101574902, + "loss": 1.0617, + "step": 6147 + }, + { + "epoch": 0.47288670102299823, + "learning_rate": 0.0016276139644322515, + "loss": 1.0008, + "step": 6148 + }, + { + "epoch": 0.4729636181832167, + "learning_rate": 0.0016272528112555, + "loss": 0.99, + "step": 6149 + }, + { + "epoch": 0.4730405353434351, + "learning_rate": 0.0016268916506483246, + "loss": 1.2492, + "step": 6150 + }, + { + "epoch": 0.47311745250365356, + "learning_rate": 0.0016265304826318134, + "loss": 1.0925, + "step": 6151 + }, + { + "epoch": 0.473194369663872, + "learning_rate": 0.0016261693072270552, + "loss": 1.3133, + "step": 6152 + }, + { + "epoch": 0.4732712868240905, + "learning_rate": 0.0016258081244551398, + "loss": 1.3241, + "step": 6153 + }, + { + "epoch": 0.4733482039843089, + "learning_rate": 0.0016254469343371573, + "loss": 1.0899, + "step": 6154 + }, + { + "epoch": 0.47342512114452734, + "learning_rate": 0.0016250857368941968, + "loss": 1.1011, + "step": 6155 + }, + { + "epoch": 0.4735020383047458, + "learning_rate": 0.0016247245321473504, + "loss": 1.2659, + "step": 6156 + }, + { + "epoch": 0.47357895546496426, + "learning_rate": 0.0016243633201177081, + "loss": 1.2514, + "step": 6157 + }, + { + "epoch": 0.47365587262518266, + "learning_rate": 0.0016240021008263623, + "loss": 1.1454, + "step": 6158 + }, + { + "epoch": 0.4737327897854011, + "learning_rate": 0.0016236408742944046, + "loss": 0.8226, + "step": 6159 + }, + { + "epoch": 0.4738097069456196, + "learning_rate": 0.001623279640542927, + "loss": 0.9157, + "step": 6160 + }, + { + "epoch": 0.47388662410583804, + "learning_rate": 0.001622918399593023, + "loss": 0.8018, + "step": 6161 + }, + { + "epoch": 0.47396354126605644, + "learning_rate": 0.0016225571514657854, + "loss": 0.9245, + "step": 6162 + }, + { + "epoch": 0.4740404584262749, + "learning_rate": 0.0016221958961823082, + "loss": 0.8069, + "step": 6163 + }, + { + "epoch": 0.47411737558649336, + "learning_rate": 0.001621834633763685, + "loss": 1.6641, + "step": 6164 + }, + { + "epoch": 0.47419429274671177, + "learning_rate": 0.0016214733642310112, + "loss": 1.115, + "step": 6165 + }, + { + "epoch": 0.4742712099069302, + "learning_rate": 0.0016211120876053803, + "loss": 0.8321, + "step": 6166 + }, + { + "epoch": 0.4743481270671487, + "learning_rate": 0.001620750803907889, + "loss": 1.2554, + "step": 6167 + }, + { + "epoch": 0.47442504422736714, + "learning_rate": 0.0016203895131596316, + "loss": 1.6233, + "step": 6168 + }, + { + "epoch": 0.47450196138758555, + "learning_rate": 0.0016200282153817059, + "loss": 1.4526, + "step": 6169 + }, + { + "epoch": 0.474578878547804, + "learning_rate": 0.0016196669105952074, + "loss": 1.1712, + "step": 6170 + }, + { + "epoch": 0.47465579570802247, + "learning_rate": 0.0016193055988212328, + "loss": 0.8857, + "step": 6171 + }, + { + "epoch": 0.4747327128682409, + "learning_rate": 0.0016189442800808806, + "loss": 1.0663, + "step": 6172 + }, + { + "epoch": 0.47480963002845933, + "learning_rate": 0.0016185829543952476, + "loss": 1.2616, + "step": 6173 + }, + { + "epoch": 0.4748865471886778, + "learning_rate": 0.0016182216217854324, + "loss": 0.761, + "step": 6174 + }, + { + "epoch": 0.47496346434889625, + "learning_rate": 0.001617860282272533, + "loss": 1.0008, + "step": 6175 + }, + { + "epoch": 0.4750403815091147, + "learning_rate": 0.0016174989358776494, + "loss": 1.1151, + "step": 6176 + }, + { + "epoch": 0.4751172986693331, + "learning_rate": 0.0016171375826218807, + "loss": 1.0568, + "step": 6177 + }, + { + "epoch": 0.47519421582955157, + "learning_rate": 0.0016167762225263262, + "loss": 1.0713, + "step": 6178 + }, + { + "epoch": 0.47527113298977003, + "learning_rate": 0.0016164148556120864, + "loss": 1.195, + "step": 6179 + }, + { + "epoch": 0.4753480501499885, + "learning_rate": 0.0016160534819002618, + "loss": 1.4284, + "step": 6180 + }, + { + "epoch": 0.4754249673102069, + "learning_rate": 0.0016156921014119532, + "loss": 1.3352, + "step": 6181 + }, + { + "epoch": 0.47550188447042535, + "learning_rate": 0.0016153307141682624, + "loss": 1.26, + "step": 6182 + }, + { + "epoch": 0.4755788016306438, + "learning_rate": 0.0016149693201902913, + "loss": 1.2073, + "step": 6183 + }, + { + "epoch": 0.4756557187908622, + "learning_rate": 0.0016146079194991416, + "loss": 1.2384, + "step": 6184 + }, + { + "epoch": 0.4757326359510807, + "learning_rate": 0.0016142465121159158, + "loss": 1.0126, + "step": 6185 + }, + { + "epoch": 0.47580955311129913, + "learning_rate": 0.0016138850980617172, + "loss": 1.0965, + "step": 6186 + }, + { + "epoch": 0.4758864702715176, + "learning_rate": 0.0016135236773576493, + "loss": 1.1451, + "step": 6187 + }, + { + "epoch": 0.475963387431736, + "learning_rate": 0.0016131622500248153, + "loss": 1.4201, + "step": 6188 + }, + { + "epoch": 0.47604030459195446, + "learning_rate": 0.00161280081608432, + "loss": 1.0825, + "step": 6189 + }, + { + "epoch": 0.4761172217521729, + "learning_rate": 0.0016124393755572667, + "loss": 1.3944, + "step": 6190 + }, + { + "epoch": 0.4761941389123914, + "learning_rate": 0.0016120779284647618, + "loss": 1.0913, + "step": 6191 + }, + { + "epoch": 0.4762710560726098, + "learning_rate": 0.0016117164748279096, + "loss": 1.2932, + "step": 6192 + }, + { + "epoch": 0.47634797323282824, + "learning_rate": 0.0016113550146678157, + "loss": 1.6216, + "step": 6193 + }, + { + "epoch": 0.4764248903930467, + "learning_rate": 0.001610993548005587, + "loss": 1.0621, + "step": 6194 + }, + { + "epoch": 0.47650180755326516, + "learning_rate": 0.0016106320748623287, + "loss": 1.107, + "step": 6195 + }, + { + "epoch": 0.47657872471348356, + "learning_rate": 0.0016102705952591493, + "loss": 1.1866, + "step": 6196 + }, + { + "epoch": 0.476655641873702, + "learning_rate": 0.0016099091092171538, + "loss": 1.0801, + "step": 6197 + }, + { + "epoch": 0.4767325590339205, + "learning_rate": 0.0016095476167574516, + "loss": 1.7075, + "step": 6198 + }, + { + "epoch": 0.47680947619413894, + "learning_rate": 0.0016091861179011492, + "loss": 1.1463, + "step": 6199 + }, + { + "epoch": 0.47688639335435734, + "learning_rate": 0.0016088246126693562, + "loss": 1.0707, + "step": 6200 + }, + { + "epoch": 0.4769633105145758, + "learning_rate": 0.0016084631010831808, + "loss": 1.2336, + "step": 6201 + }, + { + "epoch": 0.47704022767479426, + "learning_rate": 0.0016081015831637313, + "loss": 1.0221, + "step": 6202 + }, + { + "epoch": 0.4771171448350127, + "learning_rate": 0.001607740058932118, + "loss": 1.0629, + "step": 6203 + }, + { + "epoch": 0.4771940619952311, + "learning_rate": 0.0016073785284094503, + "loss": 1.1565, + "step": 6204 + }, + { + "epoch": 0.4772709791554496, + "learning_rate": 0.0016070169916168385, + "loss": 0.7952, + "step": 6205 + }, + { + "epoch": 0.47734789631566804, + "learning_rate": 0.0016066554485753928, + "loss": 1.104, + "step": 6206 + }, + { + "epoch": 0.47742481347588644, + "learning_rate": 0.001606293899306225, + "loss": 0.8004, + "step": 6207 + }, + { + "epoch": 0.4775017306361049, + "learning_rate": 0.001605932343830445, + "loss": 1.6401, + "step": 6208 + }, + { + "epoch": 0.47757864779632336, + "learning_rate": 0.001605570782169165, + "loss": 1.1741, + "step": 6209 + }, + { + "epoch": 0.4776555649565418, + "learning_rate": 0.001605209214343497, + "loss": 1.1576, + "step": 6210 + }, + { + "epoch": 0.4777324821167602, + "learning_rate": 0.0016048476403745537, + "loss": 0.8853, + "step": 6211 + }, + { + "epoch": 0.4778093992769787, + "learning_rate": 0.001604486060283447, + "loss": 1.0119, + "step": 6212 + }, + { + "epoch": 0.47788631643719715, + "learning_rate": 0.0016041244740912905, + "loss": 1.1708, + "step": 6213 + }, + { + "epoch": 0.4779632335974156, + "learning_rate": 0.0016037628818191976, + "loss": 1.12, + "step": 6214 + }, + { + "epoch": 0.478040150757634, + "learning_rate": 0.0016034012834882819, + "loss": 1.8501, + "step": 6215 + }, + { + "epoch": 0.47811706791785247, + "learning_rate": 0.0016030396791196573, + "loss": 0.99, + "step": 6216 + }, + { + "epoch": 0.4781939850780709, + "learning_rate": 0.0016026780687344385, + "loss": 1.0905, + "step": 6217 + }, + { + "epoch": 0.4782709022382894, + "learning_rate": 0.0016023164523537405, + "loss": 0.7269, + "step": 6218 + }, + { + "epoch": 0.4783478193985078, + "learning_rate": 0.0016019548299986781, + "loss": 1.0102, + "step": 6219 + }, + { + "epoch": 0.47842473655872625, + "learning_rate": 0.001601593201690367, + "loss": 1.033, + "step": 6220 + }, + { + "epoch": 0.4785016537189447, + "learning_rate": 0.0016012315674499222, + "loss": 1.0176, + "step": 6221 + }, + { + "epoch": 0.47857857087916317, + "learning_rate": 0.0016008699272984615, + "loss": 1.1314, + "step": 6222 + }, + { + "epoch": 0.47865548803938157, + "learning_rate": 0.0016005082812571008, + "loss": 1.1402, + "step": 6223 + }, + { + "epoch": 0.47873240519960003, + "learning_rate": 0.001600146629346956, + "loss": 1.4318, + "step": 6224 + }, + { + "epoch": 0.4788093223598185, + "learning_rate": 0.0015997849715891457, + "loss": 0.8946, + "step": 6225 + }, + { + "epoch": 0.4788862395200369, + "learning_rate": 0.0015994233080047866, + "loss": 1.1351, + "step": 6226 + }, + { + "epoch": 0.47896315668025535, + "learning_rate": 0.0015990616386149972, + "loss": 1.3198, + "step": 6227 + }, + { + "epoch": 0.4790400738404738, + "learning_rate": 0.0015986999634408955, + "loss": 1.1457, + "step": 6228 + }, + { + "epoch": 0.47911699100069227, + "learning_rate": 0.0015983382825035997, + "loss": 1.274, + "step": 6229 + }, + { + "epoch": 0.4791939081609107, + "learning_rate": 0.0015979765958242294, + "loss": 1.2721, + "step": 6230 + }, + { + "epoch": 0.47927082532112913, + "learning_rate": 0.0015976149034239039, + "loss": 1.0324, + "step": 6231 + }, + { + "epoch": 0.4793477424813476, + "learning_rate": 0.0015972532053237418, + "loss": 1.294, + "step": 6232 + }, + { + "epoch": 0.47942465964156605, + "learning_rate": 0.001596891501544864, + "loss": 1.0118, + "step": 6233 + }, + { + "epoch": 0.47950157680178446, + "learning_rate": 0.0015965297921083903, + "loss": 0.8374, + "step": 6234 + }, + { + "epoch": 0.4795784939620029, + "learning_rate": 0.0015961680770354416, + "loss": 1.227, + "step": 6235 + }, + { + "epoch": 0.4796554111222214, + "learning_rate": 0.0015958063563471384, + "loss": 1.3112, + "step": 6236 + }, + { + "epoch": 0.47973232828243983, + "learning_rate": 0.0015954446300646026, + "loss": 1.1344, + "step": 6237 + }, + { + "epoch": 0.47980924544265824, + "learning_rate": 0.0015950828982089552, + "loss": 0.9309, + "step": 6238 + }, + { + "epoch": 0.4798861626028767, + "learning_rate": 0.0015947211608013181, + "loss": 0.9173, + "step": 6239 + }, + { + "epoch": 0.47996307976309516, + "learning_rate": 0.0015943594178628143, + "loss": 1.324, + "step": 6240 + }, + { + "epoch": 0.4800399969233136, + "learning_rate": 0.0015939976694145648, + "loss": 1.3338, + "step": 6241 + }, + { + "epoch": 0.480116914083532, + "learning_rate": 0.0015936359154776942, + "loss": 1.3014, + "step": 6242 + }, + { + "epoch": 0.4801938312437505, + "learning_rate": 0.0015932741560733246, + "loss": 0.8919, + "step": 6243 + }, + { + "epoch": 0.48027074840396894, + "learning_rate": 0.00159291239122258, + "loss": 1.0611, + "step": 6244 + }, + { + "epoch": 0.48034766556418734, + "learning_rate": 0.001592550620946584, + "loss": 0.802, + "step": 6245 + }, + { + "epoch": 0.4804245827244058, + "learning_rate": 0.0015921888452664604, + "loss": 1.552, + "step": 6246 + }, + { + "epoch": 0.48050149988462426, + "learning_rate": 0.0015918270642033346, + "loss": 1.3838, + "step": 6247 + }, + { + "epoch": 0.4805784170448427, + "learning_rate": 0.0015914652777783303, + "loss": 1.3279, + "step": 6248 + }, + { + "epoch": 0.4806553342050611, + "learning_rate": 0.001591103486012574, + "loss": 0.9805, + "step": 6249 + }, + { + "epoch": 0.4807322513652796, + "learning_rate": 0.001590741688927189, + "loss": 1.1158, + "step": 6250 + }, + { + "epoch": 0.48080916852549804, + "learning_rate": 0.0015903798865433032, + "loss": 1.1117, + "step": 6251 + }, + { + "epoch": 0.4808860856857165, + "learning_rate": 0.0015900180788820408, + "loss": 1.1248, + "step": 6252 + }, + { + "epoch": 0.4809630028459349, + "learning_rate": 0.0015896562659645298, + "loss": 0.9676, + "step": 6253 + }, + { + "epoch": 0.48103992000615337, + "learning_rate": 0.0015892944478118956, + "loss": 1.0328, + "step": 6254 + }, + { + "epoch": 0.4811168371663718, + "learning_rate": 0.0015889326244452656, + "loss": 1.314, + "step": 6255 + }, + { + "epoch": 0.4811937543265903, + "learning_rate": 0.0015885707958857667, + "loss": 1.1711, + "step": 6256 + }, + { + "epoch": 0.4812706714868087, + "learning_rate": 0.0015882089621545269, + "loss": 1.5342, + "step": 6257 + }, + { + "epoch": 0.48134758864702715, + "learning_rate": 0.001587847123272674, + "loss": 1.2094, + "step": 6258 + }, + { + "epoch": 0.4814245058072456, + "learning_rate": 0.0015874852792613357, + "loss": 0.4793, + "step": 6259 + }, + { + "epoch": 0.48150142296746407, + "learning_rate": 0.001587123430141641, + "loss": 1.3869, + "step": 6260 + }, + { + "epoch": 0.48157834012768247, + "learning_rate": 0.0015867615759347187, + "loss": 1.2889, + "step": 6261 + }, + { + "epoch": 0.48165525728790093, + "learning_rate": 0.001586399716661697, + "loss": 1.0959, + "step": 6262 + }, + { + "epoch": 0.4817321744481194, + "learning_rate": 0.001586037852343706, + "loss": 1.1597, + "step": 6263 + }, + { + "epoch": 0.48180909160833785, + "learning_rate": 0.0015856759830018754, + "loss": 0.8337, + "step": 6264 + }, + { + "epoch": 0.48188600876855625, + "learning_rate": 0.0015853141086573346, + "loss": 1.0926, + "step": 6265 + }, + { + "epoch": 0.4819629259287747, + "learning_rate": 0.0015849522293312143, + "loss": 0.8937, + "step": 6266 + }, + { + "epoch": 0.48203984308899317, + "learning_rate": 0.0015845903450446442, + "loss": 1.1546, + "step": 6267 + }, + { + "epoch": 0.4821167602492116, + "learning_rate": 0.0015842284558187565, + "loss": 0.8617, + "step": 6268 + }, + { + "epoch": 0.48219367740943003, + "learning_rate": 0.0015838665616746811, + "loss": 1.2347, + "step": 6269 + }, + { + "epoch": 0.4822705945696485, + "learning_rate": 0.0015835046626335495, + "loss": 1.2136, + "step": 6270 + }, + { + "epoch": 0.48234751172986695, + "learning_rate": 0.001583142758716494, + "loss": 0.9565, + "step": 6271 + }, + { + "epoch": 0.48242442889008536, + "learning_rate": 0.0015827808499446461, + "loss": 1.4013, + "step": 6272 + }, + { + "epoch": 0.4825013460503038, + "learning_rate": 0.0015824189363391383, + "loss": 0.9568, + "step": 6273 + }, + { + "epoch": 0.4825782632105223, + "learning_rate": 0.0015820570179211023, + "loss": 0.9756, + "step": 6274 + }, + { + "epoch": 0.48265518037074073, + "learning_rate": 0.0015816950947116722, + "loss": 1.3315, + "step": 6275 + }, + { + "epoch": 0.48273209753095914, + "learning_rate": 0.0015813331667319799, + "loss": 1.2767, + "step": 6276 + }, + { + "epoch": 0.4828090146911776, + "learning_rate": 0.0015809712340031593, + "loss": 1.3024, + "step": 6277 + }, + { + "epoch": 0.48288593185139606, + "learning_rate": 0.0015806092965463443, + "loss": 1.3587, + "step": 6278 + }, + { + "epoch": 0.4829628490116145, + "learning_rate": 0.001580247354382668, + "loss": 1.3582, + "step": 6279 + }, + { + "epoch": 0.4830397661718329, + "learning_rate": 0.0015798854075332657, + "loss": 0.8317, + "step": 6280 + }, + { + "epoch": 0.4831166833320514, + "learning_rate": 0.00157952345601927, + "loss": 1.1435, + "step": 6281 + }, + { + "epoch": 0.48319360049226984, + "learning_rate": 0.0015791614998618179, + "loss": 1.0068, + "step": 6282 + }, + { + "epoch": 0.4832705176524883, + "learning_rate": 0.001578799539082043, + "loss": 1.4294, + "step": 6283 + }, + { + "epoch": 0.4833474348127067, + "learning_rate": 0.0015784375737010812, + "loss": 0.9777, + "step": 6284 + }, + { + "epoch": 0.48342435197292516, + "learning_rate": 0.0015780756037400672, + "loss": 1.2644, + "step": 6285 + }, + { + "epoch": 0.4835012691331436, + "learning_rate": 0.0015777136292201377, + "loss": 0.974, + "step": 6286 + }, + { + "epoch": 0.483578186293362, + "learning_rate": 0.0015773516501624283, + "loss": 0.7558, + "step": 6287 + }, + { + "epoch": 0.4836551034535805, + "learning_rate": 0.001576989666588075, + "loss": 1.4457, + "step": 6288 + }, + { + "epoch": 0.48373202061379894, + "learning_rate": 0.0015766276785182153, + "loss": 1.1475, + "step": 6289 + }, + { + "epoch": 0.4838089377740174, + "learning_rate": 0.0015762656859739854, + "loss": 1.1936, + "step": 6290 + }, + { + "epoch": 0.4838858549342358, + "learning_rate": 0.0015759036889765232, + "loss": 1.1512, + "step": 6291 + }, + { + "epoch": 0.48396277209445426, + "learning_rate": 0.0015755416875469649, + "loss": 0.9819, + "step": 6292 + }, + { + "epoch": 0.4840396892546727, + "learning_rate": 0.0015751796817064488, + "loss": 1.3392, + "step": 6293 + }, + { + "epoch": 0.4841166064148912, + "learning_rate": 0.0015748176714761126, + "loss": 1.0857, + "step": 6294 + }, + { + "epoch": 0.4841935235751096, + "learning_rate": 0.0015744556568770952, + "loss": 1.0391, + "step": 6295 + }, + { + "epoch": 0.48427044073532804, + "learning_rate": 0.0015740936379305336, + "loss": 1.0908, + "step": 6296 + }, + { + "epoch": 0.4843473578955465, + "learning_rate": 0.0015737316146575678, + "loss": 1.1738, + "step": 6297 + }, + { + "epoch": 0.48442427505576496, + "learning_rate": 0.0015733695870793357, + "loss": 1.1246, + "step": 6298 + }, + { + "epoch": 0.48450119221598337, + "learning_rate": 0.0015730075552169774, + "loss": 1.1251, + "step": 6299 + }, + { + "epoch": 0.4845781093762018, + "learning_rate": 0.001572645519091632, + "loss": 1.2144, + "step": 6300 + }, + { + "epoch": 0.4846550265364203, + "learning_rate": 0.0015722834787244384, + "loss": 1.242, + "step": 6301 + }, + { + "epoch": 0.48473194369663875, + "learning_rate": 0.0015719214341365377, + "loss": 1.0311, + "step": 6302 + }, + { + "epoch": 0.48480886085685715, + "learning_rate": 0.0015715593853490694, + "loss": 1.3294, + "step": 6303 + }, + { + "epoch": 0.4848857780170756, + "learning_rate": 0.001571197332383174, + "loss": 1.4097, + "step": 6304 + }, + { + "epoch": 0.48496269517729407, + "learning_rate": 0.001570835275259992, + "loss": 1.1468, + "step": 6305 + }, + { + "epoch": 0.48503961233751247, + "learning_rate": 0.0015704732140006646, + "loss": 1.0481, + "step": 6306 + }, + { + "epoch": 0.48511652949773093, + "learning_rate": 0.0015701111486263327, + "loss": 1.2476, + "step": 6307 + }, + { + "epoch": 0.4851934466579494, + "learning_rate": 0.001569749079158138, + "loss": 0.8719, + "step": 6308 + }, + { + "epoch": 0.48527036381816785, + "learning_rate": 0.0015693870056172213, + "loss": 0.9322, + "step": 6309 + }, + { + "epoch": 0.48534728097838625, + "learning_rate": 0.0015690249280247255, + "loss": 1.0444, + "step": 6310 + }, + { + "epoch": 0.4854241981386047, + "learning_rate": 0.0015686628464017922, + "loss": 1.0921, + "step": 6311 + }, + { + "epoch": 0.48550111529882317, + "learning_rate": 0.0015683007607695636, + "loss": 0.7263, + "step": 6312 + }, + { + "epoch": 0.48557803245904163, + "learning_rate": 0.0015679386711491825, + "loss": 1.0986, + "step": 6313 + }, + { + "epoch": 0.48565494961926003, + "learning_rate": 0.0015675765775617916, + "loss": 0.9929, + "step": 6314 + }, + { + "epoch": 0.4857318667794785, + "learning_rate": 0.0015672144800285338, + "loss": 1.319, + "step": 6315 + }, + { + "epoch": 0.48580878393969695, + "learning_rate": 0.0015668523785705526, + "loss": 1.1637, + "step": 6316 + }, + { + "epoch": 0.4858857010999154, + "learning_rate": 0.001566490273208991, + "loss": 0.7977, + "step": 6317 + }, + { + "epoch": 0.4859626182601338, + "learning_rate": 0.0015661281639649937, + "loss": 1.0797, + "step": 6318 + }, + { + "epoch": 0.4860395354203523, + "learning_rate": 0.0015657660508597035, + "loss": 1.0003, + "step": 6319 + }, + { + "epoch": 0.48611645258057073, + "learning_rate": 0.0015654039339142649, + "loss": 1.157, + "step": 6320 + }, + { + "epoch": 0.4861933697407892, + "learning_rate": 0.0015650418131498227, + "loss": 1.1297, + "step": 6321 + }, + { + "epoch": 0.4862702869010076, + "learning_rate": 0.0015646796885875217, + "loss": 1.2896, + "step": 6322 + }, + { + "epoch": 0.48634720406122606, + "learning_rate": 0.0015643175602485053, + "loss": 1.3231, + "step": 6323 + }, + { + "epoch": 0.4864241212214445, + "learning_rate": 0.0015639554281539204, + "loss": 1.3241, + "step": 6324 + }, + { + "epoch": 0.486501038381663, + "learning_rate": 0.0015635932923249106, + "loss": 1.056, + "step": 6325 + }, + { + "epoch": 0.4865779555418814, + "learning_rate": 0.001563231152782623, + "loss": 0.8163, + "step": 6326 + }, + { + "epoch": 0.48665487270209984, + "learning_rate": 0.001562869009548202, + "loss": 1.0155, + "step": 6327 + }, + { + "epoch": 0.4867317898623183, + "learning_rate": 0.0015625068626427942, + "loss": 1.0447, + "step": 6328 + }, + { + "epoch": 0.4868087070225367, + "learning_rate": 0.0015621447120875455, + "loss": 1.3451, + "step": 6329 + }, + { + "epoch": 0.48688562418275516, + "learning_rate": 0.0015617825579036023, + "loss": 1.021, + "step": 6330 + }, + { + "epoch": 0.4869625413429736, + "learning_rate": 0.0015614204001121108, + "loss": 0.8143, + "step": 6331 + }, + { + "epoch": 0.4870394585031921, + "learning_rate": 0.0015610582387342185, + "loss": 1.2286, + "step": 6332 + }, + { + "epoch": 0.4871163756634105, + "learning_rate": 0.001560696073791072, + "loss": 1.0498, + "step": 6333 + }, + { + "epoch": 0.48719329282362894, + "learning_rate": 0.0015603339053038186, + "loss": 0.8564, + "step": 6334 + }, + { + "epoch": 0.4872702099838474, + "learning_rate": 0.0015599717332936051, + "loss": 1.3406, + "step": 6335 + }, + { + "epoch": 0.48734712714406586, + "learning_rate": 0.00155960955778158, + "loss": 0.9921, + "step": 6336 + }, + { + "epoch": 0.48742404430428427, + "learning_rate": 0.001559247378788891, + "loss": 1.0311, + "step": 6337 + }, + { + "epoch": 0.4875009614645027, + "learning_rate": 0.0015588851963366853, + "loss": 1.3841, + "step": 6338 + }, + { + "epoch": 0.4875778786247212, + "learning_rate": 0.0015585230104461119, + "loss": 1.2471, + "step": 6339 + }, + { + "epoch": 0.48765479578493964, + "learning_rate": 0.0015581608211383184, + "loss": 1.1149, + "step": 6340 + }, + { + "epoch": 0.48773171294515805, + "learning_rate": 0.0015577986284344544, + "loss": 1.2192, + "step": 6341 + }, + { + "epoch": 0.4878086301053765, + "learning_rate": 0.0015574364323556683, + "loss": 1.0891, + "step": 6342 + }, + { + "epoch": 0.48788554726559497, + "learning_rate": 0.0015570742329231088, + "loss": 1.2197, + "step": 6343 + }, + { + "epoch": 0.4879624644258134, + "learning_rate": 0.0015567120301579257, + "loss": 1.225, + "step": 6344 + }, + { + "epoch": 0.48803938158603183, + "learning_rate": 0.0015563498240812675, + "loss": 1.234, + "step": 6345 + }, + { + "epoch": 0.4881162987462503, + "learning_rate": 0.001555987614714285, + "loss": 0.8916, + "step": 6346 + }, + { + "epoch": 0.48819321590646875, + "learning_rate": 0.0015556254020781268, + "loss": 1.0892, + "step": 6347 + }, + { + "epoch": 0.48827013306668715, + "learning_rate": 0.001555263186193944, + "loss": 1.1146, + "step": 6348 + }, + { + "epoch": 0.4883470502269056, + "learning_rate": 0.0015549009670828855, + "loss": 1.1832, + "step": 6349 + }, + { + "epoch": 0.48842396738712407, + "learning_rate": 0.0015545387447661028, + "loss": 1.2099, + "step": 6350 + }, + { + "epoch": 0.48850088454734253, + "learning_rate": 0.0015541765192647456, + "loss": 1.2221, + "step": 6351 + }, + { + "epoch": 0.48857780170756093, + "learning_rate": 0.0015538142905999652, + "loss": 1.2262, + "step": 6352 + }, + { + "epoch": 0.4886547188677794, + "learning_rate": 0.0015534520587929122, + "loss": 1.0718, + "step": 6353 + }, + { + "epoch": 0.48873163602799785, + "learning_rate": 0.0015530898238647374, + "loss": 1.2102, + "step": 6354 + }, + { + "epoch": 0.4888085531882163, + "learning_rate": 0.0015527275858365933, + "loss": 0.7523, + "step": 6355 + }, + { + "epoch": 0.4888854703484347, + "learning_rate": 0.0015523653447296298, + "loss": 1.2573, + "step": 6356 + }, + { + "epoch": 0.4889623875086532, + "learning_rate": 0.0015520031005649996, + "loss": 1.4722, + "step": 6357 + }, + { + "epoch": 0.48903930466887163, + "learning_rate": 0.001551640853363854, + "loss": 1.1594, + "step": 6358 + }, + { + "epoch": 0.4891162218290901, + "learning_rate": 0.0015512786031473453, + "loss": 1.0412, + "step": 6359 + }, + { + "epoch": 0.4891931389893085, + "learning_rate": 0.0015509163499366257, + "loss": 1.0809, + "step": 6360 + }, + { + "epoch": 0.48927005614952696, + "learning_rate": 0.0015505540937528473, + "loss": 1.103, + "step": 6361 + }, + { + "epoch": 0.4893469733097454, + "learning_rate": 0.0015501918346171619, + "loss": 1.2935, + "step": 6362 + }, + { + "epoch": 0.4894238904699639, + "learning_rate": 0.0015498295725507238, + "loss": 1.3264, + "step": 6363 + }, + { + "epoch": 0.4895008076301823, + "learning_rate": 0.001549467307574685, + "loss": 1.2447, + "step": 6364 + }, + { + "epoch": 0.48957772479040074, + "learning_rate": 0.001549105039710198, + "loss": 1.3307, + "step": 6365 + }, + { + "epoch": 0.4896546419506192, + "learning_rate": 0.0015487427689784168, + "loss": 1.1443, + "step": 6366 + }, + { + "epoch": 0.48973155911083766, + "learning_rate": 0.0015483804954004945, + "loss": 1.2331, + "step": 6367 + }, + { + "epoch": 0.48980847627105606, + "learning_rate": 0.0015480182189975852, + "loss": 1.1772, + "step": 6368 + }, + { + "epoch": 0.4898853934312745, + "learning_rate": 0.0015476559397908413, + "loss": 1.0711, + "step": 6369 + }, + { + "epoch": 0.489962310591493, + "learning_rate": 0.0015472936578014179, + "loss": 0.933, + "step": 6370 + }, + { + "epoch": 0.4900392277517114, + "learning_rate": 0.0015469313730504684, + "loss": 1.4777, + "step": 6371 + }, + { + "epoch": 0.49011614491192984, + "learning_rate": 0.0015465690855591472, + "loss": 1.0311, + "step": 6372 + }, + { + "epoch": 0.4901930620721483, + "learning_rate": 0.0015462067953486084, + "loss": 1.142, + "step": 6373 + }, + { + "epoch": 0.49026997923236676, + "learning_rate": 0.0015458445024400067, + "loss": 0.9315, + "step": 6374 + }, + { + "epoch": 0.49034689639258516, + "learning_rate": 0.0015454822068544968, + "loss": 1.1105, + "step": 6375 + }, + { + "epoch": 0.4904238135528036, + "learning_rate": 0.001545119908613233, + "loss": 1.25, + "step": 6376 + }, + { + "epoch": 0.4905007307130221, + "learning_rate": 0.0015447576077373716, + "loss": 1.1408, + "step": 6377 + }, + { + "epoch": 0.49057764787324054, + "learning_rate": 0.0015443953042480658, + "loss": 1.1946, + "step": 6378 + }, + { + "epoch": 0.49065456503345894, + "learning_rate": 0.0015440329981664727, + "loss": 0.9651, + "step": 6379 + }, + { + "epoch": 0.4907314821936774, + "learning_rate": 0.0015436706895137467, + "loss": 1.1148, + "step": 6380 + }, + { + "epoch": 0.49080839935389586, + "learning_rate": 0.0015433083783110435, + "loss": 1.015, + "step": 6381 + }, + { + "epoch": 0.4908853165141143, + "learning_rate": 0.0015429460645795192, + "loss": 1.265, + "step": 6382 + }, + { + "epoch": 0.4909622336743327, + "learning_rate": 0.0015425837483403297, + "loss": 0.9602, + "step": 6383 + }, + { + "epoch": 0.4910391508345512, + "learning_rate": 0.0015422214296146298, + "loss": 1.3532, + "step": 6384 + }, + { + "epoch": 0.49111606799476965, + "learning_rate": 0.0015418591084235774, + "loss": 1.2866, + "step": 6385 + }, + { + "epoch": 0.4911929851549881, + "learning_rate": 0.0015414967847883275, + "loss": 1.1965, + "step": 6386 + }, + { + "epoch": 0.4912699023152065, + "learning_rate": 0.0015411344587300375, + "loss": 1.3679, + "step": 6387 + }, + { + "epoch": 0.49134681947542497, + "learning_rate": 0.0015407721302698638, + "loss": 1.3245, + "step": 6388 + }, + { + "epoch": 0.4914237366356434, + "learning_rate": 0.0015404097994289622, + "loss": 1.0462, + "step": 6389 + }, + { + "epoch": 0.49150065379586183, + "learning_rate": 0.0015400474662284913, + "loss": 1.2234, + "step": 6390 + }, + { + "epoch": 0.4915775709560803, + "learning_rate": 0.0015396851306896061, + "loss": 1.1224, + "step": 6391 + }, + { + "epoch": 0.49165448811629875, + "learning_rate": 0.0015393227928334654, + "loss": 1.2868, + "step": 6392 + }, + { + "epoch": 0.4917314052765172, + "learning_rate": 0.0015389604526812252, + "loss": 1.1485, + "step": 6393 + }, + { + "epoch": 0.4918083224367356, + "learning_rate": 0.0015385981102540444, + "loss": 1.1772, + "step": 6394 + }, + { + "epoch": 0.49188523959695407, + "learning_rate": 0.0015382357655730793, + "loss": 1.0205, + "step": 6395 + }, + { + "epoch": 0.49196215675717253, + "learning_rate": 0.001537873418659488, + "loss": 1.3706, + "step": 6396 + }, + { + "epoch": 0.492039073917391, + "learning_rate": 0.001537511069534428, + "loss": 0.9855, + "step": 6397 + }, + { + "epoch": 0.4921159910776094, + "learning_rate": 0.0015371487182190581, + "loss": 1.1971, + "step": 6398 + }, + { + "epoch": 0.49219290823782785, + "learning_rate": 0.0015367863647345356, + "loss": 1.2389, + "step": 6399 + }, + { + "epoch": 0.4922698253980463, + "learning_rate": 0.0015364240091020188, + "loss": 1.0502, + "step": 6400 + }, + { + "epoch": 0.49234674255826477, + "learning_rate": 0.0015360616513426665, + "loss": 1.397, + "step": 6401 + }, + { + "epoch": 0.4924236597184832, + "learning_rate": 0.0015356992914776367, + "loss": 1.1998, + "step": 6402 + }, + { + "epoch": 0.49250057687870163, + "learning_rate": 0.0015353369295280877, + "loss": 1.0317, + "step": 6403 + }, + { + "epoch": 0.4925774940389201, + "learning_rate": 0.0015349745655151785, + "loss": 1.1539, + "step": 6404 + }, + { + "epoch": 0.49265441119913855, + "learning_rate": 0.0015346121994600688, + "loss": 0.9895, + "step": 6405 + }, + { + "epoch": 0.49273132835935696, + "learning_rate": 0.0015342498313839163, + "loss": 1.0169, + "step": 6406 + }, + { + "epoch": 0.4928082455195754, + "learning_rate": 0.00153388746130788, + "loss": 0.7782, + "step": 6407 + }, + { + "epoch": 0.4928851626797939, + "learning_rate": 0.0015335250892531196, + "loss": 0.9776, + "step": 6408 + }, + { + "epoch": 0.4929620798400123, + "learning_rate": 0.0015331627152407946, + "loss": 0.9775, + "step": 6409 + }, + { + "epoch": 0.49303899700023074, + "learning_rate": 0.0015328003392920642, + "loss": 1.3916, + "step": 6410 + }, + { + "epoch": 0.4931159141604492, + "learning_rate": 0.0015324379614280875, + "loss": 1.29, + "step": 6411 + }, + { + "epoch": 0.49319283132066766, + "learning_rate": 0.0015320755816700246, + "loss": 1.3708, + "step": 6412 + }, + { + "epoch": 0.49326974848088606, + "learning_rate": 0.0015317132000390351, + "loss": 1.2731, + "step": 6413 + }, + { + "epoch": 0.4933466656411045, + "learning_rate": 0.0015313508165562789, + "loss": 1.2686, + "step": 6414 + }, + { + "epoch": 0.493423582801323, + "learning_rate": 0.0015309884312429153, + "loss": 0.9308, + "step": 6415 + }, + { + "epoch": 0.49350049996154144, + "learning_rate": 0.0015306260441201056, + "loss": 0.9955, + "step": 6416 + }, + { + "epoch": 0.49357741712175984, + "learning_rate": 0.0015302636552090088, + "loss": 1.1347, + "step": 6417 + }, + { + "epoch": 0.4936543342819783, + "learning_rate": 0.001529901264530786, + "loss": 1.1132, + "step": 6418 + }, + { + "epoch": 0.49373125144219676, + "learning_rate": 0.0015295388721065968, + "loss": 0.8727, + "step": 6419 + }, + { + "epoch": 0.4938081686024152, + "learning_rate": 0.0015291764779576022, + "loss": 0.8353, + "step": 6420 + }, + { + "epoch": 0.4938850857626336, + "learning_rate": 0.0015288140821049629, + "loss": 0.89, + "step": 6421 + }, + { + "epoch": 0.4939620029228521, + "learning_rate": 0.0015284516845698389, + "loss": 1.022, + "step": 6422 + }, + { + "epoch": 0.49403892008307054, + "learning_rate": 0.0015280892853733916, + "loss": 0.8269, + "step": 6423 + }, + { + "epoch": 0.494115837243289, + "learning_rate": 0.0015277268845367814, + "loss": 1.1641, + "step": 6424 + }, + { + "epoch": 0.4941927544035074, + "learning_rate": 0.0015273644820811704, + "loss": 0.8085, + "step": 6425 + }, + { + "epoch": 0.49426967156372587, + "learning_rate": 0.0015270020780277177, + "loss": 1.4092, + "step": 6426 + }, + { + "epoch": 0.4943465887239443, + "learning_rate": 0.0015266396723975862, + "loss": 1.3734, + "step": 6427 + }, + { + "epoch": 0.4944235058841628, + "learning_rate": 0.0015262772652119357, + "loss": 1.3758, + "step": 6428 + }, + { + "epoch": 0.4945004230443812, + "learning_rate": 0.0015259148564919292, + "loss": 1.2438, + "step": 6429 + }, + { + "epoch": 0.49457734020459965, + "learning_rate": 0.0015255524462587267, + "loss": 1.1805, + "step": 6430 + }, + { + "epoch": 0.4946542573648181, + "learning_rate": 0.0015251900345334903, + "loss": 1.284, + "step": 6431 + }, + { + "epoch": 0.4947311745250365, + "learning_rate": 0.001524827621337382, + "loss": 1.5342, + "step": 6432 + }, + { + "epoch": 0.49480809168525497, + "learning_rate": 0.0015244652066915623, + "loss": 1.1638, + "step": 6433 + }, + { + "epoch": 0.49488500884547343, + "learning_rate": 0.001524102790617194, + "loss": 0.8618, + "step": 6434 + }, + { + "epoch": 0.4949619260056919, + "learning_rate": 0.0015237403731354386, + "loss": 1.1474, + "step": 6435 + }, + { + "epoch": 0.4950388431659103, + "learning_rate": 0.0015233779542674587, + "loss": 1.397, + "step": 6436 + }, + { + "epoch": 0.49511576032612875, + "learning_rate": 0.0015230155340344145, + "loss": 1.4225, + "step": 6437 + }, + { + "epoch": 0.4951926774863472, + "learning_rate": 0.0015226531124574703, + "loss": 1.0917, + "step": 6438 + }, + { + "epoch": 0.49526959464656567, + "learning_rate": 0.0015222906895577864, + "loss": 1.1323, + "step": 6439 + }, + { + "epoch": 0.4953465118067841, + "learning_rate": 0.0015219282653565265, + "loss": 1.2818, + "step": 6440 + }, + { + "epoch": 0.49542342896700253, + "learning_rate": 0.001521565839874852, + "loss": 1.0726, + "step": 6441 + }, + { + "epoch": 0.495500346127221, + "learning_rate": 0.0015212034131339258, + "loss": 1.1083, + "step": 6442 + }, + { + "epoch": 0.49557726328743945, + "learning_rate": 0.0015208409851549104, + "loss": 1.1327, + "step": 6443 + }, + { + "epoch": 0.49565418044765785, + "learning_rate": 0.001520478555958968, + "loss": 1.031, + "step": 6444 + }, + { + "epoch": 0.4957310976078763, + "learning_rate": 0.0015201161255672614, + "loss": 0.8077, + "step": 6445 + }, + { + "epoch": 0.4958080147680948, + "learning_rate": 0.001519753694000953, + "loss": 1.2604, + "step": 6446 + }, + { + "epoch": 0.49588493192831323, + "learning_rate": 0.0015193912612812062, + "loss": 1.7456, + "step": 6447 + }, + { + "epoch": 0.49596184908853164, + "learning_rate": 0.0015190288274291833, + "loss": 0.9868, + "step": 6448 + }, + { + "epoch": 0.4960387662487501, + "learning_rate": 0.0015186663924660476, + "loss": 1.3619, + "step": 6449 + }, + { + "epoch": 0.49611568340896856, + "learning_rate": 0.0015183039564129613, + "loss": 0.9619, + "step": 6450 + }, + { + "epoch": 0.49619260056918696, + "learning_rate": 0.0015179415192910879, + "loss": 1.1899, + "step": 6451 + }, + { + "epoch": 0.4962695177294054, + "learning_rate": 0.001517579081121591, + "loss": 1.2517, + "step": 6452 + }, + { + "epoch": 0.4963464348896239, + "learning_rate": 0.0015172166419256327, + "loss": 1.1316, + "step": 6453 + }, + { + "epoch": 0.49642335204984234, + "learning_rate": 0.001516854201724377, + "loss": 1.0495, + "step": 6454 + }, + { + "epoch": 0.49650026921006074, + "learning_rate": 0.001516491760538987, + "loss": 1.1894, + "step": 6455 + }, + { + "epoch": 0.4965771863702792, + "learning_rate": 0.0015161293183906259, + "loss": 1.2593, + "step": 6456 + }, + { + "epoch": 0.49665410353049766, + "learning_rate": 0.0015157668753004567, + "loss": 1.1953, + "step": 6457 + }, + { + "epoch": 0.4967310206907161, + "learning_rate": 0.0015154044312896435, + "loss": 1.6624, + "step": 6458 + }, + { + "epoch": 0.4968079378509345, + "learning_rate": 0.0015150419863793497, + "loss": 1.1629, + "step": 6459 + }, + { + "epoch": 0.496884855011153, + "learning_rate": 0.0015146795405907387, + "loss": 1.5651, + "step": 6460 + }, + { + "epoch": 0.49696177217137144, + "learning_rate": 0.0015143170939449734, + "loss": 1.0473, + "step": 6461 + }, + { + "epoch": 0.4970386893315899, + "learning_rate": 0.0015139546464632183, + "loss": 1.0656, + "step": 6462 + }, + { + "epoch": 0.4971156064918083, + "learning_rate": 0.0015135921981666372, + "loss": 0.934, + "step": 6463 + }, + { + "epoch": 0.49719252365202676, + "learning_rate": 0.001513229749076393, + "loss": 1.3089, + "step": 6464 + }, + { + "epoch": 0.4972694408122452, + "learning_rate": 0.0015128672992136506, + "loss": 1.0161, + "step": 6465 + }, + { + "epoch": 0.4973463579724637, + "learning_rate": 0.0015125048485995726, + "loss": 1.4793, + "step": 6466 + }, + { + "epoch": 0.4974232751326821, + "learning_rate": 0.0015121423972553239, + "loss": 1.2478, + "step": 6467 + }, + { + "epoch": 0.49750019229290054, + "learning_rate": 0.0015117799452020674, + "loss": 1.3143, + "step": 6468 + }, + { + "epoch": 0.497577109453119, + "learning_rate": 0.0015114174924609682, + "loss": 1.1109, + "step": 6469 + }, + { + "epoch": 0.4976540266133374, + "learning_rate": 0.0015110550390531896, + "loss": 0.9355, + "step": 6470 + }, + { + "epoch": 0.49773094377355587, + "learning_rate": 0.0015106925849998956, + "loss": 1.045, + "step": 6471 + }, + { + "epoch": 0.4978078609337743, + "learning_rate": 0.00151033013032225, + "loss": 1.1446, + "step": 6472 + }, + { + "epoch": 0.4978847780939928, + "learning_rate": 0.001509967675041418, + "loss": 1.1395, + "step": 6473 + }, + { + "epoch": 0.4979616952542112, + "learning_rate": 0.001509605219178563, + "loss": 1.1233, + "step": 6474 + }, + { + "epoch": 0.49803861241442965, + "learning_rate": 0.001509242762754849, + "loss": 1.1183, + "step": 6475 + }, + { + "epoch": 0.4981155295746481, + "learning_rate": 0.0015088803057914403, + "loss": 1.4301, + "step": 6476 + }, + { + "epoch": 0.49819244673486657, + "learning_rate": 0.0015085178483095014, + "loss": 1.1801, + "step": 6477 + }, + { + "epoch": 0.49826936389508497, + "learning_rate": 0.001508155390330197, + "loss": 0.963, + "step": 6478 + }, + { + "epoch": 0.49834628105530343, + "learning_rate": 0.00150779293187469, + "loss": 1.2661, + "step": 6479 + }, + { + "epoch": 0.4984231982155219, + "learning_rate": 0.0015074304729641463, + "loss": 1.1302, + "step": 6480 + }, + { + "epoch": 0.49850011537574035, + "learning_rate": 0.0015070680136197291, + "loss": 1.0476, + "step": 6481 + }, + { + "epoch": 0.49857703253595875, + "learning_rate": 0.0015067055538626037, + "loss": 1.2589, + "step": 6482 + }, + { + "epoch": 0.4986539496961772, + "learning_rate": 0.0015063430937139336, + "loss": 0.9057, + "step": 6483 + }, + { + "epoch": 0.49873086685639567, + "learning_rate": 0.0015059806331948838, + "loss": 0.9872, + "step": 6484 + }, + { + "epoch": 0.49880778401661413, + "learning_rate": 0.0015056181723266183, + "loss": 1.2148, + "step": 6485 + }, + { + "epoch": 0.49888470117683253, + "learning_rate": 0.001505255711130302, + "loss": 1.2279, + "step": 6486 + }, + { + "epoch": 0.498961618337051, + "learning_rate": 0.0015048932496270998, + "loss": 0.8585, + "step": 6487 + }, + { + "epoch": 0.49903853549726945, + "learning_rate": 0.0015045307878381747, + "loss": 1.2618, + "step": 6488 + }, + { + "epoch": 0.4991154526574879, + "learning_rate": 0.001504168325784693, + "loss": 1.0307, + "step": 6489 + }, + { + "epoch": 0.4991923698177063, + "learning_rate": 0.0015038058634878181, + "loss": 1.0074, + "step": 6490 + }, + { + "epoch": 0.4992692869779248, + "learning_rate": 0.001503443400968715, + "loss": 1.0053, + "step": 6491 + }, + { + "epoch": 0.49934620413814323, + "learning_rate": 0.0015030809382485478, + "loss": 1.1319, + "step": 6492 + }, + { + "epoch": 0.49942312129836164, + "learning_rate": 0.0015027184753484818, + "loss": 1.2984, + "step": 6493 + }, + { + "epoch": 0.4995000384585801, + "learning_rate": 0.0015023560122896808, + "loss": 0.7327, + "step": 6494 + }, + { + "epoch": 0.49957695561879856, + "learning_rate": 0.0015019935490933104, + "loss": 1.4445, + "step": 6495 + }, + { + "epoch": 0.499653872779017, + "learning_rate": 0.001501631085780534, + "loss": 1.1119, + "step": 6496 + }, + { + "epoch": 0.4997307899392354, + "learning_rate": 0.0015012686223725169, + "loss": 1.3338, + "step": 6497 + }, + { + "epoch": 0.4998077070994539, + "learning_rate": 0.0015009061588904236, + "loss": 1.3912, + "step": 6498 + }, + { + "epoch": 0.49988462425967234, + "learning_rate": 0.0015005436953554187, + "loss": 1.2257, + "step": 6499 + }, + { + "epoch": 0.4999615414198908, + "learning_rate": 0.001500181231788667, + "loss": 1.5399, + "step": 6500 + }, + { + "epoch": 0.5000384585801092, + "learning_rate": 0.001499818768211333, + "loss": 0.8163, + "step": 6501 + }, + { + "epoch": 0.5001153757403277, + "learning_rate": 0.0014994563046445812, + "loss": 0.9792, + "step": 6502 + }, + { + "epoch": 0.5001922929005461, + "learning_rate": 0.0014990938411095765, + "loss": 1.0609, + "step": 6503 + }, + { + "epoch": 0.5002692100607645, + "learning_rate": 0.0014987313776274834, + "loss": 1.3299, + "step": 6504 + }, + { + "epoch": 0.500346127220983, + "learning_rate": 0.0014983689142194661, + "loss": 1.3401, + "step": 6505 + }, + { + "epoch": 0.5004230443812014, + "learning_rate": 0.00149800645090669, + "loss": 1.1659, + "step": 6506 + }, + { + "epoch": 0.5004999615414198, + "learning_rate": 0.0014976439877103197, + "loss": 1.1513, + "step": 6507 + }, + { + "epoch": 0.5005768787016384, + "learning_rate": 0.0014972815246515187, + "loss": 1.3099, + "step": 6508 + }, + { + "epoch": 0.5006537958618568, + "learning_rate": 0.0014969190617514525, + "loss": 1.0427, + "step": 6509 + }, + { + "epoch": 0.5007307130220753, + "learning_rate": 0.001496556599031285, + "loss": 1.3597, + "step": 6510 + }, + { + "epoch": 0.5008076301822937, + "learning_rate": 0.0014961941365121822, + "loss": 0.9758, + "step": 6511 + }, + { + "epoch": 0.5008845473425121, + "learning_rate": 0.0014958316742153073, + "loss": 1.3063, + "step": 6512 + }, + { + "epoch": 0.5009614645027306, + "learning_rate": 0.0014954692121618251, + "loss": 1.4286, + "step": 6513 + }, + { + "epoch": 0.501038381662949, + "learning_rate": 0.0014951067503729005, + "loss": 1.4384, + "step": 6514 + }, + { + "epoch": 0.5011152988231674, + "learning_rate": 0.001494744288869698, + "loss": 1.0694, + "step": 6515 + }, + { + "epoch": 0.5011922159833859, + "learning_rate": 0.0014943818276733816, + "loss": 1.3276, + "step": 6516 + }, + { + "epoch": 0.5012691331436043, + "learning_rate": 0.0014940193668051165, + "loss": 0.8976, + "step": 6517 + }, + { + "epoch": 0.5013460503038227, + "learning_rate": 0.001493656906286067, + "loss": 1.0597, + "step": 6518 + }, + { + "epoch": 0.5014229674640412, + "learning_rate": 0.0014932944461373964, + "loss": 0.8895, + "step": 6519 + }, + { + "epoch": 0.5014998846242597, + "learning_rate": 0.0014929319863802712, + "loss": 0.9636, + "step": 6520 + }, + { + "epoch": 0.5015768017844782, + "learning_rate": 0.0014925695270358536, + "loss": 1.2442, + "step": 6521 + }, + { + "epoch": 0.5016537189446966, + "learning_rate": 0.00149220706812531, + "loss": 1.1827, + "step": 6522 + }, + { + "epoch": 0.501730636104915, + "learning_rate": 0.0014918446096698036, + "loss": 1.2397, + "step": 6523 + }, + { + "epoch": 0.5018075532651335, + "learning_rate": 0.0014914821516904984, + "loss": 1.1812, + "step": 6524 + }, + { + "epoch": 0.5018844704253519, + "learning_rate": 0.0014911196942085597, + "loss": 1.4535, + "step": 6525 + }, + { + "epoch": 0.5019613875855703, + "learning_rate": 0.0014907572372451514, + "loss": 1.0805, + "step": 6526 + }, + { + "epoch": 0.5020383047457888, + "learning_rate": 0.0014903947808214372, + "loss": 1.2827, + "step": 6527 + }, + { + "epoch": 0.5021152219060072, + "learning_rate": 0.0014900323249585823, + "loss": 1.1194, + "step": 6528 + }, + { + "epoch": 0.5021921390662257, + "learning_rate": 0.0014896698696777496, + "loss": 1.1704, + "step": 6529 + }, + { + "epoch": 0.5022690562264441, + "learning_rate": 0.0014893074150001046, + "loss": 1.235, + "step": 6530 + }, + { + "epoch": 0.5023459733866625, + "learning_rate": 0.0014889449609468111, + "loss": 1.0981, + "step": 6531 + }, + { + "epoch": 0.502422890546881, + "learning_rate": 0.0014885825075390316, + "loss": 1.2114, + "step": 6532 + }, + { + "epoch": 0.5024998077070995, + "learning_rate": 0.0014882200547979329, + "loss": 1.6672, + "step": 6533 + }, + { + "epoch": 0.5025767248673179, + "learning_rate": 0.0014878576027446766, + "loss": 0.639, + "step": 6534 + }, + { + "epoch": 0.5026536420275364, + "learning_rate": 0.0014874951514004275, + "loss": 1.3364, + "step": 6535 + }, + { + "epoch": 0.5027305591877548, + "learning_rate": 0.00148713270078635, + "loss": 1.3282, + "step": 6536 + }, + { + "epoch": 0.5028074763479732, + "learning_rate": 0.001486770250923607, + "loss": 0.9317, + "step": 6537 + }, + { + "epoch": 0.5028843935081917, + "learning_rate": 0.001486407801833363, + "loss": 0.6212, + "step": 6538 + }, + { + "epoch": 0.5029613106684101, + "learning_rate": 0.0014860453535367817, + "loss": 1.1881, + "step": 6539 + }, + { + "epoch": 0.5030382278286286, + "learning_rate": 0.0014856829060550267, + "loss": 1.2884, + "step": 6540 + }, + { + "epoch": 0.503115144988847, + "learning_rate": 0.0014853204594092618, + "loss": 1.2122, + "step": 6541 + }, + { + "epoch": 0.5031920621490654, + "learning_rate": 0.0014849580136206508, + "loss": 1.3868, + "step": 6542 + }, + { + "epoch": 0.5032689793092839, + "learning_rate": 0.0014845955687103564, + "loss": 1.3568, + "step": 6543 + }, + { + "epoch": 0.5033458964695023, + "learning_rate": 0.0014842331246995436, + "loss": 0.6795, + "step": 6544 + }, + { + "epoch": 0.5034228136297207, + "learning_rate": 0.0014838706816093748, + "loss": 1.0597, + "step": 6545 + }, + { + "epoch": 0.5034997307899393, + "learning_rate": 0.0014835082394610132, + "loss": 1.2906, + "step": 6546 + }, + { + "epoch": 0.5035766479501577, + "learning_rate": 0.0014831457982756233, + "loss": 1.0363, + "step": 6547 + }, + { + "epoch": 0.5036535651103762, + "learning_rate": 0.0014827833580743671, + "loss": 1.4559, + "step": 6548 + }, + { + "epoch": 0.5037304822705946, + "learning_rate": 0.0014824209188784093, + "loss": 0.9388, + "step": 6549 + }, + { + "epoch": 0.503807399430813, + "learning_rate": 0.0014820584807089122, + "loss": 1.0233, + "step": 6550 + }, + { + "epoch": 0.5038843165910315, + "learning_rate": 0.0014816960435870387, + "loss": 0.9559, + "step": 6551 + }, + { + "epoch": 0.5039612337512499, + "learning_rate": 0.0014813336075339529, + "loss": 1.1925, + "step": 6552 + }, + { + "epoch": 0.5040381509114683, + "learning_rate": 0.0014809711725708172, + "loss": 0.8224, + "step": 6553 + }, + { + "epoch": 0.5041150680716868, + "learning_rate": 0.0014806087387187938, + "loss": 1.1409, + "step": 6554 + }, + { + "epoch": 0.5041919852319052, + "learning_rate": 0.0014802463059990472, + "loss": 0.7648, + "step": 6555 + }, + { + "epoch": 0.5042689023921236, + "learning_rate": 0.001479883874432739, + "loss": 1.1594, + "step": 6556 + }, + { + "epoch": 0.5043458195523421, + "learning_rate": 0.0014795214440410322, + "loss": 1.0776, + "step": 6557 + }, + { + "epoch": 0.5044227367125605, + "learning_rate": 0.0014791590148450899, + "loss": 1.282, + "step": 6558 + }, + { + "epoch": 0.5044996538727791, + "learning_rate": 0.001478796586866074, + "loss": 1.106, + "step": 6559 + }, + { + "epoch": 0.5045765710329975, + "learning_rate": 0.001478434160125148, + "loss": 1.316, + "step": 6560 + }, + { + "epoch": 0.5046534881932159, + "learning_rate": 0.0014780717346434737, + "loss": 0.9918, + "step": 6561 + }, + { + "epoch": 0.5047304053534344, + "learning_rate": 0.0014777093104422136, + "loss": 0.8127, + "step": 6562 + }, + { + "epoch": 0.5048073225136528, + "learning_rate": 0.0014773468875425302, + "loss": 1.1305, + "step": 6563 + }, + { + "epoch": 0.5048842396738712, + "learning_rate": 0.001476984465965586, + "loss": 1.1134, + "step": 6564 + }, + { + "epoch": 0.5049611568340897, + "learning_rate": 0.0014766220457325418, + "loss": 1.299, + "step": 6565 + }, + { + "epoch": 0.5050380739943081, + "learning_rate": 0.0014762596268645617, + "loss": 1.4228, + "step": 6566 + }, + { + "epoch": 0.5051149911545266, + "learning_rate": 0.0014758972093828057, + "loss": 1.0898, + "step": 6567 + }, + { + "epoch": 0.505191908314745, + "learning_rate": 0.001475534793308438, + "loss": 1.192, + "step": 6568 + }, + { + "epoch": 0.5052688254749634, + "learning_rate": 0.0014751723786626187, + "loss": 1.4412, + "step": 6569 + }, + { + "epoch": 0.505345742635182, + "learning_rate": 0.0014748099654665096, + "loss": 0.7492, + "step": 6570 + }, + { + "epoch": 0.5054226597954004, + "learning_rate": 0.0014744475537412734, + "loss": 0.8845, + "step": 6571 + }, + { + "epoch": 0.5054995769556188, + "learning_rate": 0.0014740851435080713, + "loss": 1.1231, + "step": 6572 + }, + { + "epoch": 0.5055764941158373, + "learning_rate": 0.0014737227347880642, + "loss": 1.1297, + "step": 6573 + }, + { + "epoch": 0.5056534112760557, + "learning_rate": 0.0014733603276024143, + "loss": 1.6957, + "step": 6574 + }, + { + "epoch": 0.5057303284362741, + "learning_rate": 0.001472997921972283, + "loss": 1.1814, + "step": 6575 + }, + { + "epoch": 0.5058072455964926, + "learning_rate": 0.0014726355179188301, + "loss": 1.3075, + "step": 6576 + }, + { + "epoch": 0.505884162756711, + "learning_rate": 0.0014722731154632187, + "loss": 1.2026, + "step": 6577 + }, + { + "epoch": 0.5059610799169295, + "learning_rate": 0.0014719107146266082, + "loss": 1.096, + "step": 6578 + }, + { + "epoch": 0.5060379970771479, + "learning_rate": 0.0014715483154301614, + "loss": 0.7804, + "step": 6579 + }, + { + "epoch": 0.5061149142373663, + "learning_rate": 0.0014711859178950378, + "loss": 0.9692, + "step": 6580 + }, + { + "epoch": 0.5061918313975848, + "learning_rate": 0.001470823522042398, + "loss": 1.1861, + "step": 6581 + }, + { + "epoch": 0.5062687485578032, + "learning_rate": 0.0014704611278934035, + "loss": 1.0058, + "step": 6582 + }, + { + "epoch": 0.5063456657180216, + "learning_rate": 0.0014700987354692146, + "loss": 1.5371, + "step": 6583 + }, + { + "epoch": 0.5064225828782402, + "learning_rate": 0.0014697363447909912, + "loss": 1.0502, + "step": 6584 + }, + { + "epoch": 0.5064995000384586, + "learning_rate": 0.0014693739558798947, + "loss": 1.1506, + "step": 6585 + }, + { + "epoch": 0.5065764171986771, + "learning_rate": 0.0014690115687570845, + "loss": 1.1711, + "step": 6586 + }, + { + "epoch": 0.5066533343588955, + "learning_rate": 0.0014686491834437214, + "loss": 1.0638, + "step": 6587 + }, + { + "epoch": 0.5067302515191139, + "learning_rate": 0.0014682867999609654, + "loss": 1.0009, + "step": 6588 + }, + { + "epoch": 0.5068071686793324, + "learning_rate": 0.0014679244183299753, + "loss": 1.1536, + "step": 6589 + }, + { + "epoch": 0.5068840858395508, + "learning_rate": 0.0014675620385719128, + "loss": 0.9509, + "step": 6590 + }, + { + "epoch": 0.5069610029997692, + "learning_rate": 0.0014671996607079365, + "loss": 0.9983, + "step": 6591 + }, + { + "epoch": 0.5070379201599877, + "learning_rate": 0.0014668372847592055, + "loss": 1.2864, + "step": 6592 + }, + { + "epoch": 0.5071148373202061, + "learning_rate": 0.0014664749107468804, + "loss": 1.3432, + "step": 6593 + }, + { + "epoch": 0.5071917544804245, + "learning_rate": 0.0014661125386921205, + "loss": 1.1032, + "step": 6594 + }, + { + "epoch": 0.507268671640643, + "learning_rate": 0.0014657501686160842, + "loss": 1.4989, + "step": 6595 + }, + { + "epoch": 0.5073455888008614, + "learning_rate": 0.0014653878005399317, + "loss": 1.3576, + "step": 6596 + }, + { + "epoch": 0.50742250596108, + "learning_rate": 0.0014650254344848212, + "loss": 1.3586, + "step": 6597 + }, + { + "epoch": 0.5074994231212984, + "learning_rate": 0.0014646630704719123, + "loss": 1.2624, + "step": 6598 + }, + { + "epoch": 0.5075763402815168, + "learning_rate": 0.001464300708522364, + "loss": 1.0322, + "step": 6599 + }, + { + "epoch": 0.5076532574417353, + "learning_rate": 0.0014639383486573336, + "loss": 0.835, + "step": 6600 + }, + { + "epoch": 0.5077301746019537, + "learning_rate": 0.0014635759908979815, + "loss": 1.3156, + "step": 6601 + }, + { + "epoch": 0.5078070917621721, + "learning_rate": 0.001463213635265465, + "loss": 1.2155, + "step": 6602 + }, + { + "epoch": 0.5078840089223906, + "learning_rate": 0.0014628512817809422, + "loss": 1.1603, + "step": 6603 + }, + { + "epoch": 0.507960926082609, + "learning_rate": 0.001462488930465572, + "loss": 0.8567, + "step": 6604 + }, + { + "epoch": 0.5080378432428275, + "learning_rate": 0.0014621265813405122, + "loss": 1.3782, + "step": 6605 + }, + { + "epoch": 0.5081147604030459, + "learning_rate": 0.0014617642344269212, + "loss": 1.0919, + "step": 6606 + }, + { + "epoch": 0.5081916775632643, + "learning_rate": 0.001461401889745956, + "loss": 1.0697, + "step": 6607 + }, + { + "epoch": 0.5082685947234828, + "learning_rate": 0.0014610395473187746, + "loss": 1.1664, + "step": 6608 + }, + { + "epoch": 0.5083455118837013, + "learning_rate": 0.0014606772071665349, + "loss": 1.5752, + "step": 6609 + }, + { + "epoch": 0.5084224290439197, + "learning_rate": 0.0014603148693103944, + "loss": 1.3914, + "step": 6610 + }, + { + "epoch": 0.5084993462041382, + "learning_rate": 0.001459952533771509, + "loss": 1.3671, + "step": 6611 + }, + { + "epoch": 0.5085762633643566, + "learning_rate": 0.001459590200571038, + "loss": 1.3097, + "step": 6612 + }, + { + "epoch": 0.5086531805245751, + "learning_rate": 0.0014592278697301362, + "loss": 1.1959, + "step": 6613 + }, + { + "epoch": 0.5087300976847935, + "learning_rate": 0.0014588655412699626, + "loss": 1.3758, + "step": 6614 + }, + { + "epoch": 0.5088070148450119, + "learning_rate": 0.0014585032152116726, + "loss": 1.2366, + "step": 6615 + }, + { + "epoch": 0.5088839320052304, + "learning_rate": 0.0014581408915764228, + "loss": 1.2483, + "step": 6616 + }, + { + "epoch": 0.5089608491654488, + "learning_rate": 0.0014577785703853703, + "loss": 0.9804, + "step": 6617 + }, + { + "epoch": 0.5090377663256672, + "learning_rate": 0.001457416251659671, + "loss": 0.9902, + "step": 6618 + }, + { + "epoch": 0.5091146834858857, + "learning_rate": 0.001457053935420481, + "loss": 1.6953, + "step": 6619 + }, + { + "epoch": 0.5091916006461041, + "learning_rate": 0.0014566916216889566, + "loss": 1.0082, + "step": 6620 + }, + { + "epoch": 0.5092685178063225, + "learning_rate": 0.001456329310486254, + "loss": 1.1904, + "step": 6621 + }, + { + "epoch": 0.509345434966541, + "learning_rate": 0.0014559670018335273, + "loss": 1.3764, + "step": 6622 + }, + { + "epoch": 0.5094223521267595, + "learning_rate": 0.0014556046957519345, + "loss": 1.2398, + "step": 6623 + }, + { + "epoch": 0.509499269286978, + "learning_rate": 0.0014552423922626284, + "loss": 1.2004, + "step": 6624 + }, + { + "epoch": 0.5095761864471964, + "learning_rate": 0.001454880091386767, + "loss": 1.3528, + "step": 6625 + }, + { + "epoch": 0.5096531036074148, + "learning_rate": 0.0014545177931455037, + "loss": 1.0543, + "step": 6626 + }, + { + "epoch": 0.5097300207676333, + "learning_rate": 0.0014541554975599936, + "loss": 1.0902, + "step": 6627 + }, + { + "epoch": 0.5098069379278517, + "learning_rate": 0.001453793204651392, + "loss": 0.7525, + "step": 6628 + }, + { + "epoch": 0.5098838550880701, + "learning_rate": 0.0014534309144408532, + "loss": 1.1906, + "step": 6629 + }, + { + "epoch": 0.5099607722482886, + "learning_rate": 0.0014530686269495316, + "loss": 1.3126, + "step": 6630 + }, + { + "epoch": 0.510037689408507, + "learning_rate": 0.0014527063421985822, + "loss": 1.408, + "step": 6631 + }, + { + "epoch": 0.5101146065687255, + "learning_rate": 0.0014523440602091583, + "loss": 1.0331, + "step": 6632 + }, + { + "epoch": 0.5101915237289439, + "learning_rate": 0.001451981781002415, + "loss": 0.8259, + "step": 6633 + }, + { + "epoch": 0.5102684408891623, + "learning_rate": 0.0014516195045995058, + "loss": 1.2364, + "step": 6634 + }, + { + "epoch": 0.5103453580493809, + "learning_rate": 0.001451257231021583, + "loss": 1.0736, + "step": 6635 + }, + { + "epoch": 0.5104222752095993, + "learning_rate": 0.0014508949602898022, + "loss": 1.1846, + "step": 6636 + }, + { + "epoch": 0.5104991923698177, + "learning_rate": 0.0014505326924253157, + "loss": 1.4337, + "step": 6637 + }, + { + "epoch": 0.5105761095300362, + "learning_rate": 0.0014501704274492765, + "loss": 1.5369, + "step": 6638 + }, + { + "epoch": 0.5106530266902546, + "learning_rate": 0.0014498081653828382, + "loss": 1.0255, + "step": 6639 + }, + { + "epoch": 0.510729943850473, + "learning_rate": 0.0014494459062471537, + "loss": 1.2547, + "step": 6640 + }, + { + "epoch": 0.5108068610106915, + "learning_rate": 0.0014490836500633748, + "loss": 0.8558, + "step": 6641 + }, + { + "epoch": 0.5108837781709099, + "learning_rate": 0.001448721396852655, + "loss": 1.011, + "step": 6642 + }, + { + "epoch": 0.5109606953311284, + "learning_rate": 0.001448359146636146, + "loss": 0.6644, + "step": 6643 + }, + { + "epoch": 0.5110376124913468, + "learning_rate": 0.0014479968994350004, + "loss": 1.4301, + "step": 6644 + }, + { + "epoch": 0.5111145296515652, + "learning_rate": 0.0014476346552703705, + "loss": 1.2024, + "step": 6645 + }, + { + "epoch": 0.5111914468117837, + "learning_rate": 0.0014472724141634068, + "loss": 1.0296, + "step": 6646 + }, + { + "epoch": 0.5112683639720021, + "learning_rate": 0.0014469101761352627, + "loss": 1.4159, + "step": 6647 + }, + { + "epoch": 0.5113452811322206, + "learning_rate": 0.0014465479412070885, + "loss": 1.272, + "step": 6648 + }, + { + "epoch": 0.5114221982924391, + "learning_rate": 0.0014461857094000352, + "loss": 1.3374, + "step": 6649 + }, + { + "epoch": 0.5114991154526575, + "learning_rate": 0.001445823480735255, + "loss": 1.012, + "step": 6650 + }, + { + "epoch": 0.511576032612876, + "learning_rate": 0.0014454612552338977, + "loss": 1.4119, + "step": 6651 + }, + { + "epoch": 0.5116529497730944, + "learning_rate": 0.0014450990329171146, + "loss": 0.8706, + "step": 6652 + }, + { + "epoch": 0.5117298669333128, + "learning_rate": 0.0014447368138060565, + "loss": 1.0146, + "step": 6653 + }, + { + "epoch": 0.5118067840935313, + "learning_rate": 0.0014443745979218733, + "loss": 1.1647, + "step": 6654 + }, + { + "epoch": 0.5118837012537497, + "learning_rate": 0.0014440123852857152, + "loss": 1.1763, + "step": 6655 + }, + { + "epoch": 0.5119606184139681, + "learning_rate": 0.0014436501759187326, + "loss": 1.1731, + "step": 6656 + }, + { + "epoch": 0.5120375355741866, + "learning_rate": 0.0014432879698420744, + "loss": 0.9966, + "step": 6657 + }, + { + "epoch": 0.512114452734405, + "learning_rate": 0.0014429257670768912, + "loss": 1.0666, + "step": 6658 + }, + { + "epoch": 0.5121913698946234, + "learning_rate": 0.0014425635676443322, + "loss": 1.2646, + "step": 6659 + }, + { + "epoch": 0.512268287054842, + "learning_rate": 0.0014422013715655459, + "loss": 1.1913, + "step": 6660 + }, + { + "epoch": 0.5123452042150604, + "learning_rate": 0.0014418391788616819, + "loss": 1.1502, + "step": 6661 + }, + { + "epoch": 0.5124221213752789, + "learning_rate": 0.0014414769895538884, + "loss": 1.2237, + "step": 6662 + }, + { + "epoch": 0.5124990385354973, + "learning_rate": 0.001441114803663315, + "loss": 1.0391, + "step": 6663 + }, + { + "epoch": 0.5125759556957157, + "learning_rate": 0.0014407526212111095, + "loss": 1.1168, + "step": 6664 + }, + { + "epoch": 0.5126528728559342, + "learning_rate": 0.0014403904422184198, + "loss": 0.949, + "step": 6665 + }, + { + "epoch": 0.5127297900161526, + "learning_rate": 0.001440028266706395, + "loss": 1.1032, + "step": 6666 + }, + { + "epoch": 0.512806707176371, + "learning_rate": 0.001439666094696182, + "loss": 1.0011, + "step": 6667 + }, + { + "epoch": 0.5128836243365895, + "learning_rate": 0.001439303926208928, + "loss": 1.0922, + "step": 6668 + }, + { + "epoch": 0.5129605414968079, + "learning_rate": 0.0014389417612657815, + "loss": 1.4588, + "step": 6669 + }, + { + "epoch": 0.5130374586570264, + "learning_rate": 0.001438579599887889, + "loss": 0.9473, + "step": 6670 + }, + { + "epoch": 0.5131143758172448, + "learning_rate": 0.0014382174420963978, + "loss": 1.2888, + "step": 6671 + }, + { + "epoch": 0.5131912929774632, + "learning_rate": 0.0014378552879124552, + "loss": 1.0715, + "step": 6672 + }, + { + "epoch": 0.5132682101376818, + "learning_rate": 0.0014374931373572059, + "loss": 1.0945, + "step": 6673 + }, + { + "epoch": 0.5133451272979002, + "learning_rate": 0.0014371309904517985, + "loss": 1.2115, + "step": 6674 + }, + { + "epoch": 0.5134220444581186, + "learning_rate": 0.0014367688472173776, + "loss": 0.9833, + "step": 6675 + }, + { + "epoch": 0.5134989616183371, + "learning_rate": 0.0014364067076750894, + "loss": 1.0794, + "step": 6676 + }, + { + "epoch": 0.5135758787785555, + "learning_rate": 0.00143604457184608, + "loss": 0.9514, + "step": 6677 + }, + { + "epoch": 0.5136527959387739, + "learning_rate": 0.001435682439751495, + "loss": 1.0807, + "step": 6678 + }, + { + "epoch": 0.5137297130989924, + "learning_rate": 0.0014353203114124788, + "loss": 1.1926, + "step": 6679 + }, + { + "epoch": 0.5138066302592108, + "learning_rate": 0.0014349581868501775, + "loss": 1.0125, + "step": 6680 + }, + { + "epoch": 0.5138835474194293, + "learning_rate": 0.001434596066085735, + "loss": 1.0901, + "step": 6681 + }, + { + "epoch": 0.5139604645796477, + "learning_rate": 0.0014342339491402968, + "loss": 1.1668, + "step": 6682 + }, + { + "epoch": 0.5140373817398661, + "learning_rate": 0.001433871836035007, + "loss": 1.5974, + "step": 6683 + }, + { + "epoch": 0.5141142989000846, + "learning_rate": 0.0014335097267910088, + "loss": 1.4438, + "step": 6684 + }, + { + "epoch": 0.514191216060303, + "learning_rate": 0.001433147621429448, + "loss": 1.3482, + "step": 6685 + }, + { + "epoch": 0.5142681332205215, + "learning_rate": 0.0014327855199714667, + "loss": 1.2907, + "step": 6686 + }, + { + "epoch": 0.51434505038074, + "learning_rate": 0.0014324234224382087, + "loss": 1.0154, + "step": 6687 + }, + { + "epoch": 0.5144219675409584, + "learning_rate": 0.0014320613288508178, + "loss": 1.1176, + "step": 6688 + }, + { + "epoch": 0.5144988847011769, + "learning_rate": 0.0014316992392304363, + "loss": 1.0381, + "step": 6689 + }, + { + "epoch": 0.5145758018613953, + "learning_rate": 0.001431337153598208, + "loss": 1.1203, + "step": 6690 + }, + { + "epoch": 0.5146527190216137, + "learning_rate": 0.0014309750719752746, + "loss": 1.1187, + "step": 6691 + }, + { + "epoch": 0.5147296361818322, + "learning_rate": 0.0014306129943827785, + "loss": 1.0744, + "step": 6692 + }, + { + "epoch": 0.5148065533420506, + "learning_rate": 0.0014302509208418623, + "loss": 1.0064, + "step": 6693 + }, + { + "epoch": 0.514883470502269, + "learning_rate": 0.0014298888513736678, + "loss": 1.0792, + "step": 6694 + }, + { + "epoch": 0.5149603876624875, + "learning_rate": 0.0014295267859993355, + "loss": 0.9389, + "step": 6695 + }, + { + "epoch": 0.5150373048227059, + "learning_rate": 0.0014291647247400084, + "loss": 1.1881, + "step": 6696 + }, + { + "epoch": 0.5151142219829243, + "learning_rate": 0.0014288026676168267, + "loss": 0.8957, + "step": 6697 + }, + { + "epoch": 0.5151911391431429, + "learning_rate": 0.0014284406146509309, + "loss": 1.0012, + "step": 6698 + }, + { + "epoch": 0.5152680563033613, + "learning_rate": 0.0014280785658634626, + "loss": 0.9666, + "step": 6699 + }, + { + "epoch": 0.5153449734635798, + "learning_rate": 0.0014277165212755614, + "loss": 0.8031, + "step": 6700 + }, + { + "epoch": 0.5154218906237982, + "learning_rate": 0.0014273544809083683, + "loss": 1.2961, + "step": 6701 + }, + { + "epoch": 0.5154988077840166, + "learning_rate": 0.0014269924447830227, + "loss": 0.9829, + "step": 6702 + }, + { + "epoch": 0.5155757249442351, + "learning_rate": 0.0014266304129206642, + "loss": 1.4266, + "step": 6703 + }, + { + "epoch": 0.5156526421044535, + "learning_rate": 0.0014262683853424325, + "loss": 1.1295, + "step": 6704 + }, + { + "epoch": 0.5157295592646719, + "learning_rate": 0.0014259063620694669, + "loss": 0.7262, + "step": 6705 + }, + { + "epoch": 0.5158064764248904, + "learning_rate": 0.0014255443431229052, + "loss": 1.3216, + "step": 6706 + }, + { + "epoch": 0.5158833935851088, + "learning_rate": 0.0014251823285238877, + "loss": 1.3511, + "step": 6707 + }, + { + "epoch": 0.5159603107453273, + "learning_rate": 0.001424820318293551, + "loss": 1.5622, + "step": 6708 + }, + { + "epoch": 0.5160372279055457, + "learning_rate": 0.0014244583124530356, + "loss": 1.1363, + "step": 6709 + }, + { + "epoch": 0.5161141450657641, + "learning_rate": 0.0014240963110234775, + "loss": 1.114, + "step": 6710 + }, + { + "epoch": 0.5161910622259827, + "learning_rate": 0.0014237343140260145, + "loss": 1.316, + "step": 6711 + }, + { + "epoch": 0.5162679793862011, + "learning_rate": 0.0014233723214817848, + "loss": 1.119, + "step": 6712 + }, + { + "epoch": 0.5163448965464195, + "learning_rate": 0.0014230103334119252, + "loss": 1.3046, + "step": 6713 + }, + { + "epoch": 0.516421813706638, + "learning_rate": 0.001422648349837572, + "loss": 1.1005, + "step": 6714 + }, + { + "epoch": 0.5164987308668564, + "learning_rate": 0.0014222863707798626, + "loss": 1.3312, + "step": 6715 + }, + { + "epoch": 0.5165756480270748, + "learning_rate": 0.0014219243962599333, + "loss": 1.1983, + "step": 6716 + }, + { + "epoch": 0.5166525651872933, + "learning_rate": 0.001421562426298919, + "loss": 0.9307, + "step": 6717 + }, + { + "epoch": 0.5167294823475117, + "learning_rate": 0.0014212004609179573, + "loss": 1.1534, + "step": 6718 + }, + { + "epoch": 0.5168063995077302, + "learning_rate": 0.001420838500138182, + "loss": 0.843, + "step": 6719 + }, + { + "epoch": 0.5168833166679486, + "learning_rate": 0.00142047654398073, + "loss": 1.1494, + "step": 6720 + }, + { + "epoch": 0.516960233828167, + "learning_rate": 0.001420114592466735, + "loss": 1.1551, + "step": 6721 + }, + { + "epoch": 0.5170371509883855, + "learning_rate": 0.001419752645617332, + "loss": 1.0119, + "step": 6722 + }, + { + "epoch": 0.517114068148604, + "learning_rate": 0.0014193907034536562, + "loss": 1.1552, + "step": 6723 + }, + { + "epoch": 0.5171909853088223, + "learning_rate": 0.001419028765996841, + "loss": 1.1905, + "step": 6724 + }, + { + "epoch": 0.5172679024690409, + "learning_rate": 0.0014186668332680204, + "loss": 1.2755, + "step": 6725 + }, + { + "epoch": 0.5173448196292593, + "learning_rate": 0.0014183049052883282, + "loss": 0.8684, + "step": 6726 + }, + { + "epoch": 0.5174217367894778, + "learning_rate": 0.0014179429820788977, + "loss": 1.0045, + "step": 6727 + }, + { + "epoch": 0.5174986539496962, + "learning_rate": 0.001417581063660862, + "loss": 1.4103, + "step": 6728 + }, + { + "epoch": 0.5175755711099146, + "learning_rate": 0.0014172191500553544, + "loss": 1.1899, + "step": 6729 + }, + { + "epoch": 0.5176524882701331, + "learning_rate": 0.001416857241283506, + "loss": 1.3045, + "step": 6730 + }, + { + "epoch": 0.5177294054303515, + "learning_rate": 0.0014164953373664508, + "loss": 1.2077, + "step": 6731 + }, + { + "epoch": 0.5178063225905699, + "learning_rate": 0.0014161334383253196, + "loss": 1.3035, + "step": 6732 + }, + { + "epoch": 0.5178832397507884, + "learning_rate": 0.001415771544181244, + "loss": 1.2366, + "step": 6733 + }, + { + "epoch": 0.5179601569110068, + "learning_rate": 0.0014154096549553559, + "loss": 1.2346, + "step": 6734 + }, + { + "epoch": 0.5180370740712253, + "learning_rate": 0.0014150477706687862, + "loss": 0.8414, + "step": 6735 + }, + { + "epoch": 0.5181139912314437, + "learning_rate": 0.0014146858913426656, + "loss": 0.974, + "step": 6736 + }, + { + "epoch": 0.5181909083916622, + "learning_rate": 0.0014143240169981251, + "loss": 0.9438, + "step": 6737 + }, + { + "epoch": 0.5182678255518807, + "learning_rate": 0.001413962147656294, + "loss": 1.0879, + "step": 6738 + }, + { + "epoch": 0.5183447427120991, + "learning_rate": 0.001413600283338303, + "loss": 1.079, + "step": 6739 + }, + { + "epoch": 0.5184216598723175, + "learning_rate": 0.001413238424065282, + "loss": 1.144, + "step": 6740 + }, + { + "epoch": 0.518498577032536, + "learning_rate": 0.001412876569858359, + "loss": 1.3145, + "step": 6741 + }, + { + "epoch": 0.5185754941927544, + "learning_rate": 0.0014125147207386646, + "loss": 1.2907, + "step": 6742 + }, + { + "epoch": 0.5186524113529728, + "learning_rate": 0.0014121528767273267, + "loss": 1.1897, + "step": 6743 + }, + { + "epoch": 0.5187293285131913, + "learning_rate": 0.0014117910378454734, + "loss": 1.2253, + "step": 6744 + }, + { + "epoch": 0.5188062456734097, + "learning_rate": 0.0014114292041142336, + "loss": 1.3954, + "step": 6745 + }, + { + "epoch": 0.5188831628336282, + "learning_rate": 0.0014110673755547349, + "loss": 0.9285, + "step": 6746 + }, + { + "epoch": 0.5189600799938466, + "learning_rate": 0.0014107055521881049, + "loss": 1.0801, + "step": 6747 + }, + { + "epoch": 0.519036997154065, + "learning_rate": 0.0014103437340354707, + "loss": 0.8494, + "step": 6748 + }, + { + "epoch": 0.5191139143142836, + "learning_rate": 0.001409981921117959, + "loss": 1.2647, + "step": 6749 + }, + { + "epoch": 0.519190831474502, + "learning_rate": 0.0014096201134566973, + "loss": 0.828, + "step": 6750 + }, + { + "epoch": 0.5192677486347204, + "learning_rate": 0.0014092583110728114, + "loss": 1.0505, + "step": 6751 + }, + { + "epoch": 0.5193446657949389, + "learning_rate": 0.0014088965139874264, + "loss": 1.2376, + "step": 6752 + }, + { + "epoch": 0.5194215829551573, + "learning_rate": 0.00140853472222167, + "loss": 1.5328, + "step": 6753 + }, + { + "epoch": 0.5194985001153758, + "learning_rate": 0.0014081729357966655, + "loss": 1.3196, + "step": 6754 + }, + { + "epoch": 0.5195754172755942, + "learning_rate": 0.0014078111547335397, + "loss": 1.0608, + "step": 6755 + }, + { + "epoch": 0.5196523344358126, + "learning_rate": 0.0014074493790534165, + "loss": 1.0171, + "step": 6756 + }, + { + "epoch": 0.5197292515960311, + "learning_rate": 0.0014070876087774202, + "loss": 1.4048, + "step": 6757 + }, + { + "epoch": 0.5198061687562495, + "learning_rate": 0.0014067258439266757, + "loss": 1.2421, + "step": 6758 + }, + { + "epoch": 0.5198830859164679, + "learning_rate": 0.001406364084522306, + "loss": 0.8108, + "step": 6759 + }, + { + "epoch": 0.5199600030766864, + "learning_rate": 0.001406002330585435, + "loss": 1.175, + "step": 6760 + }, + { + "epoch": 0.5200369202369048, + "learning_rate": 0.0014056405821371862, + "loss": 1.3451, + "step": 6761 + }, + { + "epoch": 0.5201138373971232, + "learning_rate": 0.0014052788391986824, + "loss": 1.3186, + "step": 6762 + }, + { + "epoch": 0.5201907545573418, + "learning_rate": 0.0014049171017910449, + "loss": 0.9307, + "step": 6763 + }, + { + "epoch": 0.5202676717175602, + "learning_rate": 0.001404555369935398, + "loss": 1.0457, + "step": 6764 + }, + { + "epoch": 0.5203445888777787, + "learning_rate": 0.0014041936436528612, + "loss": 0.9869, + "step": 6765 + }, + { + "epoch": 0.5204215060379971, + "learning_rate": 0.0014038319229645586, + "loss": 1.3627, + "step": 6766 + }, + { + "epoch": 0.5204984231982155, + "learning_rate": 0.0014034702078916102, + "loss": 1.1022, + "step": 6767 + }, + { + "epoch": 0.520575340358434, + "learning_rate": 0.0014031084984551362, + "loss": 0.9449, + "step": 6768 + }, + { + "epoch": 0.5206522575186524, + "learning_rate": 0.0014027467946762585, + "loss": 0.9179, + "step": 6769 + }, + { + "epoch": 0.5207291746788708, + "learning_rate": 0.0014023850965760969, + "loss": 0.9102, + "step": 6770 + }, + { + "epoch": 0.5208060918390893, + "learning_rate": 0.0014020234041757706, + "loss": 0.8934, + "step": 6771 + }, + { + "epoch": 0.5208830089993077, + "learning_rate": 0.0014016617174964004, + "loss": 0.9073, + "step": 6772 + }, + { + "epoch": 0.5209599261595262, + "learning_rate": 0.0014013000365591046, + "loss": 1.0801, + "step": 6773 + }, + { + "epoch": 0.5210368433197446, + "learning_rate": 0.0014009383613850029, + "loss": 0.8615, + "step": 6774 + }, + { + "epoch": 0.521113760479963, + "learning_rate": 0.0014005766919952137, + "loss": 0.8591, + "step": 6775 + }, + { + "epoch": 0.5211906776401816, + "learning_rate": 0.0014002150284108542, + "loss": 0.8669, + "step": 6776 + }, + { + "epoch": 0.5212675948004, + "learning_rate": 0.0013998533706530442, + "loss": 1.0367, + "step": 6777 + }, + { + "epoch": 0.5213445119606184, + "learning_rate": 0.0013994917187429002, + "loss": 1.3112, + "step": 6778 + }, + { + "epoch": 0.5214214291208369, + "learning_rate": 0.0013991300727015385, + "loss": 1.364, + "step": 6779 + }, + { + "epoch": 0.5214983462810553, + "learning_rate": 0.0013987684325500779, + "loss": 1.0634, + "step": 6780 + }, + { + "epoch": 0.5215752634412737, + "learning_rate": 0.0013984067983096337, + "loss": 1.1844, + "step": 6781 + }, + { + "epoch": 0.5216521806014922, + "learning_rate": 0.0013980451700013222, + "loss": 1.1443, + "step": 6782 + }, + { + "epoch": 0.5217290977617106, + "learning_rate": 0.00139768354764626, + "loss": 1.3297, + "step": 6783 + }, + { + "epoch": 0.5218060149219291, + "learning_rate": 0.0013973219312655614, + "loss": 1.2687, + "step": 6784 + }, + { + "epoch": 0.5218829320821475, + "learning_rate": 0.0013969603208803427, + "loss": 1.268, + "step": 6785 + }, + { + "epoch": 0.5219598492423659, + "learning_rate": 0.0013965987165117186, + "loss": 1.2805, + "step": 6786 + }, + { + "epoch": 0.5220367664025845, + "learning_rate": 0.0013962371181808022, + "loss": 1.2949, + "step": 6787 + }, + { + "epoch": 0.5221136835628029, + "learning_rate": 0.0013958755259087098, + "loss": 1.1942, + "step": 6788 + }, + { + "epoch": 0.5221906007230213, + "learning_rate": 0.0013955139397165534, + "loss": 1.2702, + "step": 6789 + }, + { + "epoch": 0.5222675178832398, + "learning_rate": 0.0013951523596254466, + "loss": 1.4694, + "step": 6790 + }, + { + "epoch": 0.5223444350434582, + "learning_rate": 0.0013947907856565033, + "loss": 1.4483, + "step": 6791 + }, + { + "epoch": 0.5224213522036767, + "learning_rate": 0.0013944292178308352, + "loss": 1.2538, + "step": 6792 + }, + { + "epoch": 0.5224982693638951, + "learning_rate": 0.0013940676561695555, + "loss": 1.3387, + "step": 6793 + }, + { + "epoch": 0.5225751865241135, + "learning_rate": 0.0013937061006937756, + "loss": 1.1385, + "step": 6794 + }, + { + "epoch": 0.522652103684332, + "learning_rate": 0.001393344551424607, + "loss": 1.3387, + "step": 6795 + }, + { + "epoch": 0.5227290208445504, + "learning_rate": 0.0013929830083831616, + "loss": 0.9344, + "step": 6796 + }, + { + "epoch": 0.5228059380047688, + "learning_rate": 0.00139262147159055, + "loss": 1.1744, + "step": 6797 + }, + { + "epoch": 0.5228828551649873, + "learning_rate": 0.001392259941067882, + "loss": 0.8946, + "step": 6798 + }, + { + "epoch": 0.5229597723252057, + "learning_rate": 0.0013918984168362692, + "loss": 1.1324, + "step": 6799 + }, + { + "epoch": 0.5230366894854241, + "learning_rate": 0.00139153689891682, + "loss": 1.1671, + "step": 6800 + }, + { + "epoch": 0.5231136066456427, + "learning_rate": 0.0013911753873306438, + "loss": 1.1916, + "step": 6801 + }, + { + "epoch": 0.5231905238058611, + "learning_rate": 0.0013908138820988509, + "loss": 1.0871, + "step": 6802 + }, + { + "epoch": 0.5232674409660796, + "learning_rate": 0.0013904523832425487, + "loss": 1.4429, + "step": 6803 + }, + { + "epoch": 0.523344358126298, + "learning_rate": 0.0013900908907828462, + "loss": 1.2224, + "step": 6804 + }, + { + "epoch": 0.5234212752865164, + "learning_rate": 0.0013897294047408514, + "loss": 1.1442, + "step": 6805 + }, + { + "epoch": 0.5234981924467349, + "learning_rate": 0.001389367925137671, + "loss": 1.4412, + "step": 6806 + }, + { + "epoch": 0.5235751096069533, + "learning_rate": 0.0013890064519944134, + "loss": 1.1639, + "step": 6807 + }, + { + "epoch": 0.5236520267671717, + "learning_rate": 0.0013886449853321844, + "loss": 0.8159, + "step": 6808 + }, + { + "epoch": 0.5237289439273902, + "learning_rate": 0.0013882835251720906, + "loss": 1.1721, + "step": 6809 + }, + { + "epoch": 0.5238058610876086, + "learning_rate": 0.0013879220715352384, + "loss": 1.1325, + "step": 6810 + }, + { + "epoch": 0.5238827782478271, + "learning_rate": 0.001387560624442733, + "loss": 0.9891, + "step": 6811 + }, + { + "epoch": 0.5239596954080455, + "learning_rate": 0.0013871991839156804, + "loss": 1.3041, + "step": 6812 + }, + { + "epoch": 0.524036612568264, + "learning_rate": 0.0013868377499751852, + "loss": 1.0041, + "step": 6813 + }, + { + "epoch": 0.5241135297284825, + "learning_rate": 0.001386476322642351, + "loss": 1.5879, + "step": 6814 + }, + { + "epoch": 0.5241904468887009, + "learning_rate": 0.0013861149019382829, + "loss": 1.1554, + "step": 6815 + }, + { + "epoch": 0.5242673640489193, + "learning_rate": 0.0013857534878840845, + "loss": 1.5309, + "step": 6816 + }, + { + "epoch": 0.5243442812091378, + "learning_rate": 0.0013853920805008586, + "loss": 1.2539, + "step": 6817 + }, + { + "epoch": 0.5244211983693562, + "learning_rate": 0.0013850306798097092, + "loss": 1.0906, + "step": 6818 + }, + { + "epoch": 0.5244981155295746, + "learning_rate": 0.0013846692858317379, + "loss": 0.8813, + "step": 6819 + }, + { + "epoch": 0.5245750326897931, + "learning_rate": 0.0013843078985880468, + "loss": 0.9908, + "step": 6820 + }, + { + "epoch": 0.5246519498500115, + "learning_rate": 0.0013839465180997385, + "loss": 1.2657, + "step": 6821 + }, + { + "epoch": 0.52472886701023, + "learning_rate": 0.0013835851443879136, + "loss": 1.4675, + "step": 6822 + }, + { + "epoch": 0.5248057841704484, + "learning_rate": 0.0013832237774736738, + "loss": 1.2915, + "step": 6823 + }, + { + "epoch": 0.5248827013306668, + "learning_rate": 0.0013828624173781198, + "loss": 1.2075, + "step": 6824 + }, + { + "epoch": 0.5249596184908853, + "learning_rate": 0.0013825010641223502, + "loss": 1.3164, + "step": 6825 + }, + { + "epoch": 0.5250365356511038, + "learning_rate": 0.0013821397177274672, + "loss": 1.0839, + "step": 6826 + }, + { + "epoch": 0.5251134528113222, + "learning_rate": 0.0013817783782145681, + "loss": 1.4518, + "step": 6827 + }, + { + "epoch": 0.5251903699715407, + "learning_rate": 0.0013814170456047526, + "loss": 1.2361, + "step": 6828 + }, + { + "epoch": 0.5252672871317591, + "learning_rate": 0.0013810557199191197, + "loss": 1.4809, + "step": 6829 + }, + { + "epoch": 0.5253442042919776, + "learning_rate": 0.001380694401178767, + "loss": 1.3586, + "step": 6830 + }, + { + "epoch": 0.525421121452196, + "learning_rate": 0.001380333089404793, + "loss": 1.0907, + "step": 6831 + }, + { + "epoch": 0.5254980386124144, + "learning_rate": 0.0013799717846182944, + "loss": 1.0633, + "step": 6832 + }, + { + "epoch": 0.5255749557726329, + "learning_rate": 0.0013796104868403682, + "loss": 1.0793, + "step": 6833 + }, + { + "epoch": 0.5256518729328513, + "learning_rate": 0.0013792491960921114, + "loss": 1.1081, + "step": 6834 + }, + { + "epoch": 0.5257287900930697, + "learning_rate": 0.00137888791239462, + "loss": 0.8416, + "step": 6835 + }, + { + "epoch": 0.5258057072532882, + "learning_rate": 0.001378526635768989, + "loss": 1.1909, + "step": 6836 + }, + { + "epoch": 0.5258826244135066, + "learning_rate": 0.0013781653662363151, + "loss": 1.2034, + "step": 6837 + }, + { + "epoch": 0.5259595415737252, + "learning_rate": 0.0013778041038176915, + "loss": 1.0422, + "step": 6838 + }, + { + "epoch": 0.5260364587339436, + "learning_rate": 0.0013774428485342147, + "loss": 1.3183, + "step": 6839 + }, + { + "epoch": 0.526113375894162, + "learning_rate": 0.0013770816004069774, + "loss": 1.0863, + "step": 6840 + }, + { + "epoch": 0.5261902930543805, + "learning_rate": 0.001376720359457073, + "loss": 1.3833, + "step": 6841 + }, + { + "epoch": 0.5262672102145989, + "learning_rate": 0.0013763591257055958, + "loss": 1.0414, + "step": 6842 + }, + { + "epoch": 0.5263441273748173, + "learning_rate": 0.001375997899173638, + "loss": 1.3602, + "step": 6843 + }, + { + "epoch": 0.5264210445350358, + "learning_rate": 0.001375636679882292, + "loss": 1.2499, + "step": 6844 + }, + { + "epoch": 0.5264979616952542, + "learning_rate": 0.0013752754678526499, + "loss": 1.0176, + "step": 6845 + }, + { + "epoch": 0.5265748788554726, + "learning_rate": 0.0013749142631058035, + "loss": 1.4699, + "step": 6846 + }, + { + "epoch": 0.5266517960156911, + "learning_rate": 0.001374553065662843, + "loss": 1.3072, + "step": 6847 + }, + { + "epoch": 0.5267287131759095, + "learning_rate": 0.0013741918755448607, + "loss": 1.1194, + "step": 6848 + }, + { + "epoch": 0.526805630336128, + "learning_rate": 0.0013738306927729444, + "loss": 0.7833, + "step": 6849 + }, + { + "epoch": 0.5268825474963464, + "learning_rate": 0.001373469517368187, + "loss": 0.8987, + "step": 6850 + }, + { + "epoch": 0.5269594646565648, + "learning_rate": 0.0013731083493516757, + "loss": 1.1682, + "step": 6851 + }, + { + "epoch": 0.5270363818167834, + "learning_rate": 0.0013727471887445, + "loss": 1.3395, + "step": 6852 + }, + { + "epoch": 0.5271132989770018, + "learning_rate": 0.0013723860355677488, + "loss": 1.4319, + "step": 6853 + }, + { + "epoch": 0.5271902161372202, + "learning_rate": 0.00137202488984251, + "loss": 1.2318, + "step": 6854 + }, + { + "epoch": 0.5272671332974387, + "learning_rate": 0.0013716637515898708, + "loss": 1.2094, + "step": 6855 + }, + { + "epoch": 0.5273440504576571, + "learning_rate": 0.0013713026208309192, + "loss": 1.1864, + "step": 6856 + }, + { + "epoch": 0.5274209676178756, + "learning_rate": 0.0013709414975867416, + "loss": 1.1114, + "step": 6857 + }, + { + "epoch": 0.527497884778094, + "learning_rate": 0.0013705803818784246, + "loss": 1.0764, + "step": 6858 + }, + { + "epoch": 0.5275748019383124, + "learning_rate": 0.0013702192737270545, + "loss": 0.8763, + "step": 6859 + }, + { + "epoch": 0.5276517190985309, + "learning_rate": 0.0013698581731537152, + "loss": 1.092, + "step": 6860 + }, + { + "epoch": 0.5277286362587493, + "learning_rate": 0.0013694970801794938, + "loss": 1.1911, + "step": 6861 + }, + { + "epoch": 0.5278055534189677, + "learning_rate": 0.0013691359948254734, + "loss": 1.1457, + "step": 6862 + }, + { + "epoch": 0.5278824705791862, + "learning_rate": 0.0013687749171127385, + "loss": 1.11, + "step": 6863 + }, + { + "epoch": 0.5279593877394047, + "learning_rate": 0.0013684138470623731, + "loss": 1.1553, + "step": 6864 + }, + { + "epoch": 0.528036304899623, + "learning_rate": 0.0013680527846954605, + "loss": 1.1272, + "step": 6865 + }, + { + "epoch": 0.5281132220598416, + "learning_rate": 0.001367691730033083, + "loss": 1.3221, + "step": 6866 + }, + { + "epoch": 0.52819013922006, + "learning_rate": 0.0013673306830963234, + "loss": 0.8401, + "step": 6867 + }, + { + "epoch": 0.5282670563802785, + "learning_rate": 0.0013669696439062632, + "loss": 1.0266, + "step": 6868 + }, + { + "epoch": 0.5283439735404969, + "learning_rate": 0.0013666086124839844, + "loss": 1.0802, + "step": 6869 + }, + { + "epoch": 0.5284208907007153, + "learning_rate": 0.0013662475888505683, + "loss": 0.9563, + "step": 6870 + }, + { + "epoch": 0.5284978078609338, + "learning_rate": 0.0013658865730270936, + "loss": 1.1702, + "step": 6871 + }, + { + "epoch": 0.5285747250211522, + "learning_rate": 0.001365525565034643, + "loss": 1.075, + "step": 6872 + }, + { + "epoch": 0.5286516421813706, + "learning_rate": 0.0013651645648942942, + "loss": 1.3054, + "step": 6873 + }, + { + "epoch": 0.5287285593415891, + "learning_rate": 0.0013648035726271266, + "loss": 1.3343, + "step": 6874 + }, + { + "epoch": 0.5288054765018075, + "learning_rate": 0.0013644425882542199, + "loss": 1.1683, + "step": 6875 + }, + { + "epoch": 0.528882393662026, + "learning_rate": 0.0013640816117966514, + "loss": 0.9975, + "step": 6876 + }, + { + "epoch": 0.5289593108222445, + "learning_rate": 0.0013637206432754994, + "loss": 0.877, + "step": 6877 + }, + { + "epoch": 0.5290362279824629, + "learning_rate": 0.001363359682711841, + "loss": 1.0053, + "step": 6878 + }, + { + "epoch": 0.5291131451426814, + "learning_rate": 0.001362998730126753, + "loss": 1.2998, + "step": 6879 + }, + { + "epoch": 0.5291900623028998, + "learning_rate": 0.0013626377855413122, + "loss": 1.1646, + "step": 6880 + }, + { + "epoch": 0.5292669794631182, + "learning_rate": 0.0013622768489765945, + "loss": 1.3411, + "step": 6881 + }, + { + "epoch": 0.5293438966233367, + "learning_rate": 0.0013619159204536742, + "loss": 1.1243, + "step": 6882 + }, + { + "epoch": 0.5294208137835551, + "learning_rate": 0.0013615549999936281, + "loss": 1.1901, + "step": 6883 + }, + { + "epoch": 0.5294977309437735, + "learning_rate": 0.0013611940876175298, + "loss": 0.9948, + "step": 6884 + }, + { + "epoch": 0.529574648103992, + "learning_rate": 0.0013608331833464527, + "loss": 1.031, + "step": 6885 + }, + { + "epoch": 0.5296515652642104, + "learning_rate": 0.0013604722872014715, + "loss": 1.2082, + "step": 6886 + }, + { + "epoch": 0.5297284824244289, + "learning_rate": 0.0013601113992036586, + "loss": 1.1903, + "step": 6887 + }, + { + "epoch": 0.5298053995846473, + "learning_rate": 0.0013597505193740872, + "loss": 1.2638, + "step": 6888 + }, + { + "epoch": 0.5298823167448657, + "learning_rate": 0.0013593896477338292, + "loss": 1.3412, + "step": 6889 + }, + { + "epoch": 0.5299592339050843, + "learning_rate": 0.0013590287843039557, + "loss": 1.2244, + "step": 6890 + }, + { + "epoch": 0.5300361510653027, + "learning_rate": 0.0013586679291055389, + "loss": 1.0126, + "step": 6891 + }, + { + "epoch": 0.5301130682255211, + "learning_rate": 0.0013583070821596493, + "loss": 0.8857, + "step": 6892 + }, + { + "epoch": 0.5301899853857396, + "learning_rate": 0.0013579462434873558, + "loss": 0.9358, + "step": 6893 + }, + { + "epoch": 0.530266902545958, + "learning_rate": 0.0013575854131097301, + "loss": 1.2337, + "step": 6894 + }, + { + "epoch": 0.5303438197061765, + "learning_rate": 0.0013572245910478398, + "loss": 0.8602, + "step": 6895 + }, + { + "epoch": 0.5304207368663949, + "learning_rate": 0.0013568637773227552, + "loss": 1.2004, + "step": 6896 + }, + { + "epoch": 0.5304976540266133, + "learning_rate": 0.0013565029719555437, + "loss": 1.0215, + "step": 6897 + }, + { + "epoch": 0.5305745711868318, + "learning_rate": 0.0013561421749672727, + "loss": 1.0866, + "step": 6898 + }, + { + "epoch": 0.5306514883470502, + "learning_rate": 0.0013557813863790103, + "loss": 1.3846, + "step": 6899 + }, + { + "epoch": 0.5307284055072686, + "learning_rate": 0.0013554206062118233, + "loss": 1.0447, + "step": 6900 + }, + { + "epoch": 0.5308053226674871, + "learning_rate": 0.0013550598344867772, + "loss": 1.0576, + "step": 6901 + }, + { + "epoch": 0.5308822398277055, + "learning_rate": 0.001354699071224939, + "loss": 0.9923, + "step": 6902 + }, + { + "epoch": 0.530959156987924, + "learning_rate": 0.0013543383164473738, + "loss": 0.9693, + "step": 6903 + }, + { + "epoch": 0.5310360741481425, + "learning_rate": 0.0013539775701751449, + "loss": 0.7004, + "step": 6904 + }, + { + "epoch": 0.5311129913083609, + "learning_rate": 0.0013536168324293188, + "loss": 1.1429, + "step": 6905 + }, + { + "epoch": 0.5311899084685794, + "learning_rate": 0.0013532561032309576, + "loss": 0.8434, + "step": 6906 + }, + { + "epoch": 0.5312668256287978, + "learning_rate": 0.0013528953826011264, + "loss": 1.2499, + "step": 6907 + }, + { + "epoch": 0.5313437427890162, + "learning_rate": 0.0013525346705608868, + "loss": 1.3637, + "step": 6908 + }, + { + "epoch": 0.5314206599492347, + "learning_rate": 0.0013521739671313012, + "loss": 1.5843, + "step": 6909 + }, + { + "epoch": 0.5314975771094531, + "learning_rate": 0.001351813272333432, + "loss": 1.0776, + "step": 6910 + }, + { + "epoch": 0.5315744942696715, + "learning_rate": 0.0013514525861883402, + "loss": 1.336, + "step": 6911 + }, + { + "epoch": 0.53165141142989, + "learning_rate": 0.0013510919087170864, + "loss": 1.6208, + "step": 6912 + }, + { + "epoch": 0.5317283285901084, + "learning_rate": 0.0013507312399407315, + "loss": 1.1827, + "step": 6913 + }, + { + "epoch": 0.531805245750327, + "learning_rate": 0.001350370579880335, + "loss": 1.1826, + "step": 6914 + }, + { + "epoch": 0.5318821629105454, + "learning_rate": 0.0013500099285569563, + "loss": 1.0885, + "step": 6915 + }, + { + "epoch": 0.5319590800707638, + "learning_rate": 0.0013496492859916547, + "loss": 1.3705, + "step": 6916 + }, + { + "epoch": 0.5320359972309823, + "learning_rate": 0.0013492886522054868, + "loss": 1.2497, + "step": 6917 + }, + { + "epoch": 0.5321129143912007, + "learning_rate": 0.0013489280272195128, + "loss": 1.2999, + "step": 6918 + }, + { + "epoch": 0.5321898315514191, + "learning_rate": 0.0013485674110547882, + "loss": 0.7348, + "step": 6919 + }, + { + "epoch": 0.5322667487116376, + "learning_rate": 0.0013482068037323702, + "loss": 1.4769, + "step": 6920 + }, + { + "epoch": 0.532343665871856, + "learning_rate": 0.0013478462052733153, + "loss": 0.9537, + "step": 6921 + }, + { + "epoch": 0.5324205830320744, + "learning_rate": 0.0013474856156986792, + "loss": 1.2866, + "step": 6922 + }, + { + "epoch": 0.5324975001922929, + "learning_rate": 0.0013471250350295164, + "loss": 1.0531, + "step": 6923 + }, + { + "epoch": 0.5325744173525113, + "learning_rate": 0.0013467644632868825, + "loss": 1.0772, + "step": 6924 + }, + { + "epoch": 0.5326513345127298, + "learning_rate": 0.001346403900491831, + "loss": 1.0888, + "step": 6925 + }, + { + "epoch": 0.5327282516729482, + "learning_rate": 0.001346043346665416, + "loss": 1.1258, + "step": 6926 + }, + { + "epoch": 0.5328051688331666, + "learning_rate": 0.0013456828018286909, + "loss": 1.2859, + "step": 6927 + }, + { + "epoch": 0.5328820859933852, + "learning_rate": 0.001345322266002707, + "loss": 1.1934, + "step": 6928 + }, + { + "epoch": 0.5329590031536036, + "learning_rate": 0.001344961739208518, + "loss": 1.1608, + "step": 6929 + }, + { + "epoch": 0.533035920313822, + "learning_rate": 0.0013446012214671742, + "loss": 1.4867, + "step": 6930 + }, + { + "epoch": 0.5331128374740405, + "learning_rate": 0.0013442407127997268, + "loss": 0.9409, + "step": 6931 + }, + { + "epoch": 0.5331897546342589, + "learning_rate": 0.0013438802132272269, + "loss": 1.2508, + "step": 6932 + }, + { + "epoch": 0.5332666717944774, + "learning_rate": 0.0013435197227707237, + "loss": 1.0573, + "step": 6933 + }, + { + "epoch": 0.5333435889546958, + "learning_rate": 0.0013431592414512674, + "loss": 1.1462, + "step": 6934 + }, + { + "epoch": 0.5334205061149142, + "learning_rate": 0.0013427987692899062, + "loss": 1.4857, + "step": 6935 + }, + { + "epoch": 0.5334974232751327, + "learning_rate": 0.0013424383063076887, + "loss": 1.1281, + "step": 6936 + }, + { + "epoch": 0.5335743404353511, + "learning_rate": 0.0013420778525256626, + "loss": 1.1199, + "step": 6937 + }, + { + "epoch": 0.5336512575955695, + "learning_rate": 0.001341717407964876, + "loss": 0.9395, + "step": 6938 + }, + { + "epoch": 0.533728174755788, + "learning_rate": 0.0013413569726463735, + "loss": 1.01, + "step": 6939 + }, + { + "epoch": 0.5338050919160064, + "learning_rate": 0.001340996546591204, + "loss": 1.0332, + "step": 6940 + }, + { + "epoch": 0.5338820090762249, + "learning_rate": 0.0013406361298204114, + "loss": 1.2737, + "step": 6941 + }, + { + "epoch": 0.5339589262364434, + "learning_rate": 0.0013402757223550412, + "loss": 1.146, + "step": 6942 + }, + { + "epoch": 0.5340358433966618, + "learning_rate": 0.001339915324216138, + "loss": 0.9185, + "step": 6943 + }, + { + "epoch": 0.5341127605568803, + "learning_rate": 0.0013395549354247459, + "loss": 1.1744, + "step": 6944 + }, + { + "epoch": 0.5341896777170987, + "learning_rate": 0.0013391945560019083, + "loss": 0.9265, + "step": 6945 + }, + { + "epoch": 0.5342665948773171, + "learning_rate": 0.0013388341859686683, + "loss": 1.1616, + "step": 6946 + }, + { + "epoch": 0.5343435120375356, + "learning_rate": 0.001338473825346068, + "loss": 0.9231, + "step": 6947 + }, + { + "epoch": 0.534420429197754, + "learning_rate": 0.0013381134741551493, + "loss": 1.2317, + "step": 6948 + }, + { + "epoch": 0.5344973463579724, + "learning_rate": 0.0013377531324169538, + "loss": 1.4175, + "step": 6949 + }, + { + "epoch": 0.5345742635181909, + "learning_rate": 0.0013373928001525217, + "loss": 0.8812, + "step": 6950 + }, + { + "epoch": 0.5346511806784093, + "learning_rate": 0.001337032477382894, + "loss": 1.2232, + "step": 6951 + }, + { + "epoch": 0.5347280978386278, + "learning_rate": 0.0013366721641291089, + "loss": 1.4312, + "step": 6952 + }, + { + "epoch": 0.5348050149988463, + "learning_rate": 0.0013363118604122074, + "loss": 1.1784, + "step": 6953 + }, + { + "epoch": 0.5348819321590647, + "learning_rate": 0.0013359515662532265, + "loss": 1.0246, + "step": 6954 + }, + { + "epoch": 0.5349588493192832, + "learning_rate": 0.0013355912816732044, + "loss": 0.789, + "step": 6955 + }, + { + "epoch": 0.5350357664795016, + "learning_rate": 0.0013352310066931789, + "loss": 1.348, + "step": 6956 + }, + { + "epoch": 0.53511268363972, + "learning_rate": 0.0013348707413341867, + "loss": 1.016, + "step": 6957 + }, + { + "epoch": 0.5351896007999385, + "learning_rate": 0.0013345104856172635, + "loss": 1.295, + "step": 6958 + }, + { + "epoch": 0.5352665179601569, + "learning_rate": 0.0013341502395634462, + "loss": 1.066, + "step": 6959 + }, + { + "epoch": 0.5353434351203754, + "learning_rate": 0.0013337900031937687, + "loss": 1.0589, + "step": 6960 + }, + { + "epoch": 0.5354203522805938, + "learning_rate": 0.0013334297765292666, + "loss": 1.19, + "step": 6961 + }, + { + "epoch": 0.5354972694408122, + "learning_rate": 0.0013330695595909733, + "loss": 0.81, + "step": 6962 + }, + { + "epoch": 0.5355741866010307, + "learning_rate": 0.001332709352399922, + "loss": 1.1576, + "step": 6963 + }, + { + "epoch": 0.5356511037612491, + "learning_rate": 0.0013323491549771464, + "loss": 1.3964, + "step": 6964 + }, + { + "epoch": 0.5357280209214675, + "learning_rate": 0.0013319889673436788, + "loss": 1.2666, + "step": 6965 + }, + { + "epoch": 0.535804938081686, + "learning_rate": 0.0013316287895205497, + "loss": 1.2331, + "step": 6966 + }, + { + "epoch": 0.5358818552419045, + "learning_rate": 0.0013312686215287914, + "loss": 1.0216, + "step": 6967 + }, + { + "epoch": 0.5359587724021229, + "learning_rate": 0.001330908463389434, + "loss": 1.2304, + "step": 6968 + }, + { + "epoch": 0.5360356895623414, + "learning_rate": 0.0013305483151235076, + "loss": 1.0852, + "step": 6969 + }, + { + "epoch": 0.5361126067225598, + "learning_rate": 0.001330188176752042, + "loss": 0.8415, + "step": 6970 + }, + { + "epoch": 0.5361895238827783, + "learning_rate": 0.0013298280482960652, + "loss": 1.28, + "step": 6971 + }, + { + "epoch": 0.5362664410429967, + "learning_rate": 0.0013294679297766062, + "loss": 1.2948, + "step": 6972 + }, + { + "epoch": 0.5363433582032151, + "learning_rate": 0.0013291078212146928, + "loss": 0.992, + "step": 6973 + }, + { + "epoch": 0.5364202753634336, + "learning_rate": 0.0013287477226313514, + "loss": 0.9287, + "step": 6974 + }, + { + "epoch": 0.536497192523652, + "learning_rate": 0.0013283876340476094, + "loss": 1.3024, + "step": 6975 + }, + { + "epoch": 0.5365741096838704, + "learning_rate": 0.0013280275554844925, + "loss": 1.1904, + "step": 6976 + }, + { + "epoch": 0.5366510268440889, + "learning_rate": 0.0013276674869630248, + "loss": 1.2229, + "step": 6977 + }, + { + "epoch": 0.5367279440043073, + "learning_rate": 0.0013273074285042332, + "loss": 1.329, + "step": 6978 + }, + { + "epoch": 0.5368048611645259, + "learning_rate": 0.0013269473801291402, + "loss": 1.1713, + "step": 6979 + }, + { + "epoch": 0.5368817783247443, + "learning_rate": 0.0013265873418587702, + "loss": 1.2024, + "step": 6980 + }, + { + "epoch": 0.5369586954849627, + "learning_rate": 0.0013262273137141464, + "loss": 1.3499, + "step": 6981 + }, + { + "epoch": 0.5370356126451812, + "learning_rate": 0.00132586729571629, + "loss": 1.3022, + "step": 6982 + }, + { + "epoch": 0.5371125298053996, + "learning_rate": 0.0013255072878862245, + "loss": 1.024, + "step": 6983 + }, + { + "epoch": 0.537189446965618, + "learning_rate": 0.00132514729024497, + "loss": 1.078, + "step": 6984 + }, + { + "epoch": 0.5372663641258365, + "learning_rate": 0.0013247873028135473, + "loss": 1.61, + "step": 6985 + }, + { + "epoch": 0.5373432812860549, + "learning_rate": 0.0013244273256129767, + "loss": 1.3039, + "step": 6986 + }, + { + "epoch": 0.5374201984462733, + "learning_rate": 0.0013240673586642781, + "loss": 0.8892, + "step": 6987 + }, + { + "epoch": 0.5374971156064918, + "learning_rate": 0.0013237074019884687, + "loss": 1.1584, + "step": 6988 + }, + { + "epoch": 0.5375740327667102, + "learning_rate": 0.0013233474556065689, + "loss": 1.2637, + "step": 6989 + }, + { + "epoch": 0.5376509499269287, + "learning_rate": 0.0013229875195395941, + "loss": 0.8415, + "step": 6990 + }, + { + "epoch": 0.5377278670871471, + "learning_rate": 0.0013226275938085638, + "loss": 1.1912, + "step": 6991 + }, + { + "epoch": 0.5378047842473656, + "learning_rate": 0.0013222676784344924, + "loss": 1.0295, + "step": 6992 + }, + { + "epoch": 0.5378817014075841, + "learning_rate": 0.0013219077734383964, + "loss": 1.2195, + "step": 6993 + }, + { + "epoch": 0.5379586185678025, + "learning_rate": 0.0013215478788412916, + "loss": 1.1863, + "step": 6994 + }, + { + "epoch": 0.5380355357280209, + "learning_rate": 0.0013211879946641919, + "loss": 1.0444, + "step": 6995 + }, + { + "epoch": 0.5381124528882394, + "learning_rate": 0.0013208281209281113, + "loss": 1.185, + "step": 6996 + }, + { + "epoch": 0.5381893700484578, + "learning_rate": 0.0013204682576540637, + "loss": 1.2222, + "step": 6997 + }, + { + "epoch": 0.5382662872086763, + "learning_rate": 0.0013201084048630614, + "loss": 1.0717, + "step": 6998 + }, + { + "epoch": 0.5383432043688947, + "learning_rate": 0.001319748562576117, + "loss": 0.9739, + "step": 6999 + }, + { + "epoch": 0.5384201215291131, + "learning_rate": 0.0013193887308142424, + "loss": 1.1365, + "step": 7000 + }, + { + "epoch": 0.5384970386893316, + "learning_rate": 0.0013190289095984468, + "loss": 1.292, + "step": 7001 + }, + { + "epoch": 0.53857395584955, + "learning_rate": 0.001318669098949743, + "loss": 1.0203, + "step": 7002 + }, + { + "epoch": 0.5386508730097684, + "learning_rate": 0.001318309298889139, + "loss": 0.9898, + "step": 7003 + }, + { + "epoch": 0.538727790169987, + "learning_rate": 0.001317949509437644, + "loss": 1.2468, + "step": 7004 + }, + { + "epoch": 0.5388047073302054, + "learning_rate": 0.0013175897306162677, + "loss": 1.4557, + "step": 7005 + }, + { + "epoch": 0.5388816244904238, + "learning_rate": 0.0013172299624460166, + "loss": 0.8888, + "step": 7006 + }, + { + "epoch": 0.5389585416506423, + "learning_rate": 0.0013168702049478984, + "loss": 0.9741, + "step": 7007 + }, + { + "epoch": 0.5390354588108607, + "learning_rate": 0.00131651045814292, + "loss": 1.2969, + "step": 7008 + }, + { + "epoch": 0.5391123759710792, + "learning_rate": 0.0013161507220520868, + "loss": 1.4163, + "step": 7009 + }, + { + "epoch": 0.5391892931312976, + "learning_rate": 0.001315790996696405, + "loss": 1.3695, + "step": 7010 + }, + { + "epoch": 0.539266210291516, + "learning_rate": 0.0013154312820968792, + "loss": 1.3334, + "step": 7011 + }, + { + "epoch": 0.5393431274517345, + "learning_rate": 0.0013150715782745124, + "loss": 1.105, + "step": 7012 + }, + { + "epoch": 0.5394200446119529, + "learning_rate": 0.0013147118852503093, + "loss": 1.2386, + "step": 7013 + }, + { + "epoch": 0.5394969617721713, + "learning_rate": 0.0013143522030452726, + "loss": 1.2102, + "step": 7014 + }, + { + "epoch": 0.5395738789323898, + "learning_rate": 0.0013139925316804034, + "loss": 0.9835, + "step": 7015 + }, + { + "epoch": 0.5396507960926082, + "learning_rate": 0.001313632871176705, + "loss": 1.3669, + "step": 7016 + }, + { + "epoch": 0.5397277132528268, + "learning_rate": 0.0013132732215551768, + "loss": 1.003, + "step": 7017 + }, + { + "epoch": 0.5398046304130452, + "learning_rate": 0.00131291358283682, + "loss": 1.1371, + "step": 7018 + }, + { + "epoch": 0.5398815475732636, + "learning_rate": 0.0013125539550426343, + "loss": 1.3373, + "step": 7019 + }, + { + "epoch": 0.5399584647334821, + "learning_rate": 0.0013121943381936181, + "loss": 1.4093, + "step": 7020 + }, + { + "epoch": 0.5400353818937005, + "learning_rate": 0.0013118347323107706, + "loss": 1.2889, + "step": 7021 + }, + { + "epoch": 0.5401122990539189, + "learning_rate": 0.0013114751374150894, + "loss": 1.3281, + "step": 7022 + }, + { + "epoch": 0.5401892162141374, + "learning_rate": 0.0013111155535275706, + "loss": 1.0294, + "step": 7023 + }, + { + "epoch": 0.5402661333743558, + "learning_rate": 0.0013107559806692123, + "loss": 0.8113, + "step": 7024 + }, + { + "epoch": 0.5403430505345742, + "learning_rate": 0.0013103964188610092, + "loss": 1.1516, + "step": 7025 + }, + { + "epoch": 0.5404199676947927, + "learning_rate": 0.0013100368681239567, + "loss": 0.7669, + "step": 7026 + }, + { + "epoch": 0.5404968848550111, + "learning_rate": 0.0013096773284790495, + "loss": 0.9348, + "step": 7027 + }, + { + "epoch": 0.5405738020152296, + "learning_rate": 0.0013093177999472813, + "loss": 1.3785, + "step": 7028 + }, + { + "epoch": 0.540650719175448, + "learning_rate": 0.0013089582825496456, + "loss": 0.9523, + "step": 7029 + }, + { + "epoch": 0.5407276363356665, + "learning_rate": 0.0013085987763071351, + "loss": 1.2283, + "step": 7030 + }, + { + "epoch": 0.540804553495885, + "learning_rate": 0.001308239281240741, + "loss": 1.0576, + "step": 7031 + }, + { + "epoch": 0.5408814706561034, + "learning_rate": 0.0013078797973714558, + "loss": 1.2732, + "step": 7032 + }, + { + "epoch": 0.5409583878163218, + "learning_rate": 0.0013075203247202693, + "loss": 0.8747, + "step": 7033 + }, + { + "epoch": 0.5410353049765403, + "learning_rate": 0.001307160863308171, + "loss": 0.8084, + "step": 7034 + }, + { + "epoch": 0.5411122221367587, + "learning_rate": 0.0013068014131561517, + "loss": 1.3837, + "step": 7035 + }, + { + "epoch": 0.5411891392969772, + "learning_rate": 0.0013064419742851983, + "loss": 1.02, + "step": 7036 + }, + { + "epoch": 0.5412660564571956, + "learning_rate": 0.0013060825467163008, + "loss": 0.8834, + "step": 7037 + }, + { + "epoch": 0.541342973617414, + "learning_rate": 0.0013057231304704452, + "loss": 1.6106, + "step": 7038 + }, + { + "epoch": 0.5414198907776325, + "learning_rate": 0.001305363725568618, + "loss": 1.4662, + "step": 7039 + }, + { + "epoch": 0.5414968079378509, + "learning_rate": 0.001305004332031806, + "loss": 1.2767, + "step": 7040 + }, + { + "epoch": 0.5415737250980693, + "learning_rate": 0.0013046449498809942, + "loss": 1.1696, + "step": 7041 + }, + { + "epoch": 0.5416506422582879, + "learning_rate": 0.0013042855791371671, + "loss": 1.274, + "step": 7042 + }, + { + "epoch": 0.5417275594185063, + "learning_rate": 0.0013039262198213095, + "loss": 1.2978, + "step": 7043 + }, + { + "epoch": 0.5418044765787247, + "learning_rate": 0.0013035668719544047, + "loss": 1.1238, + "step": 7044 + }, + { + "epoch": 0.5418813937389432, + "learning_rate": 0.0013032075355574337, + "loss": 1.4077, + "step": 7045 + }, + { + "epoch": 0.5419583108991616, + "learning_rate": 0.0013028482106513808, + "loss": 1.2791, + "step": 7046 + }, + { + "epoch": 0.5420352280593801, + "learning_rate": 0.0013024888972572254, + "loss": 1.475, + "step": 7047 + }, + { + "epoch": 0.5421121452195985, + "learning_rate": 0.0013021295953959502, + "loss": 1.2005, + "step": 7048 + }, + { + "epoch": 0.5421890623798169, + "learning_rate": 0.0013017703050885337, + "loss": 1.0672, + "step": 7049 + }, + { + "epoch": 0.5422659795400354, + "learning_rate": 0.0013014110263559555, + "loss": 1.0102, + "step": 7050 + }, + { + "epoch": 0.5423428967002538, + "learning_rate": 0.0013010517592191946, + "loss": 1.1015, + "step": 7051 + }, + { + "epoch": 0.5424198138604722, + "learning_rate": 0.0013006925036992288, + "loss": 1.1986, + "step": 7052 + }, + { + "epoch": 0.5424967310206907, + "learning_rate": 0.0013003332598170352, + "loss": 1.0695, + "step": 7053 + }, + { + "epoch": 0.5425736481809091, + "learning_rate": 0.001299974027593591, + "loss": 1.3769, + "step": 7054 + }, + { + "epoch": 0.5426505653411277, + "learning_rate": 0.0012996148070498715, + "loss": 0.9897, + "step": 7055 + }, + { + "epoch": 0.5427274825013461, + "learning_rate": 0.0012992555982068526, + "loss": 1.3528, + "step": 7056 + }, + { + "epoch": 0.5428043996615645, + "learning_rate": 0.0012988964010855087, + "loss": 1.4178, + "step": 7057 + }, + { + "epoch": 0.542881316821783, + "learning_rate": 0.0012985372157068124, + "loss": 1.0812, + "step": 7058 + }, + { + "epoch": 0.5429582339820014, + "learning_rate": 0.0012981780420917393, + "loss": 0.9687, + "step": 7059 + }, + { + "epoch": 0.5430351511422198, + "learning_rate": 0.0012978188802612604, + "loss": 1.2151, + "step": 7060 + }, + { + "epoch": 0.5431120683024383, + "learning_rate": 0.0012974597302363472, + "loss": 0.9797, + "step": 7061 + }, + { + "epoch": 0.5431889854626567, + "learning_rate": 0.0012971005920379718, + "loss": 1.2642, + "step": 7062 + }, + { + "epoch": 0.5432659026228752, + "learning_rate": 0.001296741465687104, + "loss": 1.4976, + "step": 7063 + }, + { + "epoch": 0.5433428197830936, + "learning_rate": 0.0012963823512047142, + "loss": 1.1779, + "step": 7064 + }, + { + "epoch": 0.543419736943312, + "learning_rate": 0.0012960232486117712, + "loss": 1.2746, + "step": 7065 + }, + { + "epoch": 0.5434966541035305, + "learning_rate": 0.0012956641579292428, + "loss": 1.263, + "step": 7066 + }, + { + "epoch": 0.543573571263749, + "learning_rate": 0.0012953050791780977, + "loss": 1.3056, + "step": 7067 + }, + { + "epoch": 0.5436504884239673, + "learning_rate": 0.0012949460123793026, + "loss": 1.1749, + "step": 7068 + }, + { + "epoch": 0.5437274055841859, + "learning_rate": 0.0012945869575538226, + "loss": 1.3322, + "step": 7069 + }, + { + "epoch": 0.5438043227444043, + "learning_rate": 0.0012942279147226251, + "loss": 1.2806, + "step": 7070 + }, + { + "epoch": 0.5438812399046227, + "learning_rate": 0.0012938688839066739, + "loss": 1.1828, + "step": 7071 + }, + { + "epoch": 0.5439581570648412, + "learning_rate": 0.0012935098651269327, + "loss": 1.3037, + "step": 7072 + }, + { + "epoch": 0.5440350742250596, + "learning_rate": 0.0012931508584043663, + "loss": 1.0784, + "step": 7073 + }, + { + "epoch": 0.5441119913852781, + "learning_rate": 0.0012927918637599368, + "loss": 1.3263, + "step": 7074 + }, + { + "epoch": 0.5441889085454965, + "learning_rate": 0.001292432881214606, + "loss": 1.3612, + "step": 7075 + }, + { + "epoch": 0.5442658257057149, + "learning_rate": 0.001292073910789336, + "loss": 1.1196, + "step": 7076 + }, + { + "epoch": 0.5443427428659334, + "learning_rate": 0.0012917149525050868, + "loss": 0.9935, + "step": 7077 + }, + { + "epoch": 0.5444196600261518, + "learning_rate": 0.0012913560063828188, + "loss": 1.1576, + "step": 7078 + }, + { + "epoch": 0.5444965771863702, + "learning_rate": 0.0012909970724434914, + "loss": 0.9746, + "step": 7079 + }, + { + "epoch": 0.5445734943465887, + "learning_rate": 0.0012906381507080617, + "loss": 1.3635, + "step": 7080 + }, + { + "epoch": 0.5446504115068072, + "learning_rate": 0.0012902792411974896, + "loss": 1.0496, + "step": 7081 + }, + { + "epoch": 0.5447273286670257, + "learning_rate": 0.0012899203439327301, + "loss": 1.1322, + "step": 7082 + }, + { + "epoch": 0.5448042458272441, + "learning_rate": 0.0012895614589347416, + "loss": 0.9644, + "step": 7083 + }, + { + "epoch": 0.5448811629874625, + "learning_rate": 0.0012892025862244786, + "loss": 1.4189, + "step": 7084 + }, + { + "epoch": 0.544958080147681, + "learning_rate": 0.0012888437258228962, + "loss": 1.4761, + "step": 7085 + }, + { + "epoch": 0.5450349973078994, + "learning_rate": 0.0012884848777509486, + "loss": 1.3381, + "step": 7086 + }, + { + "epoch": 0.5451119144681178, + "learning_rate": 0.0012881260420295896, + "loss": 1.1425, + "step": 7087 + }, + { + "epoch": 0.5451888316283363, + "learning_rate": 0.0012877672186797715, + "loss": 1.0682, + "step": 7088 + }, + { + "epoch": 0.5452657487885547, + "learning_rate": 0.001287408407722447, + "loss": 1.4989, + "step": 7089 + }, + { + "epoch": 0.5453426659487731, + "learning_rate": 0.0012870496091785676, + "loss": 1.2307, + "step": 7090 + }, + { + "epoch": 0.5454195831089916, + "learning_rate": 0.0012866908230690823, + "loss": 1.219, + "step": 7091 + }, + { + "epoch": 0.54549650026921, + "learning_rate": 0.0012863320494149431, + "loss": 1.0517, + "step": 7092 + }, + { + "epoch": 0.5455734174294286, + "learning_rate": 0.0012859732882370973, + "loss": 0.8331, + "step": 7093 + }, + { + "epoch": 0.545650334589647, + "learning_rate": 0.0012856145395564952, + "loss": 1.5791, + "step": 7094 + }, + { + "epoch": 0.5457272517498654, + "learning_rate": 0.0012852558033940834, + "loss": 0.8782, + "step": 7095 + }, + { + "epoch": 0.5458041689100839, + "learning_rate": 0.0012848970797708086, + "loss": 1.0444, + "step": 7096 + }, + { + "epoch": 0.5458810860703023, + "learning_rate": 0.0012845383687076176, + "loss": 1.0185, + "step": 7097 + }, + { + "epoch": 0.5459580032305207, + "learning_rate": 0.0012841796702254562, + "loss": 1.4756, + "step": 7098 + }, + { + "epoch": 0.5460349203907392, + "learning_rate": 0.0012838209843452682, + "loss": 1.3353, + "step": 7099 + }, + { + "epoch": 0.5461118375509576, + "learning_rate": 0.0012834623110879986, + "loss": 1.3954, + "step": 7100 + }, + { + "epoch": 0.5461887547111761, + "learning_rate": 0.0012831036504745898, + "loss": 1.1106, + "step": 7101 + }, + { + "epoch": 0.5462656718713945, + "learning_rate": 0.0012827450025259858, + "loss": 0.9238, + "step": 7102 + }, + { + "epoch": 0.5463425890316129, + "learning_rate": 0.001282386367263127, + "loss": 1.2487, + "step": 7103 + }, + { + "epoch": 0.5464195061918314, + "learning_rate": 0.0012820277447069548, + "loss": 1.0021, + "step": 7104 + }, + { + "epoch": 0.5464964233520498, + "learning_rate": 0.0012816691348784103, + "loss": 1.3208, + "step": 7105 + }, + { + "epoch": 0.5465733405122682, + "learning_rate": 0.0012813105377984324, + "loss": 1.0015, + "step": 7106 + }, + { + "epoch": 0.5466502576724868, + "learning_rate": 0.0012809519534879598, + "loss": 0.9199, + "step": 7107 + }, + { + "epoch": 0.5467271748327052, + "learning_rate": 0.0012805933819679313, + "loss": 0.8717, + "step": 7108 + }, + { + "epoch": 0.5468040919929236, + "learning_rate": 0.0012802348232592835, + "loss": 0.9197, + "step": 7109 + }, + { + "epoch": 0.5468810091531421, + "learning_rate": 0.0012798762773829535, + "loss": 0.992, + "step": 7110 + }, + { + "epoch": 0.5469579263133605, + "learning_rate": 0.001279517744359877, + "loss": 1.1453, + "step": 7111 + }, + { + "epoch": 0.547034843473579, + "learning_rate": 0.0012791592242109894, + "loss": 0.9082, + "step": 7112 + }, + { + "epoch": 0.5471117606337974, + "learning_rate": 0.0012788007169572248, + "loss": 1.2377, + "step": 7113 + }, + { + "epoch": 0.5471886777940158, + "learning_rate": 0.001278442222619517, + "loss": 0.9316, + "step": 7114 + }, + { + "epoch": 0.5472655949542343, + "learning_rate": 0.0012780837412187983, + "loss": 1.4862, + "step": 7115 + }, + { + "epoch": 0.5473425121144527, + "learning_rate": 0.0012777252727760016, + "loss": 1.0822, + "step": 7116 + }, + { + "epoch": 0.5474194292746711, + "learning_rate": 0.0012773668173120581, + "loss": 1.1401, + "step": 7117 + }, + { + "epoch": 0.5474963464348896, + "learning_rate": 0.0012770083748478978, + "loss": 1.0458, + "step": 7118 + }, + { + "epoch": 0.547573263595108, + "learning_rate": 0.0012766499454044508, + "loss": 1.1473, + "step": 7119 + }, + { + "epoch": 0.5476501807553266, + "learning_rate": 0.0012762915290026463, + "loss": 1.023, + "step": 7120 + }, + { + "epoch": 0.547727097915545, + "learning_rate": 0.0012759331256634128, + "loss": 1.2007, + "step": 7121 + }, + { + "epoch": 0.5478040150757634, + "learning_rate": 0.0012755747354076775, + "loss": 1.1641, + "step": 7122 + }, + { + "epoch": 0.5478809322359819, + "learning_rate": 0.0012752163582563669, + "loss": 1.3878, + "step": 7123 + }, + { + "epoch": 0.5479578493962003, + "learning_rate": 0.0012748579942304078, + "loss": 1.1641, + "step": 7124 + }, + { + "epoch": 0.5480347665564187, + "learning_rate": 0.0012744996433507252, + "loss": 1.1184, + "step": 7125 + }, + { + "epoch": 0.5481116837166372, + "learning_rate": 0.0012741413056382431, + "loss": 1.3532, + "step": 7126 + }, + { + "epoch": 0.5481886008768556, + "learning_rate": 0.0012737829811138857, + "loss": 0.9796, + "step": 7127 + }, + { + "epoch": 0.548265518037074, + "learning_rate": 0.0012734246697985762, + "loss": 1.048, + "step": 7128 + }, + { + "epoch": 0.5483424351972925, + "learning_rate": 0.0012730663717132355, + "loss": 1.1809, + "step": 7129 + }, + { + "epoch": 0.5484193523575109, + "learning_rate": 0.0012727080868787867, + "loss": 1.5304, + "step": 7130 + }, + { + "epoch": 0.5484962695177295, + "learning_rate": 0.0012723498153161487, + "loss": 0.6619, + "step": 7131 + }, + { + "epoch": 0.5485731866779479, + "learning_rate": 0.001271991557046243, + "loss": 1.0458, + "step": 7132 + }, + { + "epoch": 0.5486501038381663, + "learning_rate": 0.0012716333120899876, + "loss": 1.1986, + "step": 7133 + }, + { + "epoch": 0.5487270209983848, + "learning_rate": 0.0012712750804683012, + "loss": 1.0602, + "step": 7134 + }, + { + "epoch": 0.5488039381586032, + "learning_rate": 0.0012709168622021011, + "loss": 0.8825, + "step": 7135 + }, + { + "epoch": 0.5488808553188216, + "learning_rate": 0.0012705586573123044, + "loss": 1.0024, + "step": 7136 + }, + { + "epoch": 0.5489577724790401, + "learning_rate": 0.0012702004658198263, + "loss": 1.1447, + "step": 7137 + }, + { + "epoch": 0.5490346896392585, + "learning_rate": 0.001269842287745583, + "loss": 1.1722, + "step": 7138 + }, + { + "epoch": 0.549111606799477, + "learning_rate": 0.001269484123110488, + "loss": 1.1509, + "step": 7139 + }, + { + "epoch": 0.5491885239596954, + "learning_rate": 0.0012691259719354554, + "loss": 1.1132, + "step": 7140 + }, + { + "epoch": 0.5492654411199138, + "learning_rate": 0.0012687678342413985, + "loss": 1.3512, + "step": 7141 + }, + { + "epoch": 0.5493423582801323, + "learning_rate": 0.0012684097100492278, + "loss": 1.3356, + "step": 7142 + }, + { + "epoch": 0.5494192754403507, + "learning_rate": 0.0012680515993798564, + "loss": 1.1692, + "step": 7143 + }, + { + "epoch": 0.5494961926005691, + "learning_rate": 0.0012676935022541934, + "loss": 1.1104, + "step": 7144 + }, + { + "epoch": 0.5495731097607877, + "learning_rate": 0.0012673354186931488, + "loss": 1.212, + "step": 7145 + }, + { + "epoch": 0.5496500269210061, + "learning_rate": 0.0012669773487176317, + "loss": 1.2503, + "step": 7146 + }, + { + "epoch": 0.5497269440812245, + "learning_rate": 0.0012666192923485502, + "loss": 1.2845, + "step": 7147 + }, + { + "epoch": 0.549803861241443, + "learning_rate": 0.0012662612496068108, + "loss": 1.7123, + "step": 7148 + }, + { + "epoch": 0.5498807784016614, + "learning_rate": 0.0012659032205133212, + "loss": 1.2454, + "step": 7149 + }, + { + "epoch": 0.5499576955618799, + "learning_rate": 0.0012655452050889862, + "loss": 1.0506, + "step": 7150 + }, + { + "epoch": 0.5500346127220983, + "learning_rate": 0.0012651872033547114, + "loss": 1.3297, + "step": 7151 + }, + { + "epoch": 0.5501115298823167, + "learning_rate": 0.0012648292153314008, + "loss": 1.1001, + "step": 7152 + }, + { + "epoch": 0.5501884470425352, + "learning_rate": 0.0012644712410399565, + "loss": 1.0706, + "step": 7153 + }, + { + "epoch": 0.5502653642027536, + "learning_rate": 0.0012641132805012827, + "loss": 1.1781, + "step": 7154 + }, + { + "epoch": 0.550342281362972, + "learning_rate": 0.00126375533373628, + "loss": 1.3003, + "step": 7155 + }, + { + "epoch": 0.5504191985231905, + "learning_rate": 0.0012633974007658494, + "loss": 1.2581, + "step": 7156 + }, + { + "epoch": 0.550496115683409, + "learning_rate": 0.0012630394816108912, + "loss": 1.2777, + "step": 7157 + }, + { + "epoch": 0.5505730328436275, + "learning_rate": 0.0012626815762923045, + "loss": 0.9019, + "step": 7158 + }, + { + "epoch": 0.5506499500038459, + "learning_rate": 0.0012623236848309882, + "loss": 1.0706, + "step": 7159 + }, + { + "epoch": 0.5507268671640643, + "learning_rate": 0.0012619658072478396, + "loss": 1.2874, + "step": 7160 + }, + { + "epoch": 0.5508037843242828, + "learning_rate": 0.0012616079435637554, + "loss": 1.0621, + "step": 7161 + }, + { + "epoch": 0.5508807014845012, + "learning_rate": 0.0012612500937996321, + "loss": 1.3226, + "step": 7162 + }, + { + "epoch": 0.5509576186447196, + "learning_rate": 0.0012608922579763649, + "loss": 1.4676, + "step": 7163 + }, + { + "epoch": 0.5510345358049381, + "learning_rate": 0.0012605344361148467, + "loss": 1.2182, + "step": 7164 + }, + { + "epoch": 0.5511114529651565, + "learning_rate": 0.0012601766282359737, + "loss": 1.0978, + "step": 7165 + }, + { + "epoch": 0.5511883701253749, + "learning_rate": 0.001259818834360637, + "loss": 1.3387, + "step": 7166 + }, + { + "epoch": 0.5512652872855934, + "learning_rate": 0.0012594610545097282, + "loss": 0.8153, + "step": 7167 + }, + { + "epoch": 0.5513422044458118, + "learning_rate": 0.0012591032887041397, + "loss": 1.6278, + "step": 7168 + }, + { + "epoch": 0.5514191216060303, + "learning_rate": 0.001258745536964761, + "loss": 1.0743, + "step": 7169 + }, + { + "epoch": 0.5514960387662488, + "learning_rate": 0.001258387799312482, + "loss": 1.4928, + "step": 7170 + }, + { + "epoch": 0.5515729559264672, + "learning_rate": 0.0012580300757681912, + "loss": 0.801, + "step": 7171 + }, + { + "epoch": 0.5516498730866857, + "learning_rate": 0.001257672366352776, + "loss": 1.5612, + "step": 7172 + }, + { + "epoch": 0.5517267902469041, + "learning_rate": 0.0012573146710871245, + "loss": 0.7027, + "step": 7173 + }, + { + "epoch": 0.5518037074071225, + "learning_rate": 0.0012569569899921224, + "loss": 1.6275, + "step": 7174 + }, + { + "epoch": 0.551880624567341, + "learning_rate": 0.001256599323088654, + "loss": 1.2391, + "step": 7175 + }, + { + "epoch": 0.5519575417275594, + "learning_rate": 0.0012562416703976059, + "loss": 1.2503, + "step": 7176 + }, + { + "epoch": 0.5520344588877779, + "learning_rate": 0.0012558840319398595, + "loss": 0.9731, + "step": 7177 + }, + { + "epoch": 0.5521113760479963, + "learning_rate": 0.0012555264077363, + "loss": 0.9332, + "step": 7178 + }, + { + "epoch": 0.5521882932082147, + "learning_rate": 0.001255168797807808, + "loss": 0.7701, + "step": 7179 + }, + { + "epoch": 0.5522652103684332, + "learning_rate": 0.0012548112021752648, + "loss": 1.1712, + "step": 7180 + }, + { + "epoch": 0.5523421275286516, + "learning_rate": 0.0012544536208595513, + "loss": 1.1259, + "step": 7181 + }, + { + "epoch": 0.55241904468887, + "learning_rate": 0.0012540960538815467, + "loss": 1.2825, + "step": 7182 + }, + { + "epoch": 0.5524959618490886, + "learning_rate": 0.0012537385012621294, + "loss": 1.1437, + "step": 7183 + }, + { + "epoch": 0.552572879009307, + "learning_rate": 0.0012533809630221782, + "loss": 1.2397, + "step": 7184 + }, + { + "epoch": 0.5526497961695255, + "learning_rate": 0.001253023439182569, + "loss": 0.9709, + "step": 7185 + }, + { + "epoch": 0.5527267133297439, + "learning_rate": 0.0012526659297641793, + "loss": 1.7203, + "step": 7186 + }, + { + "epoch": 0.5528036304899623, + "learning_rate": 0.0012523084347878838, + "loss": 0.988, + "step": 7187 + }, + { + "epoch": 0.5528805476501808, + "learning_rate": 0.001251950954274556, + "loss": 0.9961, + "step": 7188 + }, + { + "epoch": 0.5529574648103992, + "learning_rate": 0.0012515934882450714, + "loss": 1.0273, + "step": 7189 + }, + { + "epoch": 0.5530343819706176, + "learning_rate": 0.0012512360367203017, + "loss": 1.0816, + "step": 7190 + }, + { + "epoch": 0.5531112991308361, + "learning_rate": 0.0012508785997211184, + "loss": 1.0205, + "step": 7191 + }, + { + "epoch": 0.5531882162910545, + "learning_rate": 0.001250521177268394, + "loss": 1.035, + "step": 7192 + }, + { + "epoch": 0.5532651334512729, + "learning_rate": 0.001250163769382998, + "loss": 0.9753, + "step": 7193 + }, + { + "epoch": 0.5533420506114914, + "learning_rate": 0.0012498063760857994, + "loss": 1.0268, + "step": 7194 + }, + { + "epoch": 0.5534189677717098, + "learning_rate": 0.0012494489973976677, + "loss": 0.8882, + "step": 7195 + }, + { + "epoch": 0.5534958849319284, + "learning_rate": 0.0012490916333394698, + "loss": 1.1814, + "step": 7196 + }, + { + "epoch": 0.5535728020921468, + "learning_rate": 0.0012487342839320734, + "loss": 1.0913, + "step": 7197 + }, + { + "epoch": 0.5536497192523652, + "learning_rate": 0.0012483769491963442, + "loss": 1.0421, + "step": 7198 + }, + { + "epoch": 0.5537266364125837, + "learning_rate": 0.001248019629153146, + "loss": 1.2698, + "step": 7199 + }, + { + "epoch": 0.5538035535728021, + "learning_rate": 0.0012476623238233454, + "loss": 1.0751, + "step": 7200 + }, + { + "epoch": 0.5538804707330205, + "learning_rate": 0.0012473050332278044, + "loss": 0.8832, + "step": 7201 + }, + { + "epoch": 0.553957387893239, + "learning_rate": 0.0012469477573873856, + "loss": 1.2194, + "step": 7202 + }, + { + "epoch": 0.5540343050534574, + "learning_rate": 0.0012465904963229514, + "loss": 1.1259, + "step": 7203 + }, + { + "epoch": 0.5541112222136759, + "learning_rate": 0.0012462332500553616, + "loss": 1.3259, + "step": 7204 + }, + { + "epoch": 0.5541881393738943, + "learning_rate": 0.0012458760186054771, + "loss": 1.3579, + "step": 7205 + }, + { + "epoch": 0.5542650565341127, + "learning_rate": 0.001245518801994157, + "loss": 1.0214, + "step": 7206 + }, + { + "epoch": 0.5543419736943312, + "learning_rate": 0.001245161600242259, + "loss": 1.3077, + "step": 7207 + }, + { + "epoch": 0.5544188908545497, + "learning_rate": 0.0012448044133706408, + "loss": 1.1951, + "step": 7208 + }, + { + "epoch": 0.554495808014768, + "learning_rate": 0.001244447241400159, + "loss": 1.2022, + "step": 7209 + }, + { + "epoch": 0.5545727251749866, + "learning_rate": 0.0012440900843516684, + "loss": 1.1184, + "step": 7210 + }, + { + "epoch": 0.554649642335205, + "learning_rate": 0.0012437329422460256, + "loss": 1.0422, + "step": 7211 + }, + { + "epoch": 0.5547265594954234, + "learning_rate": 0.0012433758151040829, + "loss": 1.0733, + "step": 7212 + }, + { + "epoch": 0.5548034766556419, + "learning_rate": 0.0012430187029466933, + "loss": 1.2094, + "step": 7213 + }, + { + "epoch": 0.5548803938158603, + "learning_rate": 0.0012426616057947101, + "loss": 1.262, + "step": 7214 + }, + { + "epoch": 0.5549573109760788, + "learning_rate": 0.0012423045236689835, + "loss": 0.8998, + "step": 7215 + }, + { + "epoch": 0.5550342281362972, + "learning_rate": 0.0012419474565903647, + "loss": 1.2346, + "step": 7216 + }, + { + "epoch": 0.5551111452965156, + "learning_rate": 0.0012415904045797027, + "loss": 0.9243, + "step": 7217 + }, + { + "epoch": 0.5551880624567341, + "learning_rate": 0.001241233367657846, + "loss": 1.2205, + "step": 7218 + }, + { + "epoch": 0.5552649796169525, + "learning_rate": 0.001240876345845643, + "loss": 0.9868, + "step": 7219 + }, + { + "epoch": 0.5553418967771709, + "learning_rate": 0.0012405193391639405, + "loss": 0.7864, + "step": 7220 + }, + { + "epoch": 0.5554188139373895, + "learning_rate": 0.0012401623476335833, + "loss": 1.3593, + "step": 7221 + }, + { + "epoch": 0.5554957310976079, + "learning_rate": 0.0012398053712754183, + "loss": 1.4042, + "step": 7222 + }, + { + "epoch": 0.5555726482578264, + "learning_rate": 0.001239448410110288, + "loss": 1.0608, + "step": 7223 + }, + { + "epoch": 0.5556495654180448, + "learning_rate": 0.0012390914641590375, + "loss": 1.2384, + "step": 7224 + }, + { + "epoch": 0.5557264825782632, + "learning_rate": 0.0012387345334425078, + "loss": 1.219, + "step": 7225 + }, + { + "epoch": 0.5558033997384817, + "learning_rate": 0.0012383776179815409, + "loss": 1.5466, + "step": 7226 + }, + { + "epoch": 0.5558803168987001, + "learning_rate": 0.0012380207177969777, + "loss": 1.2514, + "step": 7227 + }, + { + "epoch": 0.5559572340589185, + "learning_rate": 0.001237663832909658, + "loss": 1.0605, + "step": 7228 + }, + { + "epoch": 0.556034151219137, + "learning_rate": 0.0012373069633404199, + "loss": 1.1861, + "step": 7229 + }, + { + "epoch": 0.5561110683793554, + "learning_rate": 0.0012369501091101026, + "loss": 1.3284, + "step": 7230 + }, + { + "epoch": 0.5561879855395738, + "learning_rate": 0.0012365932702395428, + "loss": 1.1109, + "step": 7231 + }, + { + "epoch": 0.5562649026997923, + "learning_rate": 0.0012362364467495754, + "loss": 1.0749, + "step": 7232 + }, + { + "epoch": 0.5563418198600107, + "learning_rate": 0.001235879638661038, + "loss": 1.0757, + "step": 7233 + }, + { + "epoch": 0.5564187370202293, + "learning_rate": 0.0012355228459947625, + "loss": 1.2782, + "step": 7234 + }, + { + "epoch": 0.5564956541804477, + "learning_rate": 0.0012351660687715846, + "loss": 0.998, + "step": 7235 + }, + { + "epoch": 0.5565725713406661, + "learning_rate": 0.0012348093070123358, + "loss": 1.2242, + "step": 7236 + }, + { + "epoch": 0.5566494885008846, + "learning_rate": 0.0012344525607378475, + "loss": 0.8711, + "step": 7237 + }, + { + "epoch": 0.556726405661103, + "learning_rate": 0.0012340958299689514, + "loss": 1.0938, + "step": 7238 + }, + { + "epoch": 0.5568033228213214, + "learning_rate": 0.0012337391147264768, + "loss": 1.3598, + "step": 7239 + }, + { + "epoch": 0.5568802399815399, + "learning_rate": 0.0012333824150312525, + "loss": 1.0687, + "step": 7240 + }, + { + "epoch": 0.5569571571417583, + "learning_rate": 0.001233025730904107, + "loss": 0.8455, + "step": 7241 + }, + { + "epoch": 0.5570340743019768, + "learning_rate": 0.001232669062365867, + "loss": 1.1195, + "step": 7242 + }, + { + "epoch": 0.5571109914621952, + "learning_rate": 0.0012323124094373598, + "loss": 1.5632, + "step": 7243 + }, + { + "epoch": 0.5571879086224136, + "learning_rate": 0.0012319557721394098, + "loss": 1.1215, + "step": 7244 + }, + { + "epoch": 0.5572648257826321, + "learning_rate": 0.001231599150492841, + "loss": 1.6228, + "step": 7245 + }, + { + "epoch": 0.5573417429428505, + "learning_rate": 0.0012312425445184783, + "loss": 1.1486, + "step": 7246 + }, + { + "epoch": 0.557418660103069, + "learning_rate": 0.0012308859542371432, + "loss": 1.2171, + "step": 7247 + }, + { + "epoch": 0.5574955772632875, + "learning_rate": 0.0012305293796696574, + "loss": 1.3485, + "step": 7248 + }, + { + "epoch": 0.5575724944235059, + "learning_rate": 0.0012301728208368424, + "loss": 1.272, + "step": 7249 + }, + { + "epoch": 0.5576494115837243, + "learning_rate": 0.0012298162777595175, + "loss": 1.0851, + "step": 7250 + }, + { + "epoch": 0.5577263287439428, + "learning_rate": 0.0012294597504585014, + "loss": 1.2688, + "step": 7251 + }, + { + "epoch": 0.5578032459041612, + "learning_rate": 0.0012291032389546127, + "loss": 1.1037, + "step": 7252 + }, + { + "epoch": 0.5578801630643797, + "learning_rate": 0.001228746743268668, + "loss": 1.1284, + "step": 7253 + }, + { + "epoch": 0.5579570802245981, + "learning_rate": 0.0012283902634214838, + "loss": 1.0962, + "step": 7254 + }, + { + "epoch": 0.5580339973848165, + "learning_rate": 0.0012280337994338751, + "loss": 1.0351, + "step": 7255 + }, + { + "epoch": 0.558110914545035, + "learning_rate": 0.0012276773513266563, + "loss": 1.6946, + "step": 7256 + }, + { + "epoch": 0.5581878317052534, + "learning_rate": 0.0012273209191206414, + "loss": 1.1502, + "step": 7257 + }, + { + "epoch": 0.5582647488654718, + "learning_rate": 0.0012269645028366416, + "loss": 1.1958, + "step": 7258 + }, + { + "epoch": 0.5583416660256904, + "learning_rate": 0.0012266081024954688, + "loss": 1.3933, + "step": 7259 + }, + { + "epoch": 0.5584185831859088, + "learning_rate": 0.0012262517181179343, + "loss": 1.235, + "step": 7260 + }, + { + "epoch": 0.5584955003461273, + "learning_rate": 0.0012258953497248468, + "loss": 1.2645, + "step": 7261 + }, + { + "epoch": 0.5585724175063457, + "learning_rate": 0.001225538997337016, + "loss": 1.2166, + "step": 7262 + }, + { + "epoch": 0.5586493346665641, + "learning_rate": 0.001225182660975249, + "loss": 1.1861, + "step": 7263 + }, + { + "epoch": 0.5587262518267826, + "learning_rate": 0.0012248263406603524, + "loss": 1.0994, + "step": 7264 + }, + { + "epoch": 0.558803168987001, + "learning_rate": 0.001224470036413133, + "loss": 0.7398, + "step": 7265 + }, + { + "epoch": 0.5588800861472194, + "learning_rate": 0.0012241137482543953, + "loss": 1.2622, + "step": 7266 + }, + { + "epoch": 0.5589570033074379, + "learning_rate": 0.0012237574762049431, + "loss": 0.9521, + "step": 7267 + }, + { + "epoch": 0.5590339204676563, + "learning_rate": 0.0012234012202855802, + "loss": 0.9145, + "step": 7268 + }, + { + "epoch": 0.5591108376278747, + "learning_rate": 0.0012230449805171084, + "loss": 0.8893, + "step": 7269 + }, + { + "epoch": 0.5591877547880932, + "learning_rate": 0.0012226887569203283, + "loss": 1.2631, + "step": 7270 + }, + { + "epoch": 0.5592646719483116, + "learning_rate": 0.001222332549516041, + "loss": 1.0852, + "step": 7271 + }, + { + "epoch": 0.5593415891085302, + "learning_rate": 0.001221976358325045, + "loss": 1.3567, + "step": 7272 + }, + { + "epoch": 0.5594185062687486, + "learning_rate": 0.0012216201833681398, + "loss": 1.3905, + "step": 7273 + }, + { + "epoch": 0.559495423428967, + "learning_rate": 0.0012212640246661221, + "loss": 0.9227, + "step": 7274 + }, + { + "epoch": 0.5595723405891855, + "learning_rate": 0.0012209078822397882, + "loss": 1.4219, + "step": 7275 + }, + { + "epoch": 0.5596492577494039, + "learning_rate": 0.001220551756109934, + "loss": 0.6822, + "step": 7276 + }, + { + "epoch": 0.5597261749096223, + "learning_rate": 0.0012201956462973545, + "loss": 1.2539, + "step": 7277 + }, + { + "epoch": 0.5598030920698408, + "learning_rate": 0.001219839552822842, + "loss": 1.2788, + "step": 7278 + }, + { + "epoch": 0.5598800092300592, + "learning_rate": 0.0012194834757071906, + "loss": 1.1444, + "step": 7279 + }, + { + "epoch": 0.5599569263902777, + "learning_rate": 0.001219127414971191, + "loss": 1.0939, + "step": 7280 + }, + { + "epoch": 0.5600338435504961, + "learning_rate": 0.0012187713706356346, + "loss": 1.2404, + "step": 7281 + }, + { + "epoch": 0.5601107607107145, + "learning_rate": 0.0012184153427213112, + "loss": 1.2991, + "step": 7282 + }, + { + "epoch": 0.560187677870933, + "learning_rate": 0.0012180593312490088, + "loss": 1.1995, + "step": 7283 + }, + { + "epoch": 0.5602645950311514, + "learning_rate": 0.0012177033362395166, + "loss": 1.3429, + "step": 7284 + }, + { + "epoch": 0.5603415121913699, + "learning_rate": 0.0012173473577136206, + "loss": 1.1776, + "step": 7285 + }, + { + "epoch": 0.5604184293515884, + "learning_rate": 0.0012169913956921064, + "loss": 1.239, + "step": 7286 + }, + { + "epoch": 0.5604953465118068, + "learning_rate": 0.00121663545019576, + "loss": 1.0809, + "step": 7287 + }, + { + "epoch": 0.5605722636720253, + "learning_rate": 0.0012162795212453648, + "loss": 1.4057, + "step": 7288 + }, + { + "epoch": 0.5606491808322437, + "learning_rate": 0.0012159236088617042, + "loss": 1.3556, + "step": 7289 + }, + { + "epoch": 0.5607260979924621, + "learning_rate": 0.00121556771306556, + "loss": 0.7773, + "step": 7290 + }, + { + "epoch": 0.5608030151526806, + "learning_rate": 0.0012152118338777133, + "loss": 1.2532, + "step": 7291 + }, + { + "epoch": 0.560879932312899, + "learning_rate": 0.0012148559713189447, + "loss": 1.1577, + "step": 7292 + }, + { + "epoch": 0.5609568494731174, + "learning_rate": 0.0012145001254100336, + "loss": 1.322, + "step": 7293 + }, + { + "epoch": 0.5610337666333359, + "learning_rate": 0.0012141442961717567, + "loss": 1.1817, + "step": 7294 + }, + { + "epoch": 0.5611106837935543, + "learning_rate": 0.001213788483624893, + "loss": 1.1667, + "step": 7295 + }, + { + "epoch": 0.5611876009537727, + "learning_rate": 0.001213432687790218, + "loss": 1.346, + "step": 7296 + }, + { + "epoch": 0.5612645181139913, + "learning_rate": 0.0012130769086885064, + "loss": 1.2367, + "step": 7297 + }, + { + "epoch": 0.5613414352742097, + "learning_rate": 0.0012127211463405336, + "loss": 0.867, + "step": 7298 + }, + { + "epoch": 0.5614183524344282, + "learning_rate": 0.0012123654007670723, + "loss": 1.1833, + "step": 7299 + }, + { + "epoch": 0.5614952695946466, + "learning_rate": 0.001212009671988895, + "loss": 1.3038, + "step": 7300 + }, + { + "epoch": 0.561572186754865, + "learning_rate": 0.0012116539600267734, + "loss": 1.0554, + "step": 7301 + }, + { + "epoch": 0.5616491039150835, + "learning_rate": 0.001211298264901477, + "loss": 1.0425, + "step": 7302 + }, + { + "epoch": 0.5617260210753019, + "learning_rate": 0.0012109425866337763, + "loss": 1.0325, + "step": 7303 + }, + { + "epoch": 0.5618029382355203, + "learning_rate": 0.0012105869252444393, + "loss": 0.7702, + "step": 7304 + }, + { + "epoch": 0.5618798553957388, + "learning_rate": 0.0012102312807542326, + "loss": 0.9658, + "step": 7305 + }, + { + "epoch": 0.5619567725559572, + "learning_rate": 0.0012098756531839245, + "loss": 1.1969, + "step": 7306 + }, + { + "epoch": 0.5620336897161757, + "learning_rate": 0.0012095200425542782, + "loss": 1.1286, + "step": 7307 + }, + { + "epoch": 0.5621106068763941, + "learning_rate": 0.0012091644488860603, + "loss": 0.6555, + "step": 7308 + }, + { + "epoch": 0.5621875240366125, + "learning_rate": 0.001208808872200033, + "loss": 0.943, + "step": 7309 + }, + { + "epoch": 0.562264441196831, + "learning_rate": 0.0012084533125169589, + "loss": 1.2899, + "step": 7310 + }, + { + "epoch": 0.5623413583570495, + "learning_rate": 0.0012080977698576, + "loss": 1.1096, + "step": 7311 + }, + { + "epoch": 0.5624182755172679, + "learning_rate": 0.0012077422442427165, + "loss": 1.2236, + "step": 7312 + }, + { + "epoch": 0.5624951926774864, + "learning_rate": 0.0012073867356930676, + "loss": 0.8613, + "step": 7313 + }, + { + "epoch": 0.5625721098377048, + "learning_rate": 0.0012070312442294124, + "loss": 1.2096, + "step": 7314 + }, + { + "epoch": 0.5626490269979232, + "learning_rate": 0.0012066757698725088, + "loss": 1.2344, + "step": 7315 + }, + { + "epoch": 0.5627259441581417, + "learning_rate": 0.0012063203126431116, + "loss": 1.5587, + "step": 7316 + }, + { + "epoch": 0.5628028613183601, + "learning_rate": 0.0012059648725619783, + "loss": 1.2057, + "step": 7317 + }, + { + "epoch": 0.5628797784785786, + "learning_rate": 0.0012056094496498615, + "loss": 1.0798, + "step": 7318 + }, + { + "epoch": 0.562956695638797, + "learning_rate": 0.0012052540439275168, + "loss": 1.092, + "step": 7319 + }, + { + "epoch": 0.5630336127990154, + "learning_rate": 0.0012048986554156953, + "loss": 0.8852, + "step": 7320 + }, + { + "epoch": 0.5631105299592339, + "learning_rate": 0.0012045432841351488, + "loss": 1.3227, + "step": 7321 + }, + { + "epoch": 0.5631874471194523, + "learning_rate": 0.0012041879301066278, + "loss": 1.0863, + "step": 7322 + }, + { + "epoch": 0.5632643642796707, + "learning_rate": 0.0012038325933508818, + "loss": 1.3352, + "step": 7323 + }, + { + "epoch": 0.5633412814398893, + "learning_rate": 0.0012034772738886593, + "loss": 0.8323, + "step": 7324 + }, + { + "epoch": 0.5634181986001077, + "learning_rate": 0.001203121971740708, + "loss": 1.1595, + "step": 7325 + }, + { + "epoch": 0.5634951157603262, + "learning_rate": 0.001202766686927774, + "loss": 0.9934, + "step": 7326 + }, + { + "epoch": 0.5635720329205446, + "learning_rate": 0.001202411419470603, + "loss": 1.0507, + "step": 7327 + }, + { + "epoch": 0.563648950080763, + "learning_rate": 0.0012020561693899397, + "loss": 1.1643, + "step": 7328 + }, + { + "epoch": 0.5637258672409815, + "learning_rate": 0.0012017009367065264, + "loss": 1.1667, + "step": 7329 + }, + { + "epoch": 0.5638027844011999, + "learning_rate": 0.001201345721441107, + "loss": 0.7244, + "step": 7330 + }, + { + "epoch": 0.5638797015614183, + "learning_rate": 0.0012009905236144223, + "loss": 1.1365, + "step": 7331 + }, + { + "epoch": 0.5639566187216368, + "learning_rate": 0.001200635343247212, + "loss": 0.9453, + "step": 7332 + }, + { + "epoch": 0.5640335358818552, + "learning_rate": 0.001200280180360216, + "loss": 0.9196, + "step": 7333 + }, + { + "epoch": 0.5641104530420736, + "learning_rate": 0.001199925034974173, + "loss": 1.1879, + "step": 7334 + }, + { + "epoch": 0.5641873702022921, + "learning_rate": 0.0011995699071098195, + "loss": 1.3436, + "step": 7335 + }, + { + "epoch": 0.5642642873625106, + "learning_rate": 0.0011992147967878925, + "loss": 1.1447, + "step": 7336 + }, + { + "epoch": 0.5643412045227291, + "learning_rate": 0.0011988597040291267, + "loss": 1.1356, + "step": 7337 + }, + { + "epoch": 0.5644181216829475, + "learning_rate": 0.001198504628854257, + "loss": 0.844, + "step": 7338 + }, + { + "epoch": 0.5644950388431659, + "learning_rate": 0.0011981495712840165, + "loss": 1.4253, + "step": 7339 + }, + { + "epoch": 0.5645719560033844, + "learning_rate": 0.0011977945313391362, + "loss": 0.9571, + "step": 7340 + }, + { + "epoch": 0.5646488731636028, + "learning_rate": 0.001197439509040349, + "loss": 1.2405, + "step": 7341 + }, + { + "epoch": 0.5647257903238212, + "learning_rate": 0.0011970845044083842, + "loss": 1.1636, + "step": 7342 + }, + { + "epoch": 0.5648027074840397, + "learning_rate": 0.0011967295174639702, + "loss": 1.3601, + "step": 7343 + }, + { + "epoch": 0.5648796246442581, + "learning_rate": 0.0011963745482278362, + "loss": 1.2052, + "step": 7344 + }, + { + "epoch": 0.5649565418044766, + "learning_rate": 0.0011960195967207084, + "loss": 1.1753, + "step": 7345 + }, + { + "epoch": 0.565033458964695, + "learning_rate": 0.0011956646629633135, + "loss": 0.7865, + "step": 7346 + }, + { + "epoch": 0.5651103761249134, + "learning_rate": 0.001195309746976376, + "loss": 1.497, + "step": 7347 + }, + { + "epoch": 0.565187293285132, + "learning_rate": 0.0011949548487806196, + "loss": 1.139, + "step": 7348 + }, + { + "epoch": 0.5652642104453504, + "learning_rate": 0.001194599968396768, + "loss": 1.2221, + "step": 7349 + }, + { + "epoch": 0.5653411276055688, + "learning_rate": 0.0011942451058455426, + "loss": 1.0699, + "step": 7350 + }, + { + "epoch": 0.5654180447657873, + "learning_rate": 0.0011938902611476634, + "loss": 1.2305, + "step": 7351 + }, + { + "epoch": 0.5654949619260057, + "learning_rate": 0.0011935354343238515, + "loss": 1.4869, + "step": 7352 + }, + { + "epoch": 0.5655718790862241, + "learning_rate": 0.001193180625394825, + "loss": 1.2495, + "step": 7353 + }, + { + "epoch": 0.5656487962464426, + "learning_rate": 0.001192825834381301, + "loss": 1.107, + "step": 7354 + }, + { + "epoch": 0.565725713406661, + "learning_rate": 0.0011924710613039973, + "loss": 1.0977, + "step": 7355 + }, + { + "epoch": 0.5658026305668795, + "learning_rate": 0.0011921163061836284, + "loss": 0.936, + "step": 7356 + }, + { + "epoch": 0.5658795477270979, + "learning_rate": 0.0011917615690409092, + "loss": 1.2081, + "step": 7357 + }, + { + "epoch": 0.5659564648873163, + "learning_rate": 0.0011914068498965536, + "loss": 0.9764, + "step": 7358 + }, + { + "epoch": 0.5660333820475348, + "learning_rate": 0.0011910521487712732, + "loss": 1.202, + "step": 7359 + }, + { + "epoch": 0.5661102992077532, + "learning_rate": 0.0011906974656857802, + "loss": 1.345, + "step": 7360 + }, + { + "epoch": 0.5661872163679716, + "learning_rate": 0.0011903428006607847, + "loss": 0.8083, + "step": 7361 + }, + { + "epoch": 0.5662641335281902, + "learning_rate": 0.001189988153716995, + "loss": 1.1411, + "step": 7362 + }, + { + "epoch": 0.5663410506884086, + "learning_rate": 0.0011896335248751209, + "loss": 1.2688, + "step": 7363 + }, + { + "epoch": 0.5664179678486271, + "learning_rate": 0.001189278914155868, + "loss": 1.2281, + "step": 7364 + }, + { + "epoch": 0.5664948850088455, + "learning_rate": 0.0011889243215799437, + "loss": 0.998, + "step": 7365 + }, + { + "epoch": 0.5665718021690639, + "learning_rate": 0.0011885697471680525, + "loss": 1.0692, + "step": 7366 + }, + { + "epoch": 0.5666487193292824, + "learning_rate": 0.001188215190940898, + "loss": 1.176, + "step": 7367 + }, + { + "epoch": 0.5667256364895008, + "learning_rate": 0.0011878606529191836, + "loss": 1.1958, + "step": 7368 + }, + { + "epoch": 0.5668025536497192, + "learning_rate": 0.0011875061331236107, + "loss": 1.1545, + "step": 7369 + }, + { + "epoch": 0.5668794708099377, + "learning_rate": 0.0011871516315748805, + "loss": 1.2527, + "step": 7370 + }, + { + "epoch": 0.5669563879701561, + "learning_rate": 0.0011867971482936927, + "loss": 1.0219, + "step": 7371 + }, + { + "epoch": 0.5670333051303745, + "learning_rate": 0.0011864426833007463, + "loss": 1.2691, + "step": 7372 + }, + { + "epoch": 0.567110222290593, + "learning_rate": 0.0011860882366167373, + "loss": 1.3616, + "step": 7373 + }, + { + "epoch": 0.5671871394508115, + "learning_rate": 0.0011857338082623644, + "loss": 1.5308, + "step": 7374 + }, + { + "epoch": 0.56726405661103, + "learning_rate": 0.0011853793982583207, + "loss": 1.2038, + "step": 7375 + }, + { + "epoch": 0.5673409737712484, + "learning_rate": 0.001185025006625303, + "loss": 0.9733, + "step": 7376 + }, + { + "epoch": 0.5674178909314668, + "learning_rate": 0.001184670633384003, + "loss": 1.5879, + "step": 7377 + }, + { + "epoch": 0.5674948080916853, + "learning_rate": 0.0011843162785551134, + "loss": 1.2297, + "step": 7378 + }, + { + "epoch": 0.5675717252519037, + "learning_rate": 0.0011839619421593253, + "loss": 1.0965, + "step": 7379 + }, + { + "epoch": 0.5676486424121221, + "learning_rate": 0.0011836076242173288, + "loss": 1.139, + "step": 7380 + }, + { + "epoch": 0.5677255595723406, + "learning_rate": 0.0011832533247498123, + "loss": 1.139, + "step": 7381 + }, + { + "epoch": 0.567802476732559, + "learning_rate": 0.001182899043777465, + "loss": 1.253, + "step": 7382 + }, + { + "epoch": 0.5678793938927775, + "learning_rate": 0.0011825447813209723, + "loss": 1.1121, + "step": 7383 + }, + { + "epoch": 0.5679563110529959, + "learning_rate": 0.0011821905374010213, + "loss": 1.2031, + "step": 7384 + }, + { + "epoch": 0.5680332282132143, + "learning_rate": 0.0011818363120382964, + "loss": 0.6744, + "step": 7385 + }, + { + "epoch": 0.5681101453734329, + "learning_rate": 0.0011814821052534795, + "loss": 0.9784, + "step": 7386 + }, + { + "epoch": 0.5681870625336513, + "learning_rate": 0.0011811279170672558, + "loss": 1.1948, + "step": 7387 + }, + { + "epoch": 0.5682639796938697, + "learning_rate": 0.0011807737475003047, + "loss": 0.8045, + "step": 7388 + }, + { + "epoch": 0.5683408968540882, + "learning_rate": 0.001180419596573307, + "loss": 0.9747, + "step": 7389 + }, + { + "epoch": 0.5684178140143066, + "learning_rate": 0.0011800654643069424, + "loss": 0.9116, + "step": 7390 + }, + { + "epoch": 0.5684947311745251, + "learning_rate": 0.0011797113507218883, + "loss": 0.9082, + "step": 7391 + }, + { + "epoch": 0.5685716483347435, + "learning_rate": 0.0011793572558388228, + "loss": 0.9714, + "step": 7392 + }, + { + "epoch": 0.5686485654949619, + "learning_rate": 0.0011790031796784213, + "loss": 1.3846, + "step": 7393 + }, + { + "epoch": 0.5687254826551804, + "learning_rate": 0.0011786491222613583, + "loss": 1.0717, + "step": 7394 + }, + { + "epoch": 0.5688023998153988, + "learning_rate": 0.0011782950836083082, + "loss": 1.1803, + "step": 7395 + }, + { + "epoch": 0.5688793169756172, + "learning_rate": 0.0011779410637399437, + "loss": 1.6949, + "step": 7396 + }, + { + "epoch": 0.5689562341358357, + "learning_rate": 0.0011775870626769354, + "loss": 0.9212, + "step": 7397 + }, + { + "epoch": 0.5690331512960541, + "learning_rate": 0.0011772330804399552, + "loss": 1.1032, + "step": 7398 + }, + { + "epoch": 0.5691100684562725, + "learning_rate": 0.0011768791170496716, + "loss": 1.0047, + "step": 7399 + }, + { + "epoch": 0.5691869856164911, + "learning_rate": 0.001176525172526753, + "loss": 1.2054, + "step": 7400 + }, + { + "epoch": 0.5692639027767095, + "learning_rate": 0.0011761712468918667, + "loss": 1.2594, + "step": 7401 + }, + { + "epoch": 0.569340819936928, + "learning_rate": 0.0011758173401656788, + "loss": 1.2848, + "step": 7402 + }, + { + "epoch": 0.5694177370971464, + "learning_rate": 0.0011754634523688543, + "loss": 1.403, + "step": 7403 + }, + { + "epoch": 0.5694946542573648, + "learning_rate": 0.0011751095835220572, + "loss": 1.1681, + "step": 7404 + }, + { + "epoch": 0.5695715714175833, + "learning_rate": 0.0011747557336459495, + "loss": 1.1703, + "step": 7405 + }, + { + "epoch": 0.5696484885778017, + "learning_rate": 0.001174401902761194, + "loss": 0.9024, + "step": 7406 + }, + { + "epoch": 0.5697254057380201, + "learning_rate": 0.0011740480908884509, + "loss": 1.1178, + "step": 7407 + }, + { + "epoch": 0.5698023228982386, + "learning_rate": 0.0011736942980483788, + "loss": 1.1575, + "step": 7408 + }, + { + "epoch": 0.569879240058457, + "learning_rate": 0.0011733405242616375, + "loss": 1.1706, + "step": 7409 + }, + { + "epoch": 0.5699561572186755, + "learning_rate": 0.0011729867695488825, + "loss": 1.3981, + "step": 7410 + }, + { + "epoch": 0.570033074378894, + "learning_rate": 0.001172633033930772, + "loss": 1.055, + "step": 7411 + }, + { + "epoch": 0.5701099915391123, + "learning_rate": 0.0011722793174279592, + "loss": 1.102, + "step": 7412 + }, + { + "epoch": 0.5701869086993309, + "learning_rate": 0.0011719256200610983, + "loss": 0.9951, + "step": 7413 + }, + { + "epoch": 0.5702638258595493, + "learning_rate": 0.001171571941850843, + "loss": 1.2232, + "step": 7414 + }, + { + "epoch": 0.5703407430197677, + "learning_rate": 0.0011712182828178437, + "loss": 1.2202, + "step": 7415 + }, + { + "epoch": 0.5704176601799862, + "learning_rate": 0.0011708646429827517, + "loss": 1.0668, + "step": 7416 + }, + { + "epoch": 0.5704945773402046, + "learning_rate": 0.0011705110223662162, + "loss": 1.3328, + "step": 7417 + }, + { + "epoch": 0.570571494500423, + "learning_rate": 0.0011701574209888857, + "loss": 1.094, + "step": 7418 + }, + { + "epoch": 0.5706484116606415, + "learning_rate": 0.0011698038388714069, + "loss": 1.6162, + "step": 7419 + }, + { + "epoch": 0.5707253288208599, + "learning_rate": 0.001169450276034426, + "loss": 1.0043, + "step": 7420 + }, + { + "epoch": 0.5708022459810784, + "learning_rate": 0.0011690967324985882, + "loss": 1.1951, + "step": 7421 + }, + { + "epoch": 0.5708791631412968, + "learning_rate": 0.0011687432082845368, + "loss": 0.7274, + "step": 7422 + }, + { + "epoch": 0.5709560803015152, + "learning_rate": 0.0011683897034129155, + "loss": 0.8897, + "step": 7423 + }, + { + "epoch": 0.5710329974617337, + "learning_rate": 0.001168036217904364, + "loss": 1.4012, + "step": 7424 + }, + { + "epoch": 0.5711099146219522, + "learning_rate": 0.0011676827517795242, + "loss": 1.2045, + "step": 7425 + }, + { + "epoch": 0.5711868317821706, + "learning_rate": 0.001167329305059035, + "loss": 1.7016, + "step": 7426 + }, + { + "epoch": 0.5712637489423891, + "learning_rate": 0.0011669758777635337, + "loss": 1.2601, + "step": 7427 + }, + { + "epoch": 0.5713406661026075, + "learning_rate": 0.0011666224699136587, + "loss": 0.7885, + "step": 7428 + }, + { + "epoch": 0.571417583262826, + "learning_rate": 0.0011662690815300446, + "loss": 1.3878, + "step": 7429 + }, + { + "epoch": 0.5714945004230444, + "learning_rate": 0.001165915712633327, + "loss": 1.3918, + "step": 7430 + }, + { + "epoch": 0.5715714175832628, + "learning_rate": 0.001165562363244139, + "loss": 1.343, + "step": 7431 + }, + { + "epoch": 0.5716483347434813, + "learning_rate": 0.0011652090333831128, + "loss": 1.037, + "step": 7432 + }, + { + "epoch": 0.5717252519036997, + "learning_rate": 0.0011648557230708805, + "loss": 1.3406, + "step": 7433 + }, + { + "epoch": 0.5718021690639181, + "learning_rate": 0.0011645024323280722, + "loss": 1.1148, + "step": 7434 + }, + { + "epoch": 0.5718790862241366, + "learning_rate": 0.0011641491611753152, + "loss": 1.1338, + "step": 7435 + }, + { + "epoch": 0.571956003384355, + "learning_rate": 0.00116379590963324, + "loss": 1.2277, + "step": 7436 + }, + { + "epoch": 0.5720329205445734, + "learning_rate": 0.0011634426777224717, + "loss": 1.5736, + "step": 7437 + }, + { + "epoch": 0.572109837704792, + "learning_rate": 0.0011630894654636357, + "loss": 1.0534, + "step": 7438 + }, + { + "epoch": 0.5721867548650104, + "learning_rate": 0.0011627362728773574, + "loss": 1.2428, + "step": 7439 + }, + { + "epoch": 0.5722636720252289, + "learning_rate": 0.0011623830999842593, + "loss": 1.3041, + "step": 7440 + }, + { + "epoch": 0.5723405891854473, + "learning_rate": 0.001162029946804964, + "loss": 1.2164, + "step": 7441 + }, + { + "epoch": 0.5724175063456657, + "learning_rate": 0.0011616768133600925, + "loss": 0.6077, + "step": 7442 + }, + { + "epoch": 0.5724944235058842, + "learning_rate": 0.0011613236996702641, + "loss": 1.0239, + "step": 7443 + }, + { + "epoch": 0.5725713406661026, + "learning_rate": 0.0011609706057560982, + "loss": 1.3251, + "step": 7444 + }, + { + "epoch": 0.572648257826321, + "learning_rate": 0.001160617531638212, + "loss": 0.7548, + "step": 7445 + }, + { + "epoch": 0.5727251749865395, + "learning_rate": 0.0011602644773372212, + "loss": 1.1027, + "step": 7446 + }, + { + "epoch": 0.5728020921467579, + "learning_rate": 0.0011599114428737425, + "loss": 1.1049, + "step": 7447 + }, + { + "epoch": 0.5728790093069764, + "learning_rate": 0.0011595584282683882, + "loss": 1.0308, + "step": 7448 + }, + { + "epoch": 0.5729559264671948, + "learning_rate": 0.0011592054335417728, + "loss": 1.1353, + "step": 7449 + }, + { + "epoch": 0.5730328436274132, + "learning_rate": 0.001158852458714507, + "loss": 0.9771, + "step": 7450 + }, + { + "epoch": 0.5731097607876318, + "learning_rate": 0.0011584995038072016, + "loss": 1.2061, + "step": 7451 + }, + { + "epoch": 0.5731866779478502, + "learning_rate": 0.0011581465688404661, + "loss": 1.3396, + "step": 7452 + }, + { + "epoch": 0.5732635951080686, + "learning_rate": 0.001157793653834909, + "loss": 1.2358, + "step": 7453 + }, + { + "epoch": 0.5733405122682871, + "learning_rate": 0.0011574407588111366, + "loss": 1.0596, + "step": 7454 + }, + { + "epoch": 0.5734174294285055, + "learning_rate": 0.0011570878837897558, + "loss": 1.2003, + "step": 7455 + }, + { + "epoch": 0.5734943465887239, + "learning_rate": 0.0011567350287913709, + "loss": 1.3258, + "step": 7456 + }, + { + "epoch": 0.5735712637489424, + "learning_rate": 0.0011563821938365844, + "loss": 1.1323, + "step": 7457 + }, + { + "epoch": 0.5736481809091608, + "learning_rate": 0.001156029378946001, + "loss": 0.9305, + "step": 7458 + }, + { + "epoch": 0.5737250980693793, + "learning_rate": 0.0011556765841402195, + "loss": 0.9148, + "step": 7459 + }, + { + "epoch": 0.5738020152295977, + "learning_rate": 0.001155323809439842, + "loss": 0.9186, + "step": 7460 + }, + { + "epoch": 0.5738789323898161, + "learning_rate": 0.0011549710548654664, + "loss": 1.4991, + "step": 7461 + }, + { + "epoch": 0.5739558495500346, + "learning_rate": 0.0011546183204376899, + "loss": 1.5976, + "step": 7462 + }, + { + "epoch": 0.574032766710253, + "learning_rate": 0.00115426560617711, + "loss": 1.1626, + "step": 7463 + }, + { + "epoch": 0.5741096838704715, + "learning_rate": 0.0011539129121043218, + "loss": 0.9012, + "step": 7464 + }, + { + "epoch": 0.57418660103069, + "learning_rate": 0.001153560238239919, + "loss": 1.1351, + "step": 7465 + }, + { + "epoch": 0.5742635181909084, + "learning_rate": 0.0011532075846044953, + "loss": 1.2098, + "step": 7466 + }, + { + "epoch": 0.5743404353511269, + "learning_rate": 0.0011528549512186414, + "loss": 1.013, + "step": 7467 + }, + { + "epoch": 0.5744173525113453, + "learning_rate": 0.0011525023381029496, + "loss": 1.0897, + "step": 7468 + }, + { + "epoch": 0.5744942696715637, + "learning_rate": 0.0011521497452780085, + "loss": 0.9073, + "step": 7469 + }, + { + "epoch": 0.5745711868317822, + "learning_rate": 0.0011517971727644052, + "loss": 1.2426, + "step": 7470 + }, + { + "epoch": 0.5746481039920006, + "learning_rate": 0.001151444620582729, + "loss": 1.2048, + "step": 7471 + }, + { + "epoch": 0.574725021152219, + "learning_rate": 0.0011510920887535644, + "loss": 1.2618, + "step": 7472 + }, + { + "epoch": 0.5748019383124375, + "learning_rate": 0.0011507395772974958, + "loss": 1.235, + "step": 7473 + }, + { + "epoch": 0.5748788554726559, + "learning_rate": 0.0011503870862351077, + "loss": 0.6297, + "step": 7474 + }, + { + "epoch": 0.5749557726328743, + "learning_rate": 0.001150034615586982, + "loss": 1.1965, + "step": 7475 + }, + { + "epoch": 0.5750326897930929, + "learning_rate": 0.0011496821653736995, + "loss": 1.3758, + "step": 7476 + }, + { + "epoch": 0.5751096069533113, + "learning_rate": 0.001149329735615841, + "loss": 1.236, + "step": 7477 + }, + { + "epoch": 0.5751865241135298, + "learning_rate": 0.001148977326333984, + "loss": 1.1907, + "step": 7478 + }, + { + "epoch": 0.5752634412737482, + "learning_rate": 0.0011486249375487073, + "loss": 1.1245, + "step": 7479 + }, + { + "epoch": 0.5753403584339666, + "learning_rate": 0.001148272569280587, + "loss": 0.888, + "step": 7480 + }, + { + "epoch": 0.5754172755941851, + "learning_rate": 0.0011479202215501968, + "loss": 1.2184, + "step": 7481 + }, + { + "epoch": 0.5754941927544035, + "learning_rate": 0.001147567894378113, + "loss": 1.2947, + "step": 7482 + }, + { + "epoch": 0.5755711099146219, + "learning_rate": 0.0011472155877849067, + "loss": 1.3256, + "step": 7483 + }, + { + "epoch": 0.5756480270748404, + "learning_rate": 0.0011468633017911495, + "loss": 1.3511, + "step": 7484 + }, + { + "epoch": 0.5757249442350588, + "learning_rate": 0.0011465110364174127, + "loss": 0.7183, + "step": 7485 + }, + { + "epoch": 0.5758018613952773, + "learning_rate": 0.0011461587916842644, + "loss": 1.0434, + "step": 7486 + }, + { + "epoch": 0.5758787785554957, + "learning_rate": 0.0011458065676122732, + "loss": 1.032, + "step": 7487 + }, + { + "epoch": 0.5759556957157141, + "learning_rate": 0.0011454543642220058, + "loss": 1.4464, + "step": 7488 + }, + { + "epoch": 0.5760326128759327, + "learning_rate": 0.0011451021815340272, + "loss": 1.2936, + "step": 7489 + }, + { + "epoch": 0.5761095300361511, + "learning_rate": 0.0011447500195689024, + "loss": 1.0794, + "step": 7490 + }, + { + "epoch": 0.5761864471963695, + "learning_rate": 0.0011443978783471946, + "loss": 1.2515, + "step": 7491 + }, + { + "epoch": 0.576263364356588, + "learning_rate": 0.0011440457578894645, + "loss": 1.0259, + "step": 7492 + }, + { + "epoch": 0.5763402815168064, + "learning_rate": 0.0011436936582162741, + "loss": 0.9607, + "step": 7493 + }, + { + "epoch": 0.5764171986770248, + "learning_rate": 0.0011433415793481823, + "loss": 1.8073, + "step": 7494 + }, + { + "epoch": 0.5764941158372433, + "learning_rate": 0.001142989521305747, + "loss": 0.8091, + "step": 7495 + }, + { + "epoch": 0.5765710329974617, + "learning_rate": 0.0011426374841095256, + "loss": 1.2961, + "step": 7496 + }, + { + "epoch": 0.5766479501576802, + "learning_rate": 0.001142285467780074, + "loss": 1.3273, + "step": 7497 + }, + { + "epoch": 0.5767248673178986, + "learning_rate": 0.001141933472337947, + "loss": 1.635, + "step": 7498 + }, + { + "epoch": 0.576801784478117, + "learning_rate": 0.0011415814978036978, + "loss": 1.0022, + "step": 7499 + }, + { + "epoch": 0.5768787016383355, + "learning_rate": 0.001141229544197878, + "loss": 1.1319, + "step": 7500 + }, + { + "epoch": 0.576955618798554, + "learning_rate": 0.0011408776115410396, + "loss": 1.3035, + "step": 7501 + }, + { + "epoch": 0.5770325359587724, + "learning_rate": 0.0011405256998537317, + "loss": 1.3327, + "step": 7502 + }, + { + "epoch": 0.5771094531189909, + "learning_rate": 0.001140173809156502, + "loss": 1.1914, + "step": 7503 + }, + { + "epoch": 0.5771863702792093, + "learning_rate": 0.0011398219394698995, + "loss": 1.0573, + "step": 7504 + }, + { + "epoch": 0.5772632874394278, + "learning_rate": 0.0011394700908144684, + "loss": 1.3627, + "step": 7505 + }, + { + "epoch": 0.5773402045996462, + "learning_rate": 0.0011391182632107554, + "loss": 0.9205, + "step": 7506 + }, + { + "epoch": 0.5774171217598646, + "learning_rate": 0.001138766456679303, + "loss": 1.3411, + "step": 7507 + }, + { + "epoch": 0.5774940389200831, + "learning_rate": 0.001138414671240653, + "loss": 1.5169, + "step": 7508 + }, + { + "epoch": 0.5775709560803015, + "learning_rate": 0.0011380629069153477, + "loss": 1.1586, + "step": 7509 + }, + { + "epoch": 0.5776478732405199, + "learning_rate": 0.0011377111637239264, + "loss": 0.977, + "step": 7510 + }, + { + "epoch": 0.5777247904007384, + "learning_rate": 0.0011373594416869274, + "loss": 0.8814, + "step": 7511 + }, + { + "epoch": 0.5778017075609568, + "learning_rate": 0.001137007740824889, + "loss": 1.3461, + "step": 7512 + }, + { + "epoch": 0.5778786247211753, + "learning_rate": 0.0011366560611583466, + "loss": 0.9412, + "step": 7513 + }, + { + "epoch": 0.5779555418813938, + "learning_rate": 0.0011363044027078357, + "loss": 1.235, + "step": 7514 + }, + { + "epoch": 0.5780324590416122, + "learning_rate": 0.0011359527654938902, + "loss": 1.1531, + "step": 7515 + }, + { + "epoch": 0.5781093762018307, + "learning_rate": 0.0011356011495370411, + "loss": 1.1962, + "step": 7516 + }, + { + "epoch": 0.5781862933620491, + "learning_rate": 0.0011352495548578216, + "loss": 1.1847, + "step": 7517 + }, + { + "epoch": 0.5782632105222675, + "learning_rate": 0.0011348979814767603, + "loss": 0.9579, + "step": 7518 + }, + { + "epoch": 0.578340127682486, + "learning_rate": 0.0011345464294143862, + "loss": 0.8197, + "step": 7519 + }, + { + "epoch": 0.5784170448427044, + "learning_rate": 0.0011341948986912272, + "loss": 1.2128, + "step": 7520 + }, + { + "epoch": 0.5784939620029228, + "learning_rate": 0.0011338433893278094, + "loss": 1.2701, + "step": 7521 + }, + { + "epoch": 0.5785708791631413, + "learning_rate": 0.0011334919013446573, + "loss": 1.0184, + "step": 7522 + }, + { + "epoch": 0.5786477963233597, + "learning_rate": 0.0011331404347622953, + "loss": 1.5434, + "step": 7523 + }, + { + "epoch": 0.5787247134835782, + "learning_rate": 0.0011327889896012454, + "loss": 0.9095, + "step": 7524 + }, + { + "epoch": 0.5788016306437966, + "learning_rate": 0.0011324375658820292, + "loss": 1.3754, + "step": 7525 + }, + { + "epoch": 0.578878547804015, + "learning_rate": 0.0011320861636251673, + "loss": 1.1569, + "step": 7526 + }, + { + "epoch": 0.5789554649642336, + "learning_rate": 0.0011317347828511764, + "loss": 1.0138, + "step": 7527 + }, + { + "epoch": 0.579032382124452, + "learning_rate": 0.0011313834235805766, + "loss": 1.212, + "step": 7528 + }, + { + "epoch": 0.5791092992846704, + "learning_rate": 0.0011310320858338822, + "loss": 1.4247, + "step": 7529 + }, + { + "epoch": 0.5791862164448889, + "learning_rate": 0.0011306807696316089, + "loss": 1.0551, + "step": 7530 + }, + { + "epoch": 0.5792631336051073, + "learning_rate": 0.0011303294749942705, + "loss": 1.189, + "step": 7531 + }, + { + "epoch": 0.5793400507653258, + "learning_rate": 0.0011299782019423789, + "loss": 1.4765, + "step": 7532 + }, + { + "epoch": 0.5794169679255442, + "learning_rate": 0.0011296269504964462, + "loss": 1.2702, + "step": 7533 + }, + { + "epoch": 0.5794938850857626, + "learning_rate": 0.0011292757206769817, + "loss": 1.2545, + "step": 7534 + }, + { + "epoch": 0.5795708022459811, + "learning_rate": 0.001128924512504494, + "loss": 1.1158, + "step": 7535 + }, + { + "epoch": 0.5796477194061995, + "learning_rate": 0.0011285733259994912, + "loss": 0.7285, + "step": 7536 + }, + { + "epoch": 0.5797246365664179, + "learning_rate": 0.0011282221611824792, + "loss": 1.5378, + "step": 7537 + }, + { + "epoch": 0.5798015537266364, + "learning_rate": 0.0011278710180739617, + "loss": 1.0237, + "step": 7538 + }, + { + "epoch": 0.5798784708868548, + "learning_rate": 0.0011275198966944442, + "loss": 1.0707, + "step": 7539 + }, + { + "epoch": 0.5799553880470733, + "learning_rate": 0.0011271687970644277, + "loss": 0.9255, + "step": 7540 + }, + { + "epoch": 0.5800323052072918, + "learning_rate": 0.0011268177192044135, + "loss": 1.1747, + "step": 7541 + }, + { + "epoch": 0.5801092223675102, + "learning_rate": 0.0011264666631349018, + "loss": 1.2695, + "step": 7542 + }, + { + "epoch": 0.5801861395277287, + "learning_rate": 0.0011261156288763909, + "loss": 1.0433, + "step": 7543 + }, + { + "epoch": 0.5802630566879471, + "learning_rate": 0.001125764616449378, + "loss": 0.8894, + "step": 7544 + }, + { + "epoch": 0.5803399738481655, + "learning_rate": 0.0011254136258743595, + "loss": 1.4501, + "step": 7545 + }, + { + "epoch": 0.580416891008384, + "learning_rate": 0.0011250626571718293, + "loss": 1.0785, + "step": 7546 + }, + { + "epoch": 0.5804938081686024, + "learning_rate": 0.0011247117103622816, + "loss": 1.2745, + "step": 7547 + }, + { + "epoch": 0.5805707253288208, + "learning_rate": 0.0011243607854662086, + "loss": 1.278, + "step": 7548 + }, + { + "epoch": 0.5806476424890393, + "learning_rate": 0.0011240098825040998, + "loss": 1.1555, + "step": 7549 + }, + { + "epoch": 0.5807245596492577, + "learning_rate": 0.001123659001496447, + "loss": 1.2342, + "step": 7550 + }, + { + "epoch": 0.5808014768094762, + "learning_rate": 0.001123308142463736, + "loss": 0.9779, + "step": 7551 + }, + { + "epoch": 0.5808783939696947, + "learning_rate": 0.0011229573054264564, + "loss": 1.3377, + "step": 7552 + }, + { + "epoch": 0.580955311129913, + "learning_rate": 0.0011226064904050924, + "loss": 0.8546, + "step": 7553 + }, + { + "epoch": 0.5810322282901316, + "learning_rate": 0.0011222556974201281, + "loss": 1.1539, + "step": 7554 + }, + { + "epoch": 0.58110914545035, + "learning_rate": 0.0011219049264920478, + "loss": 0.819, + "step": 7555 + }, + { + "epoch": 0.5811860626105684, + "learning_rate": 0.001121554177641333, + "loss": 1.2979, + "step": 7556 + }, + { + "epoch": 0.5812629797707869, + "learning_rate": 0.0011212034508884638, + "loss": 0.7625, + "step": 7557 + }, + { + "epoch": 0.5813398969310053, + "learning_rate": 0.00112085274625392, + "loss": 0.8487, + "step": 7558 + }, + { + "epoch": 0.5814168140912237, + "learning_rate": 0.0011205020637581795, + "loss": 1.195, + "step": 7559 + }, + { + "epoch": 0.5814937312514422, + "learning_rate": 0.001120151403421719, + "loss": 1.2928, + "step": 7560 + }, + { + "epoch": 0.5815706484116606, + "learning_rate": 0.0011198007652650143, + "loss": 1.268, + "step": 7561 + }, + { + "epoch": 0.5816475655718791, + "learning_rate": 0.001119450149308538, + "loss": 1.115, + "step": 7562 + }, + { + "epoch": 0.5817244827320975, + "learning_rate": 0.0011190995555727652, + "loss": 1.4845, + "step": 7563 + }, + { + "epoch": 0.5818013998923159, + "learning_rate": 0.0011187489840781663, + "loss": 0.9678, + "step": 7564 + }, + { + "epoch": 0.5818783170525345, + "learning_rate": 0.0011183984348452109, + "loss": 1.1935, + "step": 7565 + }, + { + "epoch": 0.5819552342127529, + "learning_rate": 0.0011180479078943691, + "loss": 1.3641, + "step": 7566 + }, + { + "epoch": 0.5820321513729713, + "learning_rate": 0.001117697403246108, + "loss": 1.0703, + "step": 7567 + }, + { + "epoch": 0.5821090685331898, + "learning_rate": 0.0011173469209208935, + "loss": 1.2778, + "step": 7568 + }, + { + "epoch": 0.5821859856934082, + "learning_rate": 0.0011169964609391914, + "loss": 1.0941, + "step": 7569 + }, + { + "epoch": 0.5822629028536267, + "learning_rate": 0.0011166460233214648, + "loss": 1.2988, + "step": 7570 + }, + { + "epoch": 0.5823398200138451, + "learning_rate": 0.0011162956080881768, + "loss": 1.1732, + "step": 7571 + }, + { + "epoch": 0.5824167371740635, + "learning_rate": 0.0011159452152597882, + "loss": 1.1327, + "step": 7572 + }, + { + "epoch": 0.582493654334282, + "learning_rate": 0.0011155948448567585, + "loss": 1.192, + "step": 7573 + }, + { + "epoch": 0.5825705714945004, + "learning_rate": 0.0011152444968995465, + "loss": 1.1793, + "step": 7574 + }, + { + "epoch": 0.5826474886547188, + "learning_rate": 0.0011148941714086097, + "loss": 0.7788, + "step": 7575 + }, + { + "epoch": 0.5827244058149373, + "learning_rate": 0.0011145438684044028, + "loss": 1.1694, + "step": 7576 + }, + { + "epoch": 0.5828013229751557, + "learning_rate": 0.0011141935879073817, + "loss": 0.776, + "step": 7577 + }, + { + "epoch": 0.5828782401353741, + "learning_rate": 0.0011138433299379988, + "loss": 1.3314, + "step": 7578 + }, + { + "epoch": 0.5829551572955927, + "learning_rate": 0.0011134930945167064, + "loss": 1.2636, + "step": 7579 + }, + { + "epoch": 0.5830320744558111, + "learning_rate": 0.001113142881663955, + "loss": 1.3117, + "step": 7580 + }, + { + "epoch": 0.5831089916160296, + "learning_rate": 0.0011127926914001936, + "loss": 1.4092, + "step": 7581 + }, + { + "epoch": 0.583185908776248, + "learning_rate": 0.001112442523745871, + "loss": 1.1105, + "step": 7582 + }, + { + "epoch": 0.5832628259364664, + "learning_rate": 0.0011120923787214332, + "loss": 1.3215, + "step": 7583 + }, + { + "epoch": 0.5833397430966849, + "learning_rate": 0.0011117422563473253, + "loss": 1.4006, + "step": 7584 + }, + { + "epoch": 0.5834166602569033, + "learning_rate": 0.0011113921566439922, + "loss": 1.3776, + "step": 7585 + }, + { + "epoch": 0.5834935774171217, + "learning_rate": 0.0011110420796318764, + "loss": 1.3827, + "step": 7586 + }, + { + "epoch": 0.5835704945773402, + "learning_rate": 0.0011106920253314179, + "loss": 1.1859, + "step": 7587 + }, + { + "epoch": 0.5836474117375586, + "learning_rate": 0.0011103419937630585, + "loss": 1.1144, + "step": 7588 + }, + { + "epoch": 0.5837243288977771, + "learning_rate": 0.0011099919849472358, + "loss": 1.0833, + "step": 7589 + }, + { + "epoch": 0.5838012460579955, + "learning_rate": 0.001109641998904388, + "loss": 1.1369, + "step": 7590 + }, + { + "epoch": 0.583878163218214, + "learning_rate": 0.0011092920356549504, + "loss": 0.9468, + "step": 7591 + }, + { + "epoch": 0.5839550803784325, + "learning_rate": 0.0011089420952193577, + "loss": 0.9523, + "step": 7592 + }, + { + "epoch": 0.5840319975386509, + "learning_rate": 0.0011085921776180442, + "loss": 0.9987, + "step": 7593 + }, + { + "epoch": 0.5841089146988693, + "learning_rate": 0.0011082422828714413, + "loss": 1.3263, + "step": 7594 + }, + { + "epoch": 0.5841858318590878, + "learning_rate": 0.0011078924109999791, + "loss": 0.7874, + "step": 7595 + }, + { + "epoch": 0.5842627490193062, + "learning_rate": 0.0011075425620240885, + "loss": 0.7541, + "step": 7596 + }, + { + "epoch": 0.5843396661795246, + "learning_rate": 0.0011071927359641968, + "loss": 1.4845, + "step": 7597 + }, + { + "epoch": 0.5844165833397431, + "learning_rate": 0.0011068429328407296, + "loss": 1.0559, + "step": 7598 + }, + { + "epoch": 0.5844935004999615, + "learning_rate": 0.0011064931526741143, + "loss": 1.1676, + "step": 7599 + }, + { + "epoch": 0.58457041766018, + "learning_rate": 0.0011061433954847732, + "loss": 1.2336, + "step": 7600 + }, + { + "epoch": 0.5846473348203984, + "learning_rate": 0.0011057936612931302, + "loss": 1.1726, + "step": 7601 + }, + { + "epoch": 0.5847242519806168, + "learning_rate": 0.001105443950119606, + "loss": 1.1307, + "step": 7602 + }, + { + "epoch": 0.5848011691408354, + "learning_rate": 0.0011050942619846205, + "loss": 1.2726, + "step": 7603 + }, + { + "epoch": 0.5848780863010538, + "learning_rate": 0.001104744596908593, + "loss": 0.8969, + "step": 7604 + }, + { + "epoch": 0.5849550034612722, + "learning_rate": 0.0011043949549119404, + "loss": 1.1497, + "step": 7605 + }, + { + "epoch": 0.5850319206214907, + "learning_rate": 0.001104045336015078, + "loss": 1.1684, + "step": 7606 + }, + { + "epoch": 0.5851088377817091, + "learning_rate": 0.0011036957402384218, + "loss": 1.2694, + "step": 7607 + }, + { + "epoch": 0.5851857549419276, + "learning_rate": 0.0011033461676023839, + "loss": 1.4821, + "step": 7608 + }, + { + "epoch": 0.585262672102146, + "learning_rate": 0.001102996618127377, + "loss": 0.9267, + "step": 7609 + }, + { + "epoch": 0.5853395892623644, + "learning_rate": 0.0011026470918338116, + "loss": 1.0763, + "step": 7610 + }, + { + "epoch": 0.5854165064225829, + "learning_rate": 0.0011022975887420958, + "loss": 1.2333, + "step": 7611 + }, + { + "epoch": 0.5854934235828013, + "learning_rate": 0.001101948108872639, + "loss": 1.174, + "step": 7612 + }, + { + "epoch": 0.5855703407430197, + "learning_rate": 0.0011015986522458466, + "loss": 1.4906, + "step": 7613 + }, + { + "epoch": 0.5856472579032382, + "learning_rate": 0.001101249218882124, + "loss": 0.809, + "step": 7614 + }, + { + "epoch": 0.5857241750634566, + "learning_rate": 0.0011008998088018753, + "loss": 1.0597, + "step": 7615 + }, + { + "epoch": 0.5858010922236752, + "learning_rate": 0.0011005504220255025, + "loss": 1.1544, + "step": 7616 + }, + { + "epoch": 0.5858780093838936, + "learning_rate": 0.001100201058573407, + "loss": 1.1005, + "step": 7617 + }, + { + "epoch": 0.585954926544112, + "learning_rate": 0.0010998517184659885, + "loss": 1.0385, + "step": 7618 + }, + { + "epoch": 0.5860318437043305, + "learning_rate": 0.0010995024017236448, + "loss": 1.2724, + "step": 7619 + }, + { + "epoch": 0.5861087608645489, + "learning_rate": 0.0010991531083667735, + "loss": 1.1509, + "step": 7620 + }, + { + "epoch": 0.5861856780247673, + "learning_rate": 0.0010988038384157704, + "loss": 1.073, + "step": 7621 + }, + { + "epoch": 0.5862625951849858, + "learning_rate": 0.0010984545918910282, + "loss": 1.2951, + "step": 7622 + }, + { + "epoch": 0.5863395123452042, + "learning_rate": 0.0010981053688129418, + "loss": 1.0164, + "step": 7623 + }, + { + "epoch": 0.5864164295054226, + "learning_rate": 0.0010977561692019017, + "loss": 1.2168, + "step": 7624 + }, + { + "epoch": 0.5864933466656411, + "learning_rate": 0.0010974069930782975, + "loss": 1.3536, + "step": 7625 + }, + { + "epoch": 0.5865702638258595, + "learning_rate": 0.001097057840462519, + "loss": 1.263, + "step": 7626 + }, + { + "epoch": 0.586647180986078, + "learning_rate": 0.001096708711374953, + "loss": 1.0746, + "step": 7627 + }, + { + "epoch": 0.5867240981462964, + "learning_rate": 0.0010963596058359854, + "loss": 0.9617, + "step": 7628 + }, + { + "epoch": 0.5868010153065149, + "learning_rate": 0.0010960105238660015, + "loss": 1.0653, + "step": 7629 + }, + { + "epoch": 0.5868779324667334, + "learning_rate": 0.0010956614654853837, + "loss": 1.0892, + "step": 7630 + }, + { + "epoch": 0.5869548496269518, + "learning_rate": 0.0010953124307145146, + "loss": 1.2078, + "step": 7631 + }, + { + "epoch": 0.5870317667871702, + "learning_rate": 0.0010949634195737748, + "loss": 1.3924, + "step": 7632 + }, + { + "epoch": 0.5871086839473887, + "learning_rate": 0.0010946144320835422, + "loss": 1.1187, + "step": 7633 + }, + { + "epoch": 0.5871856011076071, + "learning_rate": 0.0010942654682641962, + "loss": 1.0561, + "step": 7634 + }, + { + "epoch": 0.5872625182678256, + "learning_rate": 0.0010939165281361116, + "loss": 1.2642, + "step": 7635 + }, + { + "epoch": 0.587339435428044, + "learning_rate": 0.0010935676117196649, + "loss": 1.2873, + "step": 7636 + }, + { + "epoch": 0.5874163525882624, + "learning_rate": 0.0010932187190352287, + "loss": 1.3543, + "step": 7637 + }, + { + "epoch": 0.5874932697484809, + "learning_rate": 0.0010928698501031751, + "loss": 1.1779, + "step": 7638 + }, + { + "epoch": 0.5875701869086993, + "learning_rate": 0.0010925210049438755, + "loss": 0.9547, + "step": 7639 + }, + { + "epoch": 0.5876471040689177, + "learning_rate": 0.001092172183577699, + "loss": 1.0573, + "step": 7640 + }, + { + "epoch": 0.5877240212291363, + "learning_rate": 0.0010918233860250136, + "loss": 1.068, + "step": 7641 + }, + { + "epoch": 0.5878009383893547, + "learning_rate": 0.0010914746123061865, + "loss": 0.7642, + "step": 7642 + }, + { + "epoch": 0.5878778555495731, + "learning_rate": 0.0010911258624415826, + "loss": 1.1254, + "step": 7643 + }, + { + "epoch": 0.5879547727097916, + "learning_rate": 0.001090777136451565, + "loss": 1.0528, + "step": 7644 + }, + { + "epoch": 0.58803168987001, + "learning_rate": 0.0010904284343564974, + "loss": 1.0813, + "step": 7645 + }, + { + "epoch": 0.5881086070302285, + "learning_rate": 0.0010900797561767398, + "loss": 0.8347, + "step": 7646 + }, + { + "epoch": 0.5881855241904469, + "learning_rate": 0.0010897311019326533, + "loss": 1.137, + "step": 7647 + }, + { + "epoch": 0.5882624413506653, + "learning_rate": 0.0010893824716445952, + "loss": 0.9631, + "step": 7648 + }, + { + "epoch": 0.5883393585108838, + "learning_rate": 0.001089033865332922, + "loss": 1.2, + "step": 7649 + }, + { + "epoch": 0.5884162756711022, + "learning_rate": 0.00108868528301799, + "loss": 1.1167, + "step": 7650 + }, + { + "epoch": 0.5884931928313206, + "learning_rate": 0.0010883367247201529, + "loss": 1.158, + "step": 7651 + }, + { + "epoch": 0.5885701099915391, + "learning_rate": 0.0010879881904597634, + "loss": 1.1843, + "step": 7652 + }, + { + "epoch": 0.5886470271517575, + "learning_rate": 0.001087639680257173, + "loss": 0.872, + "step": 7653 + }, + { + "epoch": 0.588723944311976, + "learning_rate": 0.0010872911941327312, + "loss": 1.4038, + "step": 7654 + }, + { + "epoch": 0.5888008614721945, + "learning_rate": 0.001086942732106787, + "loss": 1.2358, + "step": 7655 + }, + { + "epoch": 0.5888777786324129, + "learning_rate": 0.0010865942941996874, + "loss": 1.1987, + "step": 7656 + }, + { + "epoch": 0.5889546957926314, + "learning_rate": 0.0010862458804317768, + "loss": 1.3849, + "step": 7657 + }, + { + "epoch": 0.5890316129528498, + "learning_rate": 0.0010858974908234014, + "loss": 0.9745, + "step": 7658 + }, + { + "epoch": 0.5891085301130682, + "learning_rate": 0.0010855491253949027, + "loss": 1.3895, + "step": 7659 + }, + { + "epoch": 0.5891854472732867, + "learning_rate": 0.0010852007841666224, + "loss": 1.146, + "step": 7660 + }, + { + "epoch": 0.5892623644335051, + "learning_rate": 0.0010848524671589008, + "loss": 1.1424, + "step": 7661 + }, + { + "epoch": 0.5893392815937235, + "learning_rate": 0.0010845041743920763, + "loss": 1.1147, + "step": 7662 + }, + { + "epoch": 0.589416198753942, + "learning_rate": 0.0010841559058864857, + "loss": 0.9315, + "step": 7663 + }, + { + "epoch": 0.5894931159141604, + "learning_rate": 0.0010838076616624657, + "loss": 1.2049, + "step": 7664 + }, + { + "epoch": 0.5895700330743789, + "learning_rate": 0.0010834594417403495, + "loss": 0.9668, + "step": 7665 + }, + { + "epoch": 0.5896469502345973, + "learning_rate": 0.0010831112461404712, + "loss": 0.8758, + "step": 7666 + }, + { + "epoch": 0.5897238673948157, + "learning_rate": 0.0010827630748831618, + "loss": 0.9744, + "step": 7667 + }, + { + "epoch": 0.5898007845550343, + "learning_rate": 0.0010824149279887504, + "loss": 1.1687, + "step": 7668 + }, + { + "epoch": 0.5898777017152527, + "learning_rate": 0.0010820668054775675, + "loss": 1.3187, + "step": 7669 + }, + { + "epoch": 0.5899546188754711, + "learning_rate": 0.0010817187073699388, + "loss": 0.8032, + "step": 7670 + }, + { + "epoch": 0.5900315360356896, + "learning_rate": 0.0010813706336861909, + "loss": 1.0688, + "step": 7671 + }, + { + "epoch": 0.590108453195908, + "learning_rate": 0.001081022584446648, + "loss": 0.9207, + "step": 7672 + }, + { + "epoch": 0.5901853703561265, + "learning_rate": 0.0010806745596716328, + "loss": 1.4195, + "step": 7673 + }, + { + "epoch": 0.5902622875163449, + "learning_rate": 0.0010803265593814673, + "loss": 1.1843, + "step": 7674 + }, + { + "epoch": 0.5903392046765633, + "learning_rate": 0.0010799785835964714, + "loss": 1.6493, + "step": 7675 + }, + { + "epoch": 0.5904161218367818, + "learning_rate": 0.0010796306323369634, + "loss": 1.2095, + "step": 7676 + }, + { + "epoch": 0.5904930389970002, + "learning_rate": 0.0010792827056232614, + "loss": 1.1495, + "step": 7677 + }, + { + "epoch": 0.5905699561572186, + "learning_rate": 0.0010789348034756809, + "loss": 1.1788, + "step": 7678 + }, + { + "epoch": 0.5906468733174371, + "learning_rate": 0.0010785869259145348, + "loss": 1.1645, + "step": 7679 + }, + { + "epoch": 0.5907237904776556, + "learning_rate": 0.0010782390729601385, + "loss": 1.136, + "step": 7680 + }, + { + "epoch": 0.590800707637874, + "learning_rate": 0.0010778912446328022, + "loss": 0.9883, + "step": 7681 + }, + { + "epoch": 0.5908776247980925, + "learning_rate": 0.0010775434409528353, + "loss": 1.2706, + "step": 7682 + }, + { + "epoch": 0.5909545419583109, + "learning_rate": 0.0010771956619405477, + "loss": 1.3735, + "step": 7683 + }, + { + "epoch": 0.5910314591185294, + "learning_rate": 0.0010768479076162457, + "loss": 1.0717, + "step": 7684 + }, + { + "epoch": 0.5911083762787478, + "learning_rate": 0.0010765001780002356, + "loss": 1.087, + "step": 7685 + }, + { + "epoch": 0.5911852934389662, + "learning_rate": 0.0010761524731128214, + "loss": 1.3048, + "step": 7686 + }, + { + "epoch": 0.5912622105991847, + "learning_rate": 0.0010758047929743059, + "loss": 1.2326, + "step": 7687 + }, + { + "epoch": 0.5913391277594031, + "learning_rate": 0.001075457137604991, + "loss": 0.9792, + "step": 7688 + }, + { + "epoch": 0.5914160449196215, + "learning_rate": 0.0010751095070251763, + "loss": 0.9589, + "step": 7689 + }, + { + "epoch": 0.59149296207984, + "learning_rate": 0.0010747619012551595, + "loss": 1.0688, + "step": 7690 + }, + { + "epoch": 0.5915698792400584, + "learning_rate": 0.0010744143203152395, + "loss": 1.0496, + "step": 7691 + }, + { + "epoch": 0.591646796400277, + "learning_rate": 0.0010740667642257097, + "loss": 1.4762, + "step": 7692 + }, + { + "epoch": 0.5917237135604954, + "learning_rate": 0.0010737192330068665, + "loss": 1.1987, + "step": 7693 + }, + { + "epoch": 0.5918006307207138, + "learning_rate": 0.0010733717266790013, + "loss": 0.9884, + "step": 7694 + }, + { + "epoch": 0.5918775478809323, + "learning_rate": 0.0010730242452624051, + "loss": 1.2176, + "step": 7695 + }, + { + "epoch": 0.5919544650411507, + "learning_rate": 0.0010726767887773688, + "loss": 1.1729, + "step": 7696 + }, + { + "epoch": 0.5920313822013691, + "learning_rate": 0.0010723293572441799, + "loss": 1.1773, + "step": 7697 + }, + { + "epoch": 0.5921082993615876, + "learning_rate": 0.0010719819506831252, + "loss": 1.1422, + "step": 7698 + }, + { + "epoch": 0.592185216521806, + "learning_rate": 0.0010716345691144911, + "loss": 1.1546, + "step": 7699 + }, + { + "epoch": 0.5922621336820244, + "learning_rate": 0.0010712872125585608, + "loss": 0.9597, + "step": 7700 + }, + { + "epoch": 0.5923390508422429, + "learning_rate": 0.0010709398810356163, + "loss": 1.353, + "step": 7701 + }, + { + "epoch": 0.5924159680024613, + "learning_rate": 0.0010705925745659404, + "loss": 1.2339, + "step": 7702 + }, + { + "epoch": 0.5924928851626798, + "learning_rate": 0.0010702452931698105, + "loss": 1.377, + "step": 7703 + }, + { + "epoch": 0.5925698023228982, + "learning_rate": 0.0010698980368675066, + "loss": 1.0627, + "step": 7704 + }, + { + "epoch": 0.5926467194831166, + "learning_rate": 0.0010695508056793044, + "loss": 1.0616, + "step": 7705 + }, + { + "epoch": 0.5927236366433352, + "learning_rate": 0.001069203599625479, + "loss": 1.4626, + "step": 7706 + }, + { + "epoch": 0.5928005538035536, + "learning_rate": 0.0010688564187263043, + "loss": 1.0914, + "step": 7707 + }, + { + "epoch": 0.592877470963772, + "learning_rate": 0.001068509263002053, + "loss": 1.2574, + "step": 7708 + }, + { + "epoch": 0.5929543881239905, + "learning_rate": 0.0010681621324729951, + "loss": 1.2338, + "step": 7709 + }, + { + "epoch": 0.5930313052842089, + "learning_rate": 0.0010678150271594008, + "loss": 1.4186, + "step": 7710 + }, + { + "epoch": 0.5931082224444274, + "learning_rate": 0.001067467947081537, + "loss": 1.1059, + "step": 7711 + }, + { + "epoch": 0.5931851396046458, + "learning_rate": 0.001067120892259671, + "loss": 1.0338, + "step": 7712 + }, + { + "epoch": 0.5932620567648642, + "learning_rate": 0.001066773862714067, + "loss": 0.9603, + "step": 7713 + }, + { + "epoch": 0.5933389739250827, + "learning_rate": 0.0010664268584649886, + "loss": 1.025, + "step": 7714 + }, + { + "epoch": 0.5934158910853011, + "learning_rate": 0.0010660798795326984, + "loss": 1.0605, + "step": 7715 + }, + { + "epoch": 0.5934928082455195, + "learning_rate": 0.0010657329259374558, + "loss": 0.9222, + "step": 7716 + }, + { + "epoch": 0.593569725405738, + "learning_rate": 0.00106538599769952, + "loss": 1.4801, + "step": 7717 + }, + { + "epoch": 0.5936466425659565, + "learning_rate": 0.001065039094839149, + "loss": 1.4249, + "step": 7718 + }, + { + "epoch": 0.5937235597261749, + "learning_rate": 0.0010646922173765986, + "loss": 1.2505, + "step": 7719 + }, + { + "epoch": 0.5938004768863934, + "learning_rate": 0.0010643453653321228, + "loss": 1.09, + "step": 7720 + }, + { + "epoch": 0.5938773940466118, + "learning_rate": 0.0010639985387259757, + "loss": 0.8344, + "step": 7721 + }, + { + "epoch": 0.5939543112068303, + "learning_rate": 0.001063651737578408, + "loss": 0.9064, + "step": 7722 + }, + { + "epoch": 0.5940312283670487, + "learning_rate": 0.0010633049619096703, + "loss": 1.3103, + "step": 7723 + }, + { + "epoch": 0.5941081455272671, + "learning_rate": 0.001062958211740011, + "loss": 1.344, + "step": 7724 + }, + { + "epoch": 0.5941850626874856, + "learning_rate": 0.001062611487089677, + "loss": 1.3596, + "step": 7725 + }, + { + "epoch": 0.594261979847704, + "learning_rate": 0.0010622647879789144, + "loss": 1.2105, + "step": 7726 + }, + { + "epoch": 0.5943388970079224, + "learning_rate": 0.0010619181144279673, + "loss": 1.4586, + "step": 7727 + }, + { + "epoch": 0.5944158141681409, + "learning_rate": 0.0010615714664570773, + "loss": 1.382, + "step": 7728 + }, + { + "epoch": 0.5944927313283593, + "learning_rate": 0.0010612248440864866, + "loss": 1.0237, + "step": 7729 + }, + { + "epoch": 0.5945696484885779, + "learning_rate": 0.0010608782473364344, + "loss": 1.4369, + "step": 7730 + }, + { + "epoch": 0.5946465656487963, + "learning_rate": 0.001060531676227159, + "loss": 1.1992, + "step": 7731 + }, + { + "epoch": 0.5947234828090147, + "learning_rate": 0.0010601851307788974, + "loss": 1.3042, + "step": 7732 + }, + { + "epoch": 0.5948003999692332, + "learning_rate": 0.001059838611011884, + "loss": 0.9944, + "step": 7733 + }, + { + "epoch": 0.5948773171294516, + "learning_rate": 0.001059492116946353, + "loss": 1.1956, + "step": 7734 + }, + { + "epoch": 0.59495423428967, + "learning_rate": 0.0010591456486025363, + "loss": 1.5136, + "step": 7735 + }, + { + "epoch": 0.5950311514498885, + "learning_rate": 0.0010587992060006645, + "loss": 1.3086, + "step": 7736 + }, + { + "epoch": 0.5951080686101069, + "learning_rate": 0.0010584527891609671, + "loss": 0.9448, + "step": 7737 + }, + { + "epoch": 0.5951849857703254, + "learning_rate": 0.0010581063981036712, + "loss": 1.2859, + "step": 7738 + }, + { + "epoch": 0.5952619029305438, + "learning_rate": 0.0010577600328490037, + "loss": 1.4452, + "step": 7739 + }, + { + "epoch": 0.5953388200907622, + "learning_rate": 0.001057413693417189, + "loss": 1.2173, + "step": 7740 + }, + { + "epoch": 0.5954157372509807, + "learning_rate": 0.0010570673798284492, + "loss": 1.3203, + "step": 7741 + }, + { + "epoch": 0.5954926544111991, + "learning_rate": 0.0010567210921030071, + "loss": 1.053, + "step": 7742 + }, + { + "epoch": 0.5955695715714175, + "learning_rate": 0.0010563748302610824, + "loss": 1.629, + "step": 7743 + }, + { + "epoch": 0.5956464887316361, + "learning_rate": 0.0010560285943228935, + "loss": 1.1092, + "step": 7744 + }, + { + "epoch": 0.5957234058918545, + "learning_rate": 0.001055682384308658, + "loss": 1.1787, + "step": 7745 + }, + { + "epoch": 0.5958003230520729, + "learning_rate": 0.0010553362002385912, + "loss": 1.0458, + "step": 7746 + }, + { + "epoch": 0.5958772402122914, + "learning_rate": 0.0010549900421329065, + "loss": 1.1286, + "step": 7747 + }, + { + "epoch": 0.5959541573725098, + "learning_rate": 0.0010546439100118173, + "loss": 1.1391, + "step": 7748 + }, + { + "epoch": 0.5960310745327283, + "learning_rate": 0.0010542978038955342, + "loss": 1.2542, + "step": 7749 + }, + { + "epoch": 0.5961079916929467, + "learning_rate": 0.0010539517238042672, + "loss": 1.3272, + "step": 7750 + }, + { + "epoch": 0.5961849088531651, + "learning_rate": 0.0010536056697582241, + "loss": 1.249, + "step": 7751 + }, + { + "epoch": 0.5962618260133836, + "learning_rate": 0.0010532596417776102, + "loss": 1.1186, + "step": 7752 + }, + { + "epoch": 0.596338743173602, + "learning_rate": 0.0010529136398826325, + "loss": 0.9633, + "step": 7753 + }, + { + "epoch": 0.5964156603338204, + "learning_rate": 0.001052567664093493, + "loss": 1.0394, + "step": 7754 + }, + { + "epoch": 0.596492577494039, + "learning_rate": 0.0010522217144303934, + "loss": 0.9104, + "step": 7755 + }, + { + "epoch": 0.5965694946542573, + "learning_rate": 0.0010518757909135348, + "loss": 1.4556, + "step": 7756 + }, + { + "epoch": 0.5966464118144759, + "learning_rate": 0.0010515298935631157, + "loss": 1.1109, + "step": 7757 + }, + { + "epoch": 0.5967233289746943, + "learning_rate": 0.0010511840223993338, + "loss": 0.9677, + "step": 7758 + }, + { + "epoch": 0.5968002461349127, + "learning_rate": 0.0010508381774423847, + "loss": 1.2739, + "step": 7759 + }, + { + "epoch": 0.5968771632951312, + "learning_rate": 0.001050492358712462, + "loss": 1.031, + "step": 7760 + }, + { + "epoch": 0.5969540804553496, + "learning_rate": 0.0010501465662297591, + "loss": 1.2035, + "step": 7761 + }, + { + "epoch": 0.597030997615568, + "learning_rate": 0.0010498008000144677, + "loss": 0.935, + "step": 7762 + }, + { + "epoch": 0.5971079147757865, + "learning_rate": 0.0010494550600867755, + "loss": 0.9552, + "step": 7763 + }, + { + "epoch": 0.5971848319360049, + "learning_rate": 0.001049109346466873, + "loss": 1.1457, + "step": 7764 + }, + { + "epoch": 0.5972617490962233, + "learning_rate": 0.0010487636591749454, + "loss": 1.3244, + "step": 7765 + }, + { + "epoch": 0.5973386662564418, + "learning_rate": 0.0010484179982311775, + "loss": 0.9031, + "step": 7766 + }, + { + "epoch": 0.5974155834166602, + "learning_rate": 0.0010480723636557539, + "loss": 1.2468, + "step": 7767 + }, + { + "epoch": 0.5974925005768787, + "learning_rate": 0.0010477267554688556, + "loss": 1.4144, + "step": 7768 + }, + { + "epoch": 0.5975694177370972, + "learning_rate": 0.0010473811736906634, + "loss": 1.0551, + "step": 7769 + }, + { + "epoch": 0.5976463348973156, + "learning_rate": 0.0010470356183413566, + "loss": 1.4656, + "step": 7770 + }, + { + "epoch": 0.5977232520575341, + "learning_rate": 0.0010466900894411117, + "loss": 1.3411, + "step": 7771 + }, + { + "epoch": 0.5978001692177525, + "learning_rate": 0.001046344587010105, + "loss": 1.3289, + "step": 7772 + }, + { + "epoch": 0.5978770863779709, + "learning_rate": 0.0010459991110685109, + "loss": 1.1409, + "step": 7773 + }, + { + "epoch": 0.5979540035381894, + "learning_rate": 0.0010456536616365013, + "loss": 1.211, + "step": 7774 + }, + { + "epoch": 0.5980309206984078, + "learning_rate": 0.0010453082387342487, + "loss": 1.1654, + "step": 7775 + }, + { + "epoch": 0.5981078378586263, + "learning_rate": 0.0010449628423819207, + "loss": 1.2061, + "step": 7776 + }, + { + "epoch": 0.5981847550188447, + "learning_rate": 0.001044617472599688, + "loss": 1.2899, + "step": 7777 + }, + { + "epoch": 0.5982616721790631, + "learning_rate": 0.001044272129407715, + "loss": 1.429, + "step": 7778 + }, + { + "epoch": 0.5983385893392816, + "learning_rate": 0.001043926812826167, + "loss": 1.2076, + "step": 7779 + }, + { + "epoch": 0.5984155064995, + "learning_rate": 0.0010435815228752082, + "loss": 1.0541, + "step": 7780 + }, + { + "epoch": 0.5984924236597184, + "learning_rate": 0.0010432362595749996, + "loss": 1.0445, + "step": 7781 + }, + { + "epoch": 0.598569340819937, + "learning_rate": 0.0010428910229457018, + "loss": 1.3896, + "step": 7782 + }, + { + "epoch": 0.5986462579801554, + "learning_rate": 0.0010425458130074738, + "loss": 1.277, + "step": 7783 + }, + { + "epoch": 0.5987231751403738, + "learning_rate": 0.0010422006297804729, + "loss": 0.8614, + "step": 7784 + }, + { + "epoch": 0.5988000923005923, + "learning_rate": 0.0010418554732848535, + "loss": 1.2623, + "step": 7785 + }, + { + "epoch": 0.5988770094608107, + "learning_rate": 0.0010415103435407712, + "loss": 0.9991, + "step": 7786 + }, + { + "epoch": 0.5989539266210292, + "learning_rate": 0.001041165240568377, + "loss": 1.8123, + "step": 7787 + }, + { + "epoch": 0.5990308437812476, + "learning_rate": 0.0010408201643878235, + "loss": 0.9383, + "step": 7788 + }, + { + "epoch": 0.599107760941466, + "learning_rate": 0.0010404751150192586, + "loss": 1.3833, + "step": 7789 + }, + { + "epoch": 0.5991846781016845, + "learning_rate": 0.0010401300924828306, + "loss": 1.1599, + "step": 7790 + }, + { + "epoch": 0.5992615952619029, + "learning_rate": 0.0010397850967986861, + "loss": 1.0424, + "step": 7791 + }, + { + "epoch": 0.5993385124221213, + "learning_rate": 0.001039440127986969, + "loss": 1.0356, + "step": 7792 + }, + { + "epoch": 0.5994154295823398, + "learning_rate": 0.001039095186067823, + "loss": 0.8999, + "step": 7793 + }, + { + "epoch": 0.5994923467425582, + "learning_rate": 0.0010387502710613894, + "loss": 0.7223, + "step": 7794 + }, + { + "epoch": 0.5995692639027768, + "learning_rate": 0.001038405382987808, + "loss": 0.7494, + "step": 7795 + }, + { + "epoch": 0.5996461810629952, + "learning_rate": 0.0010380605218672178, + "loss": 1.0897, + "step": 7796 + }, + { + "epoch": 0.5997230982232136, + "learning_rate": 0.0010377156877197553, + "loss": 0.9918, + "step": 7797 + }, + { + "epoch": 0.5998000153834321, + "learning_rate": 0.0010373708805655546, + "loss": 1.4431, + "step": 7798 + }, + { + "epoch": 0.5998769325436505, + "learning_rate": 0.0010370261004247513, + "loss": 0.9219, + "step": 7799 + }, + { + "epoch": 0.5999538497038689, + "learning_rate": 0.0010366813473174762, + "loss": 1.0544, + "step": 7800 + }, + { + "epoch": 0.6000307668640874, + "learning_rate": 0.00103633662126386, + "loss": 1.3724, + "step": 7801 + }, + { + "epoch": 0.6001076840243058, + "learning_rate": 0.0010359919222840317, + "loss": 1.1616, + "step": 7802 + }, + { + "epoch": 0.6001846011845242, + "learning_rate": 0.0010356472503981187, + "loss": 0.9862, + "step": 7803 + }, + { + "epoch": 0.6002615183447427, + "learning_rate": 0.0010353026056262463, + "loss": 1.2182, + "step": 7804 + }, + { + "epoch": 0.6003384355049611, + "learning_rate": 0.0010349579879885395, + "loss": 1.2488, + "step": 7805 + }, + { + "epoch": 0.6004153526651796, + "learning_rate": 0.0010346133975051204, + "loss": 1.319, + "step": 7806 + }, + { + "epoch": 0.600492269825398, + "learning_rate": 0.0010342688341961098, + "loss": 1.0387, + "step": 7807 + }, + { + "epoch": 0.6005691869856165, + "learning_rate": 0.001033924298081628, + "loss": 0.9835, + "step": 7808 + }, + { + "epoch": 0.600646104145835, + "learning_rate": 0.001033579789181791, + "loss": 0.9648, + "step": 7809 + }, + { + "epoch": 0.6007230213060534, + "learning_rate": 0.0010332353075167174, + "loss": 1.1639, + "step": 7810 + }, + { + "epoch": 0.6007999384662718, + "learning_rate": 0.0010328908531065202, + "loss": 1.1934, + "step": 7811 + }, + { + "epoch": 0.6008768556264903, + "learning_rate": 0.0010325464259713125, + "loss": 0.9779, + "step": 7812 + }, + { + "epoch": 0.6009537727867087, + "learning_rate": 0.0010322020261312066, + "loss": 1.212, + "step": 7813 + }, + { + "epoch": 0.6010306899469272, + "learning_rate": 0.0010318576536063116, + "loss": 0.8679, + "step": 7814 + }, + { + "epoch": 0.6011076071071456, + "learning_rate": 0.0010315133084167363, + "loss": 0.994, + "step": 7815 + }, + { + "epoch": 0.601184524267364, + "learning_rate": 0.0010311689905825872, + "loss": 1.0423, + "step": 7816 + }, + { + "epoch": 0.6012614414275825, + "learning_rate": 0.0010308247001239693, + "loss": 0.9352, + "step": 7817 + }, + { + "epoch": 0.6013383585878009, + "learning_rate": 0.001030480437060986, + "loss": 1.1421, + "step": 7818 + }, + { + "epoch": 0.6014152757480193, + "learning_rate": 0.00103013620141374, + "loss": 0.9548, + "step": 7819 + }, + { + "epoch": 0.6014921929082379, + "learning_rate": 0.0010297919932023296, + "loss": 1.0947, + "step": 7820 + }, + { + "epoch": 0.6015691100684563, + "learning_rate": 0.0010294478124468558, + "loss": 1.029, + "step": 7821 + }, + { + "epoch": 0.6016460272286747, + "learning_rate": 0.0010291036591674144, + "loss": 0.8608, + "step": 7822 + }, + { + "epoch": 0.6017229443888932, + "learning_rate": 0.0010287595333841007, + "loss": 1.2513, + "step": 7823 + }, + { + "epoch": 0.6017998615491116, + "learning_rate": 0.0010284154351170095, + "loss": 1.3662, + "step": 7824 + }, + { + "epoch": 0.6018767787093301, + "learning_rate": 0.0010280713643862319, + "loss": 0.9338, + "step": 7825 + }, + { + "epoch": 0.6019536958695485, + "learning_rate": 0.0010277273212118596, + "loss": 0.8625, + "step": 7826 + }, + { + "epoch": 0.6020306130297669, + "learning_rate": 0.0010273833056139812, + "loss": 1.2494, + "step": 7827 + }, + { + "epoch": 0.6021075301899854, + "learning_rate": 0.001027039317612684, + "loss": 0.8106, + "step": 7828 + }, + { + "epoch": 0.6021844473502038, + "learning_rate": 0.001026695357228054, + "loss": 1.3656, + "step": 7829 + }, + { + "epoch": 0.6022613645104222, + "learning_rate": 0.0010263514244801756, + "loss": 0.9023, + "step": 7830 + }, + { + "epoch": 0.6023382816706407, + "learning_rate": 0.0010260075193891307, + "loss": 0.9513, + "step": 7831 + }, + { + "epoch": 0.6024151988308591, + "learning_rate": 0.0010256636419750014, + "loss": 1.0575, + "step": 7832 + }, + { + "epoch": 0.6024921159910777, + "learning_rate": 0.0010253197922578652, + "loss": 0.9623, + "step": 7833 + }, + { + "epoch": 0.6025690331512961, + "learning_rate": 0.0010249759702578023, + "loss": 1.0004, + "step": 7834 + }, + { + "epoch": 0.6026459503115145, + "learning_rate": 0.001024632175994887, + "loss": 0.9848, + "step": 7835 + }, + { + "epoch": 0.602722867471733, + "learning_rate": 0.0010242884094891942, + "loss": 1.2173, + "step": 7836 + }, + { + "epoch": 0.6027997846319514, + "learning_rate": 0.0010239446707607973, + "loss": 1.2593, + "step": 7837 + }, + { + "epoch": 0.6028767017921698, + "learning_rate": 0.0010236009598297668, + "loss": 0.8232, + "step": 7838 + }, + { + "epoch": 0.6029536189523883, + "learning_rate": 0.0010232572767161728, + "loss": 1.0004, + "step": 7839 + }, + { + "epoch": 0.6030305361126067, + "learning_rate": 0.0010229136214400837, + "loss": 1.1045, + "step": 7840 + }, + { + "epoch": 0.6031074532728252, + "learning_rate": 0.001022569994021565, + "loss": 1.2416, + "step": 7841 + }, + { + "epoch": 0.6031843704330436, + "learning_rate": 0.001022226394480682, + "loss": 0.8683, + "step": 7842 + }, + { + "epoch": 0.603261287593262, + "learning_rate": 0.0010218828228374984, + "loss": 1.2691, + "step": 7843 + }, + { + "epoch": 0.6033382047534805, + "learning_rate": 0.001021539279112074, + "loss": 1.0257, + "step": 7844 + }, + { + "epoch": 0.603415121913699, + "learning_rate": 0.0010211957633244708, + "loss": 1.4517, + "step": 7845 + }, + { + "epoch": 0.6034920390739174, + "learning_rate": 0.0010208522754947453, + "loss": 0.9462, + "step": 7846 + }, + { + "epoch": 0.6035689562341359, + "learning_rate": 0.0010205088156429547, + "loss": 1.1834, + "step": 7847 + }, + { + "epoch": 0.6036458733943543, + "learning_rate": 0.0010201653837891545, + "loss": 1.2009, + "step": 7848 + }, + { + "epoch": 0.6037227905545727, + "learning_rate": 0.0010198219799533973, + "loss": 1.239, + "step": 7849 + }, + { + "epoch": 0.6037997077147912, + "learning_rate": 0.001019478604155735, + "loss": 1.2887, + "step": 7850 + }, + { + "epoch": 0.6038766248750096, + "learning_rate": 0.0010191352564162181, + "loss": 1.0743, + "step": 7851 + }, + { + "epoch": 0.6039535420352281, + "learning_rate": 0.0010187919367548944, + "loss": 0.8909, + "step": 7852 + }, + { + "epoch": 0.6040304591954465, + "learning_rate": 0.0010184486451918114, + "loss": 1.0651, + "step": 7853 + }, + { + "epoch": 0.6041073763556649, + "learning_rate": 0.0010181053817470143, + "loss": 0.8757, + "step": 7854 + }, + { + "epoch": 0.6041842935158834, + "learning_rate": 0.001017762146440545, + "loss": 1.1608, + "step": 7855 + }, + { + "epoch": 0.6042612106761018, + "learning_rate": 0.0010174189392924474, + "loss": 1.1209, + "step": 7856 + }, + { + "epoch": 0.6043381278363202, + "learning_rate": 0.0010170757603227606, + "loss": 1.1479, + "step": 7857 + }, + { + "epoch": 0.6044150449965388, + "learning_rate": 0.0010167326095515234, + "loss": 0.9983, + "step": 7858 + }, + { + "epoch": 0.6044919621567572, + "learning_rate": 0.0010163894869987728, + "loss": 1.2428, + "step": 7859 + }, + { + "epoch": 0.6045688793169757, + "learning_rate": 0.0010160463926845438, + "loss": 1.0819, + "step": 7860 + }, + { + "epoch": 0.6046457964771941, + "learning_rate": 0.0010157033266288708, + "loss": 1.0805, + "step": 7861 + }, + { + "epoch": 0.6047227136374125, + "learning_rate": 0.001015360288851785, + "loss": 1.3644, + "step": 7862 + }, + { + "epoch": 0.604799630797631, + "learning_rate": 0.001015017279373317, + "loss": 1.123, + "step": 7863 + }, + { + "epoch": 0.6048765479578494, + "learning_rate": 0.0010146742982134957, + "loss": 1.2639, + "step": 7864 + }, + { + "epoch": 0.6049534651180678, + "learning_rate": 0.0010143313453923478, + "loss": 1.0517, + "step": 7865 + }, + { + "epoch": 0.6050303822782863, + "learning_rate": 0.0010139884209298985, + "loss": 1.2348, + "step": 7866 + }, + { + "epoch": 0.6051072994385047, + "learning_rate": 0.0010136455248461725, + "loss": 1.0953, + "step": 7867 + }, + { + "epoch": 0.6051842165987231, + "learning_rate": 0.001013302657161191, + "loss": 0.6503, + "step": 7868 + }, + { + "epoch": 0.6052611337589416, + "learning_rate": 0.001012959817894974, + "loss": 1.3275, + "step": 7869 + }, + { + "epoch": 0.60533805091916, + "learning_rate": 0.0010126170070675414, + "loss": 1.2339, + "step": 7870 + }, + { + "epoch": 0.6054149680793786, + "learning_rate": 0.001012274224698909, + "loss": 1.1524, + "step": 7871 + }, + { + "epoch": 0.605491885239597, + "learning_rate": 0.0010119314708090935, + "loss": 0.8756, + "step": 7872 + }, + { + "epoch": 0.6055688023998154, + "learning_rate": 0.001011588745418108, + "loss": 1.1831, + "step": 7873 + }, + { + "epoch": 0.6056457195600339, + "learning_rate": 0.0010112460485459645, + "loss": 1.2186, + "step": 7874 + }, + { + "epoch": 0.6057226367202523, + "learning_rate": 0.0010109033802126737, + "loss": 1.281, + "step": 7875 + }, + { + "epoch": 0.6057995538804707, + "learning_rate": 0.0010105607404382443, + "loss": 1.1512, + "step": 7876 + }, + { + "epoch": 0.6058764710406892, + "learning_rate": 0.0010102181292426832, + "loss": 0.9668, + "step": 7877 + }, + { + "epoch": 0.6059533882009076, + "learning_rate": 0.001009875546645996, + "loss": 1.2598, + "step": 7878 + }, + { + "epoch": 0.6060303053611261, + "learning_rate": 0.0010095329926681864, + "loss": 1.3061, + "step": 7879 + }, + { + "epoch": 0.6061072225213445, + "learning_rate": 0.001009190467329257, + "loss": 1.2282, + "step": 7880 + }, + { + "epoch": 0.6061841396815629, + "learning_rate": 0.0010088479706492075, + "loss": 1.0917, + "step": 7881 + }, + { + "epoch": 0.6062610568417814, + "learning_rate": 0.0010085055026480365, + "loss": 1.3351, + "step": 7882 + }, + { + "epoch": 0.6063379740019998, + "learning_rate": 0.0010081630633457415, + "loss": 1.4174, + "step": 7883 + }, + { + "epoch": 0.6064148911622183, + "learning_rate": 0.0010078206527623181, + "loss": 0.8827, + "step": 7884 + }, + { + "epoch": 0.6064918083224368, + "learning_rate": 0.0010074782709177594, + "loss": 1.1687, + "step": 7885 + }, + { + "epoch": 0.6065687254826552, + "learning_rate": 0.001007135917832058, + "loss": 1.2557, + "step": 7886 + }, + { + "epoch": 0.6066456426428736, + "learning_rate": 0.001006793593525204, + "loss": 1.2157, + "step": 7887 + }, + { + "epoch": 0.6067225598030921, + "learning_rate": 0.0010064512980171856, + "loss": 1.1743, + "step": 7888 + }, + { + "epoch": 0.6067994769633105, + "learning_rate": 0.001006109031327991, + "loss": 1.2562, + "step": 7889 + }, + { + "epoch": 0.606876394123529, + "learning_rate": 0.001005766793477604, + "loss": 1.0796, + "step": 7890 + }, + { + "epoch": 0.6069533112837474, + "learning_rate": 0.0010054245844860095, + "loss": 1.1306, + "step": 7891 + }, + { + "epoch": 0.6070302284439658, + "learning_rate": 0.0010050824043731893, + "loss": 1.0124, + "step": 7892 + }, + { + "epoch": 0.6071071456041843, + "learning_rate": 0.0010047402531591224, + "loss": 1.2791, + "step": 7893 + }, + { + "epoch": 0.6071840627644027, + "learning_rate": 0.001004398130863789, + "loss": 1.3058, + "step": 7894 + }, + { + "epoch": 0.6072609799246211, + "learning_rate": 0.0010040560375071653, + "loss": 1.0634, + "step": 7895 + }, + { + "epoch": 0.6073378970848397, + "learning_rate": 0.0010037139731092258, + "loss": 1.2333, + "step": 7896 + }, + { + "epoch": 0.607414814245058, + "learning_rate": 0.001003371937689945, + "loss": 1.1136, + "step": 7897 + }, + { + "epoch": 0.6074917314052766, + "learning_rate": 0.0010030299312692942, + "loss": 1.0772, + "step": 7898 + }, + { + "epoch": 0.607568648565495, + "learning_rate": 0.0010026879538672438, + "loss": 1.3217, + "step": 7899 + }, + { + "epoch": 0.6076455657257134, + "learning_rate": 0.001002346005503762, + "loss": 0.9078, + "step": 7900 + }, + { + "epoch": 0.6077224828859319, + "learning_rate": 0.0010020040861988156, + "loss": 1.3689, + "step": 7901 + }, + { + "epoch": 0.6077994000461503, + "learning_rate": 0.0010016621959723696, + "loss": 1.2904, + "step": 7902 + }, + { + "epoch": 0.6078763172063687, + "learning_rate": 0.0010013203348443877, + "loss": 1.0072, + "step": 7903 + }, + { + "epoch": 0.6079532343665872, + "learning_rate": 0.0010009785028348302, + "loss": 0.9837, + "step": 7904 + }, + { + "epoch": 0.6080301515268056, + "learning_rate": 0.001000636699963659, + "loss": 1.3249, + "step": 7905 + }, + { + "epoch": 0.608107068687024, + "learning_rate": 0.001000294926250831, + "loss": 1.1906, + "step": 7906 + }, + { + "epoch": 0.6081839858472425, + "learning_rate": 0.0009999531817163026, + "loss": 1.3993, + "step": 7907 + }, + { + "epoch": 0.6082609030074609, + "learning_rate": 0.0009996114663800295, + "loss": 1.2132, + "step": 7908 + }, + { + "epoch": 0.6083378201676795, + "learning_rate": 0.0009992697802619639, + "loss": 0.8712, + "step": 7909 + }, + { + "epoch": 0.6084147373278979, + "learning_rate": 0.000998928123382058, + "loss": 1.2964, + "step": 7910 + }, + { + "epoch": 0.6084916544881163, + "learning_rate": 0.0009985864957602615, + "loss": 1.3279, + "step": 7911 + }, + { + "epoch": 0.6085685716483348, + "learning_rate": 0.0009982448974165216, + "loss": 1.19, + "step": 7912 + }, + { + "epoch": 0.6086454888085532, + "learning_rate": 0.0009979033283707853, + "loss": 1.4645, + "step": 7913 + }, + { + "epoch": 0.6087224059687716, + "learning_rate": 0.0009975617886429972, + "loss": 1.7533, + "step": 7914 + }, + { + "epoch": 0.6087993231289901, + "learning_rate": 0.000997220278253099, + "loss": 0.7365, + "step": 7915 + }, + { + "epoch": 0.6088762402892085, + "learning_rate": 0.0009968787972210336, + "loss": 1.1548, + "step": 7916 + }, + { + "epoch": 0.608953157449427, + "learning_rate": 0.0009965373455667389, + "loss": 0.9278, + "step": 7917 + }, + { + "epoch": 0.6090300746096454, + "learning_rate": 0.0009961959233101539, + "loss": 0.9917, + "step": 7918 + }, + { + "epoch": 0.6091069917698638, + "learning_rate": 0.0009958545304712138, + "loss": 1.0208, + "step": 7919 + }, + { + "epoch": 0.6091839089300823, + "learning_rate": 0.000995513167069853, + "loss": 1.0637, + "step": 7920 + }, + { + "epoch": 0.6092608260903007, + "learning_rate": 0.0009951718331260042, + "loss": 1.2475, + "step": 7921 + }, + { + "epoch": 0.6093377432505191, + "learning_rate": 0.000994830528659598, + "loss": 1.2193, + "step": 7922 + }, + { + "epoch": 0.6094146604107377, + "learning_rate": 0.0009944892536905637, + "loss": 1.1543, + "step": 7923 + }, + { + "epoch": 0.6094915775709561, + "learning_rate": 0.000994148008238829, + "loss": 1.5268, + "step": 7924 + }, + { + "epoch": 0.6095684947311745, + "learning_rate": 0.0009938067923243195, + "loss": 1.1637, + "step": 7925 + }, + { + "epoch": 0.609645411891393, + "learning_rate": 0.0009934656059669578, + "loss": 1.1411, + "step": 7926 + }, + { + "epoch": 0.6097223290516114, + "learning_rate": 0.0009931244491866685, + "loss": 1.1863, + "step": 7927 + }, + { + "epoch": 0.6097992462118299, + "learning_rate": 0.00099278332200337, + "loss": 0.9868, + "step": 7928 + }, + { + "epoch": 0.6098761633720483, + "learning_rate": 0.000992442224436983, + "loss": 1.2126, + "step": 7929 + }, + { + "epoch": 0.6099530805322667, + "learning_rate": 0.0009921011565074227, + "loss": 1.0634, + "step": 7930 + }, + { + "epoch": 0.6100299976924852, + "learning_rate": 0.000991760118234605, + "loss": 1.1871, + "step": 7931 + }, + { + "epoch": 0.6101069148527036, + "learning_rate": 0.000991419109638444, + "loss": 1.1458, + "step": 7932 + }, + { + "epoch": 0.610183832012922, + "learning_rate": 0.0009910781307388514, + "loss": 0.8921, + "step": 7933 + }, + { + "epoch": 0.6102607491731405, + "learning_rate": 0.0009907371815557368, + "loss": 1.3179, + "step": 7934 + }, + { + "epoch": 0.610337666333359, + "learning_rate": 0.000990396262109009, + "loss": 1.1721, + "step": 7935 + }, + { + "epoch": 0.6104145834935775, + "learning_rate": 0.0009900553724185742, + "loss": 1.1723, + "step": 7936 + }, + { + "epoch": 0.6104915006537959, + "learning_rate": 0.0009897145125043383, + "loss": 1.1278, + "step": 7937 + }, + { + "epoch": 0.6105684178140143, + "learning_rate": 0.000989373682386204, + "loss": 1.3781, + "step": 7938 + }, + { + "epoch": 0.6106453349742328, + "learning_rate": 0.0009890328820840715, + "loss": 1.0431, + "step": 7939 + }, + { + "epoch": 0.6107222521344512, + "learning_rate": 0.0009886921116178426, + "loss": 1.6225, + "step": 7940 + }, + { + "epoch": 0.6107991692946696, + "learning_rate": 0.0009883513710074138, + "loss": 1.5153, + "step": 7941 + }, + { + "epoch": 0.6108760864548881, + "learning_rate": 0.0009880106602726815, + "loss": 1.1906, + "step": 7942 + }, + { + "epoch": 0.6109530036151065, + "learning_rate": 0.0009876699794335404, + "loss": 1.0122, + "step": 7943 + }, + { + "epoch": 0.6110299207753249, + "learning_rate": 0.0009873293285098833, + "loss": 1.2095, + "step": 7944 + }, + { + "epoch": 0.6111068379355434, + "learning_rate": 0.0009869887075216006, + "loss": 1.3353, + "step": 7945 + }, + { + "epoch": 0.6111837550957618, + "learning_rate": 0.0009866481164885821, + "loss": 1.1863, + "step": 7946 + }, + { + "epoch": 0.6112606722559804, + "learning_rate": 0.000986307555430715, + "loss": 1.2304, + "step": 7947 + }, + { + "epoch": 0.6113375894161988, + "learning_rate": 0.0009859670243678855, + "loss": 0.8882, + "step": 7948 + }, + { + "epoch": 0.6114145065764172, + "learning_rate": 0.0009856265233199773, + "loss": 1.0656, + "step": 7949 + }, + { + "epoch": 0.6114914237366357, + "learning_rate": 0.0009852860523068714, + "loss": 1.01, + "step": 7950 + }, + { + "epoch": 0.6115683408968541, + "learning_rate": 0.0009849456113484505, + "loss": 0.9631, + "step": 7951 + }, + { + "epoch": 0.6116452580570725, + "learning_rate": 0.0009846052004645917, + "loss": 1.7442, + "step": 7952 + }, + { + "epoch": 0.611722175217291, + "learning_rate": 0.000984264819675172, + "loss": 1.1071, + "step": 7953 + }, + { + "epoch": 0.6117990923775094, + "learning_rate": 0.0009839244690000675, + "loss": 1.0134, + "step": 7954 + }, + { + "epoch": 0.6118760095377279, + "learning_rate": 0.0009835841484591506, + "loss": 1.2044, + "step": 7955 + }, + { + "epoch": 0.6119529266979463, + "learning_rate": 0.000983243858072294, + "loss": 1.4278, + "step": 7956 + }, + { + "epoch": 0.6120298438581647, + "learning_rate": 0.0009829035978593669, + "loss": 1.6517, + "step": 7957 + }, + { + "epoch": 0.6121067610183832, + "learning_rate": 0.000982563367840237, + "loss": 1.0621, + "step": 7958 + }, + { + "epoch": 0.6121836781786016, + "learning_rate": 0.000982223168034772, + "loss": 1.0888, + "step": 7959 + }, + { + "epoch": 0.61226059533882, + "learning_rate": 0.0009818829984628362, + "loss": 1.0146, + "step": 7960 + }, + { + "epoch": 0.6123375124990386, + "learning_rate": 0.000981542859144291, + "loss": 0.7239, + "step": 7961 + }, + { + "epoch": 0.612414429659257, + "learning_rate": 0.0009812027500989997, + "loss": 1.0811, + "step": 7962 + }, + { + "epoch": 0.6124913468194755, + "learning_rate": 0.0009808626713468194, + "loss": 0.8919, + "step": 7963 + }, + { + "epoch": 0.6125682639796939, + "learning_rate": 0.0009805226229076096, + "loss": 1.3856, + "step": 7964 + }, + { + "epoch": 0.6126451811399123, + "learning_rate": 0.000980182604801225, + "loss": 0.9222, + "step": 7965 + }, + { + "epoch": 0.6127220983001308, + "learning_rate": 0.0009798426170475197, + "loss": 1.085, + "step": 7966 + }, + { + "epoch": 0.6127990154603492, + "learning_rate": 0.0009795026596663462, + "loss": 1.1534, + "step": 7967 + }, + { + "epoch": 0.6128759326205676, + "learning_rate": 0.000979162732677555, + "loss": 1.3604, + "step": 7968 + }, + { + "epoch": 0.6129528497807861, + "learning_rate": 0.000978822836100994, + "loss": 1.1482, + "step": 7969 + }, + { + "epoch": 0.6130297669410045, + "learning_rate": 0.0009784829699565115, + "loss": 1.1764, + "step": 7970 + }, + { + "epoch": 0.6131066841012229, + "learning_rate": 0.000978143134263952, + "loss": 1.4959, + "step": 7971 + }, + { + "epoch": 0.6131836012614414, + "learning_rate": 0.000977803329043158, + "loss": 1.2298, + "step": 7972 + }, + { + "epoch": 0.6132605184216599, + "learning_rate": 0.0009774635543139726, + "loss": 0.8295, + "step": 7973 + }, + { + "epoch": 0.6133374355818784, + "learning_rate": 0.000977123810096234, + "loss": 1.0659, + "step": 7974 + }, + { + "epoch": 0.6134143527420968, + "learning_rate": 0.0009767840964097817, + "loss": 1.3188, + "step": 7975 + }, + { + "epoch": 0.6134912699023152, + "learning_rate": 0.0009764444132744513, + "loss": 1.2851, + "step": 7976 + }, + { + "epoch": 0.6135681870625337, + "learning_rate": 0.0009761047607100767, + "loss": 0.8915, + "step": 7977 + }, + { + "epoch": 0.6136451042227521, + "learning_rate": 0.0009757651387364916, + "loss": 0.7588, + "step": 7978 + }, + { + "epoch": 0.6137220213829705, + "learning_rate": 0.0009754255473735263, + "loss": 1.1247, + "step": 7979 + }, + { + "epoch": 0.613798938543189, + "learning_rate": 0.0009750859866410097, + "loss": 1.2587, + "step": 7980 + }, + { + "epoch": 0.6138758557034074, + "learning_rate": 0.0009747464565587698, + "loss": 1.2248, + "step": 7981 + }, + { + "epoch": 0.6139527728636259, + "learning_rate": 0.0009744069571466314, + "loss": 1.265, + "step": 7982 + }, + { + "epoch": 0.6140296900238443, + "learning_rate": 0.0009740674884244186, + "loss": 1.3845, + "step": 7983 + }, + { + "epoch": 0.6141066071840627, + "learning_rate": 0.0009737280504119539, + "loss": 1.1153, + "step": 7984 + }, + { + "epoch": 0.6141835243442813, + "learning_rate": 0.0009733886431290557, + "loss": 1.1525, + "step": 7985 + }, + { + "epoch": 0.6142604415044997, + "learning_rate": 0.0009730492665955443, + "loss": 1.2444, + "step": 7986 + }, + { + "epoch": 0.6143373586647181, + "learning_rate": 0.0009727099208312354, + "loss": 1.5033, + "step": 7987 + }, + { + "epoch": 0.6144142758249366, + "learning_rate": 0.0009723706058559431, + "loss": 1.3061, + "step": 7988 + }, + { + "epoch": 0.614491192985155, + "learning_rate": 0.0009720313216894813, + "loss": 1.4493, + "step": 7989 + }, + { + "epoch": 0.6145681101453734, + "learning_rate": 0.0009716920683516612, + "loss": 1.0055, + "step": 7990 + }, + { + "epoch": 0.6146450273055919, + "learning_rate": 0.0009713528458622911, + "loss": 1.2471, + "step": 7991 + }, + { + "epoch": 0.6147219444658103, + "learning_rate": 0.0009710136542411798, + "loss": 1.2029, + "step": 7992 + }, + { + "epoch": 0.6147988616260288, + "learning_rate": 0.0009706744935081323, + "loss": 0.9112, + "step": 7993 + }, + { + "epoch": 0.6148757787862472, + "learning_rate": 0.000970335363682953, + "loss": 1.1652, + "step": 7994 + }, + { + "epoch": 0.6149526959464656, + "learning_rate": 0.0009699962647854441, + "loss": 0.9244, + "step": 7995 + }, + { + "epoch": 0.6150296131066841, + "learning_rate": 0.0009696571968354048, + "loss": 1.2673, + "step": 7996 + }, + { + "epoch": 0.6151065302669025, + "learning_rate": 0.0009693181598526353, + "loss": 1.3341, + "step": 7997 + }, + { + "epoch": 0.6151834474271209, + "learning_rate": 0.0009689791538569313, + "loss": 1.091, + "step": 7998 + }, + { + "epoch": 0.6152603645873395, + "learning_rate": 0.0009686401788680875, + "loss": 1.5006, + "step": 7999 + }, + { + "epoch": 0.6153372817475579, + "learning_rate": 0.0009683012349058979, + "loss": 1.2326, + "step": 8000 + }, + { + "epoch": 0.6154141989077764, + "learning_rate": 0.0009679623219901529, + "loss": 1.1959, + "step": 8001 + }, + { + "epoch": 0.6154911160679948, + "learning_rate": 0.0009676234401406427, + "loss": 1.3407, + "step": 8002 + }, + { + "epoch": 0.6155680332282132, + "learning_rate": 0.0009672845893771546, + "loss": 1.1276, + "step": 8003 + }, + { + "epoch": 0.6156449503884317, + "learning_rate": 0.0009669457697194744, + "loss": 1.2509, + "step": 8004 + }, + { + "epoch": 0.6157218675486501, + "learning_rate": 0.0009666069811873863, + "loss": 1.461, + "step": 8005 + }, + { + "epoch": 0.6157987847088685, + "learning_rate": 0.000966268223800673, + "loss": 1.1217, + "step": 8006 + }, + { + "epoch": 0.615875701869087, + "learning_rate": 0.0009659294975791134, + "loss": 0.864, + "step": 8007 + }, + { + "epoch": 0.6159526190293054, + "learning_rate": 0.0009655908025424879, + "loss": 1.0173, + "step": 8008 + }, + { + "epoch": 0.6160295361895238, + "learning_rate": 0.0009652521387105719, + "loss": 1.4209, + "step": 8009 + }, + { + "epoch": 0.6161064533497423, + "learning_rate": 0.0009649135061031409, + "loss": 1.2029, + "step": 8010 + }, + { + "epoch": 0.6161833705099607, + "learning_rate": 0.0009645749047399679, + "loss": 1.157, + "step": 8011 + }, + { + "epoch": 0.6162602876701793, + "learning_rate": 0.000964236334640824, + "loss": 1.4025, + "step": 8012 + }, + { + "epoch": 0.6163372048303977, + "learning_rate": 0.0009638977958254792, + "loss": 1.1947, + "step": 8013 + }, + { + "epoch": 0.6164141219906161, + "learning_rate": 0.0009635592883137008, + "loss": 0.9063, + "step": 8014 + }, + { + "epoch": 0.6164910391508346, + "learning_rate": 0.0009632208121252544, + "loss": 1.0217, + "step": 8015 + }, + { + "epoch": 0.616567956311053, + "learning_rate": 0.0009628823672799045, + "loss": 0.8972, + "step": 8016 + }, + { + "epoch": 0.6166448734712714, + "learning_rate": 0.0009625439537974129, + "loss": 0.9665, + "step": 8017 + }, + { + "epoch": 0.6167217906314899, + "learning_rate": 0.0009622055716975396, + "loss": 1.0596, + "step": 8018 + }, + { + "epoch": 0.6167987077917083, + "learning_rate": 0.0009618672210000439, + "loss": 0.9571, + "step": 8019 + }, + { + "epoch": 0.6168756249519268, + "learning_rate": 0.0009615289017246816, + "loss": 1.1336, + "step": 8020 + }, + { + "epoch": 0.6169525421121452, + "learning_rate": 0.0009611906138912084, + "loss": 1.3222, + "step": 8021 + }, + { + "epoch": 0.6170294592723636, + "learning_rate": 0.0009608523575193765, + "loss": 1.5397, + "step": 8022 + }, + { + "epoch": 0.6171063764325821, + "learning_rate": 0.0009605141326289372, + "loss": 1.0812, + "step": 8023 + }, + { + "epoch": 0.6171832935928006, + "learning_rate": 0.0009601759392396401, + "loss": 1.4692, + "step": 8024 + }, + { + "epoch": 0.617260210753019, + "learning_rate": 0.0009598377773712327, + "loss": 0.9726, + "step": 8025 + }, + { + "epoch": 0.6173371279132375, + "learning_rate": 0.00095949964704346, + "loss": 0.8921, + "step": 8026 + }, + { + "epoch": 0.6174140450734559, + "learning_rate": 0.0009591615482760664, + "loss": 1.3792, + "step": 8027 + }, + { + "epoch": 0.6174909622336743, + "learning_rate": 0.0009588234810887937, + "loss": 1.0445, + "step": 8028 + }, + { + "epoch": 0.6175678793938928, + "learning_rate": 0.0009584854455013817, + "loss": 0.7489, + "step": 8029 + }, + { + "epoch": 0.6176447965541112, + "learning_rate": 0.000958147441533569, + "loss": 1.0757, + "step": 8030 + }, + { + "epoch": 0.6177217137143297, + "learning_rate": 0.0009578094692050918, + "loss": 1.0493, + "step": 8031 + }, + { + "epoch": 0.6177986308745481, + "learning_rate": 0.0009574715285356849, + "loss": 1.0729, + "step": 8032 + }, + { + "epoch": 0.6178755480347665, + "learning_rate": 0.0009571336195450813, + "loss": 1.4625, + "step": 8033 + }, + { + "epoch": 0.617952465194985, + "learning_rate": 0.0009567957422530109, + "loss": 1.0553, + "step": 8034 + }, + { + "epoch": 0.6180293823552034, + "learning_rate": 0.0009564578966792035, + "loss": 1.4732, + "step": 8035 + }, + { + "epoch": 0.6181062995154218, + "learning_rate": 0.0009561200828433858, + "loss": 1.2348, + "step": 8036 + }, + { + "epoch": 0.6181832166756404, + "learning_rate": 0.000955782300765283, + "loss": 1.2797, + "step": 8037 + }, + { + "epoch": 0.6182601338358588, + "learning_rate": 0.0009554445504646193, + "loss": 0.8568, + "step": 8038 + }, + { + "epoch": 0.6183370509960773, + "learning_rate": 0.0009551068319611153, + "loss": 1.1488, + "step": 8039 + }, + { + "epoch": 0.6184139681562957, + "learning_rate": 0.0009547691452744917, + "loss": 1.1749, + "step": 8040 + }, + { + "epoch": 0.6184908853165141, + "learning_rate": 0.0009544314904244658, + "loss": 1.3081, + "step": 8041 + }, + { + "epoch": 0.6185678024767326, + "learning_rate": 0.0009540938674307536, + "loss": 1.0842, + "step": 8042 + }, + { + "epoch": 0.618644719636951, + "learning_rate": 0.0009537562763130697, + "loss": 0.9682, + "step": 8043 + }, + { + "epoch": 0.6187216367971694, + "learning_rate": 0.0009534187170911263, + "loss": 0.7129, + "step": 8044 + }, + { + "epoch": 0.6187985539573879, + "learning_rate": 0.0009530811897846328, + "loss": 1.0388, + "step": 8045 + }, + { + "epoch": 0.6188754711176063, + "learning_rate": 0.0009527436944132996, + "loss": 1.1447, + "step": 8046 + }, + { + "epoch": 0.6189523882778247, + "learning_rate": 0.0009524062309968321, + "loss": 0.8394, + "step": 8047 + }, + { + "epoch": 0.6190293054380432, + "learning_rate": 0.000952068799554935, + "loss": 1.229, + "step": 8048 + }, + { + "epoch": 0.6191062225982616, + "learning_rate": 0.0009517314001073121, + "loss": 1.0077, + "step": 8049 + }, + { + "epoch": 0.6191831397584802, + "learning_rate": 0.000951394032673664, + "loss": 1.2297, + "step": 8050 + }, + { + "epoch": 0.6192600569186986, + "learning_rate": 0.0009510566972736901, + "loss": 1.1122, + "step": 8051 + }, + { + "epoch": 0.619336974078917, + "learning_rate": 0.000950719393927088, + "loss": 0.9952, + "step": 8052 + }, + { + "epoch": 0.6194138912391355, + "learning_rate": 0.0009503821226535525, + "loss": 0.8947, + "step": 8053 + }, + { + "epoch": 0.6194908083993539, + "learning_rate": 0.000950044883472778, + "loss": 1.2, + "step": 8054 + }, + { + "epoch": 0.6195677255595723, + "learning_rate": 0.0009497076764044561, + "loss": 1.0742, + "step": 8055 + }, + { + "epoch": 0.6196446427197908, + "learning_rate": 0.0009493705014682757, + "loss": 1.3496, + "step": 8056 + }, + { + "epoch": 0.6197215598800092, + "learning_rate": 0.0009490333586839266, + "loss": 1.0115, + "step": 8057 + }, + { + "epoch": 0.6197984770402277, + "learning_rate": 0.0009486962480710927, + "loss": 1.3346, + "step": 8058 + }, + { + "epoch": 0.6198753942004461, + "learning_rate": 0.0009483591696494606, + "loss": 0.9056, + "step": 8059 + }, + { + "epoch": 0.6199523113606645, + "learning_rate": 0.0009480221234387108, + "loss": 1.3529, + "step": 8060 + }, + { + "epoch": 0.620029228520883, + "learning_rate": 0.0009476851094585246, + "loss": 1.3914, + "step": 8061 + }, + { + "epoch": 0.6201061456811015, + "learning_rate": 0.0009473481277285805, + "loss": 1.0591, + "step": 8062 + }, + { + "epoch": 0.6201830628413199, + "learning_rate": 0.0009470111782685554, + "loss": 1.2498, + "step": 8063 + }, + { + "epoch": 0.6202599800015384, + "learning_rate": 0.0009466742610981235, + "loss": 1.0873, + "step": 8064 + }, + { + "epoch": 0.6203368971617568, + "learning_rate": 0.0009463373762369584, + "loss": 1.4817, + "step": 8065 + }, + { + "epoch": 0.6204138143219753, + "learning_rate": 0.0009460005237047306, + "loss": 1.1098, + "step": 8066 + }, + { + "epoch": 0.6204907314821937, + "learning_rate": 0.00094566370352111, + "loss": 1.1081, + "step": 8067 + }, + { + "epoch": 0.6205676486424121, + "learning_rate": 0.0009453269157057637, + "loss": 1.1561, + "step": 8068 + }, + { + "epoch": 0.6206445658026306, + "learning_rate": 0.0009449901602783558, + "loss": 0.9915, + "step": 8069 + }, + { + "epoch": 0.620721482962849, + "learning_rate": 0.000944653437258552, + "loss": 1.3577, + "step": 8070 + }, + { + "epoch": 0.6207984001230674, + "learning_rate": 0.0009443167466660124, + "loss": 1.1501, + "step": 8071 + }, + { + "epoch": 0.6208753172832859, + "learning_rate": 0.0009439800885203968, + "loss": 1.0286, + "step": 8072 + }, + { + "epoch": 0.6209522344435043, + "learning_rate": 0.0009436434628413636, + "loss": 1.3252, + "step": 8073 + }, + { + "epoch": 0.6210291516037227, + "learning_rate": 0.0009433068696485684, + "loss": 1.2765, + "step": 8074 + }, + { + "epoch": 0.6211060687639413, + "learning_rate": 0.0009429703089616649, + "loss": 1.0048, + "step": 8075 + }, + { + "epoch": 0.6211829859241597, + "learning_rate": 0.0009426337808003058, + "loss": 1.0308, + "step": 8076 + }, + { + "epoch": 0.6212599030843782, + "learning_rate": 0.0009422972851841411, + "loss": 1.1436, + "step": 8077 + }, + { + "epoch": 0.6213368202445966, + "learning_rate": 0.0009419608221328191, + "loss": 1.4121, + "step": 8078 + }, + { + "epoch": 0.621413737404815, + "learning_rate": 0.0009416243916659869, + "loss": 1.1917, + "step": 8079 + }, + { + "epoch": 0.6214906545650335, + "learning_rate": 0.0009412879938032872, + "loss": 1.535, + "step": 8080 + }, + { + "epoch": 0.6215675717252519, + "learning_rate": 0.000940951628564365, + "loss": 1.042, + "step": 8081 + }, + { + "epoch": 0.6216444888854703, + "learning_rate": 0.0009406152959688593, + "loss": 0.7926, + "step": 8082 + }, + { + "epoch": 0.6217214060456888, + "learning_rate": 0.0009402789960364093, + "loss": 0.8164, + "step": 8083 + }, + { + "epoch": 0.6217983232059072, + "learning_rate": 0.0009399427287866524, + "loss": 1.0363, + "step": 8084 + }, + { + "epoch": 0.6218752403661257, + "learning_rate": 0.0009396064942392228, + "loss": 0.9432, + "step": 8085 + }, + { + "epoch": 0.6219521575263441, + "learning_rate": 0.0009392702924137543, + "loss": 1.1427, + "step": 8086 + }, + { + "epoch": 0.6220290746865625, + "learning_rate": 0.0009389341233298779, + "loss": 1.3593, + "step": 8087 + }, + { + "epoch": 0.6221059918467811, + "learning_rate": 0.0009385979870072225, + "loss": 1.2726, + "step": 8088 + }, + { + "epoch": 0.6221829090069995, + "learning_rate": 0.000938261883465416, + "loss": 1.041, + "step": 8089 + }, + { + "epoch": 0.6222598261672179, + "learning_rate": 0.0009379258127240838, + "loss": 0.8908, + "step": 8090 + }, + { + "epoch": 0.6223367433274364, + "learning_rate": 0.0009375897748028481, + "loss": 1.0971, + "step": 8091 + }, + { + "epoch": 0.6224136604876548, + "learning_rate": 0.0009372537697213328, + "loss": 0.9342, + "step": 8092 + }, + { + "epoch": 0.6224905776478732, + "learning_rate": 0.0009369177974991558, + "loss": 1.1111, + "step": 8093 + }, + { + "epoch": 0.6225674948080917, + "learning_rate": 0.0009365818581559354, + "loss": 1.1114, + "step": 8094 + }, + { + "epoch": 0.6226444119683101, + "learning_rate": 0.0009362459517112875, + "loss": 1.1003, + "step": 8095 + }, + { + "epoch": 0.6227213291285286, + "learning_rate": 0.0009359100781848259, + "loss": 0.9977, + "step": 8096 + }, + { + "epoch": 0.622798246288747, + "learning_rate": 0.000935574237596163, + "loss": 0.8794, + "step": 8097 + }, + { + "epoch": 0.6228751634489654, + "learning_rate": 0.0009352384299649086, + "loss": 1.4536, + "step": 8098 + }, + { + "epoch": 0.622952080609184, + "learning_rate": 0.0009349026553106707, + "loss": 1.2938, + "step": 8099 + }, + { + "epoch": 0.6230289977694023, + "learning_rate": 0.0009345669136530556, + "loss": 0.8505, + "step": 8100 + }, + { + "epoch": 0.6231059149296208, + "learning_rate": 0.0009342312050116683, + "loss": 0.7771, + "step": 8101 + }, + { + "epoch": 0.6231828320898393, + "learning_rate": 0.0009338955294061095, + "loss": 0.9755, + "step": 8102 + }, + { + "epoch": 0.6232597492500577, + "learning_rate": 0.0009335598868559817, + "loss": 0.7584, + "step": 8103 + }, + { + "epoch": 0.6233366664102762, + "learning_rate": 0.0009332242773808815, + "loss": 1.0345, + "step": 8104 + }, + { + "epoch": 0.6234135835704946, + "learning_rate": 0.0009328887010004075, + "loss": 1.3528, + "step": 8105 + }, + { + "epoch": 0.623490500730713, + "learning_rate": 0.0009325531577341527, + "loss": 1.3813, + "step": 8106 + }, + { + "epoch": 0.6235674178909315, + "learning_rate": 0.0009322176476017103, + "loss": 1.2704, + "step": 8107 + }, + { + "epoch": 0.6236443350511499, + "learning_rate": 0.0009318821706226715, + "loss": 1.2574, + "step": 8108 + }, + { + "epoch": 0.6237212522113683, + "learning_rate": 0.0009315467268166247, + "loss": 0.9548, + "step": 8109 + }, + { + "epoch": 0.6237981693715868, + "learning_rate": 0.0009312113162031566, + "loss": 1.5025, + "step": 8110 + }, + { + "epoch": 0.6238750865318052, + "learning_rate": 0.0009308759388018528, + "loss": 0.971, + "step": 8111 + }, + { + "epoch": 0.6239520036920236, + "learning_rate": 0.0009305405946322966, + "loss": 1.3512, + "step": 8112 + }, + { + "epoch": 0.6240289208522422, + "learning_rate": 0.0009302052837140676, + "loss": 1.215, + "step": 8113 + }, + { + "epoch": 0.6241058380124606, + "learning_rate": 0.0009298700060667468, + "loss": 1.0269, + "step": 8114 + }, + { + "epoch": 0.6241827551726791, + "learning_rate": 0.0009295347617099094, + "loss": 1.1533, + "step": 8115 + }, + { + "epoch": 0.6242596723328975, + "learning_rate": 0.0009291995506631328, + "loss": 1.1212, + "step": 8116 + }, + { + "epoch": 0.6243365894931159, + "learning_rate": 0.0009288643729459891, + "loss": 1.1638, + "step": 8117 + }, + { + "epoch": 0.6244135066533344, + "learning_rate": 0.0009285292285780496, + "loss": 1.1108, + "step": 8118 + }, + { + "epoch": 0.6244904238135528, + "learning_rate": 0.0009281941175788844, + "loss": 1.2896, + "step": 8119 + }, + { + "epoch": 0.6245673409737712, + "learning_rate": 0.0009278590399680605, + "loss": 1.2024, + "step": 8120 + }, + { + "epoch": 0.6246442581339897, + "learning_rate": 0.0009275239957651431, + "loss": 1.2816, + "step": 8121 + }, + { + "epoch": 0.6247211752942081, + "learning_rate": 0.0009271889849896966, + "loss": 1.5103, + "step": 8122 + }, + { + "epoch": 0.6247980924544266, + "learning_rate": 0.000926854007661282, + "loss": 1.2158, + "step": 8123 + }, + { + "epoch": 0.624875009614645, + "learning_rate": 0.0009265190637994594, + "loss": 0.923, + "step": 8124 + }, + { + "epoch": 0.6249519267748634, + "learning_rate": 0.0009261841534237867, + "loss": 0.9093, + "step": 8125 + }, + { + "epoch": 0.625028843935082, + "learning_rate": 0.0009258492765538183, + "loss": 1.359, + "step": 8126 + }, + { + "epoch": 0.6251057610953004, + "learning_rate": 0.0009255144332091101, + "loss": 0.9265, + "step": 8127 + }, + { + "epoch": 0.6251826782555188, + "learning_rate": 0.0009251796234092123, + "loss": 1.3336, + "step": 8128 + }, + { + "epoch": 0.6252595954157373, + "learning_rate": 0.0009248448471736753, + "loss": 1.2168, + "step": 8129 + }, + { + "epoch": 0.6253365125759557, + "learning_rate": 0.0009245101045220472, + "loss": 1.0361, + "step": 8130 + }, + { + "epoch": 0.6254134297361741, + "learning_rate": 0.000924175395473874, + "loss": 0.9382, + "step": 8131 + }, + { + "epoch": 0.6254903468963926, + "learning_rate": 0.0009238407200486992, + "loss": 1.1063, + "step": 8132 + }, + { + "epoch": 0.625567264056611, + "learning_rate": 0.0009235060782660655, + "loss": 1.5559, + "step": 8133 + }, + { + "epoch": 0.6256441812168295, + "learning_rate": 0.0009231714701455123, + "loss": 1.0572, + "step": 8134 + }, + { + "epoch": 0.6257210983770479, + "learning_rate": 0.0009228368957065788, + "loss": 1.2197, + "step": 8135 + }, + { + "epoch": 0.6257980155372663, + "learning_rate": 0.0009225023549688005, + "loss": 1.1472, + "step": 8136 + }, + { + "epoch": 0.6258749326974848, + "learning_rate": 0.0009221678479517106, + "loss": 1.4049, + "step": 8137 + }, + { + "epoch": 0.6259518498577032, + "learning_rate": 0.0009218333746748432, + "loss": 1.0489, + "step": 8138 + }, + { + "epoch": 0.6260287670179217, + "learning_rate": 0.0009214989351577274, + "loss": 1.3332, + "step": 8139 + }, + { + "epoch": 0.6261056841781402, + "learning_rate": 0.0009211645294198912, + "loss": 1.7117, + "step": 8140 + }, + { + "epoch": 0.6261826013383586, + "learning_rate": 0.0009208301574808618, + "loss": 1.2354, + "step": 8141 + }, + { + "epoch": 0.6262595184985771, + "learning_rate": 0.0009204958193601627, + "loss": 1.0934, + "step": 8142 + }, + { + "epoch": 0.6263364356587955, + "learning_rate": 0.0009201615150773169, + "loss": 1.2804, + "step": 8143 + }, + { + "epoch": 0.6264133528190139, + "learning_rate": 0.0009198272446518449, + "loss": 0.9547, + "step": 8144 + }, + { + "epoch": 0.6264902699792324, + "learning_rate": 0.0009194930081032641, + "loss": 0.735, + "step": 8145 + }, + { + "epoch": 0.6265671871394508, + "learning_rate": 0.000919158805451092, + "loss": 1.0617, + "step": 8146 + }, + { + "epoch": 0.6266441042996692, + "learning_rate": 0.0009188246367148426, + "loss": 1.3298, + "step": 8147 + }, + { + "epoch": 0.6267210214598877, + "learning_rate": 0.0009184905019140277, + "loss": 1.2708, + "step": 8148 + }, + { + "epoch": 0.6267979386201061, + "learning_rate": 0.0009181564010681593, + "loss": 0.9733, + "step": 8149 + }, + { + "epoch": 0.6268748557803245, + "learning_rate": 0.0009178223341967447, + "loss": 1.524, + "step": 8150 + }, + { + "epoch": 0.626951772940543, + "learning_rate": 0.0009174883013192904, + "loss": 1.04, + "step": 8151 + }, + { + "epoch": 0.6270286901007615, + "learning_rate": 0.0009171543024553016, + "loss": 0.8358, + "step": 8152 + }, + { + "epoch": 0.62710560726098, + "learning_rate": 0.0009168203376242803, + "loss": 1.2077, + "step": 8153 + }, + { + "epoch": 0.6271825244211984, + "learning_rate": 0.0009164864068457275, + "loss": 1.3779, + "step": 8154 + }, + { + "epoch": 0.6272594415814168, + "learning_rate": 0.0009161525101391416, + "loss": 0.8769, + "step": 8155 + }, + { + "epoch": 0.6273363587416353, + "learning_rate": 0.0009158186475240187, + "loss": 1.0898, + "step": 8156 + }, + { + "epoch": 0.6274132759018537, + "learning_rate": 0.0009154848190198539, + "loss": 1.0027, + "step": 8157 + }, + { + "epoch": 0.6274901930620721, + "learning_rate": 0.0009151510246461403, + "loss": 1.0532, + "step": 8158 + }, + { + "epoch": 0.6275671102222906, + "learning_rate": 0.0009148172644223669, + "loss": 1.1443, + "step": 8159 + }, + { + "epoch": 0.627644027382509, + "learning_rate": 0.000914483538368024, + "loss": 1.1674, + "step": 8160 + }, + { + "epoch": 0.6277209445427275, + "learning_rate": 0.0009141498465025968, + "loss": 1.0831, + "step": 8161 + }, + { + "epoch": 0.6277978617029459, + "learning_rate": 0.0009138161888455716, + "loss": 1.1636, + "step": 8162 + }, + { + "epoch": 0.6278747788631643, + "learning_rate": 0.0009134825654164295, + "loss": 0.9123, + "step": 8163 + }, + { + "epoch": 0.6279516960233829, + "learning_rate": 0.0009131489762346515, + "loss": 1.1613, + "step": 8164 + }, + { + "epoch": 0.6280286131836013, + "learning_rate": 0.0009128154213197168, + "loss": 1.4738, + "step": 8165 + }, + { + "epoch": 0.6281055303438197, + "learning_rate": 0.0009124819006911015, + "loss": 1.2864, + "step": 8166 + }, + { + "epoch": 0.6281824475040382, + "learning_rate": 0.0009121484143682799, + "loss": 1.0098, + "step": 8167 + }, + { + "epoch": 0.6282593646642566, + "learning_rate": 0.0009118149623707254, + "loss": 1.2836, + "step": 8168 + }, + { + "epoch": 0.6283362818244751, + "learning_rate": 0.000911481544717908, + "loss": 0.5749, + "step": 8169 + }, + { + "epoch": 0.6284131989846935, + "learning_rate": 0.0009111481614292968, + "loss": 1.1781, + "step": 8170 + }, + { + "epoch": 0.6284901161449119, + "learning_rate": 0.0009108148125243584, + "loss": 0.7363, + "step": 8171 + }, + { + "epoch": 0.6285670333051304, + "learning_rate": 0.0009104814980225562, + "loss": 1.1923, + "step": 8172 + }, + { + "epoch": 0.6286439504653488, + "learning_rate": 0.0009101482179433548, + "loss": 1.1633, + "step": 8173 + }, + { + "epoch": 0.6287208676255672, + "learning_rate": 0.0009098149723062132, + "loss": 1.5104, + "step": 8174 + }, + { + "epoch": 0.6287977847857857, + "learning_rate": 0.0009094817611305902, + "loss": 1.0493, + "step": 8175 + }, + { + "epoch": 0.6288747019460041, + "learning_rate": 0.0009091485844359428, + "loss": 0.8758, + "step": 8176 + }, + { + "epoch": 0.6289516191062225, + "learning_rate": 0.0009088154422417255, + "loss": 1.1665, + "step": 8177 + }, + { + "epoch": 0.6290285362664411, + "learning_rate": 0.0009084823345673903, + "loss": 0.8523, + "step": 8178 + }, + { + "epoch": 0.6291054534266595, + "learning_rate": 0.0009081492614323882, + "loss": 1.1758, + "step": 8179 + }, + { + "epoch": 0.629182370586878, + "learning_rate": 0.0009078162228561674, + "loss": 0.9003, + "step": 8180 + }, + { + "epoch": 0.6292592877470964, + "learning_rate": 0.0009074832188581749, + "loss": 0.8993, + "step": 8181 + }, + { + "epoch": 0.6293362049073148, + "learning_rate": 0.0009071502494578545, + "loss": 1.4683, + "step": 8182 + }, + { + "epoch": 0.6294131220675333, + "learning_rate": 0.0009068173146746488, + "loss": 0.7721, + "step": 8183 + }, + { + "epoch": 0.6294900392277517, + "learning_rate": 0.0009064844145279985, + "loss": 1.0183, + "step": 8184 + }, + { + "epoch": 0.6295669563879701, + "learning_rate": 0.0009061515490373421, + "loss": 1.184, + "step": 8185 + }, + { + "epoch": 0.6296438735481886, + "learning_rate": 0.0009058187182221148, + "loss": 1.1737, + "step": 8186 + }, + { + "epoch": 0.629720790708407, + "learning_rate": 0.0009054859221017524, + "loss": 1.2641, + "step": 8187 + }, + { + "epoch": 0.6297977078686255, + "learning_rate": 0.0009051531606956863, + "loss": 1.1508, + "step": 8188 + }, + { + "epoch": 0.629874625028844, + "learning_rate": 0.0009048204340233474, + "loss": 1.3649, + "step": 8189 + }, + { + "epoch": 0.6299515421890624, + "learning_rate": 0.0009044877421041633, + "loss": 1.2373, + "step": 8190 + }, + { + "epoch": 0.6300284593492809, + "learning_rate": 0.0009041550849575605, + "loss": 1.1942, + "step": 8191 + }, + { + "epoch": 0.6301053765094993, + "learning_rate": 0.0009038224626029636, + "loss": 0.8716, + "step": 8192 + }, + { + "epoch": 0.6301822936697177, + "learning_rate": 0.0009034898750597945, + "loss": 1.1884, + "step": 8193 + }, + { + "epoch": 0.6302592108299362, + "learning_rate": 0.0009031573223474729, + "loss": 0.9751, + "step": 8194 + }, + { + "epoch": 0.6303361279901546, + "learning_rate": 0.0009028248044854175, + "loss": 1.0067, + "step": 8195 + }, + { + "epoch": 0.630413045150373, + "learning_rate": 0.0009024923214930443, + "loss": 0.8566, + "step": 8196 + }, + { + "epoch": 0.6304899623105915, + "learning_rate": 0.0009021598733897664, + "loss": 1.1764, + "step": 8197 + }, + { + "epoch": 0.6305668794708099, + "learning_rate": 0.0009018274601949975, + "loss": 1.1409, + "step": 8198 + }, + { + "epoch": 0.6306437966310284, + "learning_rate": 0.0009014950819281459, + "loss": 1.0688, + "step": 8199 + }, + { + "epoch": 0.6307207137912468, + "learning_rate": 0.0009011627386086206, + "loss": 0.8363, + "step": 8200 + }, + { + "epoch": 0.6307976309514652, + "learning_rate": 0.0009008304302558269, + "loss": 0.9203, + "step": 8201 + }, + { + "epoch": 0.6308745481116838, + "learning_rate": 0.0009004981568891688, + "loss": 1.0089, + "step": 8202 + }, + { + "epoch": 0.6309514652719022, + "learning_rate": 0.0009001659185280482, + "loss": 1.5066, + "step": 8203 + }, + { + "epoch": 0.6310283824321206, + "learning_rate": 0.0008998337151918649, + "loss": 1.5113, + "step": 8204 + }, + { + "epoch": 0.6311052995923391, + "learning_rate": 0.000899501546900016, + "loss": 1.0865, + "step": 8205 + }, + { + "epoch": 0.6311822167525575, + "learning_rate": 0.000899169413671898, + "loss": 1.0892, + "step": 8206 + }, + { + "epoch": 0.631259133912776, + "learning_rate": 0.0008988373155269037, + "loss": 1.3598, + "step": 8207 + }, + { + "epoch": 0.6313360510729944, + "learning_rate": 0.0008985052524844254, + "loss": 1.3512, + "step": 8208 + }, + { + "epoch": 0.6314129682332128, + "learning_rate": 0.0008981732245638528, + "loss": 1.119, + "step": 8209 + }, + { + "epoch": 0.6314898853934313, + "learning_rate": 0.0008978412317845717, + "loss": 0.883, + "step": 8210 + }, + { + "epoch": 0.6315668025536497, + "learning_rate": 0.0008975092741659694, + "loss": 1.031, + "step": 8211 + }, + { + "epoch": 0.6316437197138681, + "learning_rate": 0.0008971773517274284, + "loss": 0.9025, + "step": 8212 + }, + { + "epoch": 0.6317206368740866, + "learning_rate": 0.0008968454644883296, + "loss": 1.5185, + "step": 8213 + }, + { + "epoch": 0.631797554034305, + "learning_rate": 0.0008965136124680533, + "loss": 1.0785, + "step": 8214 + }, + { + "epoch": 0.6318744711945234, + "learning_rate": 0.0008961817956859759, + "loss": 1.4106, + "step": 8215 + }, + { + "epoch": 0.631951388354742, + "learning_rate": 0.0008958500141614725, + "loss": 0.8974, + "step": 8216 + }, + { + "epoch": 0.6320283055149604, + "learning_rate": 0.0008955182679139165, + "loss": 1.1247, + "step": 8217 + }, + { + "epoch": 0.6321052226751789, + "learning_rate": 0.0008951865569626788, + "loss": 1.3157, + "step": 8218 + }, + { + "epoch": 0.6321821398353973, + "learning_rate": 0.0008948548813271285, + "loss": 0.957, + "step": 8219 + }, + { + "epoch": 0.6322590569956157, + "learning_rate": 0.0008945232410266328, + "loss": 1.1534, + "step": 8220 + }, + { + "epoch": 0.6323359741558342, + "learning_rate": 0.0008941916360805551, + "loss": 1.1409, + "step": 8221 + }, + { + "epoch": 0.6324128913160526, + "learning_rate": 0.00089386006650826, + "loss": 1.2061, + "step": 8222 + }, + { + "epoch": 0.632489808476271, + "learning_rate": 0.0008935285323291069, + "loss": 1.0451, + "step": 8223 + }, + { + "epoch": 0.6325667256364895, + "learning_rate": 0.0008931970335624548, + "loss": 1.1052, + "step": 8224 + }, + { + "epoch": 0.6326436427967079, + "learning_rate": 0.0008928655702276605, + "loss": 1.0595, + "step": 8225 + }, + { + "epoch": 0.6327205599569264, + "learning_rate": 0.0008925341423440779, + "loss": 1.0323, + "step": 8226 + }, + { + "epoch": 0.6327974771171448, + "learning_rate": 0.0008922027499310602, + "loss": 0.9534, + "step": 8227 + }, + { + "epoch": 0.6328743942773633, + "learning_rate": 0.0008918713930079574, + "loss": 0.8965, + "step": 8228 + }, + { + "epoch": 0.6329513114375818, + "learning_rate": 0.0008915400715941174, + "loss": 1.2773, + "step": 8229 + }, + { + "epoch": 0.6330282285978002, + "learning_rate": 0.0008912087857088868, + "loss": 1.133, + "step": 8230 + }, + { + "epoch": 0.6331051457580186, + "learning_rate": 0.0008908775353716103, + "loss": 1.3332, + "step": 8231 + }, + { + "epoch": 0.6331820629182371, + "learning_rate": 0.0008905463206016283, + "loss": 1.1906, + "step": 8232 + }, + { + "epoch": 0.6332589800784555, + "learning_rate": 0.0008902151414182828, + "loss": 1.9048, + "step": 8233 + }, + { + "epoch": 0.6333358972386739, + "learning_rate": 0.0008898839978409103, + "loss": 1.2236, + "step": 8234 + }, + { + "epoch": 0.6334128143988924, + "learning_rate": 0.0008895528898888467, + "loss": 0.9474, + "step": 8235 + }, + { + "epoch": 0.6334897315591108, + "learning_rate": 0.0008892218175814265, + "loss": 1.0123, + "step": 8236 + }, + { + "epoch": 0.6335666487193293, + "learning_rate": 0.0008888907809379804, + "loss": 1.0698, + "step": 8237 + }, + { + "epoch": 0.6336435658795477, + "learning_rate": 0.0008885597799778389, + "loss": 1.343, + "step": 8238 + }, + { + "epoch": 0.6337204830397661, + "learning_rate": 0.0008882288147203289, + "loss": 1.156, + "step": 8239 + }, + { + "epoch": 0.6337974001999847, + "learning_rate": 0.0008878978851847759, + "loss": 1.2981, + "step": 8240 + }, + { + "epoch": 0.633874317360203, + "learning_rate": 0.0008875669913905034, + "loss": 1.0384, + "step": 8241 + }, + { + "epoch": 0.6339512345204215, + "learning_rate": 0.0008872361333568332, + "loss": 0.9956, + "step": 8242 + }, + { + "epoch": 0.63402815168064, + "learning_rate": 0.0008869053111030825, + "loss": 1.084, + "step": 8243 + }, + { + "epoch": 0.6341050688408584, + "learning_rate": 0.0008865745246485708, + "loss": 0.9186, + "step": 8244 + }, + { + "epoch": 0.6341819860010769, + "learning_rate": 0.0008862437740126107, + "loss": 0.9003, + "step": 8245 + }, + { + "epoch": 0.6342589031612953, + "learning_rate": 0.0008859130592145175, + "loss": 1.1746, + "step": 8246 + }, + { + "epoch": 0.6343358203215137, + "learning_rate": 0.0008855823802736001, + "loss": 1.2765, + "step": 8247 + }, + { + "epoch": 0.6344127374817322, + "learning_rate": 0.0008852517372091679, + "loss": 1.2865, + "step": 8248 + }, + { + "epoch": 0.6344896546419506, + "learning_rate": 0.0008849211300405278, + "loss": 1.1054, + "step": 8249 + }, + { + "epoch": 0.634566571802169, + "learning_rate": 0.0008845905587869835, + "loss": 1.122, + "step": 8250 + }, + { + "epoch": 0.6346434889623875, + "learning_rate": 0.0008842600234678379, + "loss": 0.8618, + "step": 8251 + }, + { + "epoch": 0.6347204061226059, + "learning_rate": 0.0008839295241023915, + "loss": 1.2543, + "step": 8252 + }, + { + "epoch": 0.6347973232828243, + "learning_rate": 0.0008835990607099423, + "loss": 1.0198, + "step": 8253 + }, + { + "epoch": 0.6348742404430429, + "learning_rate": 0.0008832686333097857, + "loss": 1.2164, + "step": 8254 + }, + { + "epoch": 0.6349511576032613, + "learning_rate": 0.0008829382419212171, + "loss": 1.1041, + "step": 8255 + }, + { + "epoch": 0.6350280747634798, + "learning_rate": 0.0008826078865635267, + "loss": 0.9281, + "step": 8256 + }, + { + "epoch": 0.6351049919236982, + "learning_rate": 0.0008822775672560064, + "loss": 0.991, + "step": 8257 + }, + { + "epoch": 0.6351819090839166, + "learning_rate": 0.0008819472840179422, + "loss": 0.8556, + "step": 8258 + }, + { + "epoch": 0.6352588262441351, + "learning_rate": 0.0008816170368686199, + "loss": 1.1821, + "step": 8259 + }, + { + "epoch": 0.6353357434043535, + "learning_rate": 0.0008812868258273236, + "loss": 1.1493, + "step": 8260 + }, + { + "epoch": 0.6354126605645719, + "learning_rate": 0.0008809566509133343, + "loss": 0.9212, + "step": 8261 + }, + { + "epoch": 0.6354895777247904, + "learning_rate": 0.000880626512145931, + "loss": 1.503, + "step": 8262 + }, + { + "epoch": 0.6355664948850088, + "learning_rate": 0.0008802964095443917, + "loss": 1.1168, + "step": 8263 + }, + { + "epoch": 0.6356434120452273, + "learning_rate": 0.0008799663431279903, + "loss": 1.1888, + "step": 8264 + }, + { + "epoch": 0.6357203292054457, + "learning_rate": 0.0008796363129160009, + "loss": 1.1403, + "step": 8265 + }, + { + "epoch": 0.6357972463656641, + "learning_rate": 0.0008793063189276937, + "loss": 1.4732, + "step": 8266 + }, + { + "epoch": 0.6358741635258827, + "learning_rate": 0.0008789763611823366, + "loss": 1.4032, + "step": 8267 + }, + { + "epoch": 0.6359510806861011, + "learning_rate": 0.000878646439699198, + "loss": 1.2138, + "step": 8268 + }, + { + "epoch": 0.6360279978463195, + "learning_rate": 0.000878316554497541, + "loss": 1.1688, + "step": 8269 + }, + { + "epoch": 0.636104915006538, + "learning_rate": 0.0008779867055966279, + "loss": 1.0932, + "step": 8270 + }, + { + "epoch": 0.6361818321667564, + "learning_rate": 0.0008776568930157198, + "loss": 1.4304, + "step": 8271 + }, + { + "epoch": 0.6362587493269748, + "learning_rate": 0.0008773271167740741, + "loss": 1.0514, + "step": 8272 + }, + { + "epoch": 0.6363356664871933, + "learning_rate": 0.0008769973768909467, + "loss": 0.9654, + "step": 8273 + }, + { + "epoch": 0.6364125836474117, + "learning_rate": 0.0008766676733855921, + "loss": 0.9579, + "step": 8274 + }, + { + "epoch": 0.6364895008076302, + "learning_rate": 0.0008763380062772614, + "loss": 0.7052, + "step": 8275 + }, + { + "epoch": 0.6365664179678486, + "learning_rate": 0.0008760083755852048, + "loss": 0.9583, + "step": 8276 + }, + { + "epoch": 0.636643335128067, + "learning_rate": 0.0008756787813286695, + "loss": 1.5324, + "step": 8277 + }, + { + "epoch": 0.6367202522882855, + "learning_rate": 0.0008753492235269007, + "loss": 1.0465, + "step": 8278 + }, + { + "epoch": 0.636797169448504, + "learning_rate": 0.0008750197021991417, + "loss": 0.9068, + "step": 8279 + }, + { + "epoch": 0.6368740866087224, + "learning_rate": 0.0008746902173646337, + "loss": 0.9917, + "step": 8280 + }, + { + "epoch": 0.6369510037689409, + "learning_rate": 0.000874360769042615, + "loss": 0.9032, + "step": 8281 + }, + { + "epoch": 0.6370279209291593, + "learning_rate": 0.0008740313572523236, + "loss": 1.0142, + "step": 8282 + }, + { + "epoch": 0.6371048380893778, + "learning_rate": 0.0008737019820129937, + "loss": 1.4506, + "step": 8283 + }, + { + "epoch": 0.6371817552495962, + "learning_rate": 0.0008733726433438577, + "loss": 1.026, + "step": 8284 + }, + { + "epoch": 0.6372586724098146, + "learning_rate": 0.0008730433412641461, + "loss": 0.8937, + "step": 8285 + }, + { + "epoch": 0.6373355895700331, + "learning_rate": 0.0008727140757930866, + "loss": 1.0528, + "step": 8286 + }, + { + "epoch": 0.6374125067302515, + "learning_rate": 0.0008723848469499072, + "loss": 1.2463, + "step": 8287 + }, + { + "epoch": 0.6374894238904699, + "learning_rate": 0.0008720556547538299, + "loss": 1.297, + "step": 8288 + }, + { + "epoch": 0.6375663410506884, + "learning_rate": 0.000871726499224077, + "loss": 1.1878, + "step": 8289 + }, + { + "epoch": 0.6376432582109068, + "learning_rate": 0.000871397380379869, + "loss": 1.2336, + "step": 8290 + }, + { + "epoch": 0.6377201753711254, + "learning_rate": 0.0008710682982404233, + "loss": 1.1063, + "step": 8291 + }, + { + "epoch": 0.6377970925313438, + "learning_rate": 0.0008707392528249548, + "loss": 0.9645, + "step": 8292 + }, + { + "epoch": 0.6378740096915622, + "learning_rate": 0.0008704102441526772, + "loss": 0.6542, + "step": 8293 + }, + { + "epoch": 0.6379509268517807, + "learning_rate": 0.0008700812722428011, + "loss": 1.0287, + "step": 8294 + }, + { + "epoch": 0.6380278440119991, + "learning_rate": 0.0008697523371145367, + "loss": 1.1559, + "step": 8295 + }, + { + "epoch": 0.6381047611722175, + "learning_rate": 0.0008694234387870906, + "loss": 1.142, + "step": 8296 + }, + { + "epoch": 0.638181678332436, + "learning_rate": 0.0008690945772796657, + "loss": 1.09, + "step": 8297 + }, + { + "epoch": 0.6382585954926544, + "learning_rate": 0.0008687657526114668, + "loss": 1.2456, + "step": 8298 + }, + { + "epoch": 0.6383355126528728, + "learning_rate": 0.0008684369648016935, + "loss": 1.1513, + "step": 8299 + }, + { + "epoch": 0.6384124298130913, + "learning_rate": 0.0008681082138695441, + "loss": 1.1071, + "step": 8300 + }, + { + "epoch": 0.6384893469733097, + "learning_rate": 0.0008677794998342144, + "loss": 0.9978, + "step": 8301 + }, + { + "epoch": 0.6385662641335282, + "learning_rate": 0.0008674508227148981, + "loss": 0.801, + "step": 8302 + }, + { + "epoch": 0.6386431812937466, + "learning_rate": 0.000867122182530788, + "loss": 1.0011, + "step": 8303 + }, + { + "epoch": 0.638720098453965, + "learning_rate": 0.0008667935793010733, + "loss": 1.017, + "step": 8304 + }, + { + "epoch": 0.6387970156141836, + "learning_rate": 0.0008664650130449414, + "loss": 1.3035, + "step": 8305 + }, + { + "epoch": 0.638873932774402, + "learning_rate": 0.0008661364837815777, + "loss": 1.1366, + "step": 8306 + }, + { + "epoch": 0.6389508499346204, + "learning_rate": 0.0008658079915301652, + "loss": 1.0112, + "step": 8307 + }, + { + "epoch": 0.6390277670948389, + "learning_rate": 0.0008654795363098845, + "loss": 1.348, + "step": 8308 + }, + { + "epoch": 0.6391046842550573, + "learning_rate": 0.0008651511181399163, + "loss": 1.2637, + "step": 8309 + }, + { + "epoch": 0.6391816014152758, + "learning_rate": 0.0008648227370394344, + "loss": 1.4518, + "step": 8310 + }, + { + "epoch": 0.6392585185754942, + "learning_rate": 0.0008644943930276157, + "loss": 0.9448, + "step": 8311 + }, + { + "epoch": 0.6393354357357126, + "learning_rate": 0.0008641660861236315, + "loss": 0.8564, + "step": 8312 + }, + { + "epoch": 0.6394123528959311, + "learning_rate": 0.0008638378163466524, + "loss": 1.0912, + "step": 8313 + }, + { + "epoch": 0.6394892700561495, + "learning_rate": 0.0008635095837158461, + "loss": 1.1062, + "step": 8314 + }, + { + "epoch": 0.6395661872163679, + "learning_rate": 0.0008631813882503785, + "loss": 1.1459, + "step": 8315 + }, + { + "epoch": 0.6396431043765864, + "learning_rate": 0.0008628532299694129, + "loss": 1.1181, + "step": 8316 + }, + { + "epoch": 0.6397200215368049, + "learning_rate": 0.0008625251088921114, + "loss": 0.8185, + "step": 8317 + }, + { + "epoch": 0.6397969386970233, + "learning_rate": 0.0008621970250376342, + "loss": 1.2591, + "step": 8318 + }, + { + "epoch": 0.6398738558572418, + "learning_rate": 0.0008618689784251362, + "loss": 1.5552, + "step": 8319 + }, + { + "epoch": 0.6399507730174602, + "learning_rate": 0.0008615409690737739, + "loss": 0.9654, + "step": 8320 + }, + { + "epoch": 0.6400276901776787, + "learning_rate": 0.0008612129970026991, + "loss": 0.9854, + "step": 8321 + }, + { + "epoch": 0.6401046073378971, + "learning_rate": 0.0008608850622310646, + "loss": 1.2055, + "step": 8322 + }, + { + "epoch": 0.6401815244981155, + "learning_rate": 0.0008605571647780168, + "loss": 1.1933, + "step": 8323 + }, + { + "epoch": 0.640258441658334, + "learning_rate": 0.000860229304662702, + "loss": 1.2742, + "step": 8324 + }, + { + "epoch": 0.6403353588185524, + "learning_rate": 0.0008599014819042653, + "loss": 1.0467, + "step": 8325 + }, + { + "epoch": 0.6404122759787708, + "learning_rate": 0.0008595736965218483, + "loss": 1.3372, + "step": 8326 + }, + { + "epoch": 0.6404891931389893, + "learning_rate": 0.0008592459485345905, + "loss": 0.9231, + "step": 8327 + }, + { + "epoch": 0.6405661102992077, + "learning_rate": 0.0008589182379616296, + "loss": 1.0623, + "step": 8328 + }, + { + "epoch": 0.6406430274594263, + "learning_rate": 0.0008585905648221004, + "loss": 0.6794, + "step": 8329 + }, + { + "epoch": 0.6407199446196447, + "learning_rate": 0.0008582629291351369, + "loss": 1.6995, + "step": 8330 + }, + { + "epoch": 0.6407968617798631, + "learning_rate": 0.0008579353309198705, + "loss": 0.9766, + "step": 8331 + }, + { + "epoch": 0.6408737789400816, + "learning_rate": 0.0008576077701954282, + "loss": 1.282, + "step": 8332 + }, + { + "epoch": 0.6409506961003, + "learning_rate": 0.0008572802469809381, + "loss": 1.0178, + "step": 8333 + }, + { + "epoch": 0.6410276132605184, + "learning_rate": 0.0008569527612955245, + "loss": 1.1893, + "step": 8334 + }, + { + "epoch": 0.6411045304207369, + "learning_rate": 0.000856625313158309, + "loss": 1.3076, + "step": 8335 + }, + { + "epoch": 0.6411814475809553, + "learning_rate": 0.0008562979025884121, + "loss": 1.2457, + "step": 8336 + }, + { + "epoch": 0.6412583647411737, + "learning_rate": 0.0008559705296049515, + "loss": 1.2937, + "step": 8337 + }, + { + "epoch": 0.6413352819013922, + "learning_rate": 0.0008556431942270422, + "loss": 1.1312, + "step": 8338 + }, + { + "epoch": 0.6414121990616106, + "learning_rate": 0.000855315896473799, + "loss": 1.4275, + "step": 8339 + }, + { + "epoch": 0.6414891162218291, + "learning_rate": 0.0008549886363643326, + "loss": 0.9869, + "step": 8340 + }, + { + "epoch": 0.6415660333820475, + "learning_rate": 0.0008546614139177518, + "loss": 1.3316, + "step": 8341 + }, + { + "epoch": 0.6416429505422659, + "learning_rate": 0.0008543342291531636, + "loss": 0.914, + "step": 8342 + }, + { + "epoch": 0.6417198677024845, + "learning_rate": 0.0008540070820896721, + "loss": 1.2597, + "step": 8343 + }, + { + "epoch": 0.6417967848627029, + "learning_rate": 0.0008536799727463815, + "loss": 1.0967, + "step": 8344 + }, + { + "epoch": 0.6418737020229213, + "learning_rate": 0.0008533529011423902, + "loss": 1.4358, + "step": 8345 + }, + { + "epoch": 0.6419506191831398, + "learning_rate": 0.0008530258672967963, + "loss": 1.2265, + "step": 8346 + }, + { + "epoch": 0.6420275363433582, + "learning_rate": 0.000852698871228697, + "loss": 1.0294, + "step": 8347 + }, + { + "epoch": 0.6421044535035767, + "learning_rate": 0.0008523719129571852, + "loss": 0.8969, + "step": 8348 + }, + { + "epoch": 0.6421813706637951, + "learning_rate": 0.0008520449925013524, + "loss": 1.1363, + "step": 8349 + }, + { + "epoch": 0.6422582878240135, + "learning_rate": 0.0008517181098802875, + "loss": 0.9958, + "step": 8350 + }, + { + "epoch": 0.642335204984232, + "learning_rate": 0.0008513912651130774, + "loss": 0.9373, + "step": 8351 + }, + { + "epoch": 0.6424121221444504, + "learning_rate": 0.0008510644582188078, + "loss": 1.4266, + "step": 8352 + }, + { + "epoch": 0.6424890393046688, + "learning_rate": 0.0008507376892165616, + "loss": 1.178, + "step": 8353 + }, + { + "epoch": 0.6425659564648873, + "learning_rate": 0.000850410958125417, + "loss": 1.2896, + "step": 8354 + }, + { + "epoch": 0.6426428736251057, + "learning_rate": 0.0008500842649644541, + "loss": 1.1034, + "step": 8355 + }, + { + "epoch": 0.6427197907853242, + "learning_rate": 0.0008497576097527483, + "loss": 1.2321, + "step": 8356 + }, + { + "epoch": 0.6427967079455427, + "learning_rate": 0.0008494309925093735, + "loss": 1.1474, + "step": 8357 + }, + { + "epoch": 0.6428736251057611, + "learning_rate": 0.0008491044132534008, + "loss": 1.0116, + "step": 8358 + }, + { + "epoch": 0.6429505422659796, + "learning_rate": 0.0008487778720038993, + "loss": 1.0924, + "step": 8359 + }, + { + "epoch": 0.643027459426198, + "learning_rate": 0.0008484513687799372, + "loss": 1.1276, + "step": 8360 + }, + { + "epoch": 0.6431043765864164, + "learning_rate": 0.0008481249036005786, + "loss": 0.8738, + "step": 8361 + }, + { + "epoch": 0.6431812937466349, + "learning_rate": 0.0008477984764848864, + "loss": 1.0035, + "step": 8362 + }, + { + "epoch": 0.6432582109068533, + "learning_rate": 0.0008474720874519208, + "loss": 1.0124, + "step": 8363 + }, + { + "epoch": 0.6433351280670717, + "learning_rate": 0.00084714573652074, + "loss": 1.0515, + "step": 8364 + }, + { + "epoch": 0.6434120452272902, + "learning_rate": 0.0008468194237103994, + "loss": 1.1871, + "step": 8365 + }, + { + "epoch": 0.6434889623875086, + "learning_rate": 0.0008464931490399546, + "loss": 1.1569, + "step": 8366 + }, + { + "epoch": 0.6435658795477271, + "learning_rate": 0.0008461669125284548, + "loss": 1.3059, + "step": 8367 + }, + { + "epoch": 0.6436427967079456, + "learning_rate": 0.000845840714194951, + "loss": 1.0381, + "step": 8368 + }, + { + "epoch": 0.643719713868164, + "learning_rate": 0.0008455145540584894, + "loss": 1.3407, + "step": 8369 + }, + { + "epoch": 0.6437966310283825, + "learning_rate": 0.0008451884321381151, + "loss": 1.3718, + "step": 8370 + }, + { + "epoch": 0.6438735481886009, + "learning_rate": 0.0008448623484528708, + "loss": 1.267, + "step": 8371 + }, + { + "epoch": 0.6439504653488193, + "learning_rate": 0.000844536303021797, + "loss": 0.8983, + "step": 8372 + }, + { + "epoch": 0.6440273825090378, + "learning_rate": 0.0008442102958639305, + "loss": 1.2889, + "step": 8373 + }, + { + "epoch": 0.6441042996692562, + "learning_rate": 0.0008438843269983092, + "loss": 0.8949, + "step": 8374 + }, + { + "epoch": 0.6441812168294746, + "learning_rate": 0.0008435583964439663, + "loss": 0.898, + "step": 8375 + }, + { + "epoch": 0.6442581339896931, + "learning_rate": 0.0008432325042199317, + "loss": 1.0528, + "step": 8376 + }, + { + "epoch": 0.6443350511499115, + "learning_rate": 0.0008429066503452359, + "loss": 0.7661, + "step": 8377 + }, + { + "epoch": 0.64441196831013, + "learning_rate": 0.0008425808348389054, + "loss": 1.3702, + "step": 8378 + }, + { + "epoch": 0.6444888854703484, + "learning_rate": 0.000842255057719966, + "loss": 0.8219, + "step": 8379 + }, + { + "epoch": 0.6445658026305668, + "learning_rate": 0.0008419293190074389, + "loss": 1.3809, + "step": 8380 + }, + { + "epoch": 0.6446427197907854, + "learning_rate": 0.000841603618720344, + "loss": 1.9077, + "step": 8381 + }, + { + "epoch": 0.6447196369510038, + "learning_rate": 0.0008412779568777005, + "loss": 0.8228, + "step": 8382 + }, + { + "epoch": 0.6447965541112222, + "learning_rate": 0.0008409523334985238, + "loss": 1.405, + "step": 8383 + }, + { + "epoch": 0.6448734712714407, + "learning_rate": 0.0008406267486018268, + "loss": 1.0319, + "step": 8384 + }, + { + "epoch": 0.6449503884316591, + "learning_rate": 0.0008403012022066215, + "loss": 1.0342, + "step": 8385 + }, + { + "epoch": 0.6450273055918776, + "learning_rate": 0.0008399756943319157, + "loss": 1.2664, + "step": 8386 + }, + { + "epoch": 0.645104222752096, + "learning_rate": 0.0008396502249967177, + "loss": 0.888, + "step": 8387 + }, + { + "epoch": 0.6451811399123144, + "learning_rate": 0.000839324794220032, + "loss": 1.3294, + "step": 8388 + }, + { + "epoch": 0.6452580570725329, + "learning_rate": 0.0008389994020208587, + "loss": 1.1977, + "step": 8389 + }, + { + "epoch": 0.6453349742327513, + "learning_rate": 0.0008386740484181998, + "loss": 1.0861, + "step": 8390 + }, + { + "epoch": 0.6454118913929697, + "learning_rate": 0.0008383487334310523, + "loss": 1.0666, + "step": 8391 + }, + { + "epoch": 0.6454888085531882, + "learning_rate": 0.000838023457078412, + "loss": 0.9194, + "step": 8392 + }, + { + "epoch": 0.6455657257134066, + "learning_rate": 0.0008376982193792718, + "loss": 1.4254, + "step": 8393 + }, + { + "epoch": 0.6456426428736252, + "learning_rate": 0.0008373730203526226, + "loss": 1.1746, + "step": 8394 + }, + { + "epoch": 0.6457195600338436, + "learning_rate": 0.0008370478600174535, + "loss": 0.8879, + "step": 8395 + }, + { + "epoch": 0.645796477194062, + "learning_rate": 0.0008367227383927508, + "loss": 1.0596, + "step": 8396 + }, + { + "epoch": 0.6458733943542805, + "learning_rate": 0.0008363976554974987, + "loss": 1.005, + "step": 8397 + }, + { + "epoch": 0.6459503115144989, + "learning_rate": 0.0008360726113506792, + "loss": 1.2018, + "step": 8398 + }, + { + "epoch": 0.6460272286747173, + "learning_rate": 0.0008357476059712718, + "loss": 1.1786, + "step": 8399 + }, + { + "epoch": 0.6461041458349358, + "learning_rate": 0.0008354226393782534, + "loss": 1.2608, + "step": 8400 + }, + { + "epoch": 0.6461810629951542, + "learning_rate": 0.0008350977115906007, + "loss": 1.0666, + "step": 8401 + }, + { + "epoch": 0.6462579801553726, + "learning_rate": 0.0008347728226272851, + "loss": 1.3131, + "step": 8402 + }, + { + "epoch": 0.6463348973155911, + "learning_rate": 0.000834447972507277, + "loss": 1.5519, + "step": 8403 + }, + { + "epoch": 0.6464118144758095, + "learning_rate": 0.0008341231612495461, + "loss": 1.1245, + "step": 8404 + }, + { + "epoch": 0.646488731636028, + "learning_rate": 0.000833798388873058, + "loss": 0.9058, + "step": 8405 + }, + { + "epoch": 0.6465656487962465, + "learning_rate": 0.0008334736553967759, + "loss": 1.3004, + "step": 8406 + }, + { + "epoch": 0.6466425659564649, + "learning_rate": 0.0008331489608396618, + "loss": 1.427, + "step": 8407 + }, + { + "epoch": 0.6467194831166834, + "learning_rate": 0.0008328243052206743, + "loss": 1.0121, + "step": 8408 + }, + { + "epoch": 0.6467964002769018, + "learning_rate": 0.0008324996885587714, + "loss": 1.2156, + "step": 8409 + }, + { + "epoch": 0.6468733174371202, + "learning_rate": 0.0008321751108729075, + "loss": 1.3845, + "step": 8410 + }, + { + "epoch": 0.6469502345973387, + "learning_rate": 0.0008318505721820348, + "loss": 0.8668, + "step": 8411 + }, + { + "epoch": 0.6470271517575571, + "learning_rate": 0.0008315260725051036, + "loss": 1.0659, + "step": 8412 + }, + { + "epoch": 0.6471040689177756, + "learning_rate": 0.000831201611861061, + "loss": 1.1882, + "step": 8413 + }, + { + "epoch": 0.647180986077994, + "learning_rate": 0.0008308771902688544, + "loss": 0.8096, + "step": 8414 + }, + { + "epoch": 0.6472579032382124, + "learning_rate": 0.0008305528077474255, + "loss": 0.9452, + "step": 8415 + }, + { + "epoch": 0.6473348203984309, + "learning_rate": 0.0008302284643157153, + "loss": 1.2173, + "step": 8416 + }, + { + "epoch": 0.6474117375586493, + "learning_rate": 0.0008299041599926636, + "loss": 0.9444, + "step": 8417 + }, + { + "epoch": 0.6474886547188677, + "learning_rate": 0.0008295798947972063, + "loss": 0.9073, + "step": 8418 + }, + { + "epoch": 0.6475655718790863, + "learning_rate": 0.0008292556687482778, + "loss": 1.0288, + "step": 8419 + }, + { + "epoch": 0.6476424890393047, + "learning_rate": 0.0008289314818648098, + "loss": 1.2546, + "step": 8420 + }, + { + "epoch": 0.6477194061995231, + "learning_rate": 0.0008286073341657317, + "loss": 1.1336, + "step": 8421 + }, + { + "epoch": 0.6477963233597416, + "learning_rate": 0.0008282832256699709, + "loss": 0.7539, + "step": 8422 + }, + { + "epoch": 0.64787324051996, + "learning_rate": 0.0008279591563964528, + "loss": 0.8988, + "step": 8423 + }, + { + "epoch": 0.6479501576801785, + "learning_rate": 0.0008276351263641002, + "loss": 1.029, + "step": 8424 + }, + { + "epoch": 0.6480270748403969, + "learning_rate": 0.0008273111355918332, + "loss": 0.9604, + "step": 8425 + }, + { + "epoch": 0.6481039920006153, + "learning_rate": 0.0008269871840985701, + "loss": 1.4591, + "step": 8426 + }, + { + "epoch": 0.6481809091608338, + "learning_rate": 0.000826663271903226, + "loss": 1.145, + "step": 8427 + }, + { + "epoch": 0.6482578263210522, + "learning_rate": 0.0008263393990247165, + "loss": 1.0779, + "step": 8428 + }, + { + "epoch": 0.6483347434812706, + "learning_rate": 0.000826015565481951, + "loss": 1.0784, + "step": 8429 + }, + { + "epoch": 0.6484116606414891, + "learning_rate": 0.0008256917712938382, + "loss": 1.0924, + "step": 8430 + }, + { + "epoch": 0.6484885778017075, + "learning_rate": 0.0008253680164792865, + "loss": 1.3725, + "step": 8431 + }, + { + "epoch": 0.6485654949619261, + "learning_rate": 0.0008250443010571994, + "loss": 1.3314, + "step": 8432 + }, + { + "epoch": 0.6486424121221445, + "learning_rate": 0.0008247206250464788, + "loss": 0.8478, + "step": 8433 + }, + { + "epoch": 0.6487193292823629, + "learning_rate": 0.0008243969884660247, + "loss": 1.0771, + "step": 8434 + }, + { + "epoch": 0.6487962464425814, + "learning_rate": 0.0008240733913347341, + "loss": 0.9964, + "step": 8435 + }, + { + "epoch": 0.6488731636027998, + "learning_rate": 0.000823749833671503, + "loss": 1.3977, + "step": 8436 + }, + { + "epoch": 0.6489500807630182, + "learning_rate": 0.0008234263154952245, + "loss": 0.9625, + "step": 8437 + }, + { + "epoch": 0.6490269979232367, + "learning_rate": 0.0008231028368247876, + "loss": 0.9187, + "step": 8438 + }, + { + "epoch": 0.6491039150834551, + "learning_rate": 0.0008227793976790818, + "loss": 1.3996, + "step": 8439 + }, + { + "epoch": 0.6491808322436735, + "learning_rate": 0.0008224559980769928, + "loss": 0.9809, + "step": 8440 + }, + { + "epoch": 0.649257749403892, + "learning_rate": 0.0008221326380374044, + "loss": 1.2785, + "step": 8441 + }, + { + "epoch": 0.6493346665641104, + "learning_rate": 0.0008218093175791976, + "loss": 1.2468, + "step": 8442 + }, + { + "epoch": 0.649411583724329, + "learning_rate": 0.0008214860367212509, + "loss": 1.1019, + "step": 8443 + }, + { + "epoch": 0.6494885008845473, + "learning_rate": 0.0008211627954824422, + "loss": 1.2127, + "step": 8444 + }, + { + "epoch": 0.6495654180447658, + "learning_rate": 0.0008208395938816455, + "loss": 1.3368, + "step": 8445 + }, + { + "epoch": 0.6496423352049843, + "learning_rate": 0.0008205164319377327, + "loss": 1.2483, + "step": 8446 + }, + { + "epoch": 0.6497192523652027, + "learning_rate": 0.0008201933096695735, + "loss": 1.2783, + "step": 8447 + }, + { + "epoch": 0.6497961695254211, + "learning_rate": 0.0008198702270960354, + "loss": 1.3545, + "step": 8448 + }, + { + "epoch": 0.6498730866856396, + "learning_rate": 0.000819547184235983, + "loss": 1.2456, + "step": 8449 + }, + { + "epoch": 0.649950003845858, + "learning_rate": 0.0008192241811082808, + "loss": 1.492, + "step": 8450 + }, + { + "epoch": 0.6500269210060765, + "learning_rate": 0.0008189012177317869, + "loss": 1.0569, + "step": 8451 + }, + { + "epoch": 0.6501038381662949, + "learning_rate": 0.0008185782941253615, + "loss": 1.1524, + "step": 8452 + }, + { + "epoch": 0.6501807553265133, + "learning_rate": 0.0008182554103078596, + "loss": 1.4448, + "step": 8453 + }, + { + "epoch": 0.6502576724867318, + "learning_rate": 0.0008179325662981347, + "loss": 1.024, + "step": 8454 + }, + { + "epoch": 0.6503345896469502, + "learning_rate": 0.0008176097621150382, + "loss": 1.0922, + "step": 8455 + }, + { + "epoch": 0.6504115068071686, + "learning_rate": 0.000817286997777419, + "loss": 2.074, + "step": 8456 + }, + { + "epoch": 0.6504884239673872, + "learning_rate": 0.0008169642733041226, + "loss": 1.0375, + "step": 8457 + }, + { + "epoch": 0.6505653411276056, + "learning_rate": 0.000816641588713995, + "loss": 1.3542, + "step": 8458 + }, + { + "epoch": 0.650642258287824, + "learning_rate": 0.0008163189440258777, + "loss": 0.7715, + "step": 8459 + }, + { + "epoch": 0.6507191754480425, + "learning_rate": 0.0008159963392586084, + "loss": 1.0151, + "step": 8460 + }, + { + "epoch": 0.6507960926082609, + "learning_rate": 0.0008156737744310266, + "loss": 1.2908, + "step": 8461 + }, + { + "epoch": 0.6508730097684794, + "learning_rate": 0.0008153512495619654, + "loss": 1.152, + "step": 8462 + }, + { + "epoch": 0.6509499269286978, + "learning_rate": 0.0008150287646702597, + "loss": 0.6677, + "step": 8463 + }, + { + "epoch": 0.6510268440889162, + "learning_rate": 0.0008147063197747376, + "loss": 1.2038, + "step": 8464 + }, + { + "epoch": 0.6511037612491347, + "learning_rate": 0.0008143839148942271, + "loss": 0.8156, + "step": 8465 + }, + { + "epoch": 0.6511806784093531, + "learning_rate": 0.000814061550047555, + "loss": 1.3697, + "step": 8466 + }, + { + "epoch": 0.6512575955695715, + "learning_rate": 0.0008137392252535436, + "loss": 0.9923, + "step": 8467 + }, + { + "epoch": 0.65133451272979, + "learning_rate": 0.0008134169405310144, + "loss": 1.8218, + "step": 8468 + }, + { + "epoch": 0.6514114298900084, + "learning_rate": 0.0008130946958987855, + "loss": 1.0249, + "step": 8469 + }, + { + "epoch": 0.651488347050227, + "learning_rate": 0.0008127724913756724, + "loss": 1.2627, + "step": 8470 + }, + { + "epoch": 0.6515652642104454, + "learning_rate": 0.0008124503269804905, + "loss": 0.9481, + "step": 8471 + }, + { + "epoch": 0.6516421813706638, + "learning_rate": 0.0008121282027320509, + "loss": 0.8972, + "step": 8472 + }, + { + "epoch": 0.6517190985308823, + "learning_rate": 0.0008118061186491616, + "loss": 1.292, + "step": 8473 + }, + { + "epoch": 0.6517960156911007, + "learning_rate": 0.0008114840747506307, + "loss": 1.2915, + "step": 8474 + }, + { + "epoch": 0.6518729328513191, + "learning_rate": 0.000811162071055262, + "loss": 1.2656, + "step": 8475 + }, + { + "epoch": 0.6519498500115376, + "learning_rate": 0.0008108401075818582, + "loss": 1.2514, + "step": 8476 + }, + { + "epoch": 0.652026767171756, + "learning_rate": 0.0008105181843492184, + "loss": 1.3778, + "step": 8477 + }, + { + "epoch": 0.6521036843319744, + "learning_rate": 0.0008101963013761407, + "loss": 1.355, + "step": 8478 + }, + { + "epoch": 0.6521806014921929, + "learning_rate": 0.0008098744586814192, + "loss": 1.2155, + "step": 8479 + }, + { + "epoch": 0.6522575186524113, + "learning_rate": 0.0008095526562838476, + "loss": 0.9347, + "step": 8480 + }, + { + "epoch": 0.6523344358126298, + "learning_rate": 0.0008092308942022163, + "loss": 1.192, + "step": 8481 + }, + { + "epoch": 0.6524113529728482, + "learning_rate": 0.000808909172455313, + "loss": 1.2006, + "step": 8482 + }, + { + "epoch": 0.6524882701330667, + "learning_rate": 0.0008085874910619232, + "loss": 0.7732, + "step": 8483 + }, + { + "epoch": 0.6525651872932852, + "learning_rate": 0.0008082658500408298, + "loss": 0.8882, + "step": 8484 + }, + { + "epoch": 0.6526421044535036, + "learning_rate": 0.0008079442494108155, + "loss": 0.8389, + "step": 8485 + }, + { + "epoch": 0.652719021613722, + "learning_rate": 0.0008076226891906573, + "loss": 0.9276, + "step": 8486 + }, + { + "epoch": 0.6527959387739405, + "learning_rate": 0.0008073011693991309, + "loss": 1.6867, + "step": 8487 + }, + { + "epoch": 0.6528728559341589, + "learning_rate": 0.0008069796900550119, + "loss": 1.0525, + "step": 8488 + }, + { + "epoch": 0.6529497730943774, + "learning_rate": 0.0008066582511770712, + "loss": 1.1896, + "step": 8489 + }, + { + "epoch": 0.6530266902545958, + "learning_rate": 0.0008063368527840777, + "loss": 1.534, + "step": 8490 + }, + { + "epoch": 0.6531036074148142, + "learning_rate": 0.0008060154948947982, + "loss": 1.292, + "step": 8491 + }, + { + "epoch": 0.6531805245750327, + "learning_rate": 0.0008056941775279967, + "loss": 1.1956, + "step": 8492 + }, + { + "epoch": 0.6532574417352511, + "learning_rate": 0.0008053729007024363, + "loss": 1.2427, + "step": 8493 + }, + { + "epoch": 0.6533343588954695, + "learning_rate": 0.0008050516644368769, + "loss": 1.2715, + "step": 8494 + }, + { + "epoch": 0.653411276055688, + "learning_rate": 0.0008047304687500737, + "loss": 1.4022, + "step": 8495 + }, + { + "epoch": 0.6534881932159065, + "learning_rate": 0.0008044093136607836, + "loss": 1.4004, + "step": 8496 + }, + { + "epoch": 0.6535651103761249, + "learning_rate": 0.0008040881991877588, + "loss": 1.005, + "step": 8497 + }, + { + "epoch": 0.6536420275363434, + "learning_rate": 0.0008037671253497492, + "loss": 0.9758, + "step": 8498 + }, + { + "epoch": 0.6537189446965618, + "learning_rate": 0.0008034460921655027, + "loss": 1.2197, + "step": 8499 + }, + { + "epoch": 0.6537958618567803, + "learning_rate": 0.0008031250996537642, + "loss": 1.2212, + "step": 8500 + }, + { + "epoch": 0.6538727790169987, + "learning_rate": 0.000802804147833278, + "loss": 1.3095, + "step": 8501 + }, + { + "epoch": 0.6539496961772171, + "learning_rate": 0.0008024832367227844, + "loss": 0.8511, + "step": 8502 + }, + { + "epoch": 0.6540266133374356, + "learning_rate": 0.0008021623663410215, + "loss": 0.9117, + "step": 8503 + }, + { + "epoch": 0.654103530497654, + "learning_rate": 0.0008018415367067251, + "loss": 1.19, + "step": 8504 + }, + { + "epoch": 0.6541804476578724, + "learning_rate": 0.0008015207478386291, + "loss": 1.0384, + "step": 8505 + }, + { + "epoch": 0.6542573648180909, + "learning_rate": 0.000801199999755464, + "loss": 1.12, + "step": 8506 + }, + { + "epoch": 0.6543342819783093, + "learning_rate": 0.0008008792924759602, + "loss": 0.9334, + "step": 8507 + }, + { + "epoch": 0.6544111991385279, + "learning_rate": 0.0008005586260188422, + "loss": 1.1652, + "step": 8508 + }, + { + "epoch": 0.6544881162987463, + "learning_rate": 0.0008002380004028352, + "loss": 0.9915, + "step": 8509 + }, + { + "epoch": 0.6545650334589647, + "learning_rate": 0.000799917415646661, + "loss": 1.2801, + "step": 8510 + }, + { + "epoch": 0.6546419506191832, + "learning_rate": 0.0007995968717690383, + "loss": 1.2619, + "step": 8511 + }, + { + "epoch": 0.6547188677794016, + "learning_rate": 0.0007992763687886843, + "loss": 0.9215, + "step": 8512 + }, + { + "epoch": 0.65479578493962, + "learning_rate": 0.0007989559067243133, + "loss": 1.0984, + "step": 8513 + }, + { + "epoch": 0.6548727020998385, + "learning_rate": 0.0007986354855946371, + "loss": 1.2901, + "step": 8514 + }, + { + "epoch": 0.6549496192600569, + "learning_rate": 0.0007983151054183662, + "loss": 1.2948, + "step": 8515 + }, + { + "epoch": 0.6550265364202754, + "learning_rate": 0.0007979947662142076, + "loss": 1.1017, + "step": 8516 + }, + { + "epoch": 0.6551034535804938, + "learning_rate": 0.0007976744680008665, + "loss": 1.2741, + "step": 8517 + }, + { + "epoch": 0.6551803707407122, + "learning_rate": 0.0007973542107970449, + "loss": 1.1714, + "step": 8518 + }, + { + "epoch": 0.6552572879009307, + "learning_rate": 0.0007970339946214427, + "loss": 1.087, + "step": 8519 + }, + { + "epoch": 0.6553342050611491, + "learning_rate": 0.0007967138194927594, + "loss": 1.3374, + "step": 8520 + }, + { + "epoch": 0.6554111222213675, + "learning_rate": 0.0007963936854296884, + "loss": 0.9367, + "step": 8521 + }, + { + "epoch": 0.6554880393815861, + "learning_rate": 0.0007960735924509229, + "loss": 1.2661, + "step": 8522 + }, + { + "epoch": 0.6555649565418045, + "learning_rate": 0.0007957535405751544, + "loss": 1.1345, + "step": 8523 + }, + { + "epoch": 0.6556418737020229, + "learning_rate": 0.0007954335298210708, + "loss": 1.5253, + "step": 8524 + }, + { + "epoch": 0.6557187908622414, + "learning_rate": 0.0007951135602073577, + "loss": 0.8936, + "step": 8525 + }, + { + "epoch": 0.6557957080224598, + "learning_rate": 0.0007947936317526982, + "loss": 0.944, + "step": 8526 + }, + { + "epoch": 0.6558726251826783, + "learning_rate": 0.000794473744475773, + "loss": 1.3554, + "step": 8527 + }, + { + "epoch": 0.6559495423428967, + "learning_rate": 0.0007941538983952617, + "loss": 1.6405, + "step": 8528 + }, + { + "epoch": 0.6560264595031151, + "learning_rate": 0.0007938340935298407, + "loss": 1.3003, + "step": 8529 + }, + { + "epoch": 0.6561033766633336, + "learning_rate": 0.0007935143298981816, + "loss": 1.0776, + "step": 8530 + }, + { + "epoch": 0.656180293823552, + "learning_rate": 0.0007931946075189576, + "loss": 0.9144, + "step": 8531 + }, + { + "epoch": 0.6562572109837704, + "learning_rate": 0.000792874926410837, + "loss": 1.6682, + "step": 8532 + }, + { + "epoch": 0.656334128143989, + "learning_rate": 0.0007925552865924867, + "loss": 1.0407, + "step": 8533 + }, + { + "epoch": 0.6564110453042074, + "learning_rate": 0.0007922356880825702, + "loss": 1.3451, + "step": 8534 + }, + { + "epoch": 0.6564879624644259, + "learning_rate": 0.000791916130899749, + "loss": 1.165, + "step": 8535 + }, + { + "epoch": 0.6565648796246443, + "learning_rate": 0.0007915966150626836, + "loss": 0.8988, + "step": 8536 + }, + { + "epoch": 0.6566417967848627, + "learning_rate": 0.0007912771405900299, + "loss": 1.065, + "step": 8537 + }, + { + "epoch": 0.6567187139450812, + "learning_rate": 0.0007909577075004427, + "loss": 1.1627, + "step": 8538 + }, + { + "epoch": 0.6567956311052996, + "learning_rate": 0.000790638315812574, + "loss": 1.3178, + "step": 8539 + }, + { + "epoch": 0.656872548265518, + "learning_rate": 0.000790318965545073, + "loss": 0.6202, + "step": 8540 + }, + { + "epoch": 0.6569494654257365, + "learning_rate": 0.0007899996567165869, + "loss": 1.1995, + "step": 8541 + }, + { + "epoch": 0.6570263825859549, + "learning_rate": 0.0007896803893457618, + "loss": 1.2792, + "step": 8542 + }, + { + "epoch": 0.6571032997461733, + "learning_rate": 0.0007893611634512386, + "loss": 1.3113, + "step": 8543 + }, + { + "epoch": 0.6571802169063918, + "learning_rate": 0.000789041979051657, + "loss": 1.0086, + "step": 8544 + }, + { + "epoch": 0.6572571340666102, + "learning_rate": 0.0007887228361656559, + "loss": 1.2499, + "step": 8545 + }, + { + "epoch": 0.6573340512268288, + "learning_rate": 0.0007884037348118696, + "loss": 1.4176, + "step": 8546 + }, + { + "epoch": 0.6574109683870472, + "learning_rate": 0.0007880846750089307, + "loss": 0.8643, + "step": 8547 + }, + { + "epoch": 0.6574878855472656, + "learning_rate": 0.0007877656567754697, + "loss": 1.1159, + "step": 8548 + }, + { + "epoch": 0.6575648027074841, + "learning_rate": 0.0007874466801301136, + "loss": 1.2298, + "step": 8549 + }, + { + "epoch": 0.6576417198677025, + "learning_rate": 0.000787127745091489, + "loss": 1.4595, + "step": 8550 + }, + { + "epoch": 0.6577186370279209, + "learning_rate": 0.0007868088516782184, + "loss": 1.1378, + "step": 8551 + }, + { + "epoch": 0.6577955541881394, + "learning_rate": 0.0007864899999089223, + "loss": 1.0557, + "step": 8552 + }, + { + "epoch": 0.6578724713483578, + "learning_rate": 0.0007861711898022185, + "loss": 1.337, + "step": 8553 + }, + { + "epoch": 0.6579493885085763, + "learning_rate": 0.0007858524213767225, + "loss": 1.1447, + "step": 8554 + }, + { + "epoch": 0.6580263056687947, + "learning_rate": 0.0007855336946510491, + "loss": 1.3924, + "step": 8555 + }, + { + "epoch": 0.6581032228290131, + "learning_rate": 0.0007852150096438069, + "loss": 1.3481, + "step": 8556 + }, + { + "epoch": 0.6581801399892316, + "learning_rate": 0.000784896366373605, + "loss": 0.9431, + "step": 8557 + }, + { + "epoch": 0.65825705714945, + "learning_rate": 0.00078457776485905, + "loss": 1.3165, + "step": 8558 + }, + { + "epoch": 0.6583339743096684, + "learning_rate": 0.0007842592051187448, + "loss": 1.1005, + "step": 8559 + }, + { + "epoch": 0.658410891469887, + "learning_rate": 0.0007839406871712908, + "loss": 0.8452, + "step": 8560 + }, + { + "epoch": 0.6584878086301054, + "learning_rate": 0.000783622211035286, + "loss": 0.9317, + "step": 8561 + }, + { + "epoch": 0.6585647257903238, + "learning_rate": 0.0007833037767293271, + "loss": 1.0621, + "step": 8562 + }, + { + "epoch": 0.6586416429505423, + "learning_rate": 0.000782985384272007, + "loss": 1.3057, + "step": 8563 + }, + { + "epoch": 0.6587185601107607, + "learning_rate": 0.0007826670336819181, + "loss": 0.9818, + "step": 8564 + }, + { + "epoch": 0.6587954772709792, + "learning_rate": 0.0007823487249776488, + "loss": 1.0354, + "step": 8565 + }, + { + "epoch": 0.6588723944311976, + "learning_rate": 0.0007820304581777851, + "loss": 0.7924, + "step": 8566 + }, + { + "epoch": 0.658949311591416, + "learning_rate": 0.0007817122333009114, + "loss": 1.5939, + "step": 8567 + }, + { + "epoch": 0.6590262287516345, + "learning_rate": 0.0007813940503656091, + "loss": 1.0248, + "step": 8568 + }, + { + "epoch": 0.6591031459118529, + "learning_rate": 0.0007810759093904568, + "loss": 1.3644, + "step": 8569 + }, + { + "epoch": 0.6591800630720713, + "learning_rate": 0.0007807578103940317, + "loss": 1.3638, + "step": 8570 + }, + { + "epoch": 0.6592569802322898, + "learning_rate": 0.000780439753394907, + "loss": 1.1535, + "step": 8571 + }, + { + "epoch": 0.6593338973925083, + "learning_rate": 0.0007801217384116553, + "loss": 0.7405, + "step": 8572 + }, + { + "epoch": 0.6594108145527268, + "learning_rate": 0.0007798037654628458, + "loss": 1.2405, + "step": 8573 + }, + { + "epoch": 0.6594877317129452, + "learning_rate": 0.0007794858345670448, + "loss": 1.1254, + "step": 8574 + }, + { + "epoch": 0.6595646488731636, + "learning_rate": 0.0007791679457428171, + "loss": 1.2448, + "step": 8575 + }, + { + "epoch": 0.6596415660333821, + "learning_rate": 0.0007788500990087234, + "loss": 1.2192, + "step": 8576 + }, + { + "epoch": 0.6597184831936005, + "learning_rate": 0.0007785322943833248, + "loss": 1.0965, + "step": 8577 + }, + { + "epoch": 0.6597954003538189, + "learning_rate": 0.0007782145318851779, + "loss": 1.0793, + "step": 8578 + }, + { + "epoch": 0.6598723175140374, + "learning_rate": 0.0007778968115328356, + "loss": 1.3679, + "step": 8579 + }, + { + "epoch": 0.6599492346742558, + "learning_rate": 0.0007775791333448516, + "loss": 1.194, + "step": 8580 + }, + { + "epoch": 0.6600261518344742, + "learning_rate": 0.000777261497339775, + "loss": 1.0274, + "step": 8581 + }, + { + "epoch": 0.6601030689946927, + "learning_rate": 0.0007769439035361527, + "loss": 1.1253, + "step": 8582 + }, + { + "epoch": 0.6601799861549111, + "learning_rate": 0.0007766263519525292, + "loss": 1.2471, + "step": 8583 + }, + { + "epoch": 0.6602569033151297, + "learning_rate": 0.0007763088426074466, + "loss": 1.1612, + "step": 8584 + }, + { + "epoch": 0.660333820475348, + "learning_rate": 0.0007759913755194455, + "loss": 1.2863, + "step": 8585 + }, + { + "epoch": 0.6604107376355665, + "learning_rate": 0.0007756739507070622, + "loss": 0.919, + "step": 8586 + }, + { + "epoch": 0.660487654795785, + "learning_rate": 0.0007753565681888321, + "loss": 1.322, + "step": 8587 + }, + { + "epoch": 0.6605645719560034, + "learning_rate": 0.0007750392279832869, + "loss": 0.8747, + "step": 8588 + }, + { + "epoch": 0.6606414891162218, + "learning_rate": 0.0007747219301089571, + "loss": 1.2222, + "step": 8589 + }, + { + "epoch": 0.6607184062764403, + "learning_rate": 0.000774404674584369, + "loss": 1.0856, + "step": 8590 + }, + { + "epoch": 0.6607953234366587, + "learning_rate": 0.000774087461428049, + "loss": 1.197, + "step": 8591 + }, + { + "epoch": 0.6608722405968772, + "learning_rate": 0.0007737702906585178, + "loss": 1.3886, + "step": 8592 + }, + { + "epoch": 0.6609491577570956, + "learning_rate": 0.0007734531622942965, + "loss": 1.0896, + "step": 8593 + }, + { + "epoch": 0.661026074917314, + "learning_rate": 0.0007731360763539024, + "loss": 1.1717, + "step": 8594 + }, + { + "epoch": 0.6611029920775325, + "learning_rate": 0.0007728190328558504, + "loss": 1.1633, + "step": 8595 + }, + { + "epoch": 0.6611799092377509, + "learning_rate": 0.0007725020318186529, + "loss": 1.4007, + "step": 8596 + }, + { + "epoch": 0.6612568263979693, + "learning_rate": 0.0007721850732608199, + "loss": 0.8127, + "step": 8597 + }, + { + "epoch": 0.6613337435581879, + "learning_rate": 0.0007718681572008583, + "loss": 1.1896, + "step": 8598 + }, + { + "epoch": 0.6614106607184063, + "learning_rate": 0.0007715512836572744, + "loss": 1.0697, + "step": 8599 + }, + { + "epoch": 0.6614875778786247, + "learning_rate": 0.0007712344526485711, + "loss": 1.4127, + "step": 8600 + }, + { + "epoch": 0.6615644950388432, + "learning_rate": 0.0007709176641932462, + "loss": 1.2476, + "step": 8601 + }, + { + "epoch": 0.6616414121990616, + "learning_rate": 0.0007706009183097993, + "loss": 1.0648, + "step": 8602 + }, + { + "epoch": 0.6617183293592801, + "learning_rate": 0.0007702842150167245, + "loss": 1.2622, + "step": 8603 + }, + { + "epoch": 0.6617952465194985, + "learning_rate": 0.000769967554332516, + "loss": 0.8255, + "step": 8604 + }, + { + "epoch": 0.6618721636797169, + "learning_rate": 0.0007696509362756622, + "loss": 1.1466, + "step": 8605 + }, + { + "epoch": 0.6619490808399354, + "learning_rate": 0.0007693343608646508, + "loss": 1.3083, + "step": 8606 + }, + { + "epoch": 0.6620259980001538, + "learning_rate": 0.0007690178281179682, + "loss": 1.3135, + "step": 8607 + }, + { + "epoch": 0.6621029151603722, + "learning_rate": 0.0007687013380540961, + "loss": 1.053, + "step": 8608 + }, + { + "epoch": 0.6621798323205907, + "learning_rate": 0.0007683848906915153, + "loss": 0.8834, + "step": 8609 + }, + { + "epoch": 0.6622567494808091, + "learning_rate": 0.0007680684860487034, + "loss": 1.067, + "step": 8610 + }, + { + "epoch": 0.6623336666410277, + "learning_rate": 0.0007677521241441344, + "loss": 1.2828, + "step": 8611 + }, + { + "epoch": 0.6624105838012461, + "learning_rate": 0.000767435804996283, + "loss": 1.2955, + "step": 8612 + }, + { + "epoch": 0.6624875009614645, + "learning_rate": 0.0007671195286236186, + "loss": 1.196, + "step": 8613 + }, + { + "epoch": 0.662564418121683, + "learning_rate": 0.0007668032950446077, + "loss": 1.2471, + "step": 8614 + }, + { + "epoch": 0.6626413352819014, + "learning_rate": 0.0007664871042777172, + "loss": 1.2154, + "step": 8615 + }, + { + "epoch": 0.6627182524421198, + "learning_rate": 0.0007661709563414091, + "loss": 1.3028, + "step": 8616 + }, + { + "epoch": 0.6627951696023383, + "learning_rate": 0.0007658548512541435, + "loss": 1.4589, + "step": 8617 + }, + { + "epoch": 0.6628720867625567, + "learning_rate": 0.0007655387890343785, + "loss": 1.0832, + "step": 8618 + }, + { + "epoch": 0.6629490039227752, + "learning_rate": 0.000765222769700568, + "loss": 1.2397, + "step": 8619 + }, + { + "epoch": 0.6630259210829936, + "learning_rate": 0.0007649067932711668, + "loss": 1.276, + "step": 8620 + }, + { + "epoch": 0.663102838243212, + "learning_rate": 0.0007645908597646239, + "loss": 1.0063, + "step": 8621 + }, + { + "epoch": 0.6631797554034305, + "learning_rate": 0.0007642749691993871, + "loss": 0.8444, + "step": 8622 + }, + { + "epoch": 0.663256672563649, + "learning_rate": 0.0007639591215939015, + "loss": 1.1295, + "step": 8623 + }, + { + "epoch": 0.6633335897238674, + "learning_rate": 0.00076364331696661, + "loss": 0.9346, + "step": 8624 + }, + { + "epoch": 0.6634105068840859, + "learning_rate": 0.0007633275553359519, + "loss": 1.3005, + "step": 8625 + }, + { + "epoch": 0.6634874240443043, + "learning_rate": 0.000763011836720367, + "loss": 1.2666, + "step": 8626 + }, + { + "epoch": 0.6635643412045227, + "learning_rate": 0.0007626961611382882, + "loss": 1.3495, + "step": 8627 + }, + { + "epoch": 0.6636412583647412, + "learning_rate": 0.0007623805286081485, + "loss": 1.2139, + "step": 8628 + }, + { + "epoch": 0.6637181755249596, + "learning_rate": 0.0007620649391483791, + "loss": 1.0073, + "step": 8629 + }, + { + "epoch": 0.6637950926851781, + "learning_rate": 0.0007617493927774069, + "loss": 1.8874, + "step": 8630 + }, + { + "epoch": 0.6638720098453965, + "learning_rate": 0.0007614338895136569, + "loss": 1.246, + "step": 8631 + }, + { + "epoch": 0.6639489270056149, + "learning_rate": 0.0007611184293755519, + "loss": 1.1016, + "step": 8632 + }, + { + "epoch": 0.6640258441658334, + "learning_rate": 0.0007608030123815114, + "loss": 0.963, + "step": 8633 + }, + { + "epoch": 0.6641027613260518, + "learning_rate": 0.0007604876385499538, + "loss": 0.8887, + "step": 8634 + }, + { + "epoch": 0.6641796784862702, + "learning_rate": 0.0007601723078992943, + "loss": 1.2444, + "step": 8635 + }, + { + "epoch": 0.6642565956464888, + "learning_rate": 0.0007598570204479435, + "loss": 1.2917, + "step": 8636 + }, + { + "epoch": 0.6643335128067072, + "learning_rate": 0.0007595417762143135, + "loss": 1.231, + "step": 8637 + }, + { + "epoch": 0.6644104299669257, + "learning_rate": 0.0007592265752168102, + "loss": 1.2965, + "step": 8638 + }, + { + "epoch": 0.6644873471271441, + "learning_rate": 0.0007589114174738402, + "loss": 0.9876, + "step": 8639 + }, + { + "epoch": 0.6645642642873625, + "learning_rate": 0.0007585963030038046, + "loss": 1.1763, + "step": 8640 + }, + { + "epoch": 0.664641181447581, + "learning_rate": 0.0007582812318251026, + "loss": 1.6656, + "step": 8641 + }, + { + "epoch": 0.6647180986077994, + "learning_rate": 0.0007579662039561332, + "loss": 1.1732, + "step": 8642 + }, + { + "epoch": 0.6647950157680178, + "learning_rate": 0.0007576512194152905, + "loss": 1.2624, + "step": 8643 + }, + { + "epoch": 0.6648719329282363, + "learning_rate": 0.0007573362782209668, + "loss": 0.695, + "step": 8644 + }, + { + "epoch": 0.6649488500884547, + "learning_rate": 0.0007570213803915518, + "loss": 1.1702, + "step": 8645 + }, + { + "epoch": 0.6650257672486731, + "learning_rate": 0.0007567065259454325, + "loss": 1.5539, + "step": 8646 + }, + { + "epoch": 0.6651026844088916, + "learning_rate": 0.0007563917149009935, + "loss": 0.9346, + "step": 8647 + }, + { + "epoch": 0.66517960156911, + "learning_rate": 0.0007560769472766181, + "loss": 1.0809, + "step": 8648 + }, + { + "epoch": 0.6652565187293286, + "learning_rate": 0.0007557622230906839, + "loss": 0.9848, + "step": 8649 + }, + { + "epoch": 0.665333435889547, + "learning_rate": 0.00075544754236157, + "loss": 1.0547, + "step": 8650 + }, + { + "epoch": 0.6654103530497654, + "learning_rate": 0.0007551329051076499, + "loss": 0.9818, + "step": 8651 + }, + { + "epoch": 0.6654872702099839, + "learning_rate": 0.0007548183113472958, + "loss": 1.2576, + "step": 8652 + }, + { + "epoch": 0.6655641873702023, + "learning_rate": 0.0007545037610988772, + "loss": 1.4696, + "step": 8653 + }, + { + "epoch": 0.6656411045304207, + "learning_rate": 0.0007541892543807606, + "loss": 1.1655, + "step": 8654 + }, + { + "epoch": 0.6657180216906392, + "learning_rate": 0.0007538747912113104, + "loss": 1.3117, + "step": 8655 + }, + { + "epoch": 0.6657949388508576, + "learning_rate": 0.0007535603716088892, + "loss": 1.2257, + "step": 8656 + }, + { + "epoch": 0.6658718560110761, + "learning_rate": 0.000753245995591856, + "loss": 0.9942, + "step": 8657 + }, + { + "epoch": 0.6659487731712945, + "learning_rate": 0.000752931663178567, + "loss": 1.17, + "step": 8658 + }, + { + "epoch": 0.6660256903315129, + "learning_rate": 0.000752617374387377, + "loss": 1.0842, + "step": 8659 + }, + { + "epoch": 0.6661026074917314, + "learning_rate": 0.0007523031292366369, + "loss": 1.1858, + "step": 8660 + }, + { + "epoch": 0.6661795246519499, + "learning_rate": 0.0007519889277446973, + "loss": 1.1992, + "step": 8661 + }, + { + "epoch": 0.6662564418121683, + "learning_rate": 0.0007516747699299034, + "loss": 1.3494, + "step": 8662 + }, + { + "epoch": 0.6663333589723868, + "learning_rate": 0.0007513606558105987, + "loss": 1.2976, + "step": 8663 + }, + { + "epoch": 0.6664102761326052, + "learning_rate": 0.0007510465854051264, + "loss": 0.881, + "step": 8664 + }, + { + "epoch": 0.6664871932928236, + "learning_rate": 0.0007507325587318244, + "loss": 1.3611, + "step": 8665 + }, + { + "epoch": 0.6665641104530421, + "learning_rate": 0.0007504185758090291, + "loss": 1.1902, + "step": 8666 + }, + { + "epoch": 0.6666410276132605, + "learning_rate": 0.0007501046366550746, + "loss": 1.2523, + "step": 8667 + }, + { + "epoch": 0.666717944773479, + "learning_rate": 0.0007497907412882909, + "loss": 1.0071, + "step": 8668 + }, + { + "epoch": 0.6667948619336974, + "learning_rate": 0.0007494768897270084, + "loss": 1.1752, + "step": 8669 + }, + { + "epoch": 0.6668717790939158, + "learning_rate": 0.0007491630819895532, + "loss": 1.0591, + "step": 8670 + }, + { + "epoch": 0.6669486962541343, + "learning_rate": 0.0007488493180942469, + "loss": 0.9658, + "step": 8671 + }, + { + "epoch": 0.6670256134143527, + "learning_rate": 0.0007485355980594124, + "loss": 1.0712, + "step": 8672 + }, + { + "epoch": 0.6671025305745711, + "learning_rate": 0.0007482219219033674, + "loss": 1.1305, + "step": 8673 + }, + { + "epoch": 0.6671794477347897, + "learning_rate": 0.0007479082896444278, + "loss": 1.2732, + "step": 8674 + }, + { + "epoch": 0.6672563648950081, + "learning_rate": 0.0007475947013009071, + "loss": 1.1556, + "step": 8675 + }, + { + "epoch": 0.6673332820552266, + "learning_rate": 0.0007472811568911155, + "loss": 1.1323, + "step": 8676 + }, + { + "epoch": 0.667410199215445, + "learning_rate": 0.000746967656433362, + "loss": 1.2265, + "step": 8677 + }, + { + "epoch": 0.6674871163756634, + "learning_rate": 0.0007466541999459517, + "loss": 1.0317, + "step": 8678 + }, + { + "epoch": 0.6675640335358819, + "learning_rate": 0.0007463407874471879, + "loss": 0.9159, + "step": 8679 + }, + { + "epoch": 0.6676409506961003, + "learning_rate": 0.000746027418955371, + "loss": 1.0863, + "step": 8680 + }, + { + "epoch": 0.6677178678563187, + "learning_rate": 0.0007457140944887989, + "loss": 0.951, + "step": 8681 + }, + { + "epoch": 0.6677947850165372, + "learning_rate": 0.0007454008140657663, + "loss": 0.9111, + "step": 8682 + }, + { + "epoch": 0.6678717021767556, + "learning_rate": 0.0007450875777045677, + "loss": 0.9591, + "step": 8683 + }, + { + "epoch": 0.667948619336974, + "learning_rate": 0.0007447743854234918, + "loss": 1.2639, + "step": 8684 + }, + { + "epoch": 0.6680255364971925, + "learning_rate": 0.0007444612372408258, + "loss": 0.9367, + "step": 8685 + }, + { + "epoch": 0.6681024536574109, + "learning_rate": 0.0007441481331748562, + "loss": 1.1169, + "step": 8686 + }, + { + "epoch": 0.6681793708176295, + "learning_rate": 0.000743835073243865, + "loss": 1.1504, + "step": 8687 + }, + { + "epoch": 0.6682562879778479, + "learning_rate": 0.0007435220574661316, + "loss": 1.211, + "step": 8688 + }, + { + "epoch": 0.6683332051380663, + "learning_rate": 0.0007432090858599341, + "loss": 1.1638, + "step": 8689 + }, + { + "epoch": 0.6684101222982848, + "learning_rate": 0.0007428961584435459, + "loss": 1.0729, + "step": 8690 + }, + { + "epoch": 0.6684870394585032, + "learning_rate": 0.0007425832752352407, + "loss": 1.3794, + "step": 8691 + }, + { + "epoch": 0.6685639566187216, + "learning_rate": 0.000742270436253288, + "loss": 1.3861, + "step": 8692 + }, + { + "epoch": 0.6686408737789401, + "learning_rate": 0.0007419576415159529, + "loss": 1.3238, + "step": 8693 + }, + { + "epoch": 0.6687177909391585, + "learning_rate": 0.0007416448910415018, + "loss": 1.3932, + "step": 8694 + }, + { + "epoch": 0.668794708099377, + "learning_rate": 0.0007413321848481954, + "loss": 1.407, + "step": 8695 + }, + { + "epoch": 0.6688716252595954, + "learning_rate": 0.0007410195229542945, + "loss": 1.0167, + "step": 8696 + }, + { + "epoch": 0.6689485424198138, + "learning_rate": 0.0007407069053780539, + "loss": 0.8776, + "step": 8697 + }, + { + "epoch": 0.6690254595800323, + "learning_rate": 0.000740394332137728, + "loss": 1.0069, + "step": 8698 + }, + { + "epoch": 0.6691023767402507, + "learning_rate": 0.0007400818032515692, + "loss": 0.7575, + "step": 8699 + }, + { + "epoch": 0.6691792939004692, + "learning_rate": 0.0007397693187378259, + "loss": 1.0696, + "step": 8700 + }, + { + "epoch": 0.6692562110606877, + "learning_rate": 0.0007394568786147444, + "loss": 1.0674, + "step": 8701 + }, + { + "epoch": 0.6693331282209061, + "learning_rate": 0.0007391444829005684, + "loss": 1.4395, + "step": 8702 + }, + { + "epoch": 0.6694100453811245, + "learning_rate": 0.000738832131613539, + "loss": 1.1392, + "step": 8703 + }, + { + "epoch": 0.669486962541343, + "learning_rate": 0.0007385198247718941, + "loss": 0.9501, + "step": 8704 + }, + { + "epoch": 0.6695638797015614, + "learning_rate": 0.0007382075623938712, + "loss": 1.457, + "step": 8705 + }, + { + "epoch": 0.6696407968617799, + "learning_rate": 0.0007378953444977017, + "loss": 1.1341, + "step": 8706 + }, + { + "epoch": 0.6697177140219983, + "learning_rate": 0.0007375831711016179, + "loss": 1.1324, + "step": 8707 + }, + { + "epoch": 0.6697946311822167, + "learning_rate": 0.0007372710422238471, + "loss": 1.078, + "step": 8708 + }, + { + "epoch": 0.6698715483424352, + "learning_rate": 0.0007369589578826151, + "loss": 1.3785, + "step": 8709 + }, + { + "epoch": 0.6699484655026536, + "learning_rate": 0.0007366469180961449, + "loss": 1.2788, + "step": 8710 + }, + { + "epoch": 0.670025382662872, + "learning_rate": 0.0007363349228826565, + "loss": 1.4992, + "step": 8711 + }, + { + "epoch": 0.6701022998230906, + "learning_rate": 0.0007360229722603675, + "loss": 1.3291, + "step": 8712 + }, + { + "epoch": 0.670179216983309, + "learning_rate": 0.0007357110662474938, + "loss": 1.2314, + "step": 8713 + }, + { + "epoch": 0.6702561341435275, + "learning_rate": 0.0007353992048622475, + "loss": 1.1453, + "step": 8714 + }, + { + "epoch": 0.6703330513037459, + "learning_rate": 0.0007350873881228384, + "loss": 1.0049, + "step": 8715 + }, + { + "epoch": 0.6704099684639643, + "learning_rate": 0.0007347756160474739, + "loss": 1.227, + "step": 8716 + }, + { + "epoch": 0.6704868856241828, + "learning_rate": 0.0007344638886543582, + "loss": 1.1893, + "step": 8717 + }, + { + "epoch": 0.6705638027844012, + "learning_rate": 0.0007341522059616944, + "loss": 1.3653, + "step": 8718 + }, + { + "epoch": 0.6706407199446196, + "learning_rate": 0.000733840567987682, + "loss": 1.4323, + "step": 8719 + }, + { + "epoch": 0.6707176371048381, + "learning_rate": 0.0007335289747505161, + "loss": 1.0225, + "step": 8720 + }, + { + "epoch": 0.6707945542650565, + "learning_rate": 0.0007332174262683926, + "loss": 0.6422, + "step": 8721 + }, + { + "epoch": 0.6708714714252749, + "learning_rate": 0.000732905922559503, + "loss": 0.9006, + "step": 8722 + }, + { + "epoch": 0.6709483885854934, + "learning_rate": 0.0007325944636420358, + "loss": 1.0171, + "step": 8723 + }, + { + "epoch": 0.6710253057457118, + "learning_rate": 0.0007322830495341779, + "loss": 0.9103, + "step": 8724 + }, + { + "epoch": 0.6711022229059304, + "learning_rate": 0.000731971680254112, + "loss": 1.1405, + "step": 8725 + }, + { + "epoch": 0.6711791400661488, + "learning_rate": 0.0007316603558200209, + "loss": 1.1034, + "step": 8726 + }, + { + "epoch": 0.6712560572263672, + "learning_rate": 0.0007313490762500822, + "loss": 1.3074, + "step": 8727 + }, + { + "epoch": 0.6713329743865857, + "learning_rate": 0.000731037841562472, + "loss": 0.9205, + "step": 8728 + }, + { + "epoch": 0.6714098915468041, + "learning_rate": 0.0007307266517753638, + "loss": 1.4002, + "step": 8729 + }, + { + "epoch": 0.6714868087070225, + "learning_rate": 0.0007304155069069279, + "loss": 1.4464, + "step": 8730 + }, + { + "epoch": 0.671563725867241, + "learning_rate": 0.0007301044069753322, + "loss": 1.2686, + "step": 8731 + }, + { + "epoch": 0.6716406430274594, + "learning_rate": 0.0007297933519987436, + "loss": 0.9774, + "step": 8732 + }, + { + "epoch": 0.6717175601876779, + "learning_rate": 0.0007294823419953226, + "loss": 1.4144, + "step": 8733 + }, + { + "epoch": 0.6717944773478963, + "learning_rate": 0.0007291713769832316, + "loss": 0.6688, + "step": 8734 + }, + { + "epoch": 0.6718713945081147, + "learning_rate": 0.000728860456980627, + "loss": 1.3451, + "step": 8735 + }, + { + "epoch": 0.6719483116683332, + "learning_rate": 0.0007285495820056639, + "loss": 1.0131, + "step": 8736 + }, + { + "epoch": 0.6720252288285516, + "learning_rate": 0.0007282387520764948, + "loss": 1.3469, + "step": 8737 + }, + { + "epoch": 0.67210214598877, + "learning_rate": 0.0007279279672112693, + "loss": 0.9187, + "step": 8738 + }, + { + "epoch": 0.6721790631489886, + "learning_rate": 0.0007276172274281337, + "loss": 1.411, + "step": 8739 + }, + { + "epoch": 0.672255980309207, + "learning_rate": 0.0007273065327452339, + "loss": 1.0844, + "step": 8740 + }, + { + "epoch": 0.6723328974694255, + "learning_rate": 0.0007269958831807109, + "loss": 1.2339, + "step": 8741 + }, + { + "epoch": 0.6724098146296439, + "learning_rate": 0.0007266852787527038, + "loss": 1.1215, + "step": 8742 + }, + { + "epoch": 0.6724867317898623, + "learning_rate": 0.0007263747194793493, + "loss": 0.9118, + "step": 8743 + }, + { + "epoch": 0.6725636489500808, + "learning_rate": 0.0007260642053787806, + "loss": 1.1623, + "step": 8744 + }, + { + "epoch": 0.6726405661102992, + "learning_rate": 0.0007257537364691306, + "loss": 1.271, + "step": 8745 + }, + { + "epoch": 0.6727174832705176, + "learning_rate": 0.0007254433127685263, + "loss": 1.3033, + "step": 8746 + }, + { + "epoch": 0.6727944004307361, + "learning_rate": 0.0007251329342950935, + "loss": 1.141, + "step": 8747 + }, + { + "epoch": 0.6728713175909545, + "learning_rate": 0.0007248226010669569, + "loss": 1.0358, + "step": 8748 + }, + { + "epoch": 0.6729482347511729, + "learning_rate": 0.0007245123131022366, + "loss": 1.2178, + "step": 8749 + }, + { + "epoch": 0.6730251519113915, + "learning_rate": 0.0007242020704190504, + "loss": 1.1747, + "step": 8750 + }, + { + "epoch": 0.6731020690716099, + "learning_rate": 0.0007238918730355138, + "loss": 1.2216, + "step": 8751 + }, + { + "epoch": 0.6731789862318284, + "learning_rate": 0.0007235817209697391, + "loss": 1.3866, + "step": 8752 + }, + { + "epoch": 0.6732559033920468, + "learning_rate": 0.0007232716142398374, + "loss": 1.4212, + "step": 8753 + }, + { + "epoch": 0.6733328205522652, + "learning_rate": 0.0007229615528639162, + "loss": 1.1867, + "step": 8754 + }, + { + "epoch": 0.6734097377124837, + "learning_rate": 0.0007226515368600786, + "loss": 1.7131, + "step": 8755 + }, + { + "epoch": 0.6734866548727021, + "learning_rate": 0.0007223415662464286, + "loss": 0.9497, + "step": 8756 + }, + { + "epoch": 0.6735635720329205, + "learning_rate": 0.0007220316410410647, + "loss": 1.1333, + "step": 8757 + }, + { + "epoch": 0.673640489193139, + "learning_rate": 0.0007217217612620842, + "loss": 0.8965, + "step": 8758 + }, + { + "epoch": 0.6737174063533574, + "learning_rate": 0.0007214119269275813, + "loss": 1.562, + "step": 8759 + }, + { + "epoch": 0.6737943235135759, + "learning_rate": 0.0007211021380556467, + "loss": 1.2326, + "step": 8760 + }, + { + "epoch": 0.6738712406737943, + "learning_rate": 0.0007207923946643707, + "loss": 1.2517, + "step": 8761 + }, + { + "epoch": 0.6739481578340127, + "learning_rate": 0.0007204826967718386, + "loss": 1.0405, + "step": 8762 + }, + { + "epoch": 0.6740250749942313, + "learning_rate": 0.0007201730443961345, + "loss": 1.126, + "step": 8763 + }, + { + "epoch": 0.6741019921544497, + "learning_rate": 0.0007198634375553389, + "loss": 1.2274, + "step": 8764 + }, + { + "epoch": 0.6741789093146681, + "learning_rate": 0.0007195538762675302, + "loss": 0.9536, + "step": 8765 + }, + { + "epoch": 0.6742558264748866, + "learning_rate": 0.0007192443605507836, + "loss": 0.9043, + "step": 8766 + }, + { + "epoch": 0.674332743635105, + "learning_rate": 0.0007189348904231734, + "loss": 1.4426, + "step": 8767 + }, + { + "epoch": 0.6744096607953234, + "learning_rate": 0.0007186254659027685, + "loss": 1.2582, + "step": 8768 + }, + { + "epoch": 0.6744865779555419, + "learning_rate": 0.0007183160870076365, + "loss": 0.8202, + "step": 8769 + }, + { + "epoch": 0.6745634951157603, + "learning_rate": 0.0007180067537558433, + "loss": 0.9519, + "step": 8770 + }, + { + "epoch": 0.6746404122759788, + "learning_rate": 0.0007176974661654507, + "loss": 1.1611, + "step": 8771 + }, + { + "epoch": 0.6747173294361972, + "learning_rate": 0.0007173882242545184, + "loss": 1.5816, + "step": 8772 + }, + { + "epoch": 0.6747942465964156, + "learning_rate": 0.0007170790280411033, + "loss": 1.1072, + "step": 8773 + }, + { + "epoch": 0.6748711637566341, + "learning_rate": 0.0007167698775432591, + "loss": 1.2849, + "step": 8774 + }, + { + "epoch": 0.6749480809168525, + "learning_rate": 0.0007164607727790387, + "loss": 1.091, + "step": 8775 + }, + { + "epoch": 0.675024998077071, + "learning_rate": 0.0007161517137664908, + "loss": 1.4821, + "step": 8776 + }, + { + "epoch": 0.6751019152372895, + "learning_rate": 0.0007158427005236601, + "loss": 0.7069, + "step": 8777 + }, + { + "epoch": 0.6751788323975079, + "learning_rate": 0.0007155337330685919, + "loss": 1.1757, + "step": 8778 + }, + { + "epoch": 0.6752557495577264, + "learning_rate": 0.0007152248114193261, + "loss": 0.9925, + "step": 8779 + }, + { + "epoch": 0.6753326667179448, + "learning_rate": 0.0007149159355939025, + "loss": 0.915, + "step": 8780 + }, + { + "epoch": 0.6754095838781632, + "learning_rate": 0.000714607105610355, + "loss": 1.007, + "step": 8781 + }, + { + "epoch": 0.6754865010383817, + "learning_rate": 0.0007142983214867167, + "loss": 1.2613, + "step": 8782 + }, + { + "epoch": 0.6755634181986001, + "learning_rate": 0.0007139895832410188, + "loss": 1.2682, + "step": 8783 + }, + { + "epoch": 0.6756403353588185, + "learning_rate": 0.0007136808908912881, + "loss": 1.1472, + "step": 8784 + }, + { + "epoch": 0.675717252519037, + "learning_rate": 0.00071337224445555, + "loss": 1.0913, + "step": 8785 + }, + { + "epoch": 0.6757941696792554, + "learning_rate": 0.0007130636439518259, + "loss": 1.4439, + "step": 8786 + }, + { + "epoch": 0.6758710868394738, + "learning_rate": 0.0007127550893981361, + "loss": 1.1878, + "step": 8787 + }, + { + "epoch": 0.6759480039996923, + "learning_rate": 0.0007124465808124963, + "loss": 1.1368, + "step": 8788 + }, + { + "epoch": 0.6760249211599108, + "learning_rate": 0.0007121381182129228, + "loss": 0.6754, + "step": 8789 + }, + { + "epoch": 0.6761018383201293, + "learning_rate": 0.0007118297016174244, + "loss": 1.1726, + "step": 8790 + }, + { + "epoch": 0.6761787554803477, + "learning_rate": 0.0007115213310440118, + "loss": 1.3301, + "step": 8791 + }, + { + "epoch": 0.6762556726405661, + "learning_rate": 0.0007112130065106901, + "loss": 1.1458, + "step": 8792 + }, + { + "epoch": 0.6763325898007846, + "learning_rate": 0.0007109047280354633, + "loss": 0.8939, + "step": 8793 + }, + { + "epoch": 0.676409506961003, + "learning_rate": 0.0007105964956363318, + "loss": 0.815, + "step": 8794 + }, + { + "epoch": 0.6764864241212214, + "learning_rate": 0.0007102883093312934, + "loss": 0.9993, + "step": 8795 + }, + { + "epoch": 0.6765633412814399, + "learning_rate": 0.0007099801691383433, + "loss": 1.1996, + "step": 8796 + }, + { + "epoch": 0.6766402584416583, + "learning_rate": 0.0007096720750754749, + "loss": 1.2154, + "step": 8797 + }, + { + "epoch": 0.6767171756018768, + "learning_rate": 0.0007093640271606777, + "loss": 1.2682, + "step": 8798 + }, + { + "epoch": 0.6767940927620952, + "learning_rate": 0.000709056025411939, + "loss": 1.1586, + "step": 8799 + }, + { + "epoch": 0.6768710099223136, + "learning_rate": 0.0007087480698472433, + "loss": 1.3004, + "step": 8800 + }, + { + "epoch": 0.6769479270825322, + "learning_rate": 0.0007084401604845717, + "loss": 1.5114, + "step": 8801 + }, + { + "epoch": 0.6770248442427506, + "learning_rate": 0.0007081322973419054, + "loss": 1.2009, + "step": 8802 + }, + { + "epoch": 0.677101761402969, + "learning_rate": 0.0007078244804372187, + "loss": 1.0807, + "step": 8803 + }, + { + "epoch": 0.6771786785631875, + "learning_rate": 0.0007075167097884857, + "loss": 0.7348, + "step": 8804 + }, + { + "epoch": 0.6772555957234059, + "learning_rate": 0.0007072089854136787, + "loss": 0.7291, + "step": 8805 + }, + { + "epoch": 0.6773325128836243, + "learning_rate": 0.0007069013073307651, + "loss": 1.2954, + "step": 8806 + }, + { + "epoch": 0.6774094300438428, + "learning_rate": 0.0007065936755577107, + "loss": 0.7671, + "step": 8807 + }, + { + "epoch": 0.6774863472040612, + "learning_rate": 0.0007062860901124785, + "loss": 1.0619, + "step": 8808 + }, + { + "epoch": 0.6775632643642797, + "learning_rate": 0.0007059785510130282, + "loss": 1.1243, + "step": 8809 + }, + { + "epoch": 0.6776401815244981, + "learning_rate": 0.0007056710582773184, + "loss": 1.1076, + "step": 8810 + }, + { + "epoch": 0.6777170986847165, + "learning_rate": 0.0007053636119233037, + "loss": 1.1666, + "step": 8811 + }, + { + "epoch": 0.677794015844935, + "learning_rate": 0.0007050562119689347, + "loss": 1.1684, + "step": 8812 + }, + { + "epoch": 0.6778709330051534, + "learning_rate": 0.0007047488584321628, + "loss": 1.2276, + "step": 8813 + }, + { + "epoch": 0.6779478501653718, + "learning_rate": 0.0007044415513309335, + "loss": 1.07, + "step": 8814 + }, + { + "epoch": 0.6780247673255904, + "learning_rate": 0.0007041342906831913, + "loss": 1.4927, + "step": 8815 + }, + { + "epoch": 0.6781016844858088, + "learning_rate": 0.0007038270765068774, + "loss": 1.3837, + "step": 8816 + }, + { + "epoch": 0.6781786016460273, + "learning_rate": 0.0007035199088199296, + "loss": 1.326, + "step": 8817 + }, + { + "epoch": 0.6782555188062457, + "learning_rate": 0.000703212787640285, + "loss": 1.3295, + "step": 8818 + }, + { + "epoch": 0.6783324359664641, + "learning_rate": 0.000702905712985876, + "loss": 1.4315, + "step": 8819 + }, + { + "epoch": 0.6784093531266826, + "learning_rate": 0.0007025986848746333, + "loss": 1.2604, + "step": 8820 + }, + { + "epoch": 0.678486270286901, + "learning_rate": 0.0007022917033244843, + "loss": 1.132, + "step": 8821 + }, + { + "epoch": 0.6785631874471194, + "learning_rate": 0.0007019847683533541, + "loss": 1.0873, + "step": 8822 + }, + { + "epoch": 0.6786401046073379, + "learning_rate": 0.0007016778799791643, + "loss": 1.1496, + "step": 8823 + }, + { + "epoch": 0.6787170217675563, + "learning_rate": 0.0007013710382198365, + "loss": 0.6258, + "step": 8824 + }, + { + "epoch": 0.6787939389277747, + "learning_rate": 0.0007010642430932852, + "loss": 0.93, + "step": 8825 + }, + { + "epoch": 0.6788708560879932, + "learning_rate": 0.0007007574946174249, + "loss": 1.0889, + "step": 8826 + }, + { + "epoch": 0.6789477732482117, + "learning_rate": 0.0007004507928101681, + "loss": 1.2397, + "step": 8827 + }, + { + "epoch": 0.6790246904084302, + "learning_rate": 0.0007001441376894229, + "loss": 1.3494, + "step": 8828 + }, + { + "epoch": 0.6791016075686486, + "learning_rate": 0.0006998375292730949, + "loss": 1.0884, + "step": 8829 + }, + { + "epoch": 0.679178524728867, + "learning_rate": 0.0006995309675790875, + "loss": 1.1061, + "step": 8830 + }, + { + "epoch": 0.6792554418890855, + "learning_rate": 0.0006992244526253005, + "loss": 1.322, + "step": 8831 + }, + { + "epoch": 0.6793323590493039, + "learning_rate": 0.0006989179844296328, + "loss": 1.5245, + "step": 8832 + }, + { + "epoch": 0.6794092762095223, + "learning_rate": 0.0006986115630099797, + "loss": 1.1052, + "step": 8833 + }, + { + "epoch": 0.6794861933697408, + "learning_rate": 0.0006983051883842313, + "loss": 0.8898, + "step": 8834 + }, + { + "epoch": 0.6795631105299592, + "learning_rate": 0.000697998860570279, + "loss": 1.2782, + "step": 8835 + }, + { + "epoch": 0.6796400276901777, + "learning_rate": 0.0006976925795860085, + "loss": 1.2881, + "step": 8836 + }, + { + "epoch": 0.6797169448503961, + "learning_rate": 0.0006973863454493055, + "loss": 0.9359, + "step": 8837 + }, + { + "epoch": 0.6797938620106145, + "learning_rate": 0.0006970801581780499, + "loss": 0.9424, + "step": 8838 + }, + { + "epoch": 0.679870779170833, + "learning_rate": 0.0006967740177901198, + "loss": 0.7891, + "step": 8839 + }, + { + "epoch": 0.6799476963310515, + "learning_rate": 0.0006964679243033926, + "loss": 1.3828, + "step": 8840 + }, + { + "epoch": 0.6800246134912699, + "learning_rate": 0.0006961618777357408, + "loss": 0.9813, + "step": 8841 + }, + { + "epoch": 0.6801015306514884, + "learning_rate": 0.0006958558781050348, + "loss": 1.0776, + "step": 8842 + }, + { + "epoch": 0.6801784478117068, + "learning_rate": 0.0006955499254291419, + "loss": 1.3601, + "step": 8843 + }, + { + "epoch": 0.6802553649719253, + "learning_rate": 0.0006952440197259266, + "loss": 1.0327, + "step": 8844 + }, + { + "epoch": 0.6803322821321437, + "learning_rate": 0.0006949381610132525, + "loss": 1.1075, + "step": 8845 + }, + { + "epoch": 0.6804091992923621, + "learning_rate": 0.0006946323493089785, + "loss": 1.1752, + "step": 8846 + }, + { + "epoch": 0.6804861164525806, + "learning_rate": 0.0006943265846309601, + "loss": 1.2647, + "step": 8847 + }, + { + "epoch": 0.680563033612799, + "learning_rate": 0.0006940208669970524, + "loss": 1.1476, + "step": 8848 + }, + { + "epoch": 0.6806399507730174, + "learning_rate": 0.0006937151964251065, + "loss": 1.2381, + "step": 8849 + }, + { + "epoch": 0.6807168679332359, + "learning_rate": 0.0006934095729329705, + "loss": 1.0425, + "step": 8850 + }, + { + "epoch": 0.6807937850934543, + "learning_rate": 0.0006931039965384899, + "loss": 1.1445, + "step": 8851 + }, + { + "epoch": 0.6808707022536727, + "learning_rate": 0.0006927984672595081, + "loss": 1.3931, + "step": 8852 + }, + { + "epoch": 0.6809476194138913, + "learning_rate": 0.0006924929851138644, + "loss": 1.1622, + "step": 8853 + }, + { + "epoch": 0.6810245365741097, + "learning_rate": 0.0006921875501193974, + "loss": 0.9951, + "step": 8854 + }, + { + "epoch": 0.6811014537343282, + "learning_rate": 0.0006918821622939413, + "loss": 1.3899, + "step": 8855 + }, + { + "epoch": 0.6811783708945466, + "learning_rate": 0.0006915768216553278, + "loss": 1.1107, + "step": 8856 + }, + { + "epoch": 0.681255288054765, + "learning_rate": 0.0006912715282213866, + "loss": 1.1527, + "step": 8857 + }, + { + "epoch": 0.6813322052149835, + "learning_rate": 0.0006909662820099428, + "loss": 1.335, + "step": 8858 + }, + { + "epoch": 0.6814091223752019, + "learning_rate": 0.0006906610830388221, + "loss": 1.2401, + "step": 8859 + }, + { + "epoch": 0.6814860395354203, + "learning_rate": 0.0006903559313258437, + "loss": 0.9982, + "step": 8860 + }, + { + "epoch": 0.6815629566956388, + "learning_rate": 0.0006900508268888257, + "loss": 1.1287, + "step": 8861 + }, + { + "epoch": 0.6816398738558572, + "learning_rate": 0.0006897457697455845, + "loss": 1.123, + "step": 8862 + }, + { + "epoch": 0.6817167910160757, + "learning_rate": 0.0006894407599139324, + "loss": 1.1541, + "step": 8863 + }, + { + "epoch": 0.6817937081762941, + "learning_rate": 0.000689135797411679, + "loss": 0.884, + "step": 8864 + }, + { + "epoch": 0.6818706253365125, + "learning_rate": 0.0006888308822566316, + "loss": 1.2041, + "step": 8865 + }, + { + "epoch": 0.6819475424967311, + "learning_rate": 0.0006885260144665936, + "loss": 1.3779, + "step": 8866 + }, + { + "epoch": 0.6820244596569495, + "learning_rate": 0.0006882211940593678, + "loss": 1.1393, + "step": 8867 + }, + { + "epoch": 0.6821013768171679, + "learning_rate": 0.0006879164210527528, + "loss": 0.8734, + "step": 8868 + }, + { + "epoch": 0.6821782939773864, + "learning_rate": 0.000687611695464544, + "loss": 1.1687, + "step": 8869 + }, + { + "epoch": 0.6822552111376048, + "learning_rate": 0.0006873070173125352, + "loss": 1.2036, + "step": 8870 + }, + { + "epoch": 0.6823321282978232, + "learning_rate": 0.0006870023866145165, + "loss": 1.1798, + "step": 8871 + }, + { + "epoch": 0.6824090454580417, + "learning_rate": 0.0006866978033882756, + "loss": 1.2519, + "step": 8872 + }, + { + "epoch": 0.6824859626182601, + "learning_rate": 0.0006863932676515978, + "loss": 1.0979, + "step": 8873 + }, + { + "epoch": 0.6825628797784786, + "learning_rate": 0.0006860887794222644, + "loss": 1.2568, + "step": 8874 + }, + { + "epoch": 0.682639796938697, + "learning_rate": 0.0006857843387180559, + "loss": 1.4729, + "step": 8875 + }, + { + "epoch": 0.6827167140989154, + "learning_rate": 0.0006854799455567484, + "loss": 1.1827, + "step": 8876 + }, + { + "epoch": 0.682793631259134, + "learning_rate": 0.0006851755999561159, + "loss": 1.2076, + "step": 8877 + }, + { + "epoch": 0.6828705484193524, + "learning_rate": 0.0006848713019339291, + "loss": 0.7952, + "step": 8878 + }, + { + "epoch": 0.6829474655795708, + "learning_rate": 0.0006845670515079567, + "loss": 1.3458, + "step": 8879 + }, + { + "epoch": 0.6830243827397893, + "learning_rate": 0.0006842628486959633, + "loss": 0.9705, + "step": 8880 + }, + { + "epoch": 0.6831012999000077, + "learning_rate": 0.000683958693515713, + "loss": 0.8039, + "step": 8881 + }, + { + "epoch": 0.6831782170602262, + "learning_rate": 0.000683654585984965, + "loss": 1.0404, + "step": 8882 + }, + { + "epoch": 0.6832551342204446, + "learning_rate": 0.0006833505261214765, + "loss": 1.0742, + "step": 8883 + }, + { + "epoch": 0.683332051380663, + "learning_rate": 0.0006830465139430016, + "loss": 1.5145, + "step": 8884 + }, + { + "epoch": 0.6834089685408815, + "learning_rate": 0.0006827425494672925, + "loss": 1.3263, + "step": 8885 + }, + { + "epoch": 0.6834858857010999, + "learning_rate": 0.0006824386327120975, + "loss": 1.1496, + "step": 8886 + }, + { + "epoch": 0.6835628028613183, + "learning_rate": 0.0006821347636951627, + "loss": 0.8693, + "step": 8887 + }, + { + "epoch": 0.6836397200215368, + "learning_rate": 0.000681830942434231, + "loss": 1.0798, + "step": 8888 + }, + { + "epoch": 0.6837166371817552, + "learning_rate": 0.0006815271689470437, + "loss": 1.3729, + "step": 8889 + }, + { + "epoch": 0.6837935543419736, + "learning_rate": 0.0006812234432513381, + "loss": 0.9765, + "step": 8890 + }, + { + "epoch": 0.6838704715021922, + "learning_rate": 0.000680919765364849, + "loss": 1.0981, + "step": 8891 + }, + { + "epoch": 0.6839473886624106, + "learning_rate": 0.0006806161353053084, + "loss": 1.072, + "step": 8892 + }, + { + "epoch": 0.6840243058226291, + "learning_rate": 0.0006803125530904448, + "loss": 1.1629, + "step": 8893 + }, + { + "epoch": 0.6841012229828475, + "learning_rate": 0.0006800090187379863, + "loss": 1.1544, + "step": 8894 + }, + { + "epoch": 0.6841781401430659, + "learning_rate": 0.0006797055322656562, + "loss": 1.0568, + "step": 8895 + }, + { + "epoch": 0.6842550573032844, + "learning_rate": 0.0006794020936911739, + "loss": 1.0024, + "step": 8896 + }, + { + "epoch": 0.6843319744635028, + "learning_rate": 0.0006790987030322591, + "loss": 1.1567, + "step": 8897 + }, + { + "epoch": 0.6844088916237212, + "learning_rate": 0.0006787953603066267, + "loss": 1.0324, + "step": 8898 + }, + { + "epoch": 0.6844858087839397, + "learning_rate": 0.0006784920655319887, + "loss": 1.3817, + "step": 8899 + }, + { + "epoch": 0.6845627259441581, + "learning_rate": 0.0006781888187260555, + "loss": 1.2668, + "step": 8900 + }, + { + "epoch": 0.6846396431043766, + "learning_rate": 0.0006778856199065329, + "loss": 0.9629, + "step": 8901 + }, + { + "epoch": 0.684716560264595, + "learning_rate": 0.0006775824690911264, + "loss": 1.3683, + "step": 8902 + }, + { + "epoch": 0.6847934774248134, + "learning_rate": 0.0006772793662975366, + "loss": 1.2192, + "step": 8903 + }, + { + "epoch": 0.684870394585032, + "learning_rate": 0.0006769763115434621, + "loss": 1.145, + "step": 8904 + }, + { + "epoch": 0.6849473117452504, + "learning_rate": 0.0006766733048465983, + "loss": 1.0979, + "step": 8905 + }, + { + "epoch": 0.6850242289054688, + "learning_rate": 0.0006763703462246384, + "loss": 1.3081, + "step": 8906 + }, + { + "epoch": 0.6851011460656873, + "learning_rate": 0.0006760674356952718, + "loss": 1.1022, + "step": 8907 + }, + { + "epoch": 0.6851780632259057, + "learning_rate": 0.0006757645732761873, + "loss": 0.8378, + "step": 8908 + }, + { + "epoch": 0.6852549803861241, + "learning_rate": 0.0006754617589850681, + "loss": 1.3486, + "step": 8909 + }, + { + "epoch": 0.6853318975463426, + "learning_rate": 0.0006751589928395953, + "loss": 1.3753, + "step": 8910 + }, + { + "epoch": 0.685408814706561, + "learning_rate": 0.0006748562748574491, + "loss": 1.0124, + "step": 8911 + }, + { + "epoch": 0.6854857318667795, + "learning_rate": 0.000674553605056305, + "loss": 1.3546, + "step": 8912 + }, + { + "epoch": 0.6855626490269979, + "learning_rate": 0.0006742509834538362, + "loss": 0.9445, + "step": 8913 + }, + { + "epoch": 0.6856395661872163, + "learning_rate": 0.000673948410067713, + "loss": 0.9991, + "step": 8914 + }, + { + "epoch": 0.6857164833474348, + "learning_rate": 0.0006736458849156026, + "loss": 1.0037, + "step": 8915 + }, + { + "epoch": 0.6857934005076533, + "learning_rate": 0.0006733434080151707, + "loss": 0.9724, + "step": 8916 + }, + { + "epoch": 0.6858703176678717, + "learning_rate": 0.0006730409793840791, + "loss": 1.0462, + "step": 8917 + }, + { + "epoch": 0.6859472348280902, + "learning_rate": 0.0006727385990399857, + "loss": 0.8091, + "step": 8918 + }, + { + "epoch": 0.6860241519883086, + "learning_rate": 0.000672436267000548, + "loss": 1.1152, + "step": 8919 + }, + { + "epoch": 0.6861010691485271, + "learning_rate": 0.0006721339832834186, + "loss": 1.4705, + "step": 8920 + }, + { + "epoch": 0.6861779863087455, + "learning_rate": 0.0006718317479062499, + "loss": 1.2821, + "step": 8921 + }, + { + "epoch": 0.6862549034689639, + "learning_rate": 0.0006715295608866878, + "loss": 1.1767, + "step": 8922 + }, + { + "epoch": 0.6863318206291824, + "learning_rate": 0.0006712274222423777, + "loss": 1.3717, + "step": 8923 + }, + { + "epoch": 0.6864087377894008, + "learning_rate": 0.0006709253319909627, + "loss": 1.3316, + "step": 8924 + }, + { + "epoch": 0.6864856549496192, + "learning_rate": 0.0006706232901500814, + "loss": 1.103, + "step": 8925 + }, + { + "epoch": 0.6865625721098377, + "learning_rate": 0.0006703212967373708, + "loss": 1.4027, + "step": 8926 + }, + { + "epoch": 0.6866394892700561, + "learning_rate": 0.0006700193517704641, + "loss": 1.0511, + "step": 8927 + }, + { + "epoch": 0.6867164064302745, + "learning_rate": 0.0006697174552669926, + "loss": 1.2876, + "step": 8928 + }, + { + "epoch": 0.686793323590493, + "learning_rate": 0.0006694156072445836, + "loss": 0.8889, + "step": 8929 + }, + { + "epoch": 0.6868702407507115, + "learning_rate": 0.0006691138077208641, + "loss": 0.9983, + "step": 8930 + }, + { + "epoch": 0.68694715791093, + "learning_rate": 0.0006688120567134541, + "loss": 1.0804, + "step": 8931 + }, + { + "epoch": 0.6870240750711484, + "learning_rate": 0.0006685103542399749, + "loss": 1.1354, + "step": 8932 + }, + { + "epoch": 0.6871009922313668, + "learning_rate": 0.0006682087003180427, + "loss": 1.2051, + "step": 8933 + }, + { + "epoch": 0.6871779093915853, + "learning_rate": 0.0006679070949652714, + "loss": 1.5191, + "step": 8934 + }, + { + "epoch": 0.6872548265518037, + "learning_rate": 0.0006676055381992718, + "loss": 1.4377, + "step": 8935 + }, + { + "epoch": 0.6873317437120221, + "learning_rate": 0.0006673040300376526, + "loss": 1.211, + "step": 8936 + }, + { + "epoch": 0.6874086608722406, + "learning_rate": 0.0006670025704980181, + "loss": 1.0946, + "step": 8937 + }, + { + "epoch": 0.687485578032459, + "learning_rate": 0.0006667011595979723, + "loss": 0.8467, + "step": 8938 + }, + { + "epoch": 0.6875624951926775, + "learning_rate": 0.0006663997973551144, + "loss": 1.3303, + "step": 8939 + }, + { + "epoch": 0.6876394123528959, + "learning_rate": 0.0006660984837870409, + "loss": 1.2592, + "step": 8940 + }, + { + "epoch": 0.6877163295131143, + "learning_rate": 0.000665797218911346, + "loss": 1.1655, + "step": 8941 + }, + { + "epoch": 0.6877932466733329, + "learning_rate": 0.0006654960027456204, + "loss": 1.1318, + "step": 8942 + }, + { + "epoch": 0.6878701638335513, + "learning_rate": 0.0006651948353074539, + "loss": 1.2717, + "step": 8943 + }, + { + "epoch": 0.6879470809937697, + "learning_rate": 0.0006648937166144305, + "loss": 0.7865, + "step": 8944 + }, + { + "epoch": 0.6880239981539882, + "learning_rate": 0.0006645926466841328, + "loss": 1.4333, + "step": 8945 + }, + { + "epoch": 0.6881009153142066, + "learning_rate": 0.0006642916255341415, + "loss": 1.3917, + "step": 8946 + }, + { + "epoch": 0.6881778324744251, + "learning_rate": 0.0006639906531820334, + "loss": 1.0905, + "step": 8947 + }, + { + "epoch": 0.6882547496346435, + "learning_rate": 0.0006636897296453821, + "loss": 1.5172, + "step": 8948 + }, + { + "epoch": 0.6883316667948619, + "learning_rate": 0.0006633888549417591, + "loss": 1.0846, + "step": 8949 + }, + { + "epoch": 0.6884085839550804, + "learning_rate": 0.0006630880290887323, + "loss": 1.1195, + "step": 8950 + }, + { + "epoch": 0.6884855011152988, + "learning_rate": 0.0006627872521038682, + "loss": 1.0971, + "step": 8951 + }, + { + "epoch": 0.6885624182755172, + "learning_rate": 0.0006624865240047295, + "loss": 1.135, + "step": 8952 + }, + { + "epoch": 0.6886393354357357, + "learning_rate": 0.0006621858448088745, + "loss": 1.0745, + "step": 8953 + }, + { + "epoch": 0.6887162525959541, + "learning_rate": 0.0006618852145338617, + "loss": 1.1851, + "step": 8954 + }, + { + "epoch": 0.6887931697561726, + "learning_rate": 0.0006615846331972445, + "loss": 1.2612, + "step": 8955 + }, + { + "epoch": 0.6888700869163911, + "learning_rate": 0.0006612841008165744, + "loss": 1.1767, + "step": 8956 + }, + { + "epoch": 0.6889470040766095, + "learning_rate": 0.0006609836174093998, + "loss": 1.0124, + "step": 8957 + }, + { + "epoch": 0.689023921236828, + "learning_rate": 0.0006606831829932657, + "loss": 1.03, + "step": 8958 + }, + { + "epoch": 0.6891008383970464, + "learning_rate": 0.0006603827975857157, + "loss": 1.1947, + "step": 8959 + }, + { + "epoch": 0.6891777555572648, + "learning_rate": 0.0006600824612042894, + "loss": 1.1655, + "step": 8960 + }, + { + "epoch": 0.6892546727174833, + "learning_rate": 0.0006597821738665233, + "loss": 1.2997, + "step": 8961 + }, + { + "epoch": 0.6893315898777017, + "learning_rate": 0.0006594819355899517, + "loss": 1.5447, + "step": 8962 + }, + { + "epoch": 0.6894085070379201, + "learning_rate": 0.000659181746392106, + "loss": 1.4082, + "step": 8963 + }, + { + "epoch": 0.6894854241981386, + "learning_rate": 0.0006588816062905137, + "loss": 1.1373, + "step": 8964 + }, + { + "epoch": 0.689562341358357, + "learning_rate": 0.0006585815153027021, + "loss": 1.2028, + "step": 8965 + }, + { + "epoch": 0.6896392585185755, + "learning_rate": 0.0006582814734461917, + "loss": 1.1343, + "step": 8966 + }, + { + "epoch": 0.689716175678794, + "learning_rate": 0.000657981480738504, + "loss": 1.079, + "step": 8967 + }, + { + "epoch": 0.6897930928390124, + "learning_rate": 0.0006576815371971549, + "loss": 1.0593, + "step": 8968 + }, + { + "epoch": 0.6898700099992309, + "learning_rate": 0.0006573816428396589, + "loss": 1.5121, + "step": 8969 + }, + { + "epoch": 0.6899469271594493, + "learning_rate": 0.0006570817976835269, + "loss": 0.6794, + "step": 8970 + }, + { + "epoch": 0.6900238443196677, + "learning_rate": 0.000656782001746267, + "loss": 1.1135, + "step": 8971 + }, + { + "epoch": 0.6901007614798862, + "learning_rate": 0.0006564822550453845, + "loss": 0.9261, + "step": 8972 + }, + { + "epoch": 0.6901776786401046, + "learning_rate": 0.0006561825575983825, + "loss": 1.6339, + "step": 8973 + }, + { + "epoch": 0.690254595800323, + "learning_rate": 0.0006558829094227612, + "loss": 1.1517, + "step": 8974 + }, + { + "epoch": 0.6903315129605415, + "learning_rate": 0.0006555833105360153, + "loss": 1.0657, + "step": 8975 + }, + { + "epoch": 0.6904084301207599, + "learning_rate": 0.0006552837609556404, + "loss": 1.1853, + "step": 8976 + }, + { + "epoch": 0.6904853472809784, + "learning_rate": 0.0006549842606991265, + "loss": 0.9982, + "step": 8977 + }, + { + "epoch": 0.6905622644411968, + "learning_rate": 0.0006546848097839635, + "loss": 0.9886, + "step": 8978 + }, + { + "epoch": 0.6906391816014152, + "learning_rate": 0.0006543854082276348, + "loss": 1.267, + "step": 8979 + }, + { + "epoch": 0.6907160987616338, + "learning_rate": 0.0006540860560476228, + "loss": 1.11, + "step": 8980 + }, + { + "epoch": 0.6907930159218522, + "learning_rate": 0.0006537867532614083, + "loss": 1.1821, + "step": 8981 + }, + { + "epoch": 0.6908699330820706, + "learning_rate": 0.0006534874998864671, + "loss": 1.3343, + "step": 8982 + }, + { + "epoch": 0.6909468502422891, + "learning_rate": 0.0006531882959402729, + "loss": 0.9674, + "step": 8983 + }, + { + "epoch": 0.6910237674025075, + "learning_rate": 0.0006528891414402967, + "loss": 1.2381, + "step": 8984 + }, + { + "epoch": 0.691100684562726, + "learning_rate": 0.0006525900364040059, + "loss": 1.0584, + "step": 8985 + }, + { + "epoch": 0.6911776017229444, + "learning_rate": 0.0006522909808488664, + "loss": 1.3482, + "step": 8986 + }, + { + "epoch": 0.6912545188831628, + "learning_rate": 0.0006519919747923406, + "loss": 1.2494, + "step": 8987 + }, + { + "epoch": 0.6913314360433813, + "learning_rate": 0.0006516930182518861, + "loss": 1.1126, + "step": 8988 + }, + { + "epoch": 0.6914083532035997, + "learning_rate": 0.000651394111244961, + "loss": 1.2531, + "step": 8989 + }, + { + "epoch": 0.6914852703638181, + "learning_rate": 0.0006510952537890179, + "loss": 1.2582, + "step": 8990 + }, + { + "epoch": 0.6915621875240366, + "learning_rate": 0.0006507964459015078, + "loss": 1.3781, + "step": 8991 + }, + { + "epoch": 0.691639104684255, + "learning_rate": 0.0006504976875998783, + "loss": 1.3166, + "step": 8992 + }, + { + "epoch": 0.6917160218444734, + "learning_rate": 0.0006501989789015739, + "loss": 1.2769, + "step": 8993 + }, + { + "epoch": 0.691792939004692, + "learning_rate": 0.0006499003198240364, + "loss": 1.0458, + "step": 8994 + }, + { + "epoch": 0.6918698561649104, + "learning_rate": 0.0006496017103847058, + "loss": 1.1119, + "step": 8995 + }, + { + "epoch": 0.6919467733251289, + "learning_rate": 0.0006493031506010174, + "loss": 1.2924, + "step": 8996 + }, + { + "epoch": 0.6920236904853473, + "learning_rate": 0.0006490046404904048, + "loss": 1.3174, + "step": 8997 + }, + { + "epoch": 0.6921006076455657, + "learning_rate": 0.0006487061800702981, + "loss": 0.9987, + "step": 8998 + }, + { + "epoch": 0.6921775248057842, + "learning_rate": 0.0006484077693581239, + "loss": 0.9899, + "step": 8999 + }, + { + "epoch": 0.6922544419660026, + "learning_rate": 0.0006481094083713089, + "loss": 1.1364, + "step": 9000 + }, + { + "epoch": 0.692331359126221, + "learning_rate": 0.0006478110971272729, + "loss": 1.058, + "step": 9001 + }, + { + "epoch": 0.6924082762864395, + "learning_rate": 0.0006475128356434345, + "loss": 1.2429, + "step": 9002 + }, + { + "epoch": 0.6924851934466579, + "learning_rate": 0.0006472146239372104, + "loss": 1.052, + "step": 9003 + }, + { + "epoch": 0.6925621106068764, + "learning_rate": 0.0006469164620260132, + "loss": 0.8947, + "step": 9004 + }, + { + "epoch": 0.6926390277670949, + "learning_rate": 0.0006466183499272529, + "loss": 1.4427, + "step": 9005 + }, + { + "epoch": 0.6927159449273133, + "learning_rate": 0.0006463202876583367, + "loss": 1.3991, + "step": 9006 + }, + { + "epoch": 0.6927928620875318, + "learning_rate": 0.0006460222752366679, + "loss": 0.8358, + "step": 9007 + }, + { + "epoch": 0.6928697792477502, + "learning_rate": 0.0006457243126796489, + "loss": 1.2745, + "step": 9008 + }, + { + "epoch": 0.6929466964079686, + "learning_rate": 0.0006454264000046778, + "loss": 1.1079, + "step": 9009 + }, + { + "epoch": 0.6930236135681871, + "learning_rate": 0.0006451285372291496, + "loss": 0.9257, + "step": 9010 + }, + { + "epoch": 0.6931005307284055, + "learning_rate": 0.000644830724370457, + "loss": 1.387, + "step": 9011 + }, + { + "epoch": 0.6931774478886239, + "learning_rate": 0.00064453296144599, + "loss": 1.0353, + "step": 9012 + }, + { + "epoch": 0.6932543650488424, + "learning_rate": 0.0006442352484731346, + "loss": 1.0631, + "step": 9013 + }, + { + "epoch": 0.6933312822090608, + "learning_rate": 0.000643937585469275, + "loss": 1.3631, + "step": 9014 + }, + { + "epoch": 0.6934081993692793, + "learning_rate": 0.0006436399724517915, + "loss": 0.8408, + "step": 9015 + }, + { + "epoch": 0.6934851165294977, + "learning_rate": 0.0006433424094380628, + "loss": 1.1477, + "step": 9016 + }, + { + "epoch": 0.6935620336897161, + "learning_rate": 0.0006430448964454638, + "loss": 1.5772, + "step": 9017 + }, + { + "epoch": 0.6936389508499347, + "learning_rate": 0.0006427474334913664, + "loss": 1.3088, + "step": 9018 + }, + { + "epoch": 0.6937158680101531, + "learning_rate": 0.0006424500205931395, + "loss": 1.3001, + "step": 9019 + }, + { + "epoch": 0.6937927851703715, + "learning_rate": 0.00064215265776815, + "loss": 1.0358, + "step": 9020 + }, + { + "epoch": 0.69386970233059, + "learning_rate": 0.00064185534503376, + "loss": 1.4033, + "step": 9021 + }, + { + "epoch": 0.6939466194908084, + "learning_rate": 0.0006415580824073314, + "loss": 1.3163, + "step": 9022 + }, + { + "epoch": 0.6940235366510269, + "learning_rate": 0.000641260869906221, + "loss": 1.162, + "step": 9023 + }, + { + "epoch": 0.6941004538112453, + "learning_rate": 0.0006409637075477834, + "loss": 0.9415, + "step": 9024 + }, + { + "epoch": 0.6941773709714637, + "learning_rate": 0.0006406665953493701, + "loss": 1.0079, + "step": 9025 + }, + { + "epoch": 0.6942542881316822, + "learning_rate": 0.00064036953332833, + "loss": 1.2808, + "step": 9026 + }, + { + "epoch": 0.6943312052919006, + "learning_rate": 0.0006400725215020087, + "loss": 0.6978, + "step": 9027 + }, + { + "epoch": 0.694408122452119, + "learning_rate": 0.0006397755598877488, + "loss": 0.9175, + "step": 9028 + }, + { + "epoch": 0.6944850396123375, + "learning_rate": 0.0006394786485028901, + "loss": 0.9635, + "step": 9029 + }, + { + "epoch": 0.6945619567725559, + "learning_rate": 0.0006391817873647707, + "loss": 1.3319, + "step": 9030 + }, + { + "epoch": 0.6946388739327743, + "learning_rate": 0.0006388849764907236, + "loss": 1.0064, + "step": 9031 + }, + { + "epoch": 0.6947157910929929, + "learning_rate": 0.0006385882158980802, + "loss": 1.2103, + "step": 9032 + }, + { + "epoch": 0.6947927082532113, + "learning_rate": 0.0006382915056041686, + "loss": 1.1953, + "step": 9033 + }, + { + "epoch": 0.6948696254134298, + "learning_rate": 0.0006379948456263137, + "loss": 1.0738, + "step": 9034 + }, + { + "epoch": 0.6949465425736482, + "learning_rate": 0.0006376982359818384, + "loss": 0.757, + "step": 9035 + }, + { + "epoch": 0.6950234597338666, + "learning_rate": 0.0006374016766880625, + "loss": 1.3447, + "step": 9036 + }, + { + "epoch": 0.6951003768940851, + "learning_rate": 0.0006371051677623004, + "loss": 1.0917, + "step": 9037 + }, + { + "epoch": 0.6951772940543035, + "learning_rate": 0.0006368087092218675, + "loss": 1.1666, + "step": 9038 + }, + { + "epoch": 0.6952542112145219, + "learning_rate": 0.0006365123010840737, + "loss": 1.1845, + "step": 9039 + }, + { + "epoch": 0.6953311283747404, + "learning_rate": 0.0006362159433662265, + "loss": 1.0628, + "step": 9040 + }, + { + "epoch": 0.6954080455349588, + "learning_rate": 0.0006359196360856305, + "loss": 0.8804, + "step": 9041 + }, + { + "epoch": 0.6954849626951773, + "learning_rate": 0.000635623379259587, + "loss": 1.5282, + "step": 9042 + }, + { + "epoch": 0.6955618798553957, + "learning_rate": 0.0006353271729053957, + "loss": 1.197, + "step": 9043 + }, + { + "epoch": 0.6956387970156142, + "learning_rate": 0.0006350310170403522, + "loss": 1.1779, + "step": 9044 + }, + { + "epoch": 0.6957157141758327, + "learning_rate": 0.0006347349116817487, + "loss": 1.1069, + "step": 9045 + }, + { + "epoch": 0.6957926313360511, + "learning_rate": 0.0006344388568468757, + "loss": 0.9396, + "step": 9046 + }, + { + "epoch": 0.6958695484962695, + "learning_rate": 0.0006341428525530198, + "loss": 1.3949, + "step": 9047 + }, + { + "epoch": 0.695946465656488, + "learning_rate": 0.0006338468988174646, + "loss": 1.337, + "step": 9048 + }, + { + "epoch": 0.6960233828167064, + "learning_rate": 0.0006335509956574928, + "loss": 0.8997, + "step": 9049 + }, + { + "epoch": 0.6961002999769248, + "learning_rate": 0.0006332551430903808, + "loss": 0.7967, + "step": 9050 + }, + { + "epoch": 0.6961772171371433, + "learning_rate": 0.0006329593411334037, + "loss": 1.2711, + "step": 9051 + }, + { + "epoch": 0.6962541342973617, + "learning_rate": 0.0006326635898038347, + "loss": 1.3538, + "step": 9052 + }, + { + "epoch": 0.6963310514575802, + "learning_rate": 0.0006323678891189429, + "loss": 1.4564, + "step": 9053 + }, + { + "epoch": 0.6964079686177986, + "learning_rate": 0.0006320722390959941, + "loss": 1.0628, + "step": 9054 + }, + { + "epoch": 0.696484885778017, + "learning_rate": 0.0006317766397522517, + "loss": 0.4444, + "step": 9055 + }, + { + "epoch": 0.6965618029382356, + "learning_rate": 0.0006314810911049758, + "loss": 1.34, + "step": 9056 + }, + { + "epoch": 0.696638720098454, + "learning_rate": 0.0006311855931714244, + "loss": 1.4358, + "step": 9057 + }, + { + "epoch": 0.6967156372586724, + "learning_rate": 0.0006308901459688526, + "loss": 1.3789, + "step": 9058 + }, + { + "epoch": 0.6967925544188909, + "learning_rate": 0.0006305947495145096, + "loss": 1.0355, + "step": 9059 + }, + { + "epoch": 0.6968694715791093, + "learning_rate": 0.0006302994038256458, + "loss": 1.0603, + "step": 9060 + }, + { + "epoch": 0.6969463887393278, + "learning_rate": 0.0006300041089195057, + "loss": 1.0929, + "step": 9061 + }, + { + "epoch": 0.6970233058995462, + "learning_rate": 0.0006297088648133336, + "loss": 0.8972, + "step": 9062 + }, + { + "epoch": 0.6971002230597646, + "learning_rate": 0.000629413671524367, + "loss": 1.1272, + "step": 9063 + }, + { + "epoch": 0.6971771402199831, + "learning_rate": 0.0006291185290698433, + "loss": 1.2359, + "step": 9064 + }, + { + "epoch": 0.6972540573802015, + "learning_rate": 0.0006288234374669966, + "loss": 1.0291, + "step": 9065 + }, + { + "epoch": 0.6973309745404199, + "learning_rate": 0.0006285283967330573, + "loss": 1.2536, + "step": 9066 + }, + { + "epoch": 0.6974078917006384, + "learning_rate": 0.0006282334068852531, + "loss": 1.2681, + "step": 9067 + }, + { + "epoch": 0.6974848088608568, + "learning_rate": 0.0006279384679408089, + "loss": 1.262, + "step": 9068 + }, + { + "epoch": 0.6975617260210754, + "learning_rate": 0.0006276435799169458, + "loss": 1.1998, + "step": 9069 + }, + { + "epoch": 0.6976386431812938, + "learning_rate": 0.0006273487428308837, + "loss": 1.0804, + "step": 9070 + }, + { + "epoch": 0.6977155603415122, + "learning_rate": 0.0006270539566998386, + "loss": 1.2336, + "step": 9071 + }, + { + "epoch": 0.6977924775017307, + "learning_rate": 0.0006267592215410217, + "loss": 1.2207, + "step": 9072 + }, + { + "epoch": 0.6978693946619491, + "learning_rate": 0.0006264645373716444, + "loss": 1.2066, + "step": 9073 + }, + { + "epoch": 0.6979463118221675, + "learning_rate": 0.000626169904208913, + "loss": 1.2549, + "step": 9074 + }, + { + "epoch": 0.698023228982386, + "learning_rate": 0.0006258753220700318, + "loss": 1.4719, + "step": 9075 + }, + { + "epoch": 0.6981001461426044, + "learning_rate": 0.0006255807909722015, + "loss": 1.0409, + "step": 9076 + }, + { + "epoch": 0.6981770633028228, + "learning_rate": 0.0006252863109326201, + "loss": 1.3939, + "step": 9077 + }, + { + "epoch": 0.6982539804630413, + "learning_rate": 0.000624991881968482, + "loss": 1.1164, + "step": 9078 + }, + { + "epoch": 0.6983308976232597, + "learning_rate": 0.0006246975040969804, + "loss": 1.4007, + "step": 9079 + }, + { + "epoch": 0.6984078147834782, + "learning_rate": 0.0006244031773353037, + "loss": 1.0795, + "step": 9080 + }, + { + "epoch": 0.6984847319436966, + "learning_rate": 0.0006241089017006382, + "loss": 1.3028, + "step": 9081 + }, + { + "epoch": 0.698561649103915, + "learning_rate": 0.0006238146772101665, + "loss": 1.1655, + "step": 9082 + }, + { + "epoch": 0.6986385662641336, + "learning_rate": 0.0006235205038810685, + "loss": 1.1177, + "step": 9083 + }, + { + "epoch": 0.698715483424352, + "learning_rate": 0.0006232263817305228, + "loss": 1.3024, + "step": 9084 + }, + { + "epoch": 0.6987924005845704, + "learning_rate": 0.000622932310775702, + "loss": 1.2065, + "step": 9085 + }, + { + "epoch": 0.6988693177447889, + "learning_rate": 0.000622638291033777, + "loss": 1.229, + "step": 9086 + }, + { + "epoch": 0.6989462349050073, + "learning_rate": 0.0006223443225219169, + "loss": 1.0655, + "step": 9087 + }, + { + "epoch": 0.6990231520652258, + "learning_rate": 0.0006220504052572867, + "loss": 0.8928, + "step": 9088 + }, + { + "epoch": 0.6991000692254442, + "learning_rate": 0.0006217565392570483, + "loss": 0.8444, + "step": 9089 + }, + { + "epoch": 0.6991769863856626, + "learning_rate": 0.0006214627245383609, + "loss": 1.1482, + "step": 9090 + }, + { + "epoch": 0.6992539035458811, + "learning_rate": 0.0006211689611183799, + "loss": 1.3874, + "step": 9091 + }, + { + "epoch": 0.6993308207060995, + "learning_rate": 0.0006208752490142597, + "loss": 1.3886, + "step": 9092 + }, + { + "epoch": 0.6994077378663179, + "learning_rate": 0.0006205815882431506, + "loss": 1.3982, + "step": 9093 + }, + { + "epoch": 0.6994846550265365, + "learning_rate": 0.000620287978822198, + "loss": 0.9534, + "step": 9094 + }, + { + "epoch": 0.6995615721867549, + "learning_rate": 0.0006199944207685475, + "loss": 1.1033, + "step": 9095 + }, + { + "epoch": 0.6996384893469733, + "learning_rate": 0.0006197009140993401, + "loss": 1.2295, + "step": 9096 + }, + { + "epoch": 0.6997154065071918, + "learning_rate": 0.0006194074588317137, + "loss": 0.8809, + "step": 9097 + }, + { + "epoch": 0.6997923236674102, + "learning_rate": 0.0006191140549828034, + "loss": 0.9768, + "step": 9098 + }, + { + "epoch": 0.6998692408276287, + "learning_rate": 0.000618820702569741, + "loss": 1.5388, + "step": 9099 + }, + { + "epoch": 0.6999461579878471, + "learning_rate": 0.0006185274016096567, + "loss": 1.1944, + "step": 9100 + }, + { + "epoch": 0.7000230751480655, + "learning_rate": 0.0006182341521196759, + "loss": 1.3062, + "step": 9101 + }, + { + "epoch": 0.700099992308284, + "learning_rate": 0.000617940954116922, + "loss": 1.0186, + "step": 9102 + }, + { + "epoch": 0.7001769094685024, + "learning_rate": 0.0006176478076185151, + "loss": 1.1119, + "step": 9103 + }, + { + "epoch": 0.7002538266287208, + "learning_rate": 0.0006173547126415722, + "loss": 0.9446, + "step": 9104 + }, + { + "epoch": 0.7003307437889393, + "learning_rate": 0.0006170616692032071, + "loss": 1.0974, + "step": 9105 + }, + { + "epoch": 0.7004076609491577, + "learning_rate": 0.0006167686773205322, + "loss": 0.9449, + "step": 9106 + }, + { + "epoch": 0.7004845781093763, + "learning_rate": 0.0006164757370106537, + "loss": 1.1489, + "step": 9107 + }, + { + "epoch": 0.7005614952695947, + "learning_rate": 0.0006161828482906783, + "loss": 1.2107, + "step": 9108 + }, + { + "epoch": 0.7006384124298131, + "learning_rate": 0.0006158900111777075, + "loss": 1.2626, + "step": 9109 + }, + { + "epoch": 0.7007153295900316, + "learning_rate": 0.0006155972256888402, + "loss": 1.3257, + "step": 9110 + }, + { + "epoch": 0.70079224675025, + "learning_rate": 0.0006153044918411727, + "loss": 1.0702, + "step": 9111 + }, + { + "epoch": 0.7008691639104684, + "learning_rate": 0.000615011809651798, + "loss": 0.8849, + "step": 9112 + }, + { + "epoch": 0.7009460810706869, + "learning_rate": 0.0006147191791378053, + "loss": 0.9322, + "step": 9113 + }, + { + "epoch": 0.7010229982309053, + "learning_rate": 0.0006144266003162831, + "loss": 1.2252, + "step": 9114 + }, + { + "epoch": 0.7010999153911237, + "learning_rate": 0.0006141340732043152, + "loss": 1.1744, + "step": 9115 + }, + { + "epoch": 0.7011768325513422, + "learning_rate": 0.0006138415978189808, + "loss": 1.2032, + "step": 9116 + }, + { + "epoch": 0.7012537497115606, + "learning_rate": 0.0006135491741773594, + "loss": 0.8612, + "step": 9117 + }, + { + "epoch": 0.7013306668717791, + "learning_rate": 0.0006132568022965253, + "loss": 1.3158, + "step": 9118 + }, + { + "epoch": 0.7014075840319975, + "learning_rate": 0.0006129644821935517, + "loss": 1.1177, + "step": 9119 + }, + { + "epoch": 0.701484501192216, + "learning_rate": 0.0006126722138855057, + "loss": 1.1121, + "step": 9120 + }, + { + "epoch": 0.7015614183524345, + "learning_rate": 0.0006123799973894534, + "loss": 1.1532, + "step": 9121 + }, + { + "epoch": 0.7016383355126529, + "learning_rate": 0.0006120878327224587, + "loss": 1.048, + "step": 9122 + }, + { + "epoch": 0.7017152526728713, + "learning_rate": 0.0006117957199015809, + "loss": 1.0144, + "step": 9123 + }, + { + "epoch": 0.7017921698330898, + "learning_rate": 0.0006115036589438767, + "loss": 1.1654, + "step": 9124 + }, + { + "epoch": 0.7018690869933082, + "learning_rate": 0.0006112116498663996, + "loss": 1.3741, + "step": 9125 + }, + { + "epoch": 0.7019460041535267, + "learning_rate": 0.0006109196926862001, + "loss": 0.7862, + "step": 9126 + }, + { + "epoch": 0.7020229213137451, + "learning_rate": 0.000610627787420327, + "loss": 1.3426, + "step": 9127 + }, + { + "epoch": 0.7020998384739635, + "learning_rate": 0.0006103359340858247, + "loss": 1.1919, + "step": 9128 + }, + { + "epoch": 0.702176755634182, + "learning_rate": 0.0006100441326997332, + "loss": 1.0977, + "step": 9129 + }, + { + "epoch": 0.7022536727944004, + "learning_rate": 0.0006097523832790929, + "loss": 1.0682, + "step": 9130 + }, + { + "epoch": 0.7023305899546188, + "learning_rate": 0.0006094606858409388, + "loss": 1.3242, + "step": 9131 + }, + { + "epoch": 0.7024075071148373, + "learning_rate": 0.0006091690404023034, + "loss": 1.0938, + "step": 9132 + }, + { + "epoch": 0.7024844242750558, + "learning_rate": 0.000608877446980216, + "loss": 0.9688, + "step": 9133 + }, + { + "epoch": 0.7025613414352742, + "learning_rate": 0.0006085859055917031, + "loss": 1.1817, + "step": 9134 + }, + { + "epoch": 0.7026382585954927, + "learning_rate": 0.000608294416253788, + "loss": 0.8119, + "step": 9135 + }, + { + "epoch": 0.7027151757557111, + "learning_rate": 0.0006080029789834913, + "loss": 1.3178, + "step": 9136 + }, + { + "epoch": 0.7027920929159296, + "learning_rate": 0.0006077115937978306, + "loss": 0.8258, + "step": 9137 + }, + { + "epoch": 0.702869010076148, + "learning_rate": 0.0006074202607138197, + "loss": 0.9626, + "step": 9138 + }, + { + "epoch": 0.7029459272363664, + "learning_rate": 0.0006071289797484703, + "loss": 1.0446, + "step": 9139 + }, + { + "epoch": 0.7030228443965849, + "learning_rate": 0.0006068377509187894, + "loss": 1.1184, + "step": 9140 + }, + { + "epoch": 0.7030997615568033, + "learning_rate": 0.0006065465742417842, + "loss": 1.6268, + "step": 9141 + }, + { + "epoch": 0.7031766787170217, + "learning_rate": 0.0006062554497344553, + "loss": 0.9224, + "step": 9142 + }, + { + "epoch": 0.7032535958772402, + "learning_rate": 0.0006059643774138016, + "loss": 0.7972, + "step": 9143 + }, + { + "epoch": 0.7033305130374586, + "learning_rate": 0.00060567335729682, + "loss": 0.965, + "step": 9144 + }, + { + "epoch": 0.7034074301976772, + "learning_rate": 0.0006053823894005031, + "loss": 1.191, + "step": 9145 + }, + { + "epoch": 0.7034843473578956, + "learning_rate": 0.000605091473741841, + "loss": 1.2312, + "step": 9146 + }, + { + "epoch": 0.703561264518114, + "learning_rate": 0.0006048006103378203, + "loss": 0.6918, + "step": 9147 + }, + { + "epoch": 0.7036381816783325, + "learning_rate": 0.0006045097992054245, + "loss": 0.7676, + "step": 9148 + }, + { + "epoch": 0.7037150988385509, + "learning_rate": 0.0006042190403616351, + "loss": 1.4219, + "step": 9149 + }, + { + "epoch": 0.7037920159987693, + "learning_rate": 0.0006039283338234303, + "loss": 0.998, + "step": 9150 + }, + { + "epoch": 0.7038689331589878, + "learning_rate": 0.0006036376796077826, + "loss": 1.0559, + "step": 9151 + }, + { + "epoch": 0.7039458503192062, + "learning_rate": 0.0006033470777316657, + "loss": 1.4876, + "step": 9152 + }, + { + "epoch": 0.7040227674794246, + "learning_rate": 0.0006030565282120473, + "loss": 1.4646, + "step": 9153 + }, + { + "epoch": 0.7040996846396431, + "learning_rate": 0.000602766031065893, + "loss": 1.0207, + "step": 9154 + }, + { + "epoch": 0.7041766017998615, + "learning_rate": 0.0006024755863101652, + "loss": 1.0601, + "step": 9155 + }, + { + "epoch": 0.70425351896008, + "learning_rate": 0.0006021851939618228, + "loss": 1.3219, + "step": 9156 + }, + { + "epoch": 0.7043304361202984, + "learning_rate": 0.0006018948540378232, + "loss": 1.1121, + "step": 9157 + }, + { + "epoch": 0.7044073532805168, + "learning_rate": 0.000601604566555119, + "loss": 1.3604, + "step": 9158 + }, + { + "epoch": 0.7044842704407354, + "learning_rate": 0.0006013143315306603, + "loss": 1.2011, + "step": 9159 + }, + { + "epoch": 0.7045611876009538, + "learning_rate": 0.0006010241489813945, + "loss": 1.1483, + "step": 9160 + }, + { + "epoch": 0.7046381047611722, + "learning_rate": 0.0006007340189242656, + "loss": 1.1276, + "step": 9161 + }, + { + "epoch": 0.7047150219213907, + "learning_rate": 0.0006004439413762138, + "loss": 1.2896, + "step": 9162 + }, + { + "epoch": 0.7047919390816091, + "learning_rate": 0.000600153916354179, + "loss": 0.8891, + "step": 9163 + }, + { + "epoch": 0.7048688562418276, + "learning_rate": 0.0005998639438750936, + "loss": 0.9449, + "step": 9164 + }, + { + "epoch": 0.704945773402046, + "learning_rate": 0.0005995740239558914, + "loss": 1.0639, + "step": 9165 + }, + { + "epoch": 0.7050226905622644, + "learning_rate": 0.0005992841566135002, + "loss": 1.0862, + "step": 9166 + }, + { + "epoch": 0.7050996077224829, + "learning_rate": 0.0005989943418648457, + "loss": 1.1619, + "step": 9167 + }, + { + "epoch": 0.7051765248827013, + "learning_rate": 0.0005987045797268508, + "loss": 1.0624, + "step": 9168 + }, + { + "epoch": 0.7052534420429197, + "learning_rate": 0.0005984148702164346, + "loss": 1.1822, + "step": 9169 + }, + { + "epoch": 0.7053303592031382, + "learning_rate": 0.0005981252133505134, + "loss": 1.2507, + "step": 9170 + }, + { + "epoch": 0.7054072763633567, + "learning_rate": 0.0005978356091460011, + "loss": 1.0611, + "step": 9171 + }, + { + "epoch": 0.7054841935235752, + "learning_rate": 0.0005975460576198081, + "loss": 1.2883, + "step": 9172 + }, + { + "epoch": 0.7055611106837936, + "learning_rate": 0.0005972565587888413, + "loss": 1.1024, + "step": 9173 + }, + { + "epoch": 0.705638027844012, + "learning_rate": 0.0005969671126700047, + "loss": 0.968, + "step": 9174 + }, + { + "epoch": 0.7057149450042305, + "learning_rate": 0.0005966777192801991, + "loss": 1.2252, + "step": 9175 + }, + { + "epoch": 0.7057918621644489, + "learning_rate": 0.0005963883786363239, + "loss": 1.1141, + "step": 9176 + }, + { + "epoch": 0.7058687793246673, + "learning_rate": 0.0005960990907552724, + "loss": 1.105, + "step": 9177 + }, + { + "epoch": 0.7059456964848858, + "learning_rate": 0.0005958098556539365, + "loss": 1.1531, + "step": 9178 + }, + { + "epoch": 0.7060226136451042, + "learning_rate": 0.000595520673349206, + "loss": 1.1202, + "step": 9179 + }, + { + "epoch": 0.7060995308053226, + "learning_rate": 0.000595231543857966, + "loss": 1.0094, + "step": 9180 + }, + { + "epoch": 0.7061764479655411, + "learning_rate": 0.0005949424671970991, + "loss": 1.0983, + "step": 9181 + }, + { + "epoch": 0.7062533651257595, + "learning_rate": 0.000594653443383485, + "loss": 1.1388, + "step": 9182 + }, + { + "epoch": 0.706330282285978, + "learning_rate": 0.000594364472433999, + "loss": 1.2372, + "step": 9183 + }, + { + "epoch": 0.7064071994461965, + "learning_rate": 0.0005940755543655161, + "loss": 1.3763, + "step": 9184 + }, + { + "epoch": 0.7064841166064149, + "learning_rate": 0.0005937866891949054, + "loss": 1.2527, + "step": 9185 + }, + { + "epoch": 0.7065610337666334, + "learning_rate": 0.0005934978769390345, + "loss": 0.9266, + "step": 9186 + }, + { + "epoch": 0.7066379509268518, + "learning_rate": 0.0005932091176147674, + "loss": 1.3706, + "step": 9187 + }, + { + "epoch": 0.7067148680870702, + "learning_rate": 0.0005929204112389648, + "loss": 1.2825, + "step": 9188 + }, + { + "epoch": 0.7067917852472887, + "learning_rate": 0.0005926317578284844, + "loss": 1.0201, + "step": 9189 + }, + { + "epoch": 0.7068687024075071, + "learning_rate": 0.0005923431574001821, + "loss": 0.7452, + "step": 9190 + }, + { + "epoch": 0.7069456195677256, + "learning_rate": 0.0005920546099709077, + "loss": 1.5096, + "step": 9191 + }, + { + "epoch": 0.707022536727944, + "learning_rate": 0.0005917661155575116, + "loss": 1.0856, + "step": 9192 + }, + { + "epoch": 0.7070994538881624, + "learning_rate": 0.0005914776741768385, + "loss": 0.4657, + "step": 9193 + }, + { + "epoch": 0.7071763710483809, + "learning_rate": 0.0005911892858457309, + "loss": 1.2545, + "step": 9194 + }, + { + "epoch": 0.7072532882085993, + "learning_rate": 0.000590900950581028, + "loss": 1.0548, + "step": 9195 + }, + { + "epoch": 0.7073302053688177, + "learning_rate": 0.000590612668399566, + "loss": 1.1633, + "step": 9196 + }, + { + "epoch": 0.7074071225290363, + "learning_rate": 0.0005903244393181774, + "loss": 1.0331, + "step": 9197 + }, + { + "epoch": 0.7074840396892547, + "learning_rate": 0.0005900362633536934, + "loss": 1.3973, + "step": 9198 + }, + { + "epoch": 0.7075609568494731, + "learning_rate": 0.000589748140522941, + "loss": 1.3848, + "step": 9199 + }, + { + "epoch": 0.7076378740096916, + "learning_rate": 0.0005894600708427421, + "loss": 1.1831, + "step": 9200 + }, + { + "epoch": 0.70771479116991, + "learning_rate": 0.0005891720543299193, + "loss": 1.0076, + "step": 9201 + }, + { + "epoch": 0.7077917083301285, + "learning_rate": 0.0005888840910012887, + "loss": 1.3635, + "step": 9202 + }, + { + "epoch": 0.7078686254903469, + "learning_rate": 0.0005885961808736668, + "loss": 1.1163, + "step": 9203 + }, + { + "epoch": 0.7079455426505653, + "learning_rate": 0.0005883083239638632, + "loss": 1.1741, + "step": 9204 + }, + { + "epoch": 0.7080224598107838, + "learning_rate": 0.0005880205202886863, + "loss": 1.1505, + "step": 9205 + }, + { + "epoch": 0.7080993769710022, + "learning_rate": 0.000587732769864942, + "loss": 1.0987, + "step": 9206 + }, + { + "epoch": 0.7081762941312206, + "learning_rate": 0.000587445072709432, + "loss": 0.9515, + "step": 9207 + }, + { + "epoch": 0.7082532112914391, + "learning_rate": 0.0005871574288389555, + "loss": 1.1378, + "step": 9208 + }, + { + "epoch": 0.7083301284516575, + "learning_rate": 0.0005868698382703078, + "loss": 1.099, + "step": 9209 + }, + { + "epoch": 0.7084070456118761, + "learning_rate": 0.0005865823010202815, + "loss": 1.2283, + "step": 9210 + }, + { + "epoch": 0.7084839627720945, + "learning_rate": 0.000586294817105667, + "loss": 1.3589, + "step": 9211 + }, + { + "epoch": 0.7085608799323129, + "learning_rate": 0.000586007386543251, + "loss": 1.1269, + "step": 9212 + }, + { + "epoch": 0.7086377970925314, + "learning_rate": 0.0005857200093498152, + "loss": 1.2875, + "step": 9213 + }, + { + "epoch": 0.7087147142527498, + "learning_rate": 0.0005854326855421414, + "loss": 1.111, + "step": 9214 + }, + { + "epoch": 0.7087916314129682, + "learning_rate": 0.0005851454151370061, + "loss": 1.3645, + "step": 9215 + }, + { + "epoch": 0.7088685485731867, + "learning_rate": 0.0005848581981511837, + "loss": 1.1044, + "step": 9216 + }, + { + "epoch": 0.7089454657334051, + "learning_rate": 0.0005845710346014445, + "loss": 1.2185, + "step": 9217 + }, + { + "epoch": 0.7090223828936235, + "learning_rate": 0.000584283924504557, + "loss": 1.1358, + "step": 9218 + }, + { + "epoch": 0.709099300053842, + "learning_rate": 0.0005839968678772846, + "loss": 1.1562, + "step": 9219 + }, + { + "epoch": 0.7091762172140604, + "learning_rate": 0.0005837098647363904, + "loss": 0.9352, + "step": 9220 + }, + { + "epoch": 0.709253134374279, + "learning_rate": 0.0005834229150986323, + "loss": 0.9012, + "step": 9221 + }, + { + "epoch": 0.7093300515344974, + "learning_rate": 0.0005831360189807653, + "loss": 0.9845, + "step": 9222 + }, + { + "epoch": 0.7094069686947158, + "learning_rate": 0.0005828491763995417, + "loss": 0.8973, + "step": 9223 + }, + { + "epoch": 0.7094838858549343, + "learning_rate": 0.0005825623873717098, + "loss": 1.1538, + "step": 9224 + }, + { + "epoch": 0.7095608030151527, + "learning_rate": 0.0005822756519140176, + "loss": 1.1007, + "step": 9225 + }, + { + "epoch": 0.7096377201753711, + "learning_rate": 0.0005819889700432058, + "loss": 1.1432, + "step": 9226 + }, + { + "epoch": 0.7097146373355896, + "learning_rate": 0.0005817023417760142, + "loss": 1.287, + "step": 9227 + }, + { + "epoch": 0.709791554495808, + "learning_rate": 0.0005814157671291805, + "loss": 1.0971, + "step": 9228 + }, + { + "epoch": 0.7098684716560265, + "learning_rate": 0.0005811292461194375, + "loss": 1.3493, + "step": 9229 + }, + { + "epoch": 0.7099453888162449, + "learning_rate": 0.0005808427787635154, + "loss": 0.7992, + "step": 9230 + }, + { + "epoch": 0.7100223059764633, + "learning_rate": 0.0005805563650781414, + "loss": 1.1275, + "step": 9231 + }, + { + "epoch": 0.7100992231366818, + "learning_rate": 0.000580270005080039, + "loss": 1.2756, + "step": 9232 + }, + { + "epoch": 0.7101761402969002, + "learning_rate": 0.00057998369878593, + "loss": 0.9373, + "step": 9233 + }, + { + "epoch": 0.7102530574571186, + "learning_rate": 0.0005796974462125321, + "loss": 1.2198, + "step": 9234 + }, + { + "epoch": 0.7103299746173372, + "learning_rate": 0.0005794112473765584, + "loss": 0.8973, + "step": 9235 + }, + { + "epoch": 0.7104068917775556, + "learning_rate": 0.0005791251022947218, + "loss": 0.9384, + "step": 9236 + }, + { + "epoch": 0.710483808937774, + "learning_rate": 0.0005788390109837305, + "loss": 1.0546, + "step": 9237 + }, + { + "epoch": 0.7105607260979925, + "learning_rate": 0.000578552973460289, + "loss": 1.2324, + "step": 9238 + }, + { + "epoch": 0.7106376432582109, + "learning_rate": 0.0005782669897411001, + "loss": 1.47, + "step": 9239 + }, + { + "epoch": 0.7107145604184294, + "learning_rate": 0.0005779810598428614, + "loss": 1.1954, + "step": 9240 + }, + { + "epoch": 0.7107914775786478, + "learning_rate": 0.0005776951837822702, + "loss": 0.8435, + "step": 9241 + }, + { + "epoch": 0.7108683947388662, + "learning_rate": 0.0005774093615760184, + "loss": 0.9694, + "step": 9242 + }, + { + "epoch": 0.7109453118990847, + "learning_rate": 0.0005771235932407955, + "loss": 1.1737, + "step": 9243 + }, + { + "epoch": 0.7110222290593031, + "learning_rate": 0.0005768378787932878, + "loss": 1.1425, + "step": 9244 + }, + { + "epoch": 0.7110991462195215, + "learning_rate": 0.0005765522182501784, + "loss": 1.056, + "step": 9245 + }, + { + "epoch": 0.71117606337974, + "learning_rate": 0.0005762666116281469, + "loss": 0.7269, + "step": 9246 + }, + { + "epoch": 0.7112529805399584, + "learning_rate": 0.0005759810589438715, + "loss": 1.2509, + "step": 9247 + }, + { + "epoch": 0.711329897700177, + "learning_rate": 0.0005756955602140242, + "loss": 1.0475, + "step": 9248 + }, + { + "epoch": 0.7114068148603954, + "learning_rate": 0.0005754101154552768, + "loss": 1.4002, + "step": 9249 + }, + { + "epoch": 0.7114837320206138, + "learning_rate": 0.0005751247246842965, + "loss": 1.3878, + "step": 9250 + }, + { + "epoch": 0.7115606491808323, + "learning_rate": 0.0005748393879177474, + "loss": 0.8607, + "step": 9251 + }, + { + "epoch": 0.7116375663410507, + "learning_rate": 0.0005745541051722906, + "loss": 0.9819, + "step": 9252 + }, + { + "epoch": 0.7117144835012691, + "learning_rate": 0.000574268876464584, + "loss": 0.9967, + "step": 9253 + }, + { + "epoch": 0.7117914006614876, + "learning_rate": 0.0005739837018112821, + "loss": 1.0851, + "step": 9254 + }, + { + "epoch": 0.711868317821706, + "learning_rate": 0.0005736985812290372, + "loss": 1.2694, + "step": 9255 + }, + { + "epoch": 0.7119452349819244, + "learning_rate": 0.0005734135147344983, + "loss": 1.0488, + "step": 9256 + }, + { + "epoch": 0.7120221521421429, + "learning_rate": 0.0005731285023443087, + "loss": 1.1041, + "step": 9257 + }, + { + "epoch": 0.7120990693023613, + "learning_rate": 0.0005728435440751125, + "loss": 1.4289, + "step": 9258 + }, + { + "epoch": 0.7121759864625798, + "learning_rate": 0.0005725586399435473, + "loss": 1.0462, + "step": 9259 + }, + { + "epoch": 0.7122529036227983, + "learning_rate": 0.0005722737899662509, + "loss": 0.9973, + "step": 9260 + }, + { + "epoch": 0.7123298207830167, + "learning_rate": 0.0005719889941598542, + "loss": 1.0765, + "step": 9261 + }, + { + "epoch": 0.7124067379432352, + "learning_rate": 0.0005717042525409866, + "loss": 1.0844, + "step": 9262 + }, + { + "epoch": 0.7124836551034536, + "learning_rate": 0.0005714195651262758, + "loss": 1.1386, + "step": 9263 + }, + { + "epoch": 0.712560572263672, + "learning_rate": 0.0005711349319323445, + "loss": 0.929, + "step": 9264 + }, + { + "epoch": 0.7126374894238905, + "learning_rate": 0.0005708503529758123, + "loss": 1.2356, + "step": 9265 + }, + { + "epoch": 0.7127144065841089, + "learning_rate": 0.0005705658282732963, + "loss": 0.8556, + "step": 9266 + }, + { + "epoch": 0.7127913237443274, + "learning_rate": 0.0005702813578414098, + "loss": 0.928, + "step": 9267 + }, + { + "epoch": 0.7128682409045458, + "learning_rate": 0.0005699969416967643, + "loss": 0.98, + "step": 9268 + }, + { + "epoch": 0.7129451580647642, + "learning_rate": 0.0005697125798559672, + "loss": 1.6185, + "step": 9269 + }, + { + "epoch": 0.7130220752249827, + "learning_rate": 0.0005694282723356209, + "loss": 1.1985, + "step": 9270 + }, + { + "epoch": 0.7130989923852011, + "learning_rate": 0.0005691440191523282, + "loss": 1.3479, + "step": 9271 + }, + { + "epoch": 0.7131759095454195, + "learning_rate": 0.0005688598203226862, + "loss": 1.0522, + "step": 9272 + }, + { + "epoch": 0.713252826705638, + "learning_rate": 0.0005685756758632898, + "loss": 1.0467, + "step": 9273 + }, + { + "epoch": 0.7133297438658565, + "learning_rate": 0.0005682915857907303, + "loss": 1.2088, + "step": 9274 + }, + { + "epoch": 0.7134066610260749, + "learning_rate": 0.0005680075501215963, + "loss": 1.1078, + "step": 9275 + }, + { + "epoch": 0.7134835781862934, + "learning_rate": 0.0005677235688724721, + "loss": 1.1156, + "step": 9276 + }, + { + "epoch": 0.7135604953465118, + "learning_rate": 0.0005674396420599408, + "loss": 1.2745, + "step": 9277 + }, + { + "epoch": 0.7136374125067303, + "learning_rate": 0.0005671557697005808, + "loss": 1.3483, + "step": 9278 + }, + { + "epoch": 0.7137143296669487, + "learning_rate": 0.0005668719518109675, + "loss": 1.489, + "step": 9279 + }, + { + "epoch": 0.7137912468271671, + "learning_rate": 0.0005665881884076735, + "loss": 1.2943, + "step": 9280 + }, + { + "epoch": 0.7138681639873856, + "learning_rate": 0.0005663044795072673, + "loss": 1.2696, + "step": 9281 + }, + { + "epoch": 0.713945081147604, + "learning_rate": 0.0005660208251263169, + "loss": 1.3986, + "step": 9282 + }, + { + "epoch": 0.7140219983078224, + "learning_rate": 0.0005657372252813833, + "loss": 1.054, + "step": 9283 + }, + { + "epoch": 0.7140989154680409, + "learning_rate": 0.0005654536799890261, + "loss": 1.4222, + "step": 9284 + }, + { + "epoch": 0.7141758326282593, + "learning_rate": 0.000565170189265803, + "loss": 0.6641, + "step": 9285 + }, + { + "epoch": 0.7142527497884779, + "learning_rate": 0.0005648867531282669, + "loss": 0.8956, + "step": 9286 + }, + { + "epoch": 0.7143296669486963, + "learning_rate": 0.0005646033715929678, + "loss": 0.9189, + "step": 9287 + }, + { + "epoch": 0.7144065841089147, + "learning_rate": 0.0005643200446764526, + "loss": 1.2879, + "step": 9288 + }, + { + "epoch": 0.7144835012691332, + "learning_rate": 0.0005640367723952648, + "loss": 1.2763, + "step": 9289 + }, + { + "epoch": 0.7145604184293516, + "learning_rate": 0.0005637535547659458, + "loss": 1.0203, + "step": 9290 + }, + { + "epoch": 0.71463733558957, + "learning_rate": 0.0005634703918050329, + "loss": 1.4123, + "step": 9291 + }, + { + "epoch": 0.7147142527497885, + "learning_rate": 0.0005631872835290589, + "loss": 0.8268, + "step": 9292 + }, + { + "epoch": 0.7147911699100069, + "learning_rate": 0.0005629042299545561, + "loss": 1.198, + "step": 9293 + }, + { + "epoch": 0.7148680870702254, + "learning_rate": 0.0005626212310980514, + "loss": 1.0468, + "step": 9294 + }, + { + "epoch": 0.7149450042304438, + "learning_rate": 0.0005623382869760713, + "loss": 1.1087, + "step": 9295 + }, + { + "epoch": 0.7150219213906622, + "learning_rate": 0.0005620553976051351, + "loss": 1.0472, + "step": 9296 + }, + { + "epoch": 0.7150988385508807, + "learning_rate": 0.0005617725630017611, + "loss": 1.1679, + "step": 9297 + }, + { + "epoch": 0.7151757557110991, + "learning_rate": 0.0005614897831824656, + "loss": 1.5574, + "step": 9298 + }, + { + "epoch": 0.7152526728713176, + "learning_rate": 0.0005612070581637596, + "loss": 1.0795, + "step": 9299 + }, + { + "epoch": 0.7153295900315361, + "learning_rate": 0.0005609243879621521, + "loss": 1.1924, + "step": 9300 + }, + { + "epoch": 0.7154065071917545, + "learning_rate": 0.0005606417725941481, + "loss": 1.1555, + "step": 9301 + }, + { + "epoch": 0.7154834243519729, + "learning_rate": 0.0005603592120762501, + "loss": 1.1698, + "step": 9302 + }, + { + "epoch": 0.7155603415121914, + "learning_rate": 0.0005600767064249562, + "loss": 0.9323, + "step": 9303 + }, + { + "epoch": 0.7156372586724098, + "learning_rate": 0.0005597942556567643, + "loss": 1.3894, + "step": 9304 + }, + { + "epoch": 0.7157141758326283, + "learning_rate": 0.0005595118597881646, + "loss": 1.0572, + "step": 9305 + }, + { + "epoch": 0.7157910929928467, + "learning_rate": 0.0005592295188356479, + "loss": 1.1573, + "step": 9306 + }, + { + "epoch": 0.7158680101530651, + "learning_rate": 0.0005589472328157002, + "loss": 1.0623, + "step": 9307 + }, + { + "epoch": 0.7159449273132836, + "learning_rate": 0.0005586650017448041, + "loss": 1.0767, + "step": 9308 + }, + { + "epoch": 0.716021844473502, + "learning_rate": 0.0005583828256394399, + "loss": 0.9451, + "step": 9309 + }, + { + "epoch": 0.7160987616337204, + "learning_rate": 0.0005581007045160837, + "loss": 1.0433, + "step": 9310 + }, + { + "epoch": 0.716175678793939, + "learning_rate": 0.0005578186383912085, + "loss": 0.8603, + "step": 9311 + }, + { + "epoch": 0.7162525959541574, + "learning_rate": 0.0005575366272812854, + "loss": 1.2741, + "step": 9312 + }, + { + "epoch": 0.7163295131143759, + "learning_rate": 0.0005572546712027808, + "loss": 1.3708, + "step": 9313 + }, + { + "epoch": 0.7164064302745943, + "learning_rate": 0.0005569727701721584, + "loss": 1.356, + "step": 9314 + }, + { + "epoch": 0.7164833474348127, + "learning_rate": 0.0005566909242058789, + "loss": 1.1124, + "step": 9315 + }, + { + "epoch": 0.7165602645950312, + "learning_rate": 0.0005564091333203988, + "loss": 1.0062, + "step": 9316 + }, + { + "epoch": 0.7166371817552496, + "learning_rate": 0.000556127397532174, + "loss": 1.0108, + "step": 9317 + }, + { + "epoch": 0.716714098915468, + "learning_rate": 0.0005558457168576534, + "loss": 1.0154, + "step": 9318 + }, + { + "epoch": 0.7167910160756865, + "learning_rate": 0.000555564091313285, + "loss": 1.1735, + "step": 9319 + }, + { + "epoch": 0.7168679332359049, + "learning_rate": 0.000555282520915514, + "loss": 0.9607, + "step": 9320 + }, + { + "epoch": 0.7169448503961233, + "learning_rate": 0.0005550010056807813, + "loss": 1.3898, + "step": 9321 + }, + { + "epoch": 0.7170217675563418, + "learning_rate": 0.0005547195456255246, + "loss": 1.3058, + "step": 9322 + }, + { + "epoch": 0.7170986847165602, + "learning_rate": 0.0005544381407661787, + "loss": 1.3691, + "step": 9323 + }, + { + "epoch": 0.7171756018767788, + "learning_rate": 0.0005541567911191749, + "loss": 1.0796, + "step": 9324 + }, + { + "epoch": 0.7172525190369972, + "learning_rate": 0.0005538754967009423, + "loss": 1.2049, + "step": 9325 + }, + { + "epoch": 0.7173294361972156, + "learning_rate": 0.0005535942575279053, + "loss": 1.0439, + "step": 9326 + }, + { + "epoch": 0.7174063533574341, + "learning_rate": 0.0005533130736164861, + "loss": 1.515, + "step": 9327 + }, + { + "epoch": 0.7174832705176525, + "learning_rate": 0.000553031944983103, + "loss": 0.829, + "step": 9328 + }, + { + "epoch": 0.7175601876778709, + "learning_rate": 0.0005527508716441716, + "loss": 1.0455, + "step": 9329 + }, + { + "epoch": 0.7176371048380894, + "learning_rate": 0.0005524698536161041, + "loss": 1.173, + "step": 9330 + }, + { + "epoch": 0.7177140219983078, + "learning_rate": 0.0005521888909153092, + "loss": 0.8212, + "step": 9331 + }, + { + "epoch": 0.7177909391585263, + "learning_rate": 0.0005519079835581924, + "loss": 1.1177, + "step": 9332 + }, + { + "epoch": 0.7178678563187447, + "learning_rate": 0.0005516271315611568, + "loss": 1.0467, + "step": 9333 + }, + { + "epoch": 0.7179447734789631, + "learning_rate": 0.0005513463349406015, + "loss": 1.4146, + "step": 9334 + }, + { + "epoch": 0.7180216906391816, + "learning_rate": 0.0005510655937129224, + "loss": 0.9482, + "step": 9335 + }, + { + "epoch": 0.7180986077994, + "learning_rate": 0.0005507849078945121, + "loss": 1.023, + "step": 9336 + }, + { + "epoch": 0.7181755249596184, + "learning_rate": 0.0005505042775017603, + "loss": 0.8764, + "step": 9337 + }, + { + "epoch": 0.718252442119837, + "learning_rate": 0.0005502237025510529, + "loss": 1.1324, + "step": 9338 + }, + { + "epoch": 0.7183293592800554, + "learning_rate": 0.0005499431830587738, + "loss": 0.9934, + "step": 9339 + }, + { + "epoch": 0.7184062764402738, + "learning_rate": 0.0005496627190413028, + "loss": 1.209, + "step": 9340 + }, + { + "epoch": 0.7184831936004923, + "learning_rate": 0.0005493823105150151, + "loss": 1.2592, + "step": 9341 + }, + { + "epoch": 0.7185601107607107, + "learning_rate": 0.0005491019574962856, + "loss": 1.0501, + "step": 9342 + }, + { + "epoch": 0.7186370279209292, + "learning_rate": 0.0005488216600014836, + "loss": 1.2749, + "step": 9343 + }, + { + "epoch": 0.7187139450811476, + "learning_rate": 0.0005485414180469762, + "loss": 1.1647, + "step": 9344 + }, + { + "epoch": 0.718790862241366, + "learning_rate": 0.000548261231649127, + "loss": 0.8373, + "step": 9345 + }, + { + "epoch": 0.7188677794015845, + "learning_rate": 0.0005479811008242961, + "loss": 0.9246, + "step": 9346 + }, + { + "epoch": 0.7189446965618029, + "learning_rate": 0.0005477010255888412, + "loss": 1.1222, + "step": 9347 + }, + { + "epoch": 0.7190216137220213, + "learning_rate": 0.0005474210059591158, + "loss": 1.0607, + "step": 9348 + }, + { + "epoch": 0.7190985308822399, + "learning_rate": 0.0005471410419514707, + "loss": 1.3698, + "step": 9349 + }, + { + "epoch": 0.7191754480424583, + "learning_rate": 0.0005468611335822532, + "loss": 1.5374, + "step": 9350 + }, + { + "epoch": 0.7192523652026768, + "learning_rate": 0.000546581280867807, + "loss": 1.2769, + "step": 9351 + }, + { + "epoch": 0.7193292823628952, + "learning_rate": 0.0005463014838244738, + "loss": 1.0591, + "step": 9352 + }, + { + "epoch": 0.7194061995231136, + "learning_rate": 0.0005460217424685915, + "loss": 1.668, + "step": 9353 + }, + { + "epoch": 0.7194831166833321, + "learning_rate": 0.0005457420568164928, + "loss": 1.0583, + "step": 9354 + }, + { + "epoch": 0.7195600338435505, + "learning_rate": 0.0005454624268845105, + "loss": 0.8462, + "step": 9355 + }, + { + "epoch": 0.7196369510037689, + "learning_rate": 0.0005451828526889718, + "loss": 1.3111, + "step": 9356 + }, + { + "epoch": 0.7197138681639874, + "learning_rate": 0.0005449033342462015, + "loss": 1.2163, + "step": 9357 + }, + { + "epoch": 0.7197907853242058, + "learning_rate": 0.0005446238715725209, + "loss": 1.119, + "step": 9358 + }, + { + "epoch": 0.7198677024844242, + "learning_rate": 0.0005443444646842479, + "loss": 1.292, + "step": 9359 + }, + { + "epoch": 0.7199446196446427, + "learning_rate": 0.0005440651135976973, + "loss": 1.3125, + "step": 9360 + }, + { + "epoch": 0.7200215368048611, + "learning_rate": 0.0005437858183291814, + "loss": 1.2528, + "step": 9361 + }, + { + "epoch": 0.7200984539650797, + "learning_rate": 0.000543506578895008, + "loss": 0.8609, + "step": 9362 + }, + { + "epoch": 0.7201753711252981, + "learning_rate": 0.0005432273953114825, + "loss": 1.1988, + "step": 9363 + }, + { + "epoch": 0.7202522882855165, + "learning_rate": 0.0005429482675949063, + "loss": 1.0856, + "step": 9364 + }, + { + "epoch": 0.720329205445735, + "learning_rate": 0.0005426691957615778, + "loss": 1.2129, + "step": 9365 + }, + { + "epoch": 0.7204061226059534, + "learning_rate": 0.0005423901798277936, + "loss": 1.3119, + "step": 9366 + }, + { + "epoch": 0.7204830397661718, + "learning_rate": 0.0005421112198098444, + "loss": 1.2377, + "step": 9367 + }, + { + "epoch": 0.7205599569263903, + "learning_rate": 0.0005418323157240187, + "loss": 1.0323, + "step": 9368 + }, + { + "epoch": 0.7206368740866087, + "learning_rate": 0.0005415534675866033, + "loss": 1.3434, + "step": 9369 + }, + { + "epoch": 0.7207137912468272, + "learning_rate": 0.0005412746754138799, + "loss": 1.2895, + "step": 9370 + }, + { + "epoch": 0.7207907084070456, + "learning_rate": 0.0005409959392221271, + "loss": 0.9031, + "step": 9371 + }, + { + "epoch": 0.720867625567264, + "learning_rate": 0.000540717259027621, + "loss": 1.114, + "step": 9372 + }, + { + "epoch": 0.7209445427274825, + "learning_rate": 0.0005404386348466333, + "loss": 1.2747, + "step": 9373 + }, + { + "epoch": 0.7210214598877009, + "learning_rate": 0.0005401600666954344, + "loss": 1.0093, + "step": 9374 + }, + { + "epoch": 0.7210983770479193, + "learning_rate": 0.0005398815545902899, + "loss": 1.6238, + "step": 9375 + }, + { + "epoch": 0.7211752942081379, + "learning_rate": 0.0005396030985474611, + "loss": 1.0713, + "step": 9376 + }, + { + "epoch": 0.7212522113683563, + "learning_rate": 0.0005393246985832087, + "loss": 0.9556, + "step": 9377 + }, + { + "epoch": 0.7213291285285747, + "learning_rate": 0.0005390463547137883, + "loss": 0.9452, + "step": 9378 + }, + { + "epoch": 0.7214060456887932, + "learning_rate": 0.000538768066955453, + "loss": 1.1993, + "step": 9379 + }, + { + "epoch": 0.7214829628490116, + "learning_rate": 0.0005384898353244517, + "loss": 1.2313, + "step": 9380 + }, + { + "epoch": 0.7215598800092301, + "learning_rate": 0.0005382116598370307, + "loss": 1.1112, + "step": 9381 + }, + { + "epoch": 0.7216367971694485, + "learning_rate": 0.0005379335405094337, + "loss": 1.2084, + "step": 9382 + }, + { + "epoch": 0.7217137143296669, + "learning_rate": 0.0005376554773578998, + "loss": 0.8355, + "step": 9383 + }, + { + "epoch": 0.7217906314898854, + "learning_rate": 0.0005373774703986658, + "loss": 1.0795, + "step": 9384 + }, + { + "epoch": 0.7218675486501038, + "learning_rate": 0.0005370995196479642, + "loss": 1.1485, + "step": 9385 + }, + { + "epoch": 0.7219444658103222, + "learning_rate": 0.0005368216251220253, + "loss": 1.2273, + "step": 9386 + }, + { + "epoch": 0.7220213829705407, + "learning_rate": 0.0005365437868370751, + "loss": 1.2547, + "step": 9387 + }, + { + "epoch": 0.7220983001307592, + "learning_rate": 0.000536266004809338, + "loss": 1.2919, + "step": 9388 + }, + { + "epoch": 0.7221752172909777, + "learning_rate": 0.0005359882790550325, + "loss": 1.0553, + "step": 9389 + }, + { + "epoch": 0.7222521344511961, + "learning_rate": 0.0005357106095903763, + "loss": 1.1381, + "step": 9390 + }, + { + "epoch": 0.7223290516114145, + "learning_rate": 0.0005354329964315829, + "loss": 1.1194, + "step": 9391 + }, + { + "epoch": 0.722405968771633, + "learning_rate": 0.0005351554395948619, + "loss": 1.0312, + "step": 9392 + }, + { + "epoch": 0.7224828859318514, + "learning_rate": 0.0005348779390964203, + "loss": 0.7245, + "step": 9393 + }, + { + "epoch": 0.7225598030920698, + "learning_rate": 0.0005346004949524618, + "loss": 1.0681, + "step": 9394 + }, + { + "epoch": 0.7226367202522883, + "learning_rate": 0.0005343231071791858, + "loss": 1.0064, + "step": 9395 + }, + { + "epoch": 0.7227136374125067, + "learning_rate": 0.0005340457757927906, + "loss": 1.0391, + "step": 9396 + }, + { + "epoch": 0.7227905545727252, + "learning_rate": 0.0005337685008094694, + "loss": 0.9042, + "step": 9397 + }, + { + "epoch": 0.7228674717329436, + "learning_rate": 0.0005334912822454121, + "loss": 1.4803, + "step": 9398 + }, + { + "epoch": 0.722944388893162, + "learning_rate": 0.0005332141201168063, + "loss": 1.2519, + "step": 9399 + }, + { + "epoch": 0.7230213060533806, + "learning_rate": 0.0005329370144398351, + "loss": 1.1931, + "step": 9400 + }, + { + "epoch": 0.723098223213599, + "learning_rate": 0.0005326599652306804, + "loss": 1.1531, + "step": 9401 + }, + { + "epoch": 0.7231751403738174, + "learning_rate": 0.0005323829725055181, + "loss": 1.4472, + "step": 9402 + }, + { + "epoch": 0.7232520575340359, + "learning_rate": 0.0005321060362805217, + "loss": 0.9079, + "step": 9403 + }, + { + "epoch": 0.7233289746942543, + "learning_rate": 0.0005318291565718633, + "loss": 1.2469, + "step": 9404 + }, + { + "epoch": 0.7234058918544727, + "learning_rate": 0.0005315523333957094, + "loss": 1.1492, + "step": 9405 + }, + { + "epoch": 0.7234828090146912, + "learning_rate": 0.0005312755667682239, + "loss": 0.8142, + "step": 9406 + }, + { + "epoch": 0.7235597261749096, + "learning_rate": 0.0005309988567055679, + "loss": 1.0987, + "step": 9407 + }, + { + "epoch": 0.7236366433351281, + "learning_rate": 0.0005307222032238979, + "loss": 1.2596, + "step": 9408 + }, + { + "epoch": 0.7237135604953465, + "learning_rate": 0.0005304456063393691, + "loss": 1.2135, + "step": 9409 + }, + { + "epoch": 0.7237904776555649, + "learning_rate": 0.0005301690660681322, + "loss": 1.2055, + "step": 9410 + }, + { + "epoch": 0.7238673948157834, + "learning_rate": 0.0005298925824263333, + "loss": 1.2809, + "step": 9411 + }, + { + "epoch": 0.7239443119760018, + "learning_rate": 0.0005296161554301181, + "loss": 1.1521, + "step": 9412 + }, + { + "epoch": 0.7240212291362202, + "learning_rate": 0.0005293397850956268, + "loss": 1.1096, + "step": 9413 + }, + { + "epoch": 0.7240981462964388, + "learning_rate": 0.0005290634714389972, + "loss": 1.1432, + "step": 9414 + }, + { + "epoch": 0.7241750634566572, + "learning_rate": 0.0005287872144763633, + "loss": 1.1967, + "step": 9415 + }, + { + "epoch": 0.7242519806168757, + "learning_rate": 0.0005285110142238556, + "loss": 1.3089, + "step": 9416 + }, + { + "epoch": 0.7243288977770941, + "learning_rate": 0.0005282348706976027, + "loss": 0.9607, + "step": 9417 + }, + { + "epoch": 0.7244058149373125, + "learning_rate": 0.0005279587839137286, + "loss": 1.2065, + "step": 9418 + }, + { + "epoch": 0.724482732097531, + "learning_rate": 0.0005276827538883541, + "loss": 0.6844, + "step": 9419 + }, + { + "epoch": 0.7245596492577494, + "learning_rate": 0.0005274067806375968, + "loss": 1.1616, + "step": 9420 + }, + { + "epoch": 0.7246365664179678, + "learning_rate": 0.0005271308641775713, + "loss": 1.3268, + "step": 9421 + }, + { + "epoch": 0.7247134835781863, + "learning_rate": 0.000526855004524388, + "loss": 1.3412, + "step": 9422 + }, + { + "epoch": 0.7247904007384047, + "learning_rate": 0.0005265792016941563, + "loss": 1.2578, + "step": 9423 + }, + { + "epoch": 0.7248673178986231, + "learning_rate": 0.0005263034557029788, + "loss": 1.1353, + "step": 9424 + }, + { + "epoch": 0.7249442350588416, + "learning_rate": 0.0005260277665669569, + "loss": 1.0317, + "step": 9425 + }, + { + "epoch": 0.72502115221906, + "learning_rate": 0.0005257521343021893, + "loss": 1.2119, + "step": 9426 + }, + { + "epoch": 0.7250980693792786, + "learning_rate": 0.0005254765589247697, + "loss": 1.2605, + "step": 9427 + }, + { + "epoch": 0.725174986539497, + "learning_rate": 0.0005252010404507896, + "loss": 1.4108, + "step": 9428 + }, + { + "epoch": 0.7252519036997154, + "learning_rate": 0.0005249255788963367, + "loss": 1.1841, + "step": 9429 + }, + { + "epoch": 0.7253288208599339, + "learning_rate": 0.0005246501742774948, + "loss": 1.1969, + "step": 9430 + }, + { + "epoch": 0.7254057380201523, + "learning_rate": 0.0005243748266103462, + "loss": 0.8568, + "step": 9431 + }, + { + "epoch": 0.7254826551803707, + "learning_rate": 0.000524099535910969, + "loss": 0.8923, + "step": 9432 + }, + { + "epoch": 0.7255595723405892, + "learning_rate": 0.0005238243021954358, + "loss": 1.7811, + "step": 9433 + }, + { + "epoch": 0.7256364895008076, + "learning_rate": 0.0005235491254798193, + "loss": 1.0812, + "step": 9434 + }, + { + "epoch": 0.7257134066610261, + "learning_rate": 0.0005232740057801867, + "loss": 0.8363, + "step": 9435 + }, + { + "epoch": 0.7257903238212445, + "learning_rate": 0.0005229989431126037, + "loss": 1.0442, + "step": 9436 + }, + { + "epoch": 0.7258672409814629, + "learning_rate": 0.0005227239374931302, + "loss": 1.5356, + "step": 9437 + }, + { + "epoch": 0.7259441581416815, + "learning_rate": 0.0005224489889378241, + "loss": 1.0046, + "step": 9438 + }, + { + "epoch": 0.7260210753018999, + "learning_rate": 0.0005221740974627407, + "loss": 1.0041, + "step": 9439 + }, + { + "epoch": 0.7260979924621183, + "learning_rate": 0.0005218992630839307, + "loss": 0.9842, + "step": 9440 + }, + { + "epoch": 0.7261749096223368, + "learning_rate": 0.0005216244858174422, + "loss": 1.2465, + "step": 9441 + }, + { + "epoch": 0.7262518267825552, + "learning_rate": 0.0005213497656793195, + "loss": 1.1504, + "step": 9442 + }, + { + "epoch": 0.7263287439427736, + "learning_rate": 0.0005210751026856041, + "loss": 0.7373, + "step": 9443 + }, + { + "epoch": 0.7264056611029921, + "learning_rate": 0.000520800496852333, + "loss": 1.1089, + "step": 9444 + }, + { + "epoch": 0.7264825782632105, + "learning_rate": 0.0005205259481955424, + "loss": 1.1643, + "step": 9445 + }, + { + "epoch": 0.726559495423429, + "learning_rate": 0.0005202514567312616, + "loss": 1.0695, + "step": 9446 + }, + { + "epoch": 0.7266364125836474, + "learning_rate": 0.0005199770224755197, + "loss": 1.7981, + "step": 9447 + }, + { + "epoch": 0.7267133297438658, + "learning_rate": 0.0005197026454443408, + "loss": 0.8053, + "step": 9448 + }, + { + "epoch": 0.7267902469040843, + "learning_rate": 0.0005194283256537461, + "loss": 0.9412, + "step": 9449 + }, + { + "epoch": 0.7268671640643027, + "learning_rate": 0.0005191540631197534, + "loss": 0.9117, + "step": 9450 + }, + { + "epoch": 0.7269440812245211, + "learning_rate": 0.0005188798578583772, + "loss": 1.1715, + "step": 9451 + }, + { + "epoch": 0.7270209983847397, + "learning_rate": 0.000518605709885628, + "loss": 0.8628, + "step": 9452 + }, + { + "epoch": 0.7270979155449581, + "learning_rate": 0.0005183316192175149, + "loss": 1.2291, + "step": 9453 + }, + { + "epoch": 0.7271748327051766, + "learning_rate": 0.0005180575858700413, + "loss": 1.2169, + "step": 9454 + }, + { + "epoch": 0.727251749865395, + "learning_rate": 0.0005177836098592089, + "loss": 1.1948, + "step": 9455 + }, + { + "epoch": 0.7273286670256134, + "learning_rate": 0.000517509691201015, + "loss": 1.0877, + "step": 9456 + }, + { + "epoch": 0.7274055841858319, + "learning_rate": 0.0005172358299114538, + "loss": 1.093, + "step": 9457 + }, + { + "epoch": 0.7274825013460503, + "learning_rate": 0.0005169620260065175, + "loss": 0.9695, + "step": 9458 + }, + { + "epoch": 0.7275594185062687, + "learning_rate": 0.0005166882795021926, + "loss": 0.9559, + "step": 9459 + }, + { + "epoch": 0.7276363356664872, + "learning_rate": 0.0005164145904144634, + "loss": 1.4075, + "step": 9460 + }, + { + "epoch": 0.7277132528267056, + "learning_rate": 0.0005161409587593119, + "loss": 1.3804, + "step": 9461 + }, + { + "epoch": 0.727790169986924, + "learning_rate": 0.000515867384552715, + "loss": 0.9768, + "step": 9462 + }, + { + "epoch": 0.7278670871471425, + "learning_rate": 0.0005155938678106472, + "loss": 1.2198, + "step": 9463 + }, + { + "epoch": 0.727944004307361, + "learning_rate": 0.0005153204085490794, + "loss": 1.1318, + "step": 9464 + }, + { + "epoch": 0.7280209214675795, + "learning_rate": 0.0005150470067839787, + "loss": 1.1305, + "step": 9465 + }, + { + "epoch": 0.7280978386277979, + "learning_rate": 0.0005147736625313101, + "loss": 1.1635, + "step": 9466 + }, + { + "epoch": 0.7281747557880163, + "learning_rate": 0.0005145003758070348, + "loss": 1.1236, + "step": 9467 + }, + { + "epoch": 0.7282516729482348, + "learning_rate": 0.0005142271466271086, + "loss": 1.026, + "step": 9468 + }, + { + "epoch": 0.7283285901084532, + "learning_rate": 0.000513953975007487, + "loss": 0.9184, + "step": 9469 + }, + { + "epoch": 0.7284055072686716, + "learning_rate": 0.0005136808609641206, + "loss": 1.19, + "step": 9470 + }, + { + "epoch": 0.7284824244288901, + "learning_rate": 0.0005134078045129567, + "loss": 1.1412, + "step": 9471 + }, + { + "epoch": 0.7285593415891085, + "learning_rate": 0.0005131348056699391, + "loss": 1.1946, + "step": 9472 + }, + { + "epoch": 0.728636258749327, + "learning_rate": 0.0005128618644510083, + "loss": 1.1496, + "step": 9473 + }, + { + "epoch": 0.7287131759095454, + "learning_rate": 0.0005125889808721025, + "loss": 1.2636, + "step": 9474 + }, + { + "epoch": 0.7287900930697638, + "learning_rate": 0.0005123161549491551, + "loss": 1.34, + "step": 9475 + }, + { + "epoch": 0.7288670102299823, + "learning_rate": 0.0005120433866980968, + "loss": 1.1529, + "step": 9476 + }, + { + "epoch": 0.7289439273902008, + "learning_rate": 0.0005117706761348549, + "loss": 0.9917, + "step": 9477 + }, + { + "epoch": 0.7290208445504192, + "learning_rate": 0.0005114980232753529, + "loss": 1.0118, + "step": 9478 + }, + { + "epoch": 0.7290977617106377, + "learning_rate": 0.000511225428135511, + "loss": 1.16, + "step": 9479 + }, + { + "epoch": 0.7291746788708561, + "learning_rate": 0.0005109528907312476, + "loss": 1.4642, + "step": 9480 + }, + { + "epoch": 0.7292515960310745, + "learning_rate": 0.0005106804110784758, + "loss": 1.1429, + "step": 9481 + }, + { + "epoch": 0.729328513191293, + "learning_rate": 0.0005104079891931052, + "loss": 1.1273, + "step": 9482 + }, + { + "epoch": 0.7294054303515114, + "learning_rate": 0.0005101356250910438, + "loss": 1.2409, + "step": 9483 + }, + { + "epoch": 0.7294823475117299, + "learning_rate": 0.0005098633187881948, + "loss": 0.9331, + "step": 9484 + }, + { + "epoch": 0.7295592646719483, + "learning_rate": 0.0005095910703004586, + "loss": 1.1855, + "step": 9485 + }, + { + "epoch": 0.7296361818321667, + "learning_rate": 0.0005093188796437319, + "loss": 1.0239, + "step": 9486 + }, + { + "epoch": 0.7297130989923852, + "learning_rate": 0.0005090467468339078, + "loss": 1.1713, + "step": 9487 + }, + { + "epoch": 0.7297900161526036, + "learning_rate": 0.0005087746718868774, + "loss": 1.3517, + "step": 9488 + }, + { + "epoch": 0.729866933312822, + "learning_rate": 0.000508502654818527, + "loss": 1.2247, + "step": 9489 + }, + { + "epoch": 0.7299438504730406, + "learning_rate": 0.0005082306956447399, + "loss": 1.1024, + "step": 9490 + }, + { + "epoch": 0.730020767633259, + "learning_rate": 0.0005079587943813959, + "loss": 1.0691, + "step": 9491 + }, + { + "epoch": 0.7300976847934775, + "learning_rate": 0.0005076869510443715, + "loss": 1.4135, + "step": 9492 + }, + { + "epoch": 0.7301746019536959, + "learning_rate": 0.0005074151656495406, + "loss": 1.1331, + "step": 9493 + }, + { + "epoch": 0.7302515191139143, + "learning_rate": 0.0005071434382127732, + "loss": 0.9224, + "step": 9494 + }, + { + "epoch": 0.7303284362741328, + "learning_rate": 0.0005068717687499341, + "loss": 0.9198, + "step": 9495 + }, + { + "epoch": 0.7304053534343512, + "learning_rate": 0.000506600157276888, + "loss": 0.995, + "step": 9496 + }, + { + "epoch": 0.7304822705945696, + "learning_rate": 0.0005063286038094938, + "loss": 1.0685, + "step": 9497 + }, + { + "epoch": 0.7305591877547881, + "learning_rate": 0.0005060571083636083, + "loss": 1.2064, + "step": 9498 + }, + { + "epoch": 0.7306361049150065, + "learning_rate": 0.0005057856709550839, + "loss": 0.8912, + "step": 9499 + }, + { + "epoch": 0.7307130220752249, + "learning_rate": 0.0005055142915997705, + "loss": 1.7754, + "step": 9500 + }, + { + "epoch": 0.7307899392354434, + "learning_rate": 0.0005052429703135134, + "loss": 1.4261, + "step": 9501 + }, + { + "epoch": 0.7308668563956618, + "learning_rate": 0.0005049717071121565, + "loss": 1.5207, + "step": 9502 + }, + { + "epoch": 0.7309437735558804, + "learning_rate": 0.0005047005020115385, + "loss": 1.2914, + "step": 9503 + }, + { + "epoch": 0.7310206907160988, + "learning_rate": 0.0005044293550274957, + "loss": 1.3741, + "step": 9504 + }, + { + "epoch": 0.7310976078763172, + "learning_rate": 0.0005041582661758602, + "loss": 1.3324, + "step": 9505 + }, + { + "epoch": 0.7311745250365357, + "learning_rate": 0.000503887235472461, + "loss": 1.411, + "step": 9506 + }, + { + "epoch": 0.7312514421967541, + "learning_rate": 0.0005036162629331252, + "loss": 1.2625, + "step": 9507 + }, + { + "epoch": 0.7313283593569725, + "learning_rate": 0.0005033453485736737, + "loss": 0.9098, + "step": 9508 + }, + { + "epoch": 0.731405276517191, + "learning_rate": 0.0005030744924099256, + "loss": 0.9647, + "step": 9509 + }, + { + "epoch": 0.7314821936774094, + "learning_rate": 0.0005028036944576973, + "loss": 0.6814, + "step": 9510 + }, + { + "epoch": 0.7315591108376279, + "learning_rate": 0.0005025329547328006, + "loss": 0.8911, + "step": 9511 + }, + { + "epoch": 0.7316360279978463, + "learning_rate": 0.0005022622732510439, + "loss": 0.8692, + "step": 9512 + }, + { + "epoch": 0.7317129451580647, + "learning_rate": 0.0005019916500282332, + "loss": 1.288, + "step": 9513 + }, + { + "epoch": 0.7317898623182832, + "learning_rate": 0.0005017210850801696, + "loss": 1.1947, + "step": 9514 + }, + { + "epoch": 0.7318667794785017, + "learning_rate": 0.0005014505784226527, + "loss": 1.1111, + "step": 9515 + }, + { + "epoch": 0.73194369663872, + "learning_rate": 0.0005011801300714777, + "loss": 1.0163, + "step": 9516 + }, + { + "epoch": 0.7320206137989386, + "learning_rate": 0.0005009097400424347, + "loss": 1.1756, + "step": 9517 + }, + { + "epoch": 0.732097530959157, + "learning_rate": 0.0005006394083513139, + "loss": 1.3218, + "step": 9518 + }, + { + "epoch": 0.7321744481193755, + "learning_rate": 0.0005003691350138989, + "loss": 1.1557, + "step": 9519 + }, + { + "epoch": 0.7322513652795939, + "learning_rate": 0.000500098920045973, + "loss": 1.1851, + "step": 9520 + }, + { + "epoch": 0.7323282824398123, + "learning_rate": 0.0004998287634633128, + "loss": 1.1823, + "step": 9521 + }, + { + "epoch": 0.7324051996000308, + "learning_rate": 0.0004995586652816928, + "loss": 1.0033, + "step": 9522 + }, + { + "epoch": 0.7324821167602492, + "learning_rate": 0.0004992886255168855, + "loss": 1.279, + "step": 9523 + }, + { + "epoch": 0.7325590339204676, + "learning_rate": 0.0004990186441846584, + "loss": 0.9729, + "step": 9524 + }, + { + "epoch": 0.7326359510806861, + "learning_rate": 0.0004987487213007756, + "loss": 0.7133, + "step": 9525 + }, + { + "epoch": 0.7327128682409045, + "learning_rate": 0.0004984788568809988, + "loss": 1.163, + "step": 9526 + }, + { + "epoch": 0.7327897854011229, + "learning_rate": 0.000498209050941085, + "loss": 1.2016, + "step": 9527 + }, + { + "epoch": 0.7328667025613415, + "learning_rate": 0.0004979393034967884, + "loss": 1.334, + "step": 9528 + }, + { + "epoch": 0.7329436197215599, + "learning_rate": 0.0004976696145638612, + "loss": 0.9937, + "step": 9529 + }, + { + "epoch": 0.7330205368817784, + "learning_rate": 0.0004973999841580487, + "loss": 1.1001, + "step": 9530 + }, + { + "epoch": 0.7330974540419968, + "learning_rate": 0.0004971304122950966, + "loss": 1.2798, + "step": 9531 + }, + { + "epoch": 0.7331743712022152, + "learning_rate": 0.000496860898990745, + "loss": 0.9932, + "step": 9532 + }, + { + "epoch": 0.7332512883624337, + "learning_rate": 0.0004965914442607307, + "loss": 0.8744, + "step": 9533 + }, + { + "epoch": 0.7333282055226521, + "learning_rate": 0.000496322048120788, + "loss": 1.2469, + "step": 9534 + }, + { + "epoch": 0.7334051226828705, + "learning_rate": 0.0004960527105866467, + "loss": 0.3908, + "step": 9535 + }, + { + "epoch": 0.733482039843089, + "learning_rate": 0.0004957834316740335, + "loss": 1.1258, + "step": 9536 + }, + { + "epoch": 0.7335589570033074, + "learning_rate": 0.0004955142113986727, + "loss": 0.8403, + "step": 9537 + }, + { + "epoch": 0.7336358741635259, + "learning_rate": 0.000495245049776284, + "loss": 1.2507, + "step": 9538 + }, + { + "epoch": 0.7337127913237443, + "learning_rate": 0.0004949759468225839, + "loss": 1.1635, + "step": 9539 + }, + { + "epoch": 0.7337897084839627, + "learning_rate": 0.0004947069025532859, + "loss": 1.0787, + "step": 9540 + }, + { + "epoch": 0.7338666256441813, + "learning_rate": 0.0004944379169840988, + "loss": 1.2461, + "step": 9541 + }, + { + "epoch": 0.7339435428043997, + "learning_rate": 0.0004941689901307308, + "loss": 1.0702, + "step": 9542 + }, + { + "epoch": 0.7340204599646181, + "learning_rate": 0.0004939001220088833, + "loss": 1.0311, + "step": 9543 + }, + { + "epoch": 0.7340973771248366, + "learning_rate": 0.0004936313126342556, + "loss": 1.0072, + "step": 9544 + }, + { + "epoch": 0.734174294285055, + "learning_rate": 0.0004933625620225449, + "loss": 1.4752, + "step": 9545 + }, + { + "epoch": 0.7342512114452734, + "learning_rate": 0.0004930938701894434, + "loss": 1.0207, + "step": 9546 + }, + { + "epoch": 0.7343281286054919, + "learning_rate": 0.00049282523715064, + "loss": 1.3235, + "step": 9547 + }, + { + "epoch": 0.7344050457657103, + "learning_rate": 0.0004925566629218208, + "loss": 1.133, + "step": 9548 + }, + { + "epoch": 0.7344819629259288, + "learning_rate": 0.0004922881475186673, + "loss": 1.1452, + "step": 9549 + }, + { + "epoch": 0.7345588800861472, + "learning_rate": 0.0004920196909568599, + "loss": 1.055, + "step": 9550 + }, + { + "epoch": 0.7346357972463656, + "learning_rate": 0.0004917512932520734, + "loss": 0.8454, + "step": 9551 + }, + { + "epoch": 0.7347127144065841, + "learning_rate": 0.0004914829544199788, + "loss": 1.3494, + "step": 9552 + }, + { + "epoch": 0.7347896315668025, + "learning_rate": 0.000491214674476246, + "loss": 1.3645, + "step": 9553 + }, + { + "epoch": 0.734866548727021, + "learning_rate": 0.0004909464534365399, + "loss": 1.1012, + "step": 9554 + }, + { + "epoch": 0.7349434658872395, + "learning_rate": 0.0004906782913165217, + "loss": 1.4507, + "step": 9555 + }, + { + "epoch": 0.7350203830474579, + "learning_rate": 0.0004904101881318501, + "loss": 1.1934, + "step": 9556 + }, + { + "epoch": 0.7350973002076764, + "learning_rate": 0.0004901421438981792, + "loss": 1.6996, + "step": 9557 + }, + { + "epoch": 0.7351742173678948, + "learning_rate": 0.0004898741586311615, + "loss": 1.2667, + "step": 9558 + }, + { + "epoch": 0.7352511345281132, + "learning_rate": 0.0004896062323464447, + "loss": 1.1764, + "step": 9559 + }, + { + "epoch": 0.7353280516883317, + "learning_rate": 0.0004893383650596726, + "loss": 0.8002, + "step": 9560 + }, + { + "epoch": 0.7354049688485501, + "learning_rate": 0.0004890705567864868, + "loss": 0.9024, + "step": 9561 + }, + { + "epoch": 0.7354818860087685, + "learning_rate": 0.0004888028075425248, + "loss": 1.0302, + "step": 9562 + }, + { + "epoch": 0.735558803168987, + "learning_rate": 0.0004885351173434203, + "loss": 1.1395, + "step": 9563 + }, + { + "epoch": 0.7356357203292054, + "learning_rate": 0.00048826748620480517, + "loss": 1.3611, + "step": 9564 + }, + { + "epoch": 0.7357126374894238, + "learning_rate": 0.00048799991414230545, + "loss": 1.1677, + "step": 9565 + }, + { + "epoch": 0.7357895546496424, + "learning_rate": 0.00048773240117154504, + "loss": 1.1451, + "step": 9566 + }, + { + "epoch": 0.7358664718098608, + "learning_rate": 0.0004874649473081452, + "loss": 1.0619, + "step": 9567 + }, + { + "epoch": 0.7359433889700793, + "learning_rate": 0.0004871975525677223, + "loss": 1.2173, + "step": 9568 + }, + { + "epoch": 0.7360203061302977, + "learning_rate": 0.0004869302169658898, + "loss": 0.9331, + "step": 9569 + }, + { + "epoch": 0.7360972232905161, + "learning_rate": 0.00048666294051825763, + "loss": 1.1044, + "step": 9570 + }, + { + "epoch": 0.7361741404507346, + "learning_rate": 0.0004863957232404321, + "loss": 1.0849, + "step": 9571 + }, + { + "epoch": 0.736251057610953, + "learning_rate": 0.00048612856514801704, + "loss": 1.3422, + "step": 9572 + }, + { + "epoch": 0.7363279747711714, + "learning_rate": 0.00048586146625661234, + "loss": 1.2571, + "step": 9573 + }, + { + "epoch": 0.7364048919313899, + "learning_rate": 0.00048559442658181235, + "loss": 0.9953, + "step": 9574 + }, + { + "epoch": 0.7364818090916083, + "learning_rate": 0.00048532744613921154, + "loss": 1.3342, + "step": 9575 + }, + { + "epoch": 0.7365587262518268, + "learning_rate": 0.0004850605249443982, + "loss": 0.8317, + "step": 9576 + }, + { + "epoch": 0.7366356434120452, + "learning_rate": 0.00048479366301295944, + "loss": 1.3491, + "step": 9577 + }, + { + "epoch": 0.7367125605722636, + "learning_rate": 0.00048452686036047626, + "loss": 0.9034, + "step": 9578 + }, + { + "epoch": 0.7367894777324822, + "learning_rate": 0.00048426011700252754, + "loss": 1.0369, + "step": 9579 + }, + { + "epoch": 0.7368663948927006, + "learning_rate": 0.0004839934329546896, + "loss": 1.2174, + "step": 9580 + }, + { + "epoch": 0.736943312052919, + "learning_rate": 0.0004837268082325341, + "loss": 1.0057, + "step": 9581 + }, + { + "epoch": 0.7370202292131375, + "learning_rate": 0.0004834602428516293, + "loss": 1.2124, + "step": 9582 + }, + { + "epoch": 0.7370971463733559, + "learning_rate": 0.00048319373682754026, + "loss": 1.3084, + "step": 9583 + }, + { + "epoch": 0.7371740635335743, + "learning_rate": 0.00048292729017582875, + "loss": 1.3049, + "step": 9584 + }, + { + "epoch": 0.7372509806937928, + "learning_rate": 0.0004826609029120522, + "loss": 1.1122, + "step": 9585 + }, + { + "epoch": 0.7373278978540112, + "learning_rate": 0.0004823945750517667, + "loss": 1.2862, + "step": 9586 + }, + { + "epoch": 0.7374048150142297, + "learning_rate": 0.0004821283066105216, + "loss": 1.1289, + "step": 9587 + }, + { + "epoch": 0.7374817321744481, + "learning_rate": 0.0004818620976038659, + "loss": 1.1788, + "step": 9588 + }, + { + "epoch": 0.7375586493346665, + "learning_rate": 0.00048159594804734335, + "loss": 0.842, + "step": 9589 + }, + { + "epoch": 0.737635566494885, + "learning_rate": 0.00048132985795649487, + "loss": 1.2714, + "step": 9590 + }, + { + "epoch": 0.7377124836551034, + "learning_rate": 0.00048106382734685747, + "loss": 1.1957, + "step": 9591 + }, + { + "epoch": 0.7377894008153218, + "learning_rate": 0.00048079785623396533, + "loss": 1.3048, + "step": 9592 + }, + { + "epoch": 0.7378663179755404, + "learning_rate": 0.00048053194463334775, + "loss": 1.1409, + "step": 9593 + }, + { + "epoch": 0.7379432351357588, + "learning_rate": 0.0004802660925605332, + "loss": 1.3611, + "step": 9594 + }, + { + "epoch": 0.7380201522959773, + "learning_rate": 0.00048000030003104413, + "loss": 0.7823, + "step": 9595 + }, + { + "epoch": 0.7380970694561957, + "learning_rate": 0.00047973456706040037, + "loss": 0.967, + "step": 9596 + }, + { + "epoch": 0.7381739866164141, + "learning_rate": 0.0004794688936641187, + "loss": 1.2858, + "step": 9597 + }, + { + "epoch": 0.7382509037766326, + "learning_rate": 0.0004792032798577114, + "loss": 1.0721, + "step": 9598 + }, + { + "epoch": 0.738327820936851, + "learning_rate": 0.0004789377256566895, + "loss": 1.0331, + "step": 9599 + }, + { + "epoch": 0.7384047380970694, + "learning_rate": 0.0004786722310765574, + "loss": 1.0129, + "step": 9600 + }, + { + "epoch": 0.7384816552572879, + "learning_rate": 0.00047840679613281745, + "loss": 1.1207, + "step": 9601 + }, + { + "epoch": 0.7385585724175063, + "learning_rate": 0.0004781414208409701, + "loss": 1.2421, + "step": 9602 + }, + { + "epoch": 0.7386354895777247, + "learning_rate": 0.00047787610521651, + "loss": 1.2682, + "step": 9603 + }, + { + "epoch": 0.7387124067379433, + "learning_rate": 0.0004776108492749293, + "loss": 1.2052, + "step": 9604 + }, + { + "epoch": 0.7387893238981617, + "learning_rate": 0.0004773456530317167, + "loss": 1.1606, + "step": 9605 + }, + { + "epoch": 0.7388662410583802, + "learning_rate": 0.00047708051650235684, + "loss": 1.1792, + "step": 9606 + }, + { + "epoch": 0.7389431582185986, + "learning_rate": 0.00047681543970233195, + "loss": 1.2722, + "step": 9607 + }, + { + "epoch": 0.739020075378817, + "learning_rate": 0.00047655042264712044, + "loss": 1.2043, + "step": 9608 + }, + { + "epoch": 0.7390969925390355, + "learning_rate": 0.00047628546535219554, + "loss": 1.0525, + "step": 9609 + }, + { + "epoch": 0.7391739096992539, + "learning_rate": 0.0004760205678330294, + "loss": 0.9423, + "step": 9610 + }, + { + "epoch": 0.7392508268594723, + "learning_rate": 0.0004757557301050897, + "loss": 1.1264, + "step": 9611 + }, + { + "epoch": 0.7393277440196908, + "learning_rate": 0.0004754909521838403, + "loss": 1.4684, + "step": 9612 + }, + { + "epoch": 0.7394046611799092, + "learning_rate": 0.00047522623408474186, + "loss": 1.1883, + "step": 9613 + }, + { + "epoch": 0.7394815783401277, + "learning_rate": 0.00047496157582325126, + "loss": 0.9851, + "step": 9614 + }, + { + "epoch": 0.7395584955003461, + "learning_rate": 0.0004746969774148229, + "loss": 0.6258, + "step": 9615 + }, + { + "epoch": 0.7396354126605645, + "learning_rate": 0.00047443243887490663, + "loss": 1.7361, + "step": 9616 + }, + { + "epoch": 0.739712329820783, + "learning_rate": 0.000474167960218949, + "loss": 1.109, + "step": 9617 + }, + { + "epoch": 0.7397892469810015, + "learning_rate": 0.0004739035414623933, + "loss": 1.0861, + "step": 9618 + }, + { + "epoch": 0.7398661641412199, + "learning_rate": 0.0004736391826206792, + "loss": 1.15, + "step": 9619 + }, + { + "epoch": 0.7399430813014384, + "learning_rate": 0.0004733748837092423, + "loss": 0.8956, + "step": 9620 + }, + { + "epoch": 0.7400199984616568, + "learning_rate": 0.000473110644743517, + "loss": 1.4887, + "step": 9621 + }, + { + "epoch": 0.7400969156218753, + "learning_rate": 0.00047284646573893024, + "loss": 1.4459, + "step": 9622 + }, + { + "epoch": 0.7401738327820937, + "learning_rate": 0.00047258234671090937, + "loss": 0.9748, + "step": 9623 + }, + { + "epoch": 0.7402507499423121, + "learning_rate": 0.00047231828767487605, + "loss": 0.9643, + "step": 9624 + }, + { + "epoch": 0.7403276671025306, + "learning_rate": 0.0004720542886462488, + "loss": 1.1797, + "step": 9625 + }, + { + "epoch": 0.740404584262749, + "learning_rate": 0.00047179034964044304, + "loss": 1.0901, + "step": 9626 + }, + { + "epoch": 0.7404815014229674, + "learning_rate": 0.00047152647067287036, + "loss": 1.2084, + "step": 9627 + }, + { + "epoch": 0.7405584185831859, + "learning_rate": 0.0004712626517589384, + "loss": 1.4829, + "step": 9628 + }, + { + "epoch": 0.7406353357434043, + "learning_rate": 0.00047099889291405284, + "loss": 0.8356, + "step": 9629 + }, + { + "epoch": 0.7407122529036227, + "learning_rate": 0.0004707351941536142, + "loss": 1.2059, + "step": 9630 + }, + { + "epoch": 0.7407891700638413, + "learning_rate": 0.0004704715554930203, + "loss": 1.0607, + "step": 9631 + }, + { + "epoch": 0.7408660872240597, + "learning_rate": 0.0004702079769476653, + "loss": 1.2945, + "step": 9632 + }, + { + "epoch": 0.7409430043842782, + "learning_rate": 0.00046994445853293935, + "loss": 1.1079, + "step": 9633 + }, + { + "epoch": 0.7410199215444966, + "learning_rate": 0.0004696810002642307, + "loss": 1.1573, + "step": 9634 + }, + { + "epoch": 0.741096838704715, + "learning_rate": 0.0004694176021569218, + "loss": 1.144, + "step": 9635 + }, + { + "epoch": 0.7411737558649335, + "learning_rate": 0.0004691542642263928, + "loss": 0.8916, + "step": 9636 + }, + { + "epoch": 0.7412506730251519, + "learning_rate": 0.000468890986488021, + "loss": 1.2286, + "step": 9637 + }, + { + "epoch": 0.7413275901853703, + "learning_rate": 0.00046862776895717907, + "loss": 1.212, + "step": 9638 + }, + { + "epoch": 0.7414045073455888, + "learning_rate": 0.00046836461164923677, + "loss": 1.0431, + "step": 9639 + }, + { + "epoch": 0.7414814245058072, + "learning_rate": 0.0004681015145795597, + "loss": 1.1437, + "step": 9640 + }, + { + "epoch": 0.7415583416660257, + "learning_rate": 0.00046783847776351033, + "loss": 1.2865, + "step": 9641 + }, + { + "epoch": 0.7416352588262441, + "learning_rate": 0.0004675755012164483, + "loss": 1.388, + "step": 9642 + }, + { + "epoch": 0.7417121759864626, + "learning_rate": 0.0004673125849537288, + "loss": 1.1288, + "step": 9643 + }, + { + "epoch": 0.7417890931466811, + "learning_rate": 0.0004670497289907035, + "loss": 0.9553, + "step": 9644 + }, + { + "epoch": 0.7418660103068995, + "learning_rate": 0.0004667869333427212, + "loss": 1.2675, + "step": 9645 + }, + { + "epoch": 0.7419429274671179, + "learning_rate": 0.00046652419802512674, + "loss": 1.1955, + "step": 9646 + }, + { + "epoch": 0.7420198446273364, + "learning_rate": 0.00046626152305326133, + "loss": 1.226, + "step": 9647 + }, + { + "epoch": 0.7420967617875548, + "learning_rate": 0.00046599890844246293, + "loss": 0.531, + "step": 9648 + }, + { + "epoch": 0.7421736789477732, + "learning_rate": 0.0004657363542080658, + "loss": 0.9925, + "step": 9649 + }, + { + "epoch": 0.7422505961079917, + "learning_rate": 0.00046547386036540034, + "loss": 1.5166, + "step": 9650 + }, + { + "epoch": 0.7423275132682101, + "learning_rate": 0.00046521142692979494, + "loss": 1.3091, + "step": 9651 + }, + { + "epoch": 0.7424044304284286, + "learning_rate": 0.0004649490539165727, + "loss": 1.1894, + "step": 9652 + }, + { + "epoch": 0.742481347588647, + "learning_rate": 0.0004646867413410539, + "loss": 0.5401, + "step": 9653 + }, + { + "epoch": 0.7425582647488654, + "learning_rate": 0.0004644244892185553, + "loss": 1.1338, + "step": 9654 + }, + { + "epoch": 0.742635181909084, + "learning_rate": 0.0004641622975643895, + "loss": 1.0326, + "step": 9655 + }, + { + "epoch": 0.7427120990693024, + "learning_rate": 0.00046390016639386706, + "loss": 1.4126, + "step": 9656 + }, + { + "epoch": 0.7427890162295208, + "learning_rate": 0.00046363809572229434, + "loss": 1.1692, + "step": 9657 + }, + { + "epoch": 0.7428659333897393, + "learning_rate": 0.00046337608556497236, + "loss": 1.0901, + "step": 9658 + }, + { + "epoch": 0.7429428505499577, + "learning_rate": 0.0004631141359372014, + "loss": 1.406, + "step": 9659 + }, + { + "epoch": 0.7430197677101762, + "learning_rate": 0.0004628522468542763, + "loss": 0.714, + "step": 9660 + }, + { + "epoch": 0.7430966848703946, + "learning_rate": 0.00046259041833149024, + "loss": 1.1301, + "step": 9661 + }, + { + "epoch": 0.743173602030613, + "learning_rate": 0.00046232865038413036, + "loss": 0.8453, + "step": 9662 + }, + { + "epoch": 0.7432505191908315, + "learning_rate": 0.0004620669430274816, + "loss": 1.146, + "step": 9663 + }, + { + "epoch": 0.7433274363510499, + "learning_rate": 0.00046180529627682623, + "loss": 1.0231, + "step": 9664 + }, + { + "epoch": 0.7434043535112683, + "learning_rate": 0.0004615437101474415, + "loss": 0.9878, + "step": 9665 + }, + { + "epoch": 0.7434812706714868, + "learning_rate": 0.00046128218465460176, + "loss": 1.4212, + "step": 9666 + }, + { + "epoch": 0.7435581878317052, + "learning_rate": 0.0004610207198135779, + "loss": 0.8955, + "step": 9667 + }, + { + "epoch": 0.7436351049919236, + "learning_rate": 0.0004607593156396371, + "loss": 1.2782, + "step": 9668 + }, + { + "epoch": 0.7437120221521422, + "learning_rate": 0.0004604979721480422, + "loss": 1.2521, + "step": 9669 + }, + { + "epoch": 0.7437889393123606, + "learning_rate": 0.00046023668935405516, + "loss": 1.1578, + "step": 9670 + }, + { + "epoch": 0.7438658564725791, + "learning_rate": 0.00045997546727293047, + "loss": 0.9417, + "step": 9671 + }, + { + "epoch": 0.7439427736327975, + "learning_rate": 0.0004597143059199227, + "loss": 0.9857, + "step": 9672 + }, + { + "epoch": 0.7440196907930159, + "learning_rate": 0.00045945320531028057, + "loss": 1.1483, + "step": 9673 + }, + { + "epoch": 0.7440966079532344, + "learning_rate": 0.00045919216545925023, + "loss": 1.0915, + "step": 9674 + }, + { + "epoch": 0.7441735251134528, + "learning_rate": 0.0004589311863820742, + "loss": 1.2431, + "step": 9675 + }, + { + "epoch": 0.7442504422736712, + "learning_rate": 0.00045867026809399107, + "loss": 1.2593, + "step": 9676 + }, + { + "epoch": 0.7443273594338897, + "learning_rate": 0.00045840941061023587, + "loss": 1.0995, + "step": 9677 + }, + { + "epoch": 0.7444042765941081, + "learning_rate": 0.00045814861394604105, + "loss": 0.8031, + "step": 9678 + }, + { + "epoch": 0.7444811937543266, + "learning_rate": 0.0004578878781166346, + "loss": 1.0974, + "step": 9679 + }, + { + "epoch": 0.744558110914545, + "learning_rate": 0.00045762720313724086, + "loss": 0.8156, + "step": 9680 + }, + { + "epoch": 0.7446350280747634, + "learning_rate": 0.00045736658902308114, + "loss": 1.3662, + "step": 9681 + }, + { + "epoch": 0.744711945234982, + "learning_rate": 0.0004571060357893724, + "loss": 1.1884, + "step": 9682 + }, + { + "epoch": 0.7447888623952004, + "learning_rate": 0.00045684554345133005, + "loss": 1.1542, + "step": 9683 + }, + { + "epoch": 0.7448657795554188, + "learning_rate": 0.0004565851120241632, + "loss": 1.0133, + "step": 9684 + }, + { + "epoch": 0.7449426967156373, + "learning_rate": 0.00045632474152307854, + "loss": 0.6093, + "step": 9685 + }, + { + "epoch": 0.7450196138758557, + "learning_rate": 0.0004560644319632802, + "loss": 1.0068, + "step": 9686 + }, + { + "epoch": 0.7450965310360741, + "learning_rate": 0.00045580418335996786, + "loss": 1.1714, + "step": 9687 + }, + { + "epoch": 0.7451734481962926, + "learning_rate": 0.00045554399572833744, + "loss": 1.1873, + "step": 9688 + }, + { + "epoch": 0.745250365356511, + "learning_rate": 0.00045528386908358153, + "loss": 1.0161, + "step": 9689 + }, + { + "epoch": 0.7453272825167295, + "learning_rate": 0.0004550238034408889, + "loss": 1.0107, + "step": 9690 + }, + { + "epoch": 0.7454041996769479, + "learning_rate": 0.000454763798815446, + "loss": 0.8771, + "step": 9691 + }, + { + "epoch": 0.7454811168371663, + "learning_rate": 0.0004545038552224345, + "loss": 1.2641, + "step": 9692 + }, + { + "epoch": 0.7455580339973849, + "learning_rate": 0.00045424397267703154, + "loss": 1.0957, + "step": 9693 + }, + { + "epoch": 0.7456349511576033, + "learning_rate": 0.0004539841511944134, + "loss": 1.4517, + "step": 9694 + }, + { + "epoch": 0.7457118683178217, + "learning_rate": 0.0004537243907897509, + "loss": 1.0689, + "step": 9695 + }, + { + "epoch": 0.7457887854780402, + "learning_rate": 0.0004534646914782115, + "loss": 1.4997, + "step": 9696 + }, + { + "epoch": 0.7458657026382586, + "learning_rate": 0.0004532050532749595, + "loss": 1.5104, + "step": 9697 + }, + { + "epoch": 0.7459426197984771, + "learning_rate": 0.0004529454761951549, + "loss": 0.7274, + "step": 9698 + }, + { + "epoch": 0.7460195369586955, + "learning_rate": 0.00045268596025395554, + "loss": 1.0045, + "step": 9699 + }, + { + "epoch": 0.7460964541189139, + "learning_rate": 0.0004524265054665146, + "loss": 1.2495, + "step": 9700 + }, + { + "epoch": 0.7461733712791324, + "learning_rate": 0.0004521671118479818, + "loss": 1.1395, + "step": 9701 + }, + { + "epoch": 0.7462502884393508, + "learning_rate": 0.00045190777941350307, + "loss": 1.5124, + "step": 9702 + }, + { + "epoch": 0.7463272055995692, + "learning_rate": 0.00045164850817822164, + "loss": 1.4483, + "step": 9703 + }, + { + "epoch": 0.7464041227597877, + "learning_rate": 0.0004513892981572759, + "loss": 1.1922, + "step": 9704 + }, + { + "epoch": 0.7464810399200061, + "learning_rate": 0.0004511301493658027, + "loss": 0.8614, + "step": 9705 + }, + { + "epoch": 0.7465579570802245, + "learning_rate": 0.0004508710618189329, + "loss": 0.9824, + "step": 9706 + }, + { + "epoch": 0.7466348742404431, + "learning_rate": 0.00045061203553179455, + "loss": 0.982, + "step": 9707 + }, + { + "epoch": 0.7467117914006615, + "learning_rate": 0.00045035307051951353, + "loss": 1.5379, + "step": 9708 + }, + { + "epoch": 0.74678870856088, + "learning_rate": 0.00045009416679721076, + "loss": 1.152, + "step": 9709 + }, + { + "epoch": 0.7468656257210984, + "learning_rate": 0.0004498353243800037, + "loss": 1.3142, + "step": 9710 + }, + { + "epoch": 0.7469425428813168, + "learning_rate": 0.0004495765432830065, + "loss": 1.0859, + "step": 9711 + }, + { + "epoch": 0.7470194600415353, + "learning_rate": 0.00044931782352132917, + "loss": 1.1589, + "step": 9712 + }, + { + "epoch": 0.7470963772017537, + "learning_rate": 0.0004490591651100797, + "loss": 1.077, + "step": 9713 + }, + { + "epoch": 0.7471732943619721, + "learning_rate": 0.00044880056806436124, + "loss": 1.435, + "step": 9714 + }, + { + "epoch": 0.7472502115221906, + "learning_rate": 0.0004485420323992722, + "loss": 1.2256, + "step": 9715 + }, + { + "epoch": 0.747327128682409, + "learning_rate": 0.0004482835581299102, + "loss": 1.1679, + "step": 9716 + }, + { + "epoch": 0.7474040458426275, + "learning_rate": 0.0004480251452713669, + "loss": 1.0517, + "step": 9717 + }, + { + "epoch": 0.7474809630028459, + "learning_rate": 0.0004477667938387325, + "loss": 1.2843, + "step": 9718 + }, + { + "epoch": 0.7475578801630643, + "learning_rate": 0.00044750850384709116, + "loss": 1.4952, + "step": 9719 + }, + { + "epoch": 0.7476347973232829, + "learning_rate": 0.0004472502753115248, + "loss": 1.2533, + "step": 9720 + }, + { + "epoch": 0.7477117144835013, + "learning_rate": 0.0004469921082471124, + "loss": 1.1568, + "step": 9721 + }, + { + "epoch": 0.7477886316437197, + "learning_rate": 0.0004467340026689283, + "loss": 1.26, + "step": 9722 + }, + { + "epoch": 0.7478655488039382, + "learning_rate": 0.00044647595859204336, + "loss": 0.9277, + "step": 9723 + }, + { + "epoch": 0.7479424659641566, + "learning_rate": 0.0004462179760315253, + "loss": 1.4459, + "step": 9724 + }, + { + "epoch": 0.7480193831243751, + "learning_rate": 0.00044596005500243736, + "loss": 1.577, + "step": 9725 + }, + { + "epoch": 0.7480963002845935, + "learning_rate": 0.00044570219551984063, + "loss": 1.0829, + "step": 9726 + }, + { + "epoch": 0.7481732174448119, + "learning_rate": 0.00044544439759879197, + "loss": 1.2658, + "step": 9727 + }, + { + "epoch": 0.7482501346050304, + "learning_rate": 0.0004451866612543431, + "loss": 1.3286, + "step": 9728 + }, + { + "epoch": 0.7483270517652488, + "learning_rate": 0.0004449289865015448, + "loss": 0.9882, + "step": 9729 + }, + { + "epoch": 0.7484039689254672, + "learning_rate": 0.00044467137335544264, + "loss": 1.5653, + "step": 9730 + }, + { + "epoch": 0.7484808860856857, + "learning_rate": 0.0004444138218310787, + "loss": 1.2296, + "step": 9731 + }, + { + "epoch": 0.7485578032459042, + "learning_rate": 0.0004441563319434919, + "loss": 0.8052, + "step": 9732 + }, + { + "epoch": 0.7486347204061226, + "learning_rate": 0.00044389890370771714, + "loss": 1.3293, + "step": 9733 + }, + { + "epoch": 0.7487116375663411, + "learning_rate": 0.00044364153713878564, + "loss": 1.0459, + "step": 9734 + }, + { + "epoch": 0.7487885547265595, + "learning_rate": 0.00044338423225172627, + "loss": 1.6434, + "step": 9735 + }, + { + "epoch": 0.748865471886778, + "learning_rate": 0.00044312698906156264, + "loss": 1.0665, + "step": 9736 + }, + { + "epoch": 0.7489423890469964, + "learning_rate": 0.0004428698075833156, + "loss": 1.1757, + "step": 9737 + }, + { + "epoch": 0.7490193062072148, + "learning_rate": 0.00044261268783200207, + "loss": 1.2172, + "step": 9738 + }, + { + "epoch": 0.7490962233674333, + "learning_rate": 0.0004423556298226353, + "loss": 0.7575, + "step": 9739 + }, + { + "epoch": 0.7491731405276517, + "learning_rate": 0.0004420986335702266, + "loss": 1.3136, + "step": 9740 + }, + { + "epoch": 0.7492500576878701, + "learning_rate": 0.00044184169908978064, + "loss": 1.3724, + "step": 9741 + }, + { + "epoch": 0.7493269748480886, + "learning_rate": 0.00044158482639630015, + "loss": 0.932, + "step": 9742 + }, + { + "epoch": 0.749403892008307, + "learning_rate": 0.00044132801550478523, + "loss": 1.5715, + "step": 9743 + }, + { + "epoch": 0.7494808091685256, + "learning_rate": 0.00044107126643023087, + "loss": 1.2626, + "step": 9744 + }, + { + "epoch": 0.749557726328744, + "learning_rate": 0.0004408145791876287, + "loss": 1.1827, + "step": 9745 + }, + { + "epoch": 0.7496346434889624, + "learning_rate": 0.00044055795379196734, + "loss": 1.2491, + "step": 9746 + }, + { + "epoch": 0.7497115606491809, + "learning_rate": 0.0004403013902582305, + "loss": 0.8887, + "step": 9747 + }, + { + "epoch": 0.7497884778093993, + "learning_rate": 0.0004400448886014005, + "loss": 1.1813, + "step": 9748 + }, + { + "epoch": 0.7498653949696177, + "learning_rate": 0.00043978844883645437, + "loss": 1.2962, + "step": 9749 + }, + { + "epoch": 0.7499423121298362, + "learning_rate": 0.0004395320709783649, + "loss": 1.2228, + "step": 9750 + }, + { + "epoch": 0.7500192292900546, + "learning_rate": 0.00043927575504210326, + "loss": 0.9294, + "step": 9751 + }, + { + "epoch": 0.750096146450273, + "learning_rate": 0.000439019501042636, + "loss": 0.7966, + "step": 9752 + }, + { + "epoch": 0.7501730636104915, + "learning_rate": 0.0004387633089949256, + "loss": 1.073, + "step": 9753 + }, + { + "epoch": 0.7502499807707099, + "learning_rate": 0.00043850717891393167, + "loss": 0.8838, + "step": 9754 + }, + { + "epoch": 0.7503268979309284, + "learning_rate": 0.00043825111081460935, + "loss": 1.0305, + "step": 9755 + }, + { + "epoch": 0.7504038150911468, + "learning_rate": 0.00043799510471191166, + "loss": 1.3044, + "step": 9756 + }, + { + "epoch": 0.7504807322513652, + "learning_rate": 0.00043773916062078656, + "loss": 0.8356, + "step": 9757 + }, + { + "epoch": 0.7505576494115838, + "learning_rate": 0.00043748327855617906, + "loss": 0.9336, + "step": 9758 + }, + { + "epoch": 0.7506345665718022, + "learning_rate": 0.00043722745853303017, + "loss": 1.0255, + "step": 9759 + }, + { + "epoch": 0.7507114837320206, + "learning_rate": 0.00043697170056627744, + "loss": 1.1667, + "step": 9760 + }, + { + "epoch": 0.7507884008922391, + "learning_rate": 0.00043671600467085466, + "loss": 1.3942, + "step": 9761 + }, + { + "epoch": 0.7508653180524575, + "learning_rate": 0.00043646037086169347, + "loss": 1.1702, + "step": 9762 + }, + { + "epoch": 0.750942235212676, + "learning_rate": 0.0004362047991537185, + "loss": 1.3565, + "step": 9763 + }, + { + "epoch": 0.7510191523728944, + "learning_rate": 0.0004359492895618546, + "loss": 1.3172, + "step": 9764 + }, + { + "epoch": 0.7510960695331128, + "learning_rate": 0.0004356938421010206, + "loss": 0.9548, + "step": 9765 + }, + { + "epoch": 0.7511729866933313, + "learning_rate": 0.00043543845678613226, + "loss": 1.3256, + "step": 9766 + }, + { + "epoch": 0.7512499038535497, + "learning_rate": 0.00043518313363210203, + "loss": 1.2936, + "step": 9767 + }, + { + "epoch": 0.7513268210137681, + "learning_rate": 0.00043492787265383823, + "loss": 1.2388, + "step": 9768 + }, + { + "epoch": 0.7514037381739866, + "learning_rate": 0.0004346726738662455, + "loss": 1.1373, + "step": 9769 + }, + { + "epoch": 0.751480655334205, + "learning_rate": 0.0004344175372842261, + "loss": 1.0863, + "step": 9770 + }, + { + "epoch": 0.7515575724944235, + "learning_rate": 0.00043416246292267717, + "loss": 1.2035, + "step": 9771 + }, + { + "epoch": 0.751634489654642, + "learning_rate": 0.0004339074507964928, + "loss": 1.0746, + "step": 9772 + }, + { + "epoch": 0.7517114068148604, + "learning_rate": 0.0004336525009205635, + "loss": 1.2591, + "step": 9773 + }, + { + "epoch": 0.7517883239750789, + "learning_rate": 0.00043339761330977544, + "loss": 0.8002, + "step": 9774 + }, + { + "epoch": 0.7518652411352973, + "learning_rate": 0.00043314278797901333, + "loss": 0.9797, + "step": 9775 + }, + { + "epoch": 0.7519421582955157, + "learning_rate": 0.0004328880249431553, + "loss": 0.9351, + "step": 9776 + }, + { + "epoch": 0.7520190754557342, + "learning_rate": 0.00043263332421707694, + "loss": 0.9714, + "step": 9777 + }, + { + "epoch": 0.7520959926159526, + "learning_rate": 0.00043237868581565185, + "loss": 1.349, + "step": 9778 + }, + { + "epoch": 0.752172909776171, + "learning_rate": 0.00043212410975374785, + "loss": 1.3253, + "step": 9779 + }, + { + "epoch": 0.7522498269363895, + "learning_rate": 0.00043186959604622993, + "loss": 1.2438, + "step": 9780 + }, + { + "epoch": 0.7523267440966079, + "learning_rate": 0.0004316151447079595, + "loss": 1.2926, + "step": 9781 + }, + { + "epoch": 0.7524036612568265, + "learning_rate": 0.00043136075575379374, + "loss": 1.4467, + "step": 9782 + }, + { + "epoch": 0.7524805784170449, + "learning_rate": 0.00043110642919858764, + "loss": 1.8067, + "step": 9783 + }, + { + "epoch": 0.7525574955772633, + "learning_rate": 0.00043085216505719124, + "loss": 1.134, + "step": 9784 + }, + { + "epoch": 0.7526344127374818, + "learning_rate": 0.000430597963344451, + "loss": 1.2367, + "step": 9785 + }, + { + "epoch": 0.7527113298977002, + "learning_rate": 0.00043034382407521017, + "loss": 1.1981, + "step": 9786 + }, + { + "epoch": 0.7527882470579186, + "learning_rate": 0.00043008974726430836, + "loss": 1.1371, + "step": 9787 + }, + { + "epoch": 0.7528651642181371, + "learning_rate": 0.00042983573292658113, + "loss": 1.1948, + "step": 9788 + }, + { + "epoch": 0.7529420813783555, + "learning_rate": 0.00042958178107686086, + "loss": 0.8531, + "step": 9789 + }, + { + "epoch": 0.7530189985385739, + "learning_rate": 0.00042932789172997585, + "loss": 1.3793, + "step": 9790 + }, + { + "epoch": 0.7530959156987924, + "learning_rate": 0.00042907406490075067, + "loss": 1.1198, + "step": 9791 + }, + { + "epoch": 0.7531728328590108, + "learning_rate": 0.0004288203006040074, + "loss": 1.5074, + "step": 9792 + }, + { + "epoch": 0.7532497500192293, + "learning_rate": 0.0004285665988545631, + "loss": 1.0186, + "step": 9793 + }, + { + "epoch": 0.7533266671794477, + "learning_rate": 0.00042831295966723167, + "loss": 1.3088, + "step": 9794 + }, + { + "epoch": 0.7534035843396661, + "learning_rate": 0.0004280593830568235, + "loss": 1.0121, + "step": 9795 + }, + { + "epoch": 0.7534805014998847, + "learning_rate": 0.0004278058690381445, + "loss": 1.3308, + "step": 9796 + }, + { + "epoch": 0.7535574186601031, + "learning_rate": 0.00042755241762599877, + "loss": 1.2559, + "step": 9797 + }, + { + "epoch": 0.7536343358203215, + "learning_rate": 0.00042729902883518567, + "loss": 1.0505, + "step": 9798 + }, + { + "epoch": 0.75371125298054, + "learning_rate": 0.00042704570268049905, + "loss": 1.103, + "step": 9799 + }, + { + "epoch": 0.7537881701407584, + "learning_rate": 0.00042679243917673264, + "loss": 0.8859, + "step": 9800 + }, + { + "epoch": 0.7538650873009769, + "learning_rate": 0.00042653923833867424, + "loss": 1.5809, + "step": 9801 + }, + { + "epoch": 0.7539420044611953, + "learning_rate": 0.0004262861001811083, + "loss": 1.2575, + "step": 9802 + }, + { + "epoch": 0.7540189216214137, + "learning_rate": 0.000426033024718816, + "loss": 1.1196, + "step": 9803 + }, + { + "epoch": 0.7540958387816322, + "learning_rate": 0.00042578001196657417, + "loss": 1.2707, + "step": 9804 + }, + { + "epoch": 0.7541727559418506, + "learning_rate": 0.0004255270619391575, + "loss": 1.1006, + "step": 9805 + }, + { + "epoch": 0.754249673102069, + "learning_rate": 0.00042527417465133556, + "loss": 1.1542, + "step": 9806 + }, + { + "epoch": 0.7543265902622875, + "learning_rate": 0.00042502135011787445, + "loss": 1.5866, + "step": 9807 + }, + { + "epoch": 0.754403507422506, + "learning_rate": 0.0004247685883535371, + "loss": 1.1532, + "step": 9808 + }, + { + "epoch": 0.7544804245827244, + "learning_rate": 0.0004245158893730825, + "loss": 1.181, + "step": 9809 + }, + { + "epoch": 0.7545573417429429, + "learning_rate": 0.00042426325319126534, + "loss": 1.0649, + "step": 9810 + }, + { + "epoch": 0.7546342589031613, + "learning_rate": 0.00042401067982283903, + "loss": 1.1536, + "step": 9811 + }, + { + "epoch": 0.7547111760633798, + "learning_rate": 0.0004237581692825494, + "loss": 1.0421, + "step": 9812 + }, + { + "epoch": 0.7547880932235982, + "learning_rate": 0.0004235057215851425, + "loss": 1.1205, + "step": 9813 + }, + { + "epoch": 0.7548650103838166, + "learning_rate": 0.0004232533367453584, + "loss": 0.8551, + "step": 9814 + }, + { + "epoch": 0.7549419275440351, + "learning_rate": 0.0004230010147779341, + "loss": 0.9364, + "step": 9815 + }, + { + "epoch": 0.7550188447042535, + "learning_rate": 0.0004227487556976028, + "loss": 1.0157, + "step": 9816 + }, + { + "epoch": 0.7550957618644719, + "learning_rate": 0.0004224965595190946, + "loss": 1.1261, + "step": 9817 + }, + { + "epoch": 0.7551726790246904, + "learning_rate": 0.00042224442625713474, + "loss": 1.1662, + "step": 9818 + }, + { + "epoch": 0.7552495961849088, + "learning_rate": 0.00042199235592644635, + "loss": 1.5206, + "step": 9819 + }, + { + "epoch": 0.7553265133451273, + "learning_rate": 0.00042174034854174775, + "loss": 0.9527, + "step": 9820 + }, + { + "epoch": 0.7554034305053458, + "learning_rate": 0.00042148840411775415, + "loss": 0.7728, + "step": 9821 + }, + { + "epoch": 0.7554803476655642, + "learning_rate": 0.00042123652266917657, + "loss": 0.9655, + "step": 9822 + }, + { + "epoch": 0.7555572648257827, + "learning_rate": 0.0004209847042107221, + "loss": 0.8613, + "step": 9823 + }, + { + "epoch": 0.7556341819860011, + "learning_rate": 0.0004207329487570963, + "loss": 1.3195, + "step": 9824 + }, + { + "epoch": 0.7557110991462195, + "learning_rate": 0.0004204812563229978, + "loss": 1.0556, + "step": 9825 + }, + { + "epoch": 0.755788016306438, + "learning_rate": 0.0004202296269231235, + "loss": 1.2988, + "step": 9826 + }, + { + "epoch": 0.7558649334666564, + "learning_rate": 0.00041997806057216707, + "loss": 1.2244, + "step": 9827 + }, + { + "epoch": 0.7559418506268748, + "learning_rate": 0.0004197265572848172, + "loss": 1.0913, + "step": 9828 + }, + { + "epoch": 0.7560187677870933, + "learning_rate": 0.00041947511707575934, + "loss": 0.9811, + "step": 9829 + }, + { + "epoch": 0.7560956849473117, + "learning_rate": 0.0004192237399596757, + "loss": 0.7383, + "step": 9830 + }, + { + "epoch": 0.7561726021075302, + "learning_rate": 0.00041897242595124366, + "loss": 1.1587, + "step": 9831 + }, + { + "epoch": 0.7562495192677486, + "learning_rate": 0.00041872117506513886, + "loss": 0.8737, + "step": 9832 + }, + { + "epoch": 0.756326436427967, + "learning_rate": 0.00041846998731603197, + "loss": 0.924, + "step": 9833 + }, + { + "epoch": 0.7564033535881856, + "learning_rate": 0.00041821886271858867, + "loss": 1.2262, + "step": 9834 + }, + { + "epoch": 0.756480270748404, + "learning_rate": 0.00041796780128747386, + "loss": 1.234, + "step": 9835 + }, + { + "epoch": 0.7565571879086224, + "learning_rate": 0.00041771680303734685, + "loss": 0.8367, + "step": 9836 + }, + { + "epoch": 0.7566341050688409, + "learning_rate": 0.00041746586798286354, + "loss": 0.9922, + "step": 9837 + }, + { + "epoch": 0.7567110222290593, + "learning_rate": 0.00041721499613867654, + "loss": 1.1963, + "step": 9838 + }, + { + "epoch": 0.7567879393892778, + "learning_rate": 0.00041696418751943366, + "loss": 1.0821, + "step": 9839 + }, + { + "epoch": 0.7568648565494962, + "learning_rate": 0.0004167134421397812, + "loss": 1.1669, + "step": 9840 + }, + { + "epoch": 0.7569417737097146, + "learning_rate": 0.0004164627600143599, + "loss": 1.3612, + "step": 9841 + }, + { + "epoch": 0.7570186908699331, + "learning_rate": 0.00041621214115780724, + "loss": 1.1686, + "step": 9842 + }, + { + "epoch": 0.7570956080301515, + "learning_rate": 0.0004159615855847571, + "loss": 0.9845, + "step": 9843 + }, + { + "epoch": 0.7571725251903699, + "learning_rate": 0.00041571109330983976, + "loss": 1.3389, + "step": 9844 + }, + { + "epoch": 0.7572494423505884, + "learning_rate": 0.0004154606643476811, + "loss": 1.1986, + "step": 9845 + }, + { + "epoch": 0.7573263595108068, + "learning_rate": 0.0004152102987129057, + "loss": 1.526, + "step": 9846 + }, + { + "epoch": 0.7574032766710254, + "learning_rate": 0.00041495999642013044, + "loss": 0.8463, + "step": 9847 + }, + { + "epoch": 0.7574801938312438, + "learning_rate": 0.0004147097574839723, + "loss": 1.1937, + "step": 9848 + }, + { + "epoch": 0.7575571109914622, + "learning_rate": 0.00041445958191904236, + "loss": 1.1596, + "step": 9849 + }, + { + "epoch": 0.7576340281516807, + "learning_rate": 0.0004142094697399489, + "loss": 0.7386, + "step": 9850 + }, + { + "epoch": 0.7577109453118991, + "learning_rate": 0.00041395942096129613, + "loss": 1.2507, + "step": 9851 + }, + { + "epoch": 0.7577878624721175, + "learning_rate": 0.00041370943559768445, + "loss": 1.2967, + "step": 9852 + }, + { + "epoch": 0.757864779632336, + "learning_rate": 0.0004134595136637106, + "loss": 1.1242, + "step": 9853 + }, + { + "epoch": 0.7579416967925544, + "learning_rate": 0.00041320965517396846, + "loss": 1.9654, + "step": 9854 + }, + { + "epoch": 0.7580186139527728, + "learning_rate": 0.00041295986014304773, + "loss": 1.4288, + "step": 9855 + }, + { + "epoch": 0.7580955311129913, + "learning_rate": 0.0004127101285855326, + "loss": 1.4664, + "step": 9856 + }, + { + "epoch": 0.7581724482732097, + "learning_rate": 0.00041246046051600675, + "loss": 1.5966, + "step": 9857 + }, + { + "epoch": 0.7582493654334282, + "learning_rate": 0.00041221085594904763, + "loss": 1.1626, + "step": 9858 + }, + { + "epoch": 0.7583262825936467, + "learning_rate": 0.00041196131489923117, + "loss": 0.9965, + "step": 9859 + }, + { + "epoch": 0.758403199753865, + "learning_rate": 0.00041171183738112694, + "loss": 1.3445, + "step": 9860 + }, + { + "epoch": 0.7584801169140836, + "learning_rate": 0.0004114624234093023, + "loss": 1.1673, + "step": 9861 + }, + { + "epoch": 0.758557034074302, + "learning_rate": 0.00041121307299832175, + "loss": 1.1812, + "step": 9862 + }, + { + "epoch": 0.7586339512345204, + "learning_rate": 0.0004109637861627444, + "loss": 1.6752, + "step": 9863 + }, + { + "epoch": 0.7587108683947389, + "learning_rate": 0.0004107145629171266, + "loss": 1.0776, + "step": 9864 + }, + { + "epoch": 0.7587877855549573, + "learning_rate": 0.00041046540327602075, + "loss": 1.1185, + "step": 9865 + }, + { + "epoch": 0.7588647027151758, + "learning_rate": 0.00041021630725397495, + "loss": 0.8098, + "step": 9866 + }, + { + "epoch": 0.7589416198753942, + "learning_rate": 0.00040996727486553535, + "loss": 1.1833, + "step": 9867 + }, + { + "epoch": 0.7590185370356126, + "learning_rate": 0.0004097183061252429, + "loss": 1.0936, + "step": 9868 + }, + { + "epoch": 0.7590954541958311, + "learning_rate": 0.000409469401047634, + "loss": 1.0916, + "step": 9869 + }, + { + "epoch": 0.7591723713560495, + "learning_rate": 0.0004092205596472436, + "loss": 1.0741, + "step": 9870 + }, + { + "epoch": 0.7592492885162679, + "learning_rate": 0.0004089717819386016, + "loss": 1.1404, + "step": 9871 + }, + { + "epoch": 0.7593262056764865, + "learning_rate": 0.00040872306793623416, + "loss": 1.2892, + "step": 9872 + }, + { + "epoch": 0.7594031228367049, + "learning_rate": 0.0004084744176546642, + "loss": 1.1543, + "step": 9873 + }, + { + "epoch": 0.7594800399969233, + "learning_rate": 0.0004082258311084106, + "loss": 1.1737, + "step": 9874 + }, + { + "epoch": 0.7595569571571418, + "learning_rate": 0.0004079773083119881, + "loss": 1.3703, + "step": 9875 + }, + { + "epoch": 0.7596338743173602, + "learning_rate": 0.00040772884927990904, + "loss": 1.1074, + "step": 9876 + }, + { + "epoch": 0.7597107914775787, + "learning_rate": 0.0004074804540266807, + "loss": 0.8953, + "step": 9877 + }, + { + "epoch": 0.7597877086377971, + "learning_rate": 0.0004072321225668075, + "loss": 1.2146, + "step": 9878 + }, + { + "epoch": 0.7598646257980155, + "learning_rate": 0.0004069838549147892, + "loss": 0.986, + "step": 9879 + }, + { + "epoch": 0.759941542958234, + "learning_rate": 0.0004067356510851226, + "loss": 1.4233, + "step": 9880 + }, + { + "epoch": 0.7600184601184524, + "learning_rate": 0.0004064875110923014, + "loss": 0.9556, + "step": 9881 + }, + { + "epoch": 0.7600953772786708, + "learning_rate": 0.00040623943495081356, + "loss": 1.2152, + "step": 9882 + }, + { + "epoch": 0.7601722944388893, + "learning_rate": 0.0004059914226751447, + "loss": 1.191, + "step": 9883 + }, + { + "epoch": 0.7602492115991077, + "learning_rate": 0.00040574347427977726, + "loss": 0.8316, + "step": 9884 + }, + { + "epoch": 0.7603261287593263, + "learning_rate": 0.0004054955897791888, + "loss": 1.2691, + "step": 9885 + }, + { + "epoch": 0.7604030459195447, + "learning_rate": 0.00040524776918785335, + "loss": 1.0548, + "step": 9886 + }, + { + "epoch": 0.7604799630797631, + "learning_rate": 0.0004050000125202417, + "loss": 0.7575, + "step": 9887 + }, + { + "epoch": 0.7605568802399816, + "learning_rate": 0.00040475231979082, + "loss": 0.9693, + "step": 9888 + }, + { + "epoch": 0.7606337974002, + "learning_rate": 0.00040450469101405227, + "loss": 1.1548, + "step": 9889 + }, + { + "epoch": 0.7607107145604184, + "learning_rate": 0.00040425712620439767, + "loss": 1.2662, + "step": 9890 + }, + { + "epoch": 0.7607876317206369, + "learning_rate": 0.0004040096253763106, + "loss": 1.2688, + "step": 9891 + }, + { + "epoch": 0.7608645488808553, + "learning_rate": 0.00040376218854424424, + "loss": 1.1372, + "step": 9892 + }, + { + "epoch": 0.7609414660410737, + "learning_rate": 0.0004035148157226463, + "loss": 0.8423, + "step": 9893 + }, + { + "epoch": 0.7610183832012922, + "learning_rate": 0.00040326750692596064, + "loss": 1.0708, + "step": 9894 + }, + { + "epoch": 0.7610953003615106, + "learning_rate": 0.0004030202621686285, + "loss": 1.1025, + "step": 9895 + }, + { + "epoch": 0.7611722175217291, + "learning_rate": 0.000402773081465086, + "loss": 1.1715, + "step": 9896 + }, + { + "epoch": 0.7612491346819475, + "learning_rate": 0.00040252596482976706, + "loss": 1.245, + "step": 9897 + }, + { + "epoch": 0.761326051842166, + "learning_rate": 0.000402278912277101, + "loss": 1.3009, + "step": 9898 + }, + { + "epoch": 0.7614029690023845, + "learning_rate": 0.00040203192382151317, + "loss": 1.3432, + "step": 9899 + }, + { + "epoch": 0.7614798861626029, + "learning_rate": 0.00040178499947742566, + "loss": 1.1632, + "step": 9900 + }, + { + "epoch": 0.7615568033228213, + "learning_rate": 0.00040153813925925653, + "loss": 1.2224, + "step": 9901 + }, + { + "epoch": 0.7616337204830398, + "learning_rate": 0.0004012913431814196, + "loss": 1.1971, + "step": 9902 + }, + { + "epoch": 0.7617106376432582, + "learning_rate": 0.0004010446112583274, + "loss": 1.3722, + "step": 9903 + }, + { + "epoch": 0.7617875548034767, + "learning_rate": 0.00040079794350438453, + "loss": 0.9624, + "step": 9904 + }, + { + "epoch": 0.7618644719636951, + "learning_rate": 0.0004005513399339958, + "loss": 1.1057, + "step": 9905 + }, + { + "epoch": 0.7619413891239135, + "learning_rate": 0.00040030480056156017, + "loss": 1.4653, + "step": 9906 + }, + { + "epoch": 0.762018306284132, + "learning_rate": 0.0004000583254014733, + "loss": 1.0507, + "step": 9907 + }, + { + "epoch": 0.7620952234443504, + "learning_rate": 0.0003998119144681272, + "loss": 1.3143, + "step": 9908 + }, + { + "epoch": 0.7621721406045688, + "learning_rate": 0.00039956556777591, + "loss": 1.2349, + "step": 9909 + }, + { + "epoch": 0.7622490577647874, + "learning_rate": 0.0003993192853392056, + "loss": 1.2431, + "step": 9910 + }, + { + "epoch": 0.7623259749250058, + "learning_rate": 0.00039907306717239574, + "loss": 1.3292, + "step": 9911 + }, + { + "epoch": 0.7624028920852242, + "learning_rate": 0.0003988269132898573, + "loss": 1.0823, + "step": 9912 + }, + { + "epoch": 0.7624798092454427, + "learning_rate": 0.00039858082370596205, + "loss": 1.2541, + "step": 9913 + }, + { + "epoch": 0.7625567264056611, + "learning_rate": 0.00039833479843508093, + "loss": 0.9036, + "step": 9914 + }, + { + "epoch": 0.7626336435658796, + "learning_rate": 0.00039808883749157853, + "loss": 1.2089, + "step": 9915 + }, + { + "epoch": 0.762710560726098, + "learning_rate": 0.00039784294088981824, + "loss": 1.0709, + "step": 9916 + }, + { + "epoch": 0.7627874778863164, + "learning_rate": 0.0003975971086441568, + "loss": 1.2823, + "step": 9917 + }, + { + "epoch": 0.7628643950465349, + "learning_rate": 0.0003973513407689486, + "loss": 1.1153, + "step": 9918 + }, + { + "epoch": 0.7629413122067533, + "learning_rate": 0.00039710563727854523, + "loss": 1.0983, + "step": 9919 + }, + { + "epoch": 0.7630182293669717, + "learning_rate": 0.0003968599981872932, + "loss": 1.4162, + "step": 9920 + }, + { + "epoch": 0.7630951465271902, + "learning_rate": 0.0003966144235095354, + "loss": 1.1128, + "step": 9921 + }, + { + "epoch": 0.7631720636874086, + "learning_rate": 0.00039636891325961146, + "loss": 1.2088, + "step": 9922 + }, + { + "epoch": 0.7632489808476272, + "learning_rate": 0.00039612346745185626, + "loss": 1.3528, + "step": 9923 + }, + { + "epoch": 0.7633258980078456, + "learning_rate": 0.0003958780861006028, + "loss": 0.713, + "step": 9924 + }, + { + "epoch": 0.763402815168064, + "learning_rate": 0.0003956327692201789, + "loss": 1.2216, + "step": 9925 + }, + { + "epoch": 0.7634797323282825, + "learning_rate": 0.0003953875168249078, + "loss": 0.7214, + "step": 9926 + }, + { + "epoch": 0.7635566494885009, + "learning_rate": 0.00039514232892911113, + "loss": 0.9942, + "step": 9927 + }, + { + "epoch": 0.7636335666487193, + "learning_rate": 0.0003948972055471054, + "loss": 0.9765, + "step": 9928 + }, + { + "epoch": 0.7637104838089378, + "learning_rate": 0.00039465214669320354, + "loss": 1.1006, + "step": 9929 + }, + { + "epoch": 0.7637874009691562, + "learning_rate": 0.00039440715238171477, + "loss": 1.1188, + "step": 9930 + }, + { + "epoch": 0.7638643181293746, + "learning_rate": 0.00039416222262694464, + "loss": 1.2701, + "step": 9931 + }, + { + "epoch": 0.7639412352895931, + "learning_rate": 0.0003939173574431943, + "loss": 1.4445, + "step": 9932 + }, + { + "epoch": 0.7640181524498115, + "learning_rate": 0.00039367255684476265, + "loss": 1.0762, + "step": 9933 + }, + { + "epoch": 0.76409506961003, + "learning_rate": 0.0003934278208459435, + "loss": 1.1465, + "step": 9934 + }, + { + "epoch": 0.7641719867702484, + "learning_rate": 0.000393183149461027, + "loss": 0.9773, + "step": 9935 + }, + { + "epoch": 0.7642489039304668, + "learning_rate": 0.0003929385427043, + "loss": 1.2645, + "step": 9936 + }, + { + "epoch": 0.7643258210906854, + "learning_rate": 0.00039269400059004474, + "loss": 1.2645, + "step": 9937 + }, + { + "epoch": 0.7644027382509038, + "learning_rate": 0.0003924495231325416, + "loss": 0.9997, + "step": 9938 + }, + { + "epoch": 0.7644796554111222, + "learning_rate": 0.00039220511034606477, + "loss": 1.63, + "step": 9939 + }, + { + "epoch": 0.7645565725713407, + "learning_rate": 0.0003919607622448856, + "loss": 1.1389, + "step": 9940 + }, + { + "epoch": 0.7646334897315591, + "learning_rate": 0.0003917164788432726, + "loss": 1.1261, + "step": 9941 + }, + { + "epoch": 0.7647104068917776, + "learning_rate": 0.0003914722601554894, + "loss": 1.0346, + "step": 9942 + }, + { + "epoch": 0.764787324051996, + "learning_rate": 0.0003912281061957965, + "loss": 1.2263, + "step": 9943 + }, + { + "epoch": 0.7648642412122144, + "learning_rate": 0.0003909840169784498, + "loss": 1.0811, + "step": 9944 + }, + { + "epoch": 0.7649411583724329, + "learning_rate": 0.00039073999251770173, + "loss": 1.1438, + "step": 9945 + }, + { + "epoch": 0.7650180755326513, + "learning_rate": 0.0003904960328278019, + "loss": 1.187, + "step": 9946 + }, + { + "epoch": 0.7650949926928697, + "learning_rate": 0.00039025213792299513, + "loss": 1.1532, + "step": 9947 + }, + { + "epoch": 0.7651719098530883, + "learning_rate": 0.0003900083078175226, + "loss": 1.0514, + "step": 9948 + }, + { + "epoch": 0.7652488270133067, + "learning_rate": 0.0003897645425256216, + "loss": 0.8389, + "step": 9949 + }, + { + "epoch": 0.7653257441735252, + "learning_rate": 0.00038952084206152573, + "loss": 1.209, + "step": 9950 + }, + { + "epoch": 0.7654026613337436, + "learning_rate": 0.0003892772064394657, + "loss": 0.9865, + "step": 9951 + }, + { + "epoch": 0.765479578493962, + "learning_rate": 0.00038903363567366746, + "loss": 1.1961, + "step": 9952 + }, + { + "epoch": 0.7655564956541805, + "learning_rate": 0.0003887901297783523, + "loss": 1.4159, + "step": 9953 + }, + { + "epoch": 0.7656334128143989, + "learning_rate": 0.0003885466887677398, + "loss": 1.653, + "step": 9954 + }, + { + "epoch": 0.7657103299746173, + "learning_rate": 0.00038830331265604466, + "loss": 0.9447, + "step": 9955 + }, + { + "epoch": 0.7657872471348358, + "learning_rate": 0.0003880600014574776, + "loss": 1.2132, + "step": 9956 + }, + { + "epoch": 0.7658641642950542, + "learning_rate": 0.00038781675518624587, + "loss": 1.2148, + "step": 9957 + }, + { + "epoch": 0.7659410814552726, + "learning_rate": 0.00038757357385655293, + "loss": 1.2414, + "step": 9958 + }, + { + "epoch": 0.7660179986154911, + "learning_rate": 0.00038733045748259783, + "loss": 1.5285, + "step": 9959 + }, + { + "epoch": 0.7660949157757095, + "learning_rate": 0.00038708740607857744, + "loss": 1.4117, + "step": 9960 + }, + { + "epoch": 0.766171832935928, + "learning_rate": 0.00038684441965868336, + "loss": 1.2747, + "step": 9961 + }, + { + "epoch": 0.7662487500961465, + "learning_rate": 0.00038660149823710346, + "loss": 0.9704, + "step": 9962 + }, + { + "epoch": 0.7663256672563649, + "learning_rate": 0.0003863586418280226, + "loss": 1.1634, + "step": 9963 + }, + { + "epoch": 0.7664025844165834, + "learning_rate": 0.0003861158504456208, + "loss": 0.9502, + "step": 9964 + }, + { + "epoch": 0.7664795015768018, + "learning_rate": 0.0003858731241040762, + "loss": 1.151, + "step": 9965 + }, + { + "epoch": 0.7665564187370202, + "learning_rate": 0.0003856304628175606, + "loss": 1.2631, + "step": 9966 + }, + { + "epoch": 0.7666333358972387, + "learning_rate": 0.0003853878666002431, + "loss": 1.0425, + "step": 9967 + }, + { + "epoch": 0.7667102530574571, + "learning_rate": 0.00038514533546629026, + "loss": 0.8835, + "step": 9968 + }, + { + "epoch": 0.7667871702176756, + "learning_rate": 0.00038490286942986313, + "loss": 1.1304, + "step": 9969 + }, + { + "epoch": 0.766864087377894, + "learning_rate": 0.00038466046850511957, + "loss": 1.0838, + "step": 9970 + }, + { + "epoch": 0.7669410045381124, + "learning_rate": 0.00038441813270621367, + "loss": 0.9887, + "step": 9971 + }, + { + "epoch": 0.7670179216983309, + "learning_rate": 0.00038417586204729504, + "loss": 1.152, + "step": 9972 + }, + { + "epoch": 0.7670948388585493, + "learning_rate": 0.0003839336565425113, + "loss": 1.3359, + "step": 9973 + }, + { + "epoch": 0.7671717560187677, + "learning_rate": 0.0003836915162060049, + "loss": 1.1356, + "step": 9974 + }, + { + "epoch": 0.7672486731789863, + "learning_rate": 0.0003834494410519134, + "loss": 0.9708, + "step": 9975 + }, + { + "epoch": 0.7673255903392047, + "learning_rate": 0.00038320743109437314, + "loss": 1.2937, + "step": 9976 + }, + { + "epoch": 0.7674025074994231, + "learning_rate": 0.00038296548634751465, + "loss": 1.3575, + "step": 9977 + }, + { + "epoch": 0.7674794246596416, + "learning_rate": 0.00038272360682546576, + "loss": 1.1872, + "step": 9978 + }, + { + "epoch": 0.76755634181986, + "learning_rate": 0.0003824817925423498, + "loss": 1.1093, + "step": 9979 + }, + { + "epoch": 0.7676332589800785, + "learning_rate": 0.0003822400435122863, + "loss": 1.3454, + "step": 9980 + }, + { + "epoch": 0.7677101761402969, + "learning_rate": 0.0003819983597493919, + "loss": 1.2304, + "step": 9981 + }, + { + "epoch": 0.7677870933005153, + "learning_rate": 0.0003817567412677784, + "loss": 1.2618, + "step": 9982 + }, + { + "epoch": 0.7678640104607338, + "learning_rate": 0.0003815151880815542, + "loss": 0.9594, + "step": 9983 + }, + { + "epoch": 0.7679409276209522, + "learning_rate": 0.0003812737002048238, + "loss": 1.2799, + "step": 9984 + }, + { + "epoch": 0.7680178447811706, + "learning_rate": 0.0003810322776516881, + "loss": 1.0407, + "step": 9985 + }, + { + "epoch": 0.7680947619413891, + "learning_rate": 0.0003807909204362432, + "loss": 1.2365, + "step": 9986 + }, + { + "epoch": 0.7681716791016076, + "learning_rate": 0.0003805496285725839, + "loss": 1.1432, + "step": 9987 + }, + { + "epoch": 0.7682485962618261, + "learning_rate": 0.00038030840207479766, + "loss": 1.511, + "step": 9988 + }, + { + "epoch": 0.7683255134220445, + "learning_rate": 0.00038006724095697106, + "loss": 1.3152, + "step": 9989 + }, + { + "epoch": 0.7684024305822629, + "learning_rate": 0.00037982614523318544, + "loss": 1.0888, + "step": 9990 + }, + { + "epoch": 0.7684793477424814, + "learning_rate": 0.0003795851149175186, + "loss": 1.0369, + "step": 9991 + }, + { + "epoch": 0.7685562649026998, + "learning_rate": 0.00037934415002404456, + "loss": 1.7193, + "step": 9992 + }, + { + "epoch": 0.7686331820629182, + "learning_rate": 0.00037910325056683344, + "loss": 0.9679, + "step": 9993 + }, + { + "epoch": 0.7687100992231367, + "learning_rate": 0.0003788624165599513, + "loss": 0.9817, + "step": 9994 + }, + { + "epoch": 0.7687870163833551, + "learning_rate": 0.00037862164801746157, + "loss": 1.168, + "step": 9995 + }, + { + "epoch": 0.7688639335435735, + "learning_rate": 0.000378380944953423, + "loss": 1.3463, + "step": 9996 + }, + { + "epoch": 0.768940850703792, + "learning_rate": 0.00037814030738188905, + "loss": 1.5224, + "step": 9997 + }, + { + "epoch": 0.7690177678640104, + "learning_rate": 0.0003778997353169121, + "loss": 1.3013, + "step": 9998 + }, + { + "epoch": 0.769094685024229, + "learning_rate": 0.00037765922877253853, + "loss": 1.2762, + "step": 9999 + }, + { + "epoch": 0.7691716021844474, + "learning_rate": 0.00037741878776281323, + "loss": 1.1189, + "step": 10000 + }, + { + "epoch": 0.7692485193446658, + "learning_rate": 0.0003771784123017744, + "loss": 0.9666, + "step": 10001 + }, + { + "epoch": 0.7693254365048843, + "learning_rate": 0.00037693810240345784, + "loss": 1.1382, + "step": 10002 + }, + { + "epoch": 0.7694023536651027, + "learning_rate": 0.00037669785808189645, + "loss": 1.1876, + "step": 10003 + }, + { + "epoch": 0.7694792708253211, + "learning_rate": 0.00037645767935111753, + "loss": 1.2285, + "step": 10004 + }, + { + "epoch": 0.7695561879855396, + "learning_rate": 0.00037621756622514585, + "loss": 1.4077, + "step": 10005 + }, + { + "epoch": 0.769633105145758, + "learning_rate": 0.00037597751871800154, + "loss": 0.8036, + "step": 10006 + }, + { + "epoch": 0.7697100223059765, + "learning_rate": 0.0003757375368437011, + "loss": 1.0401, + "step": 10007 + }, + { + "epoch": 0.7697869394661949, + "learning_rate": 0.00037549762061625774, + "loss": 1.4036, + "step": 10008 + }, + { + "epoch": 0.7698638566264133, + "learning_rate": 0.0003752577700496808, + "loss": 1.0638, + "step": 10009 + }, + { + "epoch": 0.7699407737866318, + "learning_rate": 0.0003750179851579741, + "loss": 1.105, + "step": 10010 + }, + { + "epoch": 0.7700176909468502, + "learning_rate": 0.00037477826595513984, + "loss": 1.0871, + "step": 10011 + }, + { + "epoch": 0.7700946081070686, + "learning_rate": 0.00037453861245517544, + "loss": 0.9914, + "step": 10012 + }, + { + "epoch": 0.7701715252672872, + "learning_rate": 0.00037429902467207447, + "loss": 0.9106, + "step": 10013 + }, + { + "epoch": 0.7702484424275056, + "learning_rate": 0.00037405950261982645, + "loss": 1.1255, + "step": 10014 + }, + { + "epoch": 0.770325359587724, + "learning_rate": 0.0003738200463124177, + "loss": 1.3214, + "step": 10015 + }, + { + "epoch": 0.7704022767479425, + "learning_rate": 0.0003735806557638296, + "loss": 1.2187, + "step": 10016 + }, + { + "epoch": 0.7704791939081609, + "learning_rate": 0.00037334133098804124, + "loss": 1.4611, + "step": 10017 + }, + { + "epoch": 0.7705561110683794, + "learning_rate": 0.0003731020719990269, + "loss": 1.0869, + "step": 10018 + }, + { + "epoch": 0.7706330282285978, + "learning_rate": 0.0003728628788107569, + "loss": 1.4927, + "step": 10019 + }, + { + "epoch": 0.7707099453888162, + "learning_rate": 0.0003726237514371983, + "loss": 1.0439, + "step": 10020 + }, + { + "epoch": 0.7707868625490347, + "learning_rate": 0.00037238468989231316, + "loss": 0.9544, + "step": 10021 + }, + { + "epoch": 0.7708637797092531, + "learning_rate": 0.00037214569419006204, + "loss": 1.4451, + "step": 10022 + }, + { + "epoch": 0.7709406968694715, + "learning_rate": 0.0003719067643443988, + "loss": 1.0804, + "step": 10023 + }, + { + "epoch": 0.77101761402969, + "learning_rate": 0.00037166790036927475, + "loss": 1.1747, + "step": 10024 + }, + { + "epoch": 0.7710945311899084, + "learning_rate": 0.0003714291022786385, + "loss": 1.2461, + "step": 10025 + }, + { + "epoch": 0.771171448350127, + "learning_rate": 0.00037119037008643305, + "loss": 1.1186, + "step": 10026 + }, + { + "epoch": 0.7712483655103454, + "learning_rate": 0.0003709517038065984, + "loss": 0.9967, + "step": 10027 + }, + { + "epoch": 0.7713252826705638, + "learning_rate": 0.00037071310345307044, + "loss": 1.1569, + "step": 10028 + }, + { + "epoch": 0.7714021998307823, + "learning_rate": 0.00037047456903978066, + "loss": 1.2187, + "step": 10029 + }, + { + "epoch": 0.7714791169910007, + "learning_rate": 0.0003702361005806585, + "loss": 1.0911, + "step": 10030 + }, + { + "epoch": 0.7715560341512191, + "learning_rate": 0.0003699976980896284, + "loss": 1.1432, + "step": 10031 + }, + { + "epoch": 0.7716329513114376, + "learning_rate": 0.0003697593615806093, + "loss": 1.4127, + "step": 10032 + }, + { + "epoch": 0.771709868471656, + "learning_rate": 0.00036952109106751956, + "loss": 0.9756, + "step": 10033 + }, + { + "epoch": 0.7717867856318744, + "learning_rate": 0.00036928288656427145, + "loss": 0.9238, + "step": 10034 + }, + { + "epoch": 0.7718637027920929, + "learning_rate": 0.00036904474808477394, + "loss": 1.3014, + "step": 10035 + }, + { + "epoch": 0.7719406199523113, + "learning_rate": 0.0003688066756429323, + "loss": 0.9536, + "step": 10036 + }, + { + "epoch": 0.7720175371125299, + "learning_rate": 0.0003685686692526472, + "loss": 1.2614, + "step": 10037 + }, + { + "epoch": 0.7720944542727483, + "learning_rate": 0.00036833072892781724, + "loss": 1.1682, + "step": 10038 + }, + { + "epoch": 0.7721713714329667, + "learning_rate": 0.00036809285468233536, + "loss": 1.233, + "step": 10039 + }, + { + "epoch": 0.7722482885931852, + "learning_rate": 0.0003678550465300913, + "loss": 1.2303, + "step": 10040 + }, + { + "epoch": 0.7723252057534036, + "learning_rate": 0.0003676173044849711, + "loss": 1.0965, + "step": 10041 + }, + { + "epoch": 0.772402122913622, + "learning_rate": 0.00036737962856085645, + "loss": 1.1599, + "step": 10042 + }, + { + "epoch": 0.7724790400738405, + "learning_rate": 0.00036714201877162536, + "loss": 0.7442, + "step": 10043 + }, + { + "epoch": 0.7725559572340589, + "learning_rate": 0.00036690447513115337, + "loss": 1.0247, + "step": 10044 + }, + { + "epoch": 0.7726328743942774, + "learning_rate": 0.00036666699765330895, + "loss": 1.0928, + "step": 10045 + }, + { + "epoch": 0.7727097915544958, + "learning_rate": 0.0003664295863519602, + "loss": 1.065, + "step": 10046 + }, + { + "epoch": 0.7727867087147142, + "learning_rate": 0.00036619224124096925, + "loss": 1.0316, + "step": 10047 + }, + { + "epoch": 0.7728636258749327, + "learning_rate": 0.0003659549623341949, + "loss": 1.351, + "step": 10048 + }, + { + "epoch": 0.7729405430351511, + "learning_rate": 0.00036571774964549227, + "loss": 1.4493, + "step": 10049 + }, + { + "epoch": 0.7730174601953695, + "learning_rate": 0.00036548060318871225, + "loss": 1.2986, + "step": 10050 + }, + { + "epoch": 0.7730943773555881, + "learning_rate": 0.00036524352297770174, + "loss": 1.2326, + "step": 10051 + }, + { + "epoch": 0.7731712945158065, + "learning_rate": 0.00036500650902630507, + "loss": 1.1512, + "step": 10052 + }, + { + "epoch": 0.7732482116760249, + "learning_rate": 0.0003647695613483616, + "loss": 1.2492, + "step": 10053 + }, + { + "epoch": 0.7733251288362434, + "learning_rate": 0.00036453267995770563, + "loss": 1.3303, + "step": 10054 + }, + { + "epoch": 0.7734020459964618, + "learning_rate": 0.0003642958648681704, + "loss": 1.3181, + "step": 10055 + }, + { + "epoch": 0.7734789631566803, + "learning_rate": 0.00036405911609358274, + "loss": 1.6058, + "step": 10056 + }, + { + "epoch": 0.7735558803168987, + "learning_rate": 0.00036382243364776784, + "loss": 1.1234, + "step": 10057 + }, + { + "epoch": 0.7736327974771171, + "learning_rate": 0.0003635858175445449, + "loss": 0.9767, + "step": 10058 + }, + { + "epoch": 0.7737097146373356, + "learning_rate": 0.00036334926779772996, + "loss": 1.3137, + "step": 10059 + }, + { + "epoch": 0.773786631797554, + "learning_rate": 0.0003631127844211362, + "loss": 1.3275, + "step": 10060 + }, + { + "epoch": 0.7738635489577724, + "learning_rate": 0.0003628763674285719, + "loss": 1.0674, + "step": 10061 + }, + { + "epoch": 0.7739404661179909, + "learning_rate": 0.00036264001683384146, + "loss": 1.3303, + "step": 10062 + }, + { + "epoch": 0.7740173832782093, + "learning_rate": 0.00036240373265074586, + "loss": 1.3485, + "step": 10063 + }, + { + "epoch": 0.7740943004384279, + "learning_rate": 0.00036216751489308145, + "loss": 1.1655, + "step": 10064 + }, + { + "epoch": 0.7741712175986463, + "learning_rate": 0.00036193136357464203, + "loss": 1.159, + "step": 10065 + }, + { + "epoch": 0.7742481347588647, + "learning_rate": 0.0003616952787092168, + "loss": 1.4024, + "step": 10066 + }, + { + "epoch": 0.7743250519190832, + "learning_rate": 0.0003614592603105898, + "loss": 0.7212, + "step": 10067 + }, + { + "epoch": 0.7744019690793016, + "learning_rate": 0.0003612233083925436, + "loss": 1.4691, + "step": 10068 + }, + { + "epoch": 0.77447888623952, + "learning_rate": 0.000360987422968855, + "loss": 0.9413, + "step": 10069 + }, + { + "epoch": 0.7745558033997385, + "learning_rate": 0.0003607516040532981, + "loss": 1.1579, + "step": 10070 + }, + { + "epoch": 0.7746327205599569, + "learning_rate": 0.00036051585165964217, + "loss": 1.2264, + "step": 10071 + }, + { + "epoch": 0.7747096377201754, + "learning_rate": 0.0003602801658016527, + "loss": 1.645, + "step": 10072 + }, + { + "epoch": 0.7747865548803938, + "learning_rate": 0.00036004454649309276, + "loss": 1.1012, + "step": 10073 + }, + { + "epoch": 0.7748634720406122, + "learning_rate": 0.00035980899374771957, + "loss": 1.0654, + "step": 10074 + }, + { + "epoch": 0.7749403892008307, + "learning_rate": 0.0003595735075792877, + "loss": 1.3011, + "step": 10075 + }, + { + "epoch": 0.7750173063610492, + "learning_rate": 0.00035933808800154725, + "loss": 0.8951, + "step": 10076 + }, + { + "epoch": 0.7750942235212676, + "learning_rate": 0.00035910273502824447, + "loss": 1.1847, + "step": 10077 + }, + { + "epoch": 0.7751711406814861, + "learning_rate": 0.0003588674486731217, + "loss": 1.1407, + "step": 10078 + }, + { + "epoch": 0.7752480578417045, + "learning_rate": 0.00035863222894991865, + "loss": 0.9388, + "step": 10079 + }, + { + "epoch": 0.7753249750019229, + "learning_rate": 0.0003583970758723688, + "loss": 1.1442, + "step": 10080 + }, + { + "epoch": 0.7754018921621414, + "learning_rate": 0.00035816198945420316, + "loss": 1.0684, + "step": 10081 + }, + { + "epoch": 0.7754788093223598, + "learning_rate": 0.0003579269697091492, + "loss": 1.2297, + "step": 10082 + }, + { + "epoch": 0.7755557264825783, + "learning_rate": 0.00035769201665092977, + "loss": 0.8241, + "step": 10083 + }, + { + "epoch": 0.7756326436427967, + "learning_rate": 0.0003574571302932641, + "loss": 1.2009, + "step": 10084 + }, + { + "epoch": 0.7757095608030151, + "learning_rate": 0.0003572223106498672, + "loss": 1.2966, + "step": 10085 + }, + { + "epoch": 0.7757864779632336, + "learning_rate": 0.0003569875577344501, + "loss": 1.3863, + "step": 10086 + }, + { + "epoch": 0.775863395123452, + "learning_rate": 0.0003567528715607212, + "loss": 1.1067, + "step": 10087 + }, + { + "epoch": 0.7759403122836704, + "learning_rate": 0.0003565182521423837, + "loss": 1.3285, + "step": 10088 + }, + { + "epoch": 0.776017229443889, + "learning_rate": 0.00035628369949313696, + "loss": 1.2468, + "step": 10089 + }, + { + "epoch": 0.7760941466041074, + "learning_rate": 0.00035604921362667716, + "loss": 0.8892, + "step": 10090 + }, + { + "epoch": 0.7761710637643259, + "learning_rate": 0.0003558147945566954, + "loss": 1.2645, + "step": 10091 + }, + { + "epoch": 0.7762479809245443, + "learning_rate": 0.0003555804422968811, + "loss": 1.2205, + "step": 10092 + }, + { + "epoch": 0.7763248980847627, + "learning_rate": 0.00035534615686091703, + "loss": 1.3176, + "step": 10093 + }, + { + "epoch": 0.7764018152449812, + "learning_rate": 0.00035511193826248346, + "loss": 0.812, + "step": 10094 + }, + { + "epoch": 0.7764787324051996, + "learning_rate": 0.00035487778651525724, + "loss": 1.1748, + "step": 10095 + }, + { + "epoch": 0.776555649565418, + "learning_rate": 0.0003546437016329107, + "loss": 1.0933, + "step": 10096 + }, + { + "epoch": 0.7766325667256365, + "learning_rate": 0.000354409683629112, + "loss": 1.075, + "step": 10097 + }, + { + "epoch": 0.7767094838858549, + "learning_rate": 0.0003541757325175258, + "loss": 1.2219, + "step": 10098 + }, + { + "epoch": 0.7767864010460733, + "learning_rate": 0.00035394184831181254, + "loss": 1.5471, + "step": 10099 + }, + { + "epoch": 0.7768633182062918, + "learning_rate": 0.0003537080310256289, + "loss": 1.1618, + "step": 10100 + }, + { + "epoch": 0.7769402353665102, + "learning_rate": 0.00035347428067262837, + "loss": 1.2653, + "step": 10101 + }, + { + "epoch": 0.7770171525267288, + "learning_rate": 0.0003532405972664595, + "loss": 1.2221, + "step": 10102 + }, + { + "epoch": 0.7770940696869472, + "learning_rate": 0.0003530069808207672, + "loss": 0.8411, + "step": 10103 + }, + { + "epoch": 0.7771709868471656, + "learning_rate": 0.0003527734313491927, + "loss": 1.1979, + "step": 10104 + }, + { + "epoch": 0.7772479040073841, + "learning_rate": 0.00035253994886537307, + "loss": 1.4187, + "step": 10105 + }, + { + "epoch": 0.7773248211676025, + "learning_rate": 0.00035230653338294186, + "loss": 1.0576, + "step": 10106 + }, + { + "epoch": 0.7774017383278209, + "learning_rate": 0.00035207318491552815, + "loss": 1.108, + "step": 10107 + }, + { + "epoch": 0.7774786554880394, + "learning_rate": 0.0003518399034767571, + "loss": 0.9282, + "step": 10108 + }, + { + "epoch": 0.7775555726482578, + "learning_rate": 0.0003516066890802511, + "loss": 1.2714, + "step": 10109 + }, + { + "epoch": 0.7776324898084763, + "learning_rate": 0.0003513735417396273, + "loss": 1.3496, + "step": 10110 + }, + { + "epoch": 0.7777094069686947, + "learning_rate": 0.0003511404614684995, + "loss": 1.0805, + "step": 10111 + }, + { + "epoch": 0.7777863241289131, + "learning_rate": 0.00035090744828047754, + "loss": 0.7998, + "step": 10112 + }, + { + "epoch": 0.7778632412891316, + "learning_rate": 0.0003506745021891668, + "loss": 1.0188, + "step": 10113 + }, + { + "epoch": 0.77794015844935, + "learning_rate": 0.00035044162320817, + "loss": 0.9875, + "step": 10114 + }, + { + "epoch": 0.7780170756095685, + "learning_rate": 0.00035020881135108535, + "loss": 1.2888, + "step": 10115 + }, + { + "epoch": 0.778093992769787, + "learning_rate": 0.00034997606663150554, + "loss": 1.2168, + "step": 10116 + }, + { + "epoch": 0.7781709099300054, + "learning_rate": 0.00034974338906302227, + "loss": 1.3169, + "step": 10117 + }, + { + "epoch": 0.7782478270902238, + "learning_rate": 0.00034951077865922134, + "loss": 1.2438, + "step": 10118 + }, + { + "epoch": 0.7783247442504423, + "learning_rate": 0.0003492782354336848, + "loss": 1.2502, + "step": 10119 + }, + { + "epoch": 0.7784016614106607, + "learning_rate": 0.00034904575939999154, + "loss": 1.1903, + "step": 10120 + }, + { + "epoch": 0.7784785785708792, + "learning_rate": 0.00034881335057171555, + "loss": 1.1976, + "step": 10121 + }, + { + "epoch": 0.7785554957310976, + "learning_rate": 0.00034858100896242806, + "loss": 1.09, + "step": 10122 + }, + { + "epoch": 0.778632412891316, + "learning_rate": 0.00034834873458569545, + "loss": 1.3113, + "step": 10123 + }, + { + "epoch": 0.7787093300515345, + "learning_rate": 0.0003481165274550803, + "loss": 0.787, + "step": 10124 + }, + { + "epoch": 0.7787862472117529, + "learning_rate": 0.0003478843875841417, + "loss": 1.4309, + "step": 10125 + }, + { + "epoch": 0.7788631643719713, + "learning_rate": 0.0003476523149864344, + "loss": 1.1621, + "step": 10126 + }, + { + "epoch": 0.7789400815321899, + "learning_rate": 0.00034742030967550906, + "loss": 1.36, + "step": 10127 + }, + { + "epoch": 0.7790169986924083, + "learning_rate": 0.00034718837166491375, + "loss": 1.4446, + "step": 10128 + }, + { + "epoch": 0.7790939158526268, + "learning_rate": 0.00034695650096819004, + "loss": 1.1737, + "step": 10129 + }, + { + "epoch": 0.7791708330128452, + "learning_rate": 0.0003467246975988785, + "loss": 1.2886, + "step": 10130 + }, + { + "epoch": 0.7792477501730636, + "learning_rate": 0.0003464929615705136, + "loss": 1.0131, + "step": 10131 + }, + { + "epoch": 0.7793246673332821, + "learning_rate": 0.0003462612928966268, + "loss": 1.0757, + "step": 10132 + }, + { + "epoch": 0.7794015844935005, + "learning_rate": 0.0003460296915907456, + "loss": 1.1423, + "step": 10133 + }, + { + "epoch": 0.7794785016537189, + "learning_rate": 0.0003457981576663932, + "loss": 1.2314, + "step": 10134 + }, + { + "epoch": 0.7795554188139374, + "learning_rate": 0.000345566691137089, + "loss": 1.246, + "step": 10135 + }, + { + "epoch": 0.7796323359741558, + "learning_rate": 0.0003453352920163491, + "loss": 1.1795, + "step": 10136 + }, + { + "epoch": 0.7797092531343742, + "learning_rate": 0.00034510396031768543, + "loss": 0.9148, + "step": 10137 + }, + { + "epoch": 0.7797861702945927, + "learning_rate": 0.0003448726960546042, + "loss": 1.1476, + "step": 10138 + }, + { + "epoch": 0.7798630874548111, + "learning_rate": 0.0003446414992406106, + "loss": 1.1403, + "step": 10139 + }, + { + "epoch": 0.7799400046150297, + "learning_rate": 0.0003444103698892034, + "loss": 1.257, + "step": 10140 + }, + { + "epoch": 0.7800169217752481, + "learning_rate": 0.0003441793080138799, + "loss": 1.0433, + "step": 10141 + }, + { + "epoch": 0.7800938389354665, + "learning_rate": 0.0003439483136281308, + "loss": 1.2958, + "step": 10142 + }, + { + "epoch": 0.780170756095685, + "learning_rate": 0.000343717386745444, + "loss": 1.4031, + "step": 10143 + }, + { + "epoch": 0.7802476732559034, + "learning_rate": 0.00034348652737930456, + "loss": 1.0079, + "step": 10144 + }, + { + "epoch": 0.7803245904161218, + "learning_rate": 0.00034325573554319205, + "loss": 1.8904, + "step": 10145 + }, + { + "epoch": 0.7804015075763403, + "learning_rate": 0.0003430250112505825, + "loss": 1.1857, + "step": 10146 + }, + { + "epoch": 0.7804784247365587, + "learning_rate": 0.00034279435451494843, + "loss": 1.5649, + "step": 10147 + }, + { + "epoch": 0.7805553418967772, + "learning_rate": 0.00034256376534975763, + "loss": 1.0356, + "step": 10148 + }, + { + "epoch": 0.7806322590569956, + "learning_rate": 0.0003423332437684753, + "loss": 1.4342, + "step": 10149 + }, + { + "epoch": 0.780709176217214, + "learning_rate": 0.00034210278978456164, + "loss": 1.1224, + "step": 10150 + }, + { + "epoch": 0.7807860933774325, + "learning_rate": 0.00034187240341147215, + "loss": 1.1346, + "step": 10151 + }, + { + "epoch": 0.780863010537651, + "learning_rate": 0.0003416420846626601, + "loss": 1.0649, + "step": 10152 + }, + { + "epoch": 0.7809399276978694, + "learning_rate": 0.0003414118335515742, + "loss": 1.3088, + "step": 10153 + }, + { + "epoch": 0.7810168448580879, + "learning_rate": 0.00034118165009165875, + "loss": 1.2056, + "step": 10154 + }, + { + "epoch": 0.7810937620183063, + "learning_rate": 0.0003409515342963544, + "loss": 1.0821, + "step": 10155 + }, + { + "epoch": 0.7811706791785247, + "learning_rate": 0.0003407214861790979, + "loss": 1.1166, + "step": 10156 + }, + { + "epoch": 0.7812475963387432, + "learning_rate": 0.0003404915057533216, + "loss": 1.2853, + "step": 10157 + }, + { + "epoch": 0.7813245134989616, + "learning_rate": 0.0003402615930324552, + "loss": 0.8062, + "step": 10158 + }, + { + "epoch": 0.7814014306591801, + "learning_rate": 0.00034003174802992286, + "loss": 0.9184, + "step": 10159 + }, + { + "epoch": 0.7814783478193985, + "learning_rate": 0.00033980197075914574, + "loss": 1.4058, + "step": 10160 + }, + { + "epoch": 0.7815552649796169, + "learning_rate": 0.0003395722612335406, + "loss": 1.0509, + "step": 10161 + }, + { + "epoch": 0.7816321821398354, + "learning_rate": 0.0003393426194665201, + "loss": 1.1489, + "step": 10162 + }, + { + "epoch": 0.7817090993000538, + "learning_rate": 0.0003391130454714945, + "loss": 0.9629, + "step": 10163 + }, + { + "epoch": 0.7817860164602722, + "learning_rate": 0.0003388835392618676, + "loss": 0.9332, + "step": 10164 + }, + { + "epoch": 0.7818629336204908, + "learning_rate": 0.0003386541008510406, + "loss": 0.8313, + "step": 10165 + }, + { + "epoch": 0.7819398507807092, + "learning_rate": 0.00033842473025241126, + "loss": 1.1998, + "step": 10166 + }, + { + "epoch": 0.7820167679409277, + "learning_rate": 0.0003381954274793726, + "loss": 1.1993, + "step": 10167 + }, + { + "epoch": 0.7820936851011461, + "learning_rate": 0.0003379661925453137, + "loss": 1.1401, + "step": 10168 + }, + { + "epoch": 0.7821706022613645, + "learning_rate": 0.00033773702546361983, + "loss": 1.111, + "step": 10169 + }, + { + "epoch": 0.782247519421583, + "learning_rate": 0.00033750792624767184, + "loss": 1.2488, + "step": 10170 + }, + { + "epoch": 0.7823244365818014, + "learning_rate": 0.00033727889491084816, + "loss": 1.0438, + "step": 10171 + }, + { + "epoch": 0.7824013537420198, + "learning_rate": 0.00033704993146652177, + "loss": 1.3586, + "step": 10172 + }, + { + "epoch": 0.7824782709022383, + "learning_rate": 0.0003368210359280612, + "loss": 1.3086, + "step": 10173 + }, + { + "epoch": 0.7825551880624567, + "learning_rate": 0.00033659220830883294, + "loss": 1.0261, + "step": 10174 + }, + { + "epoch": 0.7826321052226752, + "learning_rate": 0.0003363634486221978, + "loss": 1.0639, + "step": 10175 + }, + { + "epoch": 0.7827090223828936, + "learning_rate": 0.00033613475688151436, + "loss": 1.049, + "step": 10176 + }, + { + "epoch": 0.782785939543112, + "learning_rate": 0.0003359061331001349, + "loss": 0.6166, + "step": 10177 + }, + { + "epoch": 0.7828628567033306, + "learning_rate": 0.0003356775772914091, + "loss": 1.0824, + "step": 10178 + }, + { + "epoch": 0.782939773863549, + "learning_rate": 0.0003354490894686835, + "loss": 1.1148, + "step": 10179 + }, + { + "epoch": 0.7830166910237674, + "learning_rate": 0.00033522066964529923, + "loss": 1.1823, + "step": 10180 + }, + { + "epoch": 0.7830936081839859, + "learning_rate": 0.0003349923178345941, + "loss": 1.5071, + "step": 10181 + }, + { + "epoch": 0.7831705253442043, + "learning_rate": 0.00033476403404990144, + "loss": 1.1846, + "step": 10182 + }, + { + "epoch": 0.7832474425044227, + "learning_rate": 0.0003345358183045513, + "loss": 1.2911, + "step": 10183 + }, + { + "epoch": 0.7833243596646412, + "learning_rate": 0.000334307670611869, + "loss": 1.3045, + "step": 10184 + }, + { + "epoch": 0.7834012768248596, + "learning_rate": 0.0003340795909851774, + "loss": 1.4123, + "step": 10185 + }, + { + "epoch": 0.7834781939850781, + "learning_rate": 0.00033385157943779275, + "loss": 0.7913, + "step": 10186 + }, + { + "epoch": 0.7835551111452965, + "learning_rate": 0.0003336236359830302, + "loss": 1.3126, + "step": 10187 + }, + { + "epoch": 0.7836320283055149, + "learning_rate": 0.000333395760634199, + "loss": 1.0296, + "step": 10188 + }, + { + "epoch": 0.7837089454657334, + "learning_rate": 0.00033316795340460525, + "loss": 1.1184, + "step": 10189 + }, + { + "epoch": 0.7837858626259518, + "learning_rate": 0.0003329402143075506, + "loss": 0.9802, + "step": 10190 + }, + { + "epoch": 0.7838627797861702, + "learning_rate": 0.00033271254335633327, + "loss": 1.1399, + "step": 10191 + }, + { + "epoch": 0.7839396969463888, + "learning_rate": 0.00033248494056424663, + "loss": 1.2731, + "step": 10192 + }, + { + "epoch": 0.7840166141066072, + "learning_rate": 0.00033225740594458147, + "loss": 0.9823, + "step": 10193 + }, + { + "epoch": 0.7840935312668257, + "learning_rate": 0.0003320299395106233, + "loss": 1.3699, + "step": 10194 + }, + { + "epoch": 0.7841704484270441, + "learning_rate": 0.0003318025412756542, + "loss": 1.0667, + "step": 10195 + }, + { + "epoch": 0.7842473655872625, + "learning_rate": 0.0003315752112529523, + "loss": 1.3758, + "step": 10196 + }, + { + "epoch": 0.784324282747481, + "learning_rate": 0.0003313479494557909, + "loss": 1.1681, + "step": 10197 + }, + { + "epoch": 0.7844011999076994, + "learning_rate": 0.0003311207558974416, + "loss": 1.1172, + "step": 10198 + }, + { + "epoch": 0.7844781170679178, + "learning_rate": 0.00033089363059116914, + "loss": 1.0151, + "step": 10199 + }, + { + "epoch": 0.7845550342281363, + "learning_rate": 0.0003306665735502356, + "loss": 1.2513, + "step": 10200 + }, + { + "epoch": 0.7846319513883547, + "learning_rate": 0.0003304395847878998, + "loss": 1.4673, + "step": 10201 + }, + { + "epoch": 0.7847088685485731, + "learning_rate": 0.0003302126643174157, + "loss": 1.2164, + "step": 10202 + }, + { + "epoch": 0.7847857857087916, + "learning_rate": 0.00032998581215203316, + "loss": 1.2011, + "step": 10203 + }, + { + "epoch": 0.78486270286901, + "learning_rate": 0.0003297590283049984, + "loss": 1.0943, + "step": 10204 + }, + { + "epoch": 0.7849396200292286, + "learning_rate": 0.0003295323127895531, + "loss": 1.1943, + "step": 10205 + }, + { + "epoch": 0.785016537189447, + "learning_rate": 0.0003293056656189364, + "loss": 1.2752, + "step": 10206 + }, + { + "epoch": 0.7850934543496654, + "learning_rate": 0.00032907908680638237, + "loss": 1.6692, + "step": 10207 + }, + { + "epoch": 0.7851703715098839, + "learning_rate": 0.00032885257636511983, + "loss": 1.376, + "step": 10208 + }, + { + "epoch": 0.7852472886701023, + "learning_rate": 0.00032862613430837653, + "loss": 1.1115, + "step": 10209 + }, + { + "epoch": 0.7853242058303207, + "learning_rate": 0.00032839976064937383, + "loss": 1.011, + "step": 10210 + }, + { + "epoch": 0.7854011229905392, + "learning_rate": 0.0003281734554013302, + "loss": 1.2191, + "step": 10211 + }, + { + "epoch": 0.7854780401507576, + "learning_rate": 0.0003279472185774597, + "loss": 1.3628, + "step": 10212 + }, + { + "epoch": 0.7855549573109761, + "learning_rate": 0.0003277210501909722, + "loss": 0.9716, + "step": 10213 + }, + { + "epoch": 0.7856318744711945, + "learning_rate": 0.0003274949502550747, + "loss": 1.1441, + "step": 10214 + }, + { + "epoch": 0.7857087916314129, + "learning_rate": 0.000327268918782969, + "loss": 0.9489, + "step": 10215 + }, + { + "epoch": 0.7857857087916315, + "learning_rate": 0.00032704295578785337, + "loss": 0.8836, + "step": 10216 + }, + { + "epoch": 0.7858626259518499, + "learning_rate": 0.00032681706128292196, + "loss": 0.9046, + "step": 10217 + }, + { + "epoch": 0.7859395431120683, + "learning_rate": 0.00032659123528136497, + "loss": 1.2939, + "step": 10218 + }, + { + "epoch": 0.7860164602722868, + "learning_rate": 0.0003263654777963683, + "loss": 1.0273, + "step": 10219 + }, + { + "epoch": 0.7860933774325052, + "learning_rate": 0.0003261397888411153, + "loss": 1.094, + "step": 10220 + }, + { + "epoch": 0.7861702945927236, + "learning_rate": 0.000325914168428783, + "loss": 0.9821, + "step": 10221 + }, + { + "epoch": 0.7862472117529421, + "learning_rate": 0.0003256886165725456, + "loss": 1.2025, + "step": 10222 + }, + { + "epoch": 0.7863241289131605, + "learning_rate": 0.0003254631332855742, + "loss": 0.9703, + "step": 10223 + }, + { + "epoch": 0.786401046073379, + "learning_rate": 0.00032523771858103433, + "loss": 0.7507, + "step": 10224 + }, + { + "epoch": 0.7864779632335974, + "learning_rate": 0.00032501237247208856, + "loss": 1.0983, + "step": 10225 + }, + { + "epoch": 0.7865548803938158, + "learning_rate": 0.00032478709497189476, + "loss": 1.1567, + "step": 10226 + }, + { + "epoch": 0.7866317975540343, + "learning_rate": 0.0003245618860936069, + "loss": 1.1132, + "step": 10227 + }, + { + "epoch": 0.7867087147142527, + "learning_rate": 0.0003243367458503761, + "loss": 1.0461, + "step": 10228 + }, + { + "epoch": 0.7867856318744711, + "learning_rate": 0.0003241116742553482, + "loss": 1.5757, + "step": 10229 + }, + { + "epoch": 0.7868625490346897, + "learning_rate": 0.0003238866713216643, + "loss": 1.1834, + "step": 10230 + }, + { + "epoch": 0.7869394661949081, + "learning_rate": 0.0003236617370624638, + "loss": 1.2352, + "step": 10231 + }, + { + "epoch": 0.7870163833551266, + "learning_rate": 0.00032343687149088003, + "loss": 2.163, + "step": 10232 + }, + { + "epoch": 0.787093300515345, + "learning_rate": 0.0003232120746200444, + "loss": 1.071, + "step": 10233 + }, + { + "epoch": 0.7871702176755634, + "learning_rate": 0.00032298734646308164, + "loss": 0.7298, + "step": 10234 + }, + { + "epoch": 0.7872471348357819, + "learning_rate": 0.00032276268703311403, + "loss": 1.3017, + "step": 10235 + }, + { + "epoch": 0.7873240519960003, + "learning_rate": 0.00032253809634326034, + "loss": 1.0415, + "step": 10236 + }, + { + "epoch": 0.7874009691562187, + "learning_rate": 0.0003223135744066344, + "loss": 1.0901, + "step": 10237 + }, + { + "epoch": 0.7874778863164372, + "learning_rate": 0.0003220891212363461, + "loss": 1.1322, + "step": 10238 + }, + { + "epoch": 0.7875548034766556, + "learning_rate": 0.00032186473684550176, + "loss": 1.0514, + "step": 10239 + }, + { + "epoch": 0.787631720636874, + "learning_rate": 0.0003216404212472031, + "loss": 1.2795, + "step": 10240 + }, + { + "epoch": 0.7877086377970925, + "learning_rate": 0.00032141617445454796, + "loss": 1.1256, + "step": 10241 + }, + { + "epoch": 0.787785554957311, + "learning_rate": 0.0003211919964806312, + "loss": 1.6298, + "step": 10242 + }, + { + "epoch": 0.7878624721175295, + "learning_rate": 0.0003209678873385421, + "loss": 1.0925, + "step": 10243 + }, + { + "epoch": 0.7879393892777479, + "learning_rate": 0.0003207438470413669, + "loss": 1.3912, + "step": 10244 + }, + { + "epoch": 0.7880163064379663, + "learning_rate": 0.0003205198756021876, + "loss": 1.4718, + "step": 10245 + }, + { + "epoch": 0.7880932235981848, + "learning_rate": 0.0003202959730340818, + "loss": 1.2329, + "step": 10246 + }, + { + "epoch": 0.7881701407584032, + "learning_rate": 0.0003200721393501236, + "loss": 1.2347, + "step": 10247 + }, + { + "epoch": 0.7882470579186216, + "learning_rate": 0.00031984837456338293, + "loss": 0.858, + "step": 10248 + }, + { + "epoch": 0.7883239750788401, + "learning_rate": 0.0003196246786869253, + "loss": 1.2483, + "step": 10249 + }, + { + "epoch": 0.7884008922390585, + "learning_rate": 0.000319401051733813, + "loss": 1.0947, + "step": 10250 + }, + { + "epoch": 0.788477809399277, + "learning_rate": 0.0003191774937171038, + "loss": 0.9954, + "step": 10251 + }, + { + "epoch": 0.7885547265594954, + "learning_rate": 0.0003189540046498514, + "loss": 0.6918, + "step": 10252 + }, + { + "epoch": 0.7886316437197138, + "learning_rate": 0.00031873058454510555, + "loss": 0.8014, + "step": 10253 + }, + { + "epoch": 0.7887085608799324, + "learning_rate": 0.0003185072334159115, + "loss": 1.3529, + "step": 10254 + }, + { + "epoch": 0.7887854780401508, + "learning_rate": 0.0003182839512753119, + "loss": 1.1859, + "step": 10255 + }, + { + "epoch": 0.7888623952003692, + "learning_rate": 0.0003180607381363442, + "loss": 1.2967, + "step": 10256 + }, + { + "epoch": 0.7889393123605877, + "learning_rate": 0.00031783759401204107, + "loss": 1.4974, + "step": 10257 + }, + { + "epoch": 0.7890162295208061, + "learning_rate": 0.00031761451891543327, + "loss": 0.9802, + "step": 10258 + }, + { + "epoch": 0.7890931466810245, + "learning_rate": 0.00031739151285954594, + "loss": 0.807, + "step": 10259 + }, + { + "epoch": 0.789170063841243, + "learning_rate": 0.0003171685758574005, + "loss": 0.7691, + "step": 10260 + }, + { + "epoch": 0.7892469810014614, + "learning_rate": 0.00031694570792201485, + "loss": 1.1775, + "step": 10261 + }, + { + "epoch": 0.7893238981616799, + "learning_rate": 0.0003167229090664018, + "loss": 1.1752, + "step": 10262 + }, + { + "epoch": 0.7894008153218983, + "learning_rate": 0.0003165001793035716, + "loss": 1.0563, + "step": 10263 + }, + { + "epoch": 0.7894777324821167, + "learning_rate": 0.0003162775186465295, + "loss": 1.1318, + "step": 10264 + }, + { + "epoch": 0.7895546496423352, + "learning_rate": 0.00031605492710827656, + "loss": 1.0815, + "step": 10265 + }, + { + "epoch": 0.7896315668025536, + "learning_rate": 0.0003158324047018104, + "loss": 1.6239, + "step": 10266 + }, + { + "epoch": 0.789708483962772, + "learning_rate": 0.0003156099514401242, + "loss": 1.1011, + "step": 10267 + }, + { + "epoch": 0.7897854011229906, + "learning_rate": 0.0003153875673362069, + "loss": 1.0138, + "step": 10268 + }, + { + "epoch": 0.789862318283209, + "learning_rate": 0.0003151652524030448, + "loss": 1.3059, + "step": 10269 + }, + { + "epoch": 0.7899392354434275, + "learning_rate": 0.0003149430066536178, + "loss": 0.821, + "step": 10270 + }, + { + "epoch": 0.7900161526036459, + "learning_rate": 0.00031472083010090397, + "loss": 1.2222, + "step": 10271 + }, + { + "epoch": 0.7900930697638643, + "learning_rate": 0.00031449872275787625, + "loss": 1.3313, + "step": 10272 + }, + { + "epoch": 0.7901699869240828, + "learning_rate": 0.00031427668463750353, + "loss": 1.3168, + "step": 10273 + }, + { + "epoch": 0.7902469040843012, + "learning_rate": 0.0003140547157527509, + "loss": 1.4956, + "step": 10274 + }, + { + "epoch": 0.7903238212445196, + "learning_rate": 0.00031383281611657937, + "loss": 1.1706, + "step": 10275 + }, + { + "epoch": 0.7904007384047381, + "learning_rate": 0.00031361098574194556, + "loss": 1.2118, + "step": 10276 + }, + { + "epoch": 0.7904776555649565, + "learning_rate": 0.00031338922464180306, + "loss": 0.9576, + "step": 10277 + }, + { + "epoch": 0.7905545727251749, + "learning_rate": 0.0003131675328291009, + "loss": 1.2488, + "step": 10278 + }, + { + "epoch": 0.7906314898853934, + "learning_rate": 0.0003129459103167824, + "loss": 1.4225, + "step": 10279 + }, + { + "epoch": 0.7907084070456118, + "learning_rate": 0.0003127243571177899, + "loss": 1.0132, + "step": 10280 + }, + { + "epoch": 0.7907853242058304, + "learning_rate": 0.0003125028732450591, + "loss": 1.1335, + "step": 10281 + }, + { + "epoch": 0.7908622413660488, + "learning_rate": 0.00031228145871152387, + "loss": 0.9534, + "step": 10282 + }, + { + "epoch": 0.7909391585262672, + "learning_rate": 0.00031206011353011165, + "loss": 1.1116, + "step": 10283 + }, + { + "epoch": 0.7910160756864857, + "learning_rate": 0.00031183883771374716, + "loss": 1.123, + "step": 10284 + }, + { + "epoch": 0.7910929928467041, + "learning_rate": 0.0003116176312753515, + "loss": 1.1737, + "step": 10285 + }, + { + "epoch": 0.7911699100069225, + "learning_rate": 0.00031139649422784075, + "loss": 1.0924, + "step": 10286 + }, + { + "epoch": 0.791246827167141, + "learning_rate": 0.00031117542658412774, + "loss": 1.2451, + "step": 10287 + }, + { + "epoch": 0.7913237443273594, + "learning_rate": 0.00031095442835712023, + "loss": 0.9532, + "step": 10288 + }, + { + "epoch": 0.7914006614875779, + "learning_rate": 0.00031073349955972263, + "loss": 1.1028, + "step": 10289 + }, + { + "epoch": 0.7914775786477963, + "learning_rate": 0.0003105126402048358, + "loss": 1.0861, + "step": 10290 + }, + { + "epoch": 0.7915544958080147, + "learning_rate": 0.00031029185030535583, + "loss": 1.0757, + "step": 10291 + }, + { + "epoch": 0.7916314129682333, + "learning_rate": 0.00031007112987417384, + "loss": 0.8651, + "step": 10292 + }, + { + "epoch": 0.7917083301284517, + "learning_rate": 0.00030985047892417915, + "loss": 1.0436, + "step": 10293 + }, + { + "epoch": 0.7917852472886701, + "learning_rate": 0.0003096298974682553, + "loss": 1.1169, + "step": 10294 + }, + { + "epoch": 0.7918621644488886, + "learning_rate": 0.0003094093855192823, + "loss": 0.919, + "step": 10295 + }, + { + "epoch": 0.791939081609107, + "learning_rate": 0.0003091889430901362, + "loss": 1.201, + "step": 10296 + }, + { + "epoch": 0.7920159987693255, + "learning_rate": 0.00030896857019368823, + "loss": 1.2314, + "step": 10297 + }, + { + "epoch": 0.7920929159295439, + "learning_rate": 0.0003087482668428072, + "loss": 0.8611, + "step": 10298 + }, + { + "epoch": 0.7921698330897623, + "learning_rate": 0.00030852803305035637, + "loss": 1.1182, + "step": 10299 + }, + { + "epoch": 0.7922467502499808, + "learning_rate": 0.00030830786882919524, + "loss": 1.2974, + "step": 10300 + }, + { + "epoch": 0.7923236674101992, + "learning_rate": 0.0003080877741921798, + "loss": 0.9761, + "step": 10301 + }, + { + "epoch": 0.7924005845704176, + "learning_rate": 0.0003078677491521613, + "loss": 1.2451, + "step": 10302 + }, + { + "epoch": 0.7924775017306361, + "learning_rate": 0.000307647793721987, + "loss": 1.1248, + "step": 10303 + }, + { + "epoch": 0.7925544188908545, + "learning_rate": 0.0003074279079145013, + "loss": 1.1212, + "step": 10304 + }, + { + "epoch": 0.7926313360510729, + "learning_rate": 0.00030720809174254266, + "loss": 1.3907, + "step": 10305 + }, + { + "epoch": 0.7927082532112915, + "learning_rate": 0.00030698834521894627, + "loss": 1.2372, + "step": 10306 + }, + { + "epoch": 0.7927851703715099, + "learning_rate": 0.00030676866835654417, + "loss": 1.4555, + "step": 10307 + }, + { + "epoch": 0.7928620875317284, + "learning_rate": 0.00030654906116816296, + "loss": 0.7656, + "step": 10308 + }, + { + "epoch": 0.7929390046919468, + "learning_rate": 0.0003063295236666259, + "loss": 1.0952, + "step": 10309 + }, + { + "epoch": 0.7930159218521652, + "learning_rate": 0.0003061100558647518, + "loss": 1.3046, + "step": 10310 + }, + { + "epoch": 0.7930928390123837, + "learning_rate": 0.0003058906577753556, + "loss": 0.8801, + "step": 10311 + }, + { + "epoch": 0.7931697561726021, + "learning_rate": 0.0003056713294112487, + "loss": 1.1586, + "step": 10312 + }, + { + "epoch": 0.7932466733328205, + "learning_rate": 0.00030545207078523797, + "loss": 1.227, + "step": 10313 + }, + { + "epoch": 0.793323590493039, + "learning_rate": 0.00030523288191012496, + "loss": 1.1585, + "step": 10314 + }, + { + "epoch": 0.7934005076532574, + "learning_rate": 0.0003050137627987096, + "loss": 1.5249, + "step": 10315 + }, + { + "epoch": 0.7934774248134759, + "learning_rate": 0.00030479471346378544, + "loss": 1.1025, + "step": 10316 + }, + { + "epoch": 0.7935543419736943, + "learning_rate": 0.00030457573391814454, + "loss": 0.9091, + "step": 10317 + }, + { + "epoch": 0.7936312591339127, + "learning_rate": 0.0003043568241745719, + "loss": 1.1504, + "step": 10318 + }, + { + "epoch": 0.7937081762941313, + "learning_rate": 0.00030413798424585, + "loss": 1.0784, + "step": 10319 + }, + { + "epoch": 0.7937850934543497, + "learning_rate": 0.0003039192141447579, + "loss": 1.3788, + "step": 10320 + }, + { + "epoch": 0.7938620106145681, + "learning_rate": 0.0003037005138840693, + "loss": 1.3925, + "step": 10321 + }, + { + "epoch": 0.7939389277747866, + "learning_rate": 0.00030348188347655464, + "loss": 1.193, + "step": 10322 + }, + { + "epoch": 0.794015844935005, + "learning_rate": 0.00030326332293497966, + "loss": 0.8348, + "step": 10323 + }, + { + "epoch": 0.7940927620952234, + "learning_rate": 0.0003030448322721065, + "loss": 1.3422, + "step": 10324 + }, + { + "epoch": 0.7941696792554419, + "learning_rate": 0.00030282641150069264, + "loss": 1.512, + "step": 10325 + }, + { + "epoch": 0.7942465964156603, + "learning_rate": 0.000302608060633493, + "loss": 1.129, + "step": 10326 + }, + { + "epoch": 0.7943235135758788, + "learning_rate": 0.0003023897796832559, + "loss": 1.0456, + "step": 10327 + }, + { + "epoch": 0.7944004307360972, + "learning_rate": 0.000302171568662728, + "loss": 1.1939, + "step": 10328 + }, + { + "epoch": 0.7944773478963156, + "learning_rate": 0.0003019534275846505, + "loss": 1.1195, + "step": 10329 + }, + { + "epoch": 0.7945542650565341, + "learning_rate": 0.00030173535646176107, + "loss": 1.2152, + "step": 10330 + }, + { + "epoch": 0.7946311822167526, + "learning_rate": 0.00030151735530679295, + "loss": 1.0945, + "step": 10331 + }, + { + "epoch": 0.794708099376971, + "learning_rate": 0.00030129942413247544, + "loss": 1.1961, + "step": 10332 + }, + { + "epoch": 0.7947850165371895, + "learning_rate": 0.0003010815629515334, + "loss": 1.0837, + "step": 10333 + }, + { + "epoch": 0.7948619336974079, + "learning_rate": 0.0003008637717766889, + "loss": 1.3923, + "step": 10334 + }, + { + "epoch": 0.7949388508576264, + "learning_rate": 0.0003006460506206583, + "loss": 0.8961, + "step": 10335 + }, + { + "epoch": 0.7950157680178448, + "learning_rate": 0.0003004283994961548, + "loss": 1.2444, + "step": 10336 + }, + { + "epoch": 0.7950926851780632, + "learning_rate": 0.0003002108184158872, + "loss": 1.0956, + "step": 10337 + }, + { + "epoch": 0.7951696023382817, + "learning_rate": 0.00029999330739256005, + "loss": 1.4607, + "step": 10338 + }, + { + "epoch": 0.7952465194985001, + "learning_rate": 0.0002997758664388749, + "loss": 1.1263, + "step": 10339 + }, + { + "epoch": 0.7953234366587185, + "learning_rate": 0.00029955849556752745, + "loss": 1.27, + "step": 10340 + }, + { + "epoch": 0.795400353818937, + "learning_rate": 0.0002993411947912101, + "loss": 1.7407, + "step": 10341 + }, + { + "epoch": 0.7954772709791554, + "learning_rate": 0.0002991239641226122, + "loss": 1.254, + "step": 10342 + }, + { + "epoch": 0.7955541881393738, + "learning_rate": 0.0002989068035744174, + "loss": 1.2726, + "step": 10343 + }, + { + "epoch": 0.7956311052995924, + "learning_rate": 0.00029868971315930617, + "loss": 0.993, + "step": 10344 + }, + { + "epoch": 0.7957080224598108, + "learning_rate": 0.00029847269288995444, + "loss": 1.3377, + "step": 10345 + }, + { + "epoch": 0.7957849396200293, + "learning_rate": 0.0002982557427790341, + "loss": 1.064, + "step": 10346 + }, + { + "epoch": 0.7958618567802477, + "learning_rate": 0.00029803886283921374, + "loss": 0.9866, + "step": 10347 + }, + { + "epoch": 0.7959387739404661, + "learning_rate": 0.0002978220530831571, + "loss": 1.3031, + "step": 10348 + }, + { + "epoch": 0.7960156911006846, + "learning_rate": 0.00029760531352352306, + "loss": 1.13, + "step": 10349 + }, + { + "epoch": 0.796092608260903, + "learning_rate": 0.00029738864417296816, + "loss": 1.2601, + "step": 10350 + }, + { + "epoch": 0.7961695254211214, + "learning_rate": 0.00029717204504414377, + "loss": 1.0851, + "step": 10351 + }, + { + "epoch": 0.7962464425813399, + "learning_rate": 0.0002969555161496971, + "loss": 1.4252, + "step": 10352 + }, + { + "epoch": 0.7963233597415583, + "learning_rate": 0.0002967390575022717, + "loss": 1.2639, + "step": 10353 + }, + { + "epoch": 0.7964002769017768, + "learning_rate": 0.0002965226691145064, + "loss": 1.2395, + "step": 10354 + }, + { + "epoch": 0.7964771940619952, + "learning_rate": 0.00029630635099903683, + "loss": 1.0853, + "step": 10355 + }, + { + "epoch": 0.7965541112222136, + "learning_rate": 0.0002960901031684942, + "loss": 1.3469, + "step": 10356 + }, + { + "epoch": 0.7966310283824322, + "learning_rate": 0.00029587392563550487, + "loss": 1.2555, + "step": 10357 + }, + { + "epoch": 0.7967079455426506, + "learning_rate": 0.00029565781841269217, + "loss": 1.2899, + "step": 10358 + }, + { + "epoch": 0.796784862702869, + "learning_rate": 0.0002954417815126744, + "loss": 1.1721, + "step": 10359 + }, + { + "epoch": 0.7968617798630875, + "learning_rate": 0.000295225814948066, + "loss": 1.0623, + "step": 10360 + }, + { + "epoch": 0.7969386970233059, + "learning_rate": 0.00029500991873147864, + "loss": 1.3483, + "step": 10361 + }, + { + "epoch": 0.7970156141835243, + "learning_rate": 0.00029479409287551763, + "loss": 1.4211, + "step": 10362 + }, + { + "epoch": 0.7970925313437428, + "learning_rate": 0.0002945783373927852, + "loss": 1.2193, + "step": 10363 + }, + { + "epoch": 0.7971694485039612, + "learning_rate": 0.00029436265229588023, + "loss": 1.2345, + "step": 10364 + }, + { + "epoch": 0.7972463656641797, + "learning_rate": 0.00029414703759739636, + "loss": 1.1392, + "step": 10365 + }, + { + "epoch": 0.7973232828243981, + "learning_rate": 0.0002939314933099239, + "loss": 1.3871, + "step": 10366 + }, + { + "epoch": 0.7974001999846165, + "learning_rate": 0.0002937160194460483, + "loss": 1.1328, + "step": 10367 + }, + { + "epoch": 0.797477117144835, + "learning_rate": 0.00029350061601835116, + "loss": 1.1989, + "step": 10368 + }, + { + "epoch": 0.7975540343050534, + "learning_rate": 0.0002932852830394108, + "loss": 0.9767, + "step": 10369 + }, + { + "epoch": 0.7976309514652719, + "learning_rate": 0.00029307002052180093, + "loss": 0.736, + "step": 10370 + }, + { + "epoch": 0.7977078686254904, + "learning_rate": 0.0002928548284780894, + "loss": 1.5841, + "step": 10371 + }, + { + "epoch": 0.7977847857857088, + "learning_rate": 0.0002926397069208429, + "loss": 1.3683, + "step": 10372 + }, + { + "epoch": 0.7978617029459273, + "learning_rate": 0.00029242465586262187, + "loss": 1.1142, + "step": 10373 + }, + { + "epoch": 0.7979386201061457, + "learning_rate": 0.00029220967531598454, + "loss": 1.3105, + "step": 10374 + }, + { + "epoch": 0.7980155372663641, + "learning_rate": 0.00029199476529348257, + "loss": 1.1201, + "step": 10375 + }, + { + "epoch": 0.7980924544265826, + "learning_rate": 0.0002917799258076648, + "loss": 1.2185, + "step": 10376 + }, + { + "epoch": 0.798169371586801, + "learning_rate": 0.0002915651568710766, + "loss": 1.6242, + "step": 10377 + }, + { + "epoch": 0.7982462887470194, + "learning_rate": 0.00029135045849625833, + "loss": 1.3077, + "step": 10378 + }, + { + "epoch": 0.7983232059072379, + "learning_rate": 0.0002911358306957465, + "loss": 0.8829, + "step": 10379 + }, + { + "epoch": 0.7984001230674563, + "learning_rate": 0.0002909212734820734, + "loss": 0.7368, + "step": 10380 + }, + { + "epoch": 0.7984770402276747, + "learning_rate": 0.0002907067868677671, + "loss": 0.8756, + "step": 10381 + }, + { + "epoch": 0.7985539573878933, + "learning_rate": 0.00029049237086535145, + "loss": 0.8449, + "step": 10382 + }, + { + "epoch": 0.7986308745481117, + "learning_rate": 0.0002902780254873477, + "loss": 1.052, + "step": 10383 + }, + { + "epoch": 0.7987077917083302, + "learning_rate": 0.00029006375074626986, + "loss": 1.1996, + "step": 10384 + }, + { + "epoch": 0.7987847088685486, + "learning_rate": 0.00028984954665463097, + "loss": 1.2681, + "step": 10385 + }, + { + "epoch": 0.798861626028767, + "learning_rate": 0.00028963541322493826, + "loss": 1.1129, + "step": 10386 + }, + { + "epoch": 0.7989385431889855, + "learning_rate": 0.00028942135046969514, + "loss": 1.2128, + "step": 10387 + }, + { + "epoch": 0.7990154603492039, + "learning_rate": 0.00028920735840140115, + "loss": 0.8635, + "step": 10388 + }, + { + "epoch": 0.7990923775094223, + "learning_rate": 0.00028899343703255113, + "loss": 1.1278, + "step": 10389 + }, + { + "epoch": 0.7991692946696408, + "learning_rate": 0.0002887795863756363, + "loss": 1.3238, + "step": 10390 + }, + { + "epoch": 0.7992462118298592, + "learning_rate": 0.0002885658064431439, + "loss": 1.143, + "step": 10391 + }, + { + "epoch": 0.7993231289900777, + "learning_rate": 0.00028835209724755665, + "loss": 0.8206, + "step": 10392 + }, + { + "epoch": 0.7994000461502961, + "learning_rate": 0.00028813845880135315, + "loss": 0.9658, + "step": 10393 + }, + { + "epoch": 0.7994769633105145, + "learning_rate": 0.00028792489111700815, + "loss": 1.4044, + "step": 10394 + }, + { + "epoch": 0.7995538804707331, + "learning_rate": 0.0002877113942069914, + "loss": 0.9718, + "step": 10395 + }, + { + "epoch": 0.7996307976309515, + "learning_rate": 0.00028749796808377054, + "loss": 0.5927, + "step": 10396 + }, + { + "epoch": 0.7997077147911699, + "learning_rate": 0.00028728461275980666, + "loss": 1.3003, + "step": 10397 + }, + { + "epoch": 0.7997846319513884, + "learning_rate": 0.0002870713282475578, + "loss": 0.9377, + "step": 10398 + }, + { + "epoch": 0.7998615491116068, + "learning_rate": 0.0002868581145594784, + "loss": 0.8702, + "step": 10399 + }, + { + "epoch": 0.7999384662718253, + "learning_rate": 0.00028664497170801794, + "loss": 1.1248, + "step": 10400 + }, + { + "epoch": 0.8000153834320437, + "learning_rate": 0.00028643189970562226, + "loss": 1.0356, + "step": 10401 + }, + { + "epoch": 0.8000923005922621, + "learning_rate": 0.00028621889856473274, + "loss": 1.3393, + "step": 10402 + }, + { + "epoch": 0.8001692177524806, + "learning_rate": 0.0002860059682977861, + "loss": 0.9831, + "step": 10403 + }, + { + "epoch": 0.800246134912699, + "learning_rate": 0.0002857931089172168, + "loss": 0.8914, + "step": 10404 + }, + { + "epoch": 0.8003230520729174, + "learning_rate": 0.00028558032043545325, + "loss": 1.3209, + "step": 10405 + }, + { + "epoch": 0.8003999692331359, + "learning_rate": 0.0002853676028649202, + "loss": 1.1725, + "step": 10406 + }, + { + "epoch": 0.8004768863933543, + "learning_rate": 0.00028515495621803886, + "loss": 1.2159, + "step": 10407 + }, + { + "epoch": 0.8005538035535728, + "learning_rate": 0.0002849423805072257, + "loss": 1.2795, + "step": 10408 + }, + { + "epoch": 0.8006307207137913, + "learning_rate": 0.00028472987574489295, + "loss": 1.0802, + "step": 10409 + }, + { + "epoch": 0.8007076378740097, + "learning_rate": 0.00028451744194344985, + "loss": 1.1759, + "step": 10410 + }, + { + "epoch": 0.8007845550342282, + "learning_rate": 0.0002843050791152993, + "loss": 0.8882, + "step": 10411 + }, + { + "epoch": 0.8008614721944466, + "learning_rate": 0.0002840927872728427, + "loss": 1.1256, + "step": 10412 + }, + { + "epoch": 0.800938389354665, + "learning_rate": 0.0002838805664284754, + "loss": 1.0765, + "step": 10413 + }, + { + "epoch": 0.8010153065148835, + "learning_rate": 0.00028366841659458906, + "loss": 1.4075, + "step": 10414 + }, + { + "epoch": 0.8010922236751019, + "learning_rate": 0.0002834563377835717, + "loss": 1.1554, + "step": 10415 + }, + { + "epoch": 0.8011691408353203, + "learning_rate": 0.00028324433000780655, + "loss": 1.1122, + "step": 10416 + }, + { + "epoch": 0.8012460579955388, + "learning_rate": 0.0002830323932796728, + "loss": 0.8693, + "step": 10417 + }, + { + "epoch": 0.8013229751557572, + "learning_rate": 0.00028282052761154607, + "loss": 1.1575, + "step": 10418 + }, + { + "epoch": 0.8013998923159757, + "learning_rate": 0.00028260873301579737, + "loss": 1.001, + "step": 10419 + }, + { + "epoch": 0.8014768094761942, + "learning_rate": 0.0002823970095047934, + "loss": 1.0212, + "step": 10420 + }, + { + "epoch": 0.8015537266364126, + "learning_rate": 0.000282185357090897, + "loss": 1.3056, + "step": 10421 + }, + { + "epoch": 0.8016306437966311, + "learning_rate": 0.00028197377578646657, + "loss": 1.321, + "step": 10422 + }, + { + "epoch": 0.8017075609568495, + "learning_rate": 0.0002817622656038574, + "loss": 1.0864, + "step": 10423 + }, + { + "epoch": 0.8017844781170679, + "learning_rate": 0.00028155082655541877, + "loss": 0.9638, + "step": 10424 + }, + { + "epoch": 0.8018613952772864, + "learning_rate": 0.0002813394586534968, + "loss": 1.136, + "step": 10425 + }, + { + "epoch": 0.8019383124375048, + "learning_rate": 0.00028112816191043434, + "loss": 0.99, + "step": 10426 + }, + { + "epoch": 0.8020152295977232, + "learning_rate": 0.0002809169363385686, + "loss": 1.1861, + "step": 10427 + }, + { + "epoch": 0.8020921467579417, + "learning_rate": 0.00028070578195023364, + "loss": 0.9569, + "step": 10428 + }, + { + "epoch": 0.8021690639181601, + "learning_rate": 0.0002804946987577589, + "loss": 1.2844, + "step": 10429 + }, + { + "epoch": 0.8022459810783786, + "learning_rate": 0.00028028368677346897, + "loss": 1.193, + "step": 10430 + }, + { + "epoch": 0.802322898238597, + "learning_rate": 0.0002800727460096864, + "loss": 1.2253, + "step": 10431 + }, + { + "epoch": 0.8023998153988154, + "learning_rate": 0.0002798618764787278, + "loss": 1.1087, + "step": 10432 + }, + { + "epoch": 0.802476732559034, + "learning_rate": 0.00027965107819290496, + "loss": 1.1771, + "step": 10433 + }, + { + "epoch": 0.8025536497192524, + "learning_rate": 0.00027944035116452795, + "loss": 1.2884, + "step": 10434 + }, + { + "epoch": 0.8026305668794708, + "learning_rate": 0.000279229695405901, + "loss": 1.1639, + "step": 10435 + }, + { + "epoch": 0.8027074840396893, + "learning_rate": 0.0002790191109293243, + "loss": 1.49, + "step": 10436 + }, + { + "epoch": 0.8027844011999077, + "learning_rate": 0.00027880859774709403, + "loss": 1.5186, + "step": 10437 + }, + { + "epoch": 0.8028613183601262, + "learning_rate": 0.00027859815587150225, + "loss": 1.1989, + "step": 10438 + }, + { + "epoch": 0.8029382355203446, + "learning_rate": 0.00027838778531483747, + "loss": 0.9743, + "step": 10439 + }, + { + "epoch": 0.803015152680563, + "learning_rate": 0.0002781774860893829, + "loss": 0.9675, + "step": 10440 + }, + { + "epoch": 0.8030920698407815, + "learning_rate": 0.00027796725820741834, + "loss": 1.2377, + "step": 10441 + }, + { + "epoch": 0.8031689870009999, + "learning_rate": 0.00027775710168121917, + "loss": 1.145, + "step": 10442 + }, + { + "epoch": 0.8032459041612183, + "learning_rate": 0.0002775470165230566, + "loss": 1.266, + "step": 10443 + }, + { + "epoch": 0.8033228213214368, + "learning_rate": 0.00027733700274519736, + "loss": 1.2017, + "step": 10444 + }, + { + "epoch": 0.8033997384816552, + "learning_rate": 0.0002771270603599055, + "loss": 1.1181, + "step": 10445 + }, + { + "epoch": 0.8034766556418736, + "learning_rate": 0.0002769171893794387, + "loss": 0.6137, + "step": 10446 + }, + { + "epoch": 0.8035535728020922, + "learning_rate": 0.0002767073898160514, + "loss": 1.1542, + "step": 10447 + }, + { + "epoch": 0.8036304899623106, + "learning_rate": 0.00027649766168199495, + "loss": 1.2338, + "step": 10448 + }, + { + "epoch": 0.8037074071225291, + "learning_rate": 0.00027628800498951517, + "loss": 0.9504, + "step": 10449 + }, + { + "epoch": 0.8037843242827475, + "learning_rate": 0.0002760784197508539, + "loss": 1.4597, + "step": 10450 + }, + { + "epoch": 0.8038612414429659, + "learning_rate": 0.00027586890597824953, + "loss": 1.0946, + "step": 10451 + }, + { + "epoch": 0.8039381586031844, + "learning_rate": 0.000275659463683935, + "loss": 1.1617, + "step": 10452 + }, + { + "epoch": 0.8040150757634028, + "learning_rate": 0.0002754500928801407, + "loss": 1.0196, + "step": 10453 + }, + { + "epoch": 0.8040919929236212, + "learning_rate": 0.00027524079357909203, + "loss": 1.1086, + "step": 10454 + }, + { + "epoch": 0.8041689100838397, + "learning_rate": 0.00027503156579300914, + "loss": 1.1331, + "step": 10455 + }, + { + "epoch": 0.8042458272440581, + "learning_rate": 0.0002748224095341101, + "loss": 1.089, + "step": 10456 + }, + { + "epoch": 0.8043227444042766, + "learning_rate": 0.0002746133248146071, + "loss": 0.9148, + "step": 10457 + }, + { + "epoch": 0.804399661564495, + "learning_rate": 0.00027440431164670975, + "loss": 1.2087, + "step": 10458 + }, + { + "epoch": 0.8044765787247135, + "learning_rate": 0.0002741953700426218, + "loss": 1.1022, + "step": 10459 + }, + { + "epoch": 0.804553495884932, + "learning_rate": 0.0002739865000145431, + "loss": 1.1847, + "step": 10460 + }, + { + "epoch": 0.8046304130451504, + "learning_rate": 0.0002737777015746708, + "loss": 1.2637, + "step": 10461 + }, + { + "epoch": 0.8047073302053688, + "learning_rate": 0.00027356897473519644, + "loss": 1.1909, + "step": 10462 + }, + { + "epoch": 0.8047842473655873, + "learning_rate": 0.00027336031950830794, + "loss": 1.1146, + "step": 10463 + }, + { + "epoch": 0.8048611645258057, + "learning_rate": 0.00027315173590618854, + "loss": 1.2136, + "step": 10464 + }, + { + "epoch": 0.8049380816860241, + "learning_rate": 0.000272943223941018, + "loss": 1.2451, + "step": 10465 + }, + { + "epoch": 0.8050149988462426, + "learning_rate": 0.0002727347836249711, + "loss": 1.2267, + "step": 10466 + }, + { + "epoch": 0.805091916006461, + "learning_rate": 0.0002725264149702198, + "loss": 1.2632, + "step": 10467 + }, + { + "epoch": 0.8051688331666795, + "learning_rate": 0.0002723181179889298, + "loss": 1.211, + "step": 10468 + }, + { + "epoch": 0.8052457503268979, + "learning_rate": 0.00027210989269326456, + "loss": 1.2043, + "step": 10469 + }, + { + "epoch": 0.8053226674871163, + "learning_rate": 0.0002719017390953824, + "loss": 1.116, + "step": 10470 + }, + { + "epoch": 0.8053995846473349, + "learning_rate": 0.0002716936572074377, + "loss": 1.2985, + "step": 10471 + }, + { + "epoch": 0.8054765018075533, + "learning_rate": 0.00027148564704158044, + "loss": 0.9905, + "step": 10472 + }, + { + "epoch": 0.8055534189677717, + "learning_rate": 0.0002712777086099564, + "loss": 1.1847, + "step": 10473 + }, + { + "epoch": 0.8056303361279902, + "learning_rate": 0.0002710698419247074, + "loss": 1.1079, + "step": 10474 + }, + { + "epoch": 0.8057072532882086, + "learning_rate": 0.0002708620469979712, + "loss": 1.5432, + "step": 10475 + }, + { + "epoch": 0.8057841704484271, + "learning_rate": 0.0002706543238418812, + "loss": 1.3157, + "step": 10476 + }, + { + "epoch": 0.8058610876086455, + "learning_rate": 0.0002704466724685665, + "loss": 0.9149, + "step": 10477 + }, + { + "epoch": 0.8059380047688639, + "learning_rate": 0.00027023909289015176, + "loss": 1.751, + "step": 10478 + }, + { + "epoch": 0.8060149219290824, + "learning_rate": 0.00027003158511875787, + "loss": 1.4602, + "step": 10479 + }, + { + "epoch": 0.8060918390893008, + "learning_rate": 0.0002698241491665021, + "loss": 1.0065, + "step": 10480 + }, + { + "epoch": 0.8061687562495192, + "learning_rate": 0.00026961678504549614, + "loss": 0.8948, + "step": 10481 + }, + { + "epoch": 0.8062456734097377, + "learning_rate": 0.0002694094927678479, + "loss": 1.0458, + "step": 10482 + }, + { + "epoch": 0.8063225905699561, + "learning_rate": 0.0002692022723456623, + "loss": 1.2696, + "step": 10483 + }, + { + "epoch": 0.8063995077301745, + "learning_rate": 0.00026899512379103873, + "loss": 1.0358, + "step": 10484 + }, + { + "epoch": 0.8064764248903931, + "learning_rate": 0.00026878804711607276, + "loss": 1.123, + "step": 10485 + }, + { + "epoch": 0.8065533420506115, + "learning_rate": 0.0002685810423328558, + "loss": 1.144, + "step": 10486 + }, + { + "epoch": 0.80663025921083, + "learning_rate": 0.0002683741094534749, + "loss": 1.4369, + "step": 10487 + }, + { + "epoch": 0.8067071763710484, + "learning_rate": 0.00026816724849001354, + "loss": 1.2492, + "step": 10488 + }, + { + "epoch": 0.8067840935312668, + "learning_rate": 0.00026796045945455077, + "loss": 1.2483, + "step": 10489 + }, + { + "epoch": 0.8068610106914853, + "learning_rate": 0.00026775374235916006, + "loss": 0.9917, + "step": 10490 + }, + { + "epoch": 0.8069379278517037, + "learning_rate": 0.0002675470972159129, + "loss": 1.4104, + "step": 10491 + }, + { + "epoch": 0.8070148450119221, + "learning_rate": 0.0002673405240368752, + "loss": 1.2326, + "step": 10492 + }, + { + "epoch": 0.8070917621721406, + "learning_rate": 0.00026713402283410877, + "loss": 0.9989, + "step": 10493 + }, + { + "epoch": 0.807168679332359, + "learning_rate": 0.0002669275936196718, + "loss": 1.417, + "step": 10494 + }, + { + "epoch": 0.8072455964925775, + "learning_rate": 0.0002667212364056173, + "loss": 1.1293, + "step": 10495 + }, + { + "epoch": 0.807322513652796, + "learning_rate": 0.0002665149512039955, + "loss": 0.891, + "step": 10496 + }, + { + "epoch": 0.8073994308130144, + "learning_rate": 0.0002663087380268511, + "loss": 1.0769, + "step": 10497 + }, + { + "epoch": 0.8074763479732329, + "learning_rate": 0.00026610259688622523, + "loss": 0.8323, + "step": 10498 + }, + { + "epoch": 0.8075532651334513, + "learning_rate": 0.00026589652779415485, + "loss": 1.1398, + "step": 10499 + }, + { + "epoch": 0.8076301822936697, + "learning_rate": 0.0002656905307626721, + "loss": 0.9009, + "step": 10500 + }, + { + "epoch": 0.8077070994538882, + "learning_rate": 0.00026548460580380547, + "loss": 1.4085, + "step": 10501 + }, + { + "epoch": 0.8077840166141066, + "learning_rate": 0.0002652787529295799, + "loss": 1.3816, + "step": 10502 + }, + { + "epoch": 0.8078609337743251, + "learning_rate": 0.0002650729721520141, + "loss": 1.3336, + "step": 10503 + }, + { + "epoch": 0.8079378509345435, + "learning_rate": 0.00026486726348312465, + "loss": 0.9549, + "step": 10504 + }, + { + "epoch": 0.8080147680947619, + "learning_rate": 0.000264661626934923, + "loss": 1.2421, + "step": 10505 + }, + { + "epoch": 0.8080916852549804, + "learning_rate": 0.0002644560625194165, + "loss": 0.9791, + "step": 10506 + }, + { + "epoch": 0.8081686024151988, + "learning_rate": 0.00026425057024860804, + "loss": 1.113, + "step": 10507 + }, + { + "epoch": 0.8082455195754172, + "learning_rate": 0.0002640451501344967, + "loss": 1.2074, + "step": 10508 + }, + { + "epoch": 0.8083224367356358, + "learning_rate": 0.00026383980218907686, + "loss": 0.786, + "step": 10509 + }, + { + "epoch": 0.8083993538958542, + "learning_rate": 0.0002636345264243395, + "loss": 1.3403, + "step": 10510 + }, + { + "epoch": 0.8084762710560726, + "learning_rate": 0.0002634293228522711, + "loss": 1.1428, + "step": 10511 + }, + { + "epoch": 0.8085531882162911, + "learning_rate": 0.0002632241914848525, + "loss": 1.1962, + "step": 10512 + }, + { + "epoch": 0.8086301053765095, + "learning_rate": 0.00026301913233406257, + "loss": 1.6917, + "step": 10513 + }, + { + "epoch": 0.808707022536728, + "learning_rate": 0.0002628141454118744, + "loss": 1.0791, + "step": 10514 + }, + { + "epoch": 0.8087839396969464, + "learning_rate": 0.0002626092307302582, + "loss": 0.9986, + "step": 10515 + }, + { + "epoch": 0.8088608568571648, + "learning_rate": 0.00026240438830117853, + "loss": 1.1085, + "step": 10516 + }, + { + "epoch": 0.8089377740173833, + "learning_rate": 0.00026219961813659585, + "loss": 1.2558, + "step": 10517 + }, + { + "epoch": 0.8090146911776017, + "learning_rate": 0.0002619949202484679, + "loss": 1.0907, + "step": 10518 + }, + { + "epoch": 0.8090916083378201, + "learning_rate": 0.00026179029464874667, + "loss": 0.9909, + "step": 10519 + }, + { + "epoch": 0.8091685254980386, + "learning_rate": 0.00026158574134938064, + "loss": 1.5166, + "step": 10520 + }, + { + "epoch": 0.809245442658257, + "learning_rate": 0.0002613812603623137, + "loss": 1.1873, + "step": 10521 + }, + { + "epoch": 0.8093223598184756, + "learning_rate": 0.00026117685169948563, + "loss": 1.0856, + "step": 10522 + }, + { + "epoch": 0.809399276978694, + "learning_rate": 0.00026097251537283257, + "loss": 1.3302, + "step": 10523 + }, + { + "epoch": 0.8094761941389124, + "learning_rate": 0.000260768251394286, + "loss": 1.1469, + "step": 10524 + }, + { + "epoch": 0.8095531112991309, + "learning_rate": 0.0002605640597757721, + "loss": 0.9616, + "step": 10525 + }, + { + "epoch": 0.8096300284593493, + "learning_rate": 0.00026035994052921456, + "loss": 0.9726, + "step": 10526 + }, + { + "epoch": 0.8097069456195677, + "learning_rate": 0.0002601558936665323, + "loss": 1.0679, + "step": 10527 + }, + { + "epoch": 0.8097838627797862, + "learning_rate": 0.0002599519191996394, + "loss": 0.9663, + "step": 10528 + }, + { + "epoch": 0.8098607799400046, + "learning_rate": 0.0002597480171404463, + "loss": 1.2546, + "step": 10529 + }, + { + "epoch": 0.809937697100223, + "learning_rate": 0.0002595441875008591, + "loss": 1.072, + "step": 10530 + }, + { + "epoch": 0.8100146142604415, + "learning_rate": 0.0002593404302927791, + "loss": 1.074, + "step": 10531 + }, + { + "epoch": 0.8100915314206599, + "learning_rate": 0.0002591367455281047, + "loss": 1.1389, + "step": 10532 + }, + { + "epoch": 0.8101684485808784, + "learning_rate": 0.00025893313321872894, + "loss": 1.0699, + "step": 10533 + }, + { + "epoch": 0.8102453657410968, + "learning_rate": 0.00025872959337654096, + "loss": 1.2549, + "step": 10534 + }, + { + "epoch": 0.8103222829013152, + "learning_rate": 0.0002585261260134257, + "loss": 1.3641, + "step": 10535 + }, + { + "epoch": 0.8103992000615338, + "learning_rate": 0.0002583227311412633, + "loss": 0.9831, + "step": 10536 + }, + { + "epoch": 0.8104761172217522, + "learning_rate": 0.0002581194087719315, + "loss": 1.3057, + "step": 10537 + }, + { + "epoch": 0.8105530343819706, + "learning_rate": 0.0002579161589173013, + "loss": 1.1308, + "step": 10538 + }, + { + "epoch": 0.8106299515421891, + "learning_rate": 0.00025771298158924066, + "loss": 1.0699, + "step": 10539 + }, + { + "epoch": 0.8107068687024075, + "learning_rate": 0.00025750987679961415, + "loss": 1.2459, + "step": 10540 + }, + { + "epoch": 0.810783785862626, + "learning_rate": 0.0002573068445602809, + "loss": 1.464, + "step": 10541 + }, + { + "epoch": 0.8108607030228444, + "learning_rate": 0.0002571038848830959, + "loss": 1.5286, + "step": 10542 + }, + { + "epoch": 0.8109376201830628, + "learning_rate": 0.0002569009977799104, + "loss": 1.2353, + "step": 10543 + }, + { + "epoch": 0.8110145373432813, + "learning_rate": 0.0002566981832625709, + "loss": 1.4405, + "step": 10544 + }, + { + "epoch": 0.8110914545034997, + "learning_rate": 0.0002564954413429204, + "loss": 1.166, + "step": 10545 + }, + { + "epoch": 0.8111683716637181, + "learning_rate": 0.00025629277203279715, + "loss": 0.9393, + "step": 10546 + }, + { + "epoch": 0.8112452888239366, + "learning_rate": 0.0002560901753440352, + "loss": 0.853, + "step": 10547 + }, + { + "epoch": 0.811322205984155, + "learning_rate": 0.00025588765128846413, + "loss": 1.2401, + "step": 10548 + }, + { + "epoch": 0.8113991231443735, + "learning_rate": 0.00025568519987790976, + "loss": 1.0539, + "step": 10549 + }, + { + "epoch": 0.811476040304592, + "learning_rate": 0.0002554828211241933, + "loss": 1.314, + "step": 10550 + }, + { + "epoch": 0.8115529574648104, + "learning_rate": 0.0002552805150391321, + "loss": 1.0729, + "step": 10551 + }, + { + "epoch": 0.8116298746250289, + "learning_rate": 0.00025507828163453834, + "loss": 1.1186, + "step": 10552 + }, + { + "epoch": 0.8117067917852473, + "learning_rate": 0.00025487612092222147, + "loss": 1.188, + "step": 10553 + }, + { + "epoch": 0.8117837089454657, + "learning_rate": 0.0002546740329139855, + "loss": 1.4008, + "step": 10554 + }, + { + "epoch": 0.8118606261056842, + "learning_rate": 0.0002544720176216307, + "loss": 1.1238, + "step": 10555 + }, + { + "epoch": 0.8119375432659026, + "learning_rate": 0.00025427007505695275, + "loss": 1.4683, + "step": 10556 + }, + { + "epoch": 0.812014460426121, + "learning_rate": 0.00025406820523174364, + "loss": 1.5652, + "step": 10557 + }, + { + "epoch": 0.8120913775863395, + "learning_rate": 0.00025386640815778985, + "loss": 1.0449, + "step": 10558 + }, + { + "epoch": 0.8121682947465579, + "learning_rate": 0.0002536646838468756, + "loss": 1.2106, + "step": 10559 + }, + { + "epoch": 0.8122452119067765, + "learning_rate": 0.00025346303231077943, + "loss": 1.4836, + "step": 10560 + }, + { + "epoch": 0.8123221290669949, + "learning_rate": 0.000253261453561276, + "loss": 0.9808, + "step": 10561 + }, + { + "epoch": 0.8123990462272133, + "learning_rate": 0.0002530599476101354, + "loss": 0.9315, + "step": 10562 + }, + { + "epoch": 0.8124759633874318, + "learning_rate": 0.00025285851446912405, + "loss": 0.9325, + "step": 10563 + }, + { + "epoch": 0.8125528805476502, + "learning_rate": 0.0002526571541500039, + "loss": 0.9049, + "step": 10564 + }, + { + "epoch": 0.8126297977078686, + "learning_rate": 0.00025245586666453236, + "loss": 1.0214, + "step": 10565 + }, + { + "epoch": 0.8127067148680871, + "learning_rate": 0.0002522546520244627, + "loss": 1.0878, + "step": 10566 + }, + { + "epoch": 0.8127836320283055, + "learning_rate": 0.0002520535102415445, + "loss": 1.2121, + "step": 10567 + }, + { + "epoch": 0.8128605491885239, + "learning_rate": 0.0002518524413275224, + "loss": 1.1159, + "step": 10568 + }, + { + "epoch": 0.8129374663487424, + "learning_rate": 0.00025165144529413716, + "loss": 0.9496, + "step": 10569 + }, + { + "epoch": 0.8130143835089608, + "learning_rate": 0.00025145052215312484, + "loss": 1.2616, + "step": 10570 + }, + { + "epoch": 0.8130913006691793, + "learning_rate": 0.0002512496719162175, + "loss": 1.0543, + "step": 10571 + }, + { + "epoch": 0.8131682178293977, + "learning_rate": 0.0002510488945951437, + "loss": 1.1819, + "step": 10572 + }, + { + "epoch": 0.8132451349896161, + "learning_rate": 0.0002508481902016269, + "loss": 1.0739, + "step": 10573 + }, + { + "epoch": 0.8133220521498347, + "learning_rate": 0.00025064755874738517, + "loss": 0.873, + "step": 10574 + }, + { + "epoch": 0.8133989693100531, + "learning_rate": 0.000250447000244135, + "loss": 1.2985, + "step": 10575 + }, + { + "epoch": 0.8134758864702715, + "learning_rate": 0.0002502465147035868, + "loss": 1.3009, + "step": 10576 + }, + { + "epoch": 0.81355280363049, + "learning_rate": 0.00025004610213744707, + "loss": 1.106, + "step": 10577 + }, + { + "epoch": 0.8136297207907084, + "learning_rate": 0.00024984576255741815, + "loss": 0.9423, + "step": 10578 + }, + { + "epoch": 0.8137066379509269, + "learning_rate": 0.00024964549597519754, + "loss": 1.2203, + "step": 10579 + }, + { + "epoch": 0.8137835551111453, + "learning_rate": 0.00024944530240247984, + "loss": 0.921, + "step": 10580 + }, + { + "epoch": 0.8138604722713637, + "learning_rate": 0.0002492451818509543, + "loss": 0.9101, + "step": 10581 + }, + { + "epoch": 0.8139373894315822, + "learning_rate": 0.0002490451343323061, + "loss": 1.046, + "step": 10582 + }, + { + "epoch": 0.8140143065918006, + "learning_rate": 0.0002488451598582164, + "loss": 0.9688, + "step": 10583 + }, + { + "epoch": 0.814091223752019, + "learning_rate": 0.00024864525844036153, + "loss": 0.7727, + "step": 10584 + }, + { + "epoch": 0.8141681409122375, + "learning_rate": 0.0002484454300904137, + "loss": 1.0728, + "step": 10585 + }, + { + "epoch": 0.814245058072456, + "learning_rate": 0.0002482456748200425, + "loss": 0.9359, + "step": 10586 + }, + { + "epoch": 0.8143219752326744, + "learning_rate": 0.00024804599264091046, + "loss": 1.4848, + "step": 10587 + }, + { + "epoch": 0.8143988923928929, + "learning_rate": 0.00024784638356467717, + "loss": 0.8717, + "step": 10588 + }, + { + "epoch": 0.8144758095531113, + "learning_rate": 0.00024764684760299893, + "loss": 1.0713, + "step": 10589 + }, + { + "epoch": 0.8145527267133298, + "learning_rate": 0.00024744738476752647, + "loss": 1.663, + "step": 10590 + }, + { + "epoch": 0.8146296438735482, + "learning_rate": 0.00024724799506990665, + "loss": 0.8926, + "step": 10591 + }, + { + "epoch": 0.8147065610337666, + "learning_rate": 0.00024704867852178193, + "loss": 0.9111, + "step": 10592 + }, + { + "epoch": 0.8147834781939851, + "learning_rate": 0.00024684943513479034, + "loss": 1.673, + "step": 10593 + }, + { + "epoch": 0.8148603953542035, + "learning_rate": 0.00024665026492056666, + "loss": 0.9266, + "step": 10594 + }, + { + "epoch": 0.8149373125144219, + "learning_rate": 0.00024645116789074044, + "loss": 1.0953, + "step": 10595 + }, + { + "epoch": 0.8150142296746404, + "learning_rate": 0.00024625214405693617, + "loss": 1.0139, + "step": 10596 + }, + { + "epoch": 0.8150911468348588, + "learning_rate": 0.00024605319343077637, + "loss": 0.9901, + "step": 10597 + }, + { + "epoch": 0.8151680639950774, + "learning_rate": 0.0002458543160238769, + "loss": 1.2695, + "step": 10598 + }, + { + "epoch": 0.8152449811552958, + "learning_rate": 0.0002456555118478519, + "loss": 1.116, + "step": 10599 + }, + { + "epoch": 0.8153218983155142, + "learning_rate": 0.00024545678091430817, + "loss": 1.2579, + "step": 10600 + }, + { + "epoch": 0.8153988154757327, + "learning_rate": 0.00024525812323485026, + "loss": 1.2549, + "step": 10601 + }, + { + "epoch": 0.8154757326359511, + "learning_rate": 0.0002450595388210787, + "loss": 1.6772, + "step": 10602 + }, + { + "epoch": 0.8155526497961695, + "learning_rate": 0.0002448610276845883, + "loss": 1.288, + "step": 10603 + }, + { + "epoch": 0.815629566956388, + "learning_rate": 0.00024466258983697064, + "loss": 1.0606, + "step": 10604 + }, + { + "epoch": 0.8157064841166064, + "learning_rate": 0.00024446422528981264, + "loss": 1.4465, + "step": 10605 + }, + { + "epoch": 0.8157834012768248, + "learning_rate": 0.00024426593405469694, + "loss": 1.2004, + "step": 10606 + }, + { + "epoch": 0.8158603184370433, + "learning_rate": 0.0002440677161432019, + "loss": 1.357, + "step": 10607 + }, + { + "epoch": 0.8159372355972617, + "learning_rate": 0.00024386957156690225, + "loss": 1.0785, + "step": 10608 + }, + { + "epoch": 0.8160141527574802, + "learning_rate": 0.00024367150033736696, + "loss": 1.1325, + "step": 10609 + }, + { + "epoch": 0.8160910699176986, + "learning_rate": 0.00024347350246616223, + "loss": 1.4211, + "step": 10610 + }, + { + "epoch": 0.816167987077917, + "learning_rate": 0.00024327557796484934, + "loss": 1.2836, + "step": 10611 + }, + { + "epoch": 0.8162449042381356, + "learning_rate": 0.0002430777268449852, + "loss": 1.1152, + "step": 10612 + }, + { + "epoch": 0.816321821398354, + "learning_rate": 0.00024287994911812255, + "loss": 1.3357, + "step": 10613 + }, + { + "epoch": 0.8163987385585724, + "learning_rate": 0.00024268224479580969, + "loss": 1.209, + "step": 10614 + }, + { + "epoch": 0.8164756557187909, + "learning_rate": 0.00024248461388959053, + "loss": 1.2637, + "step": 10615 + }, + { + "epoch": 0.8165525728790093, + "learning_rate": 0.00024228705641100585, + "loss": 0.9578, + "step": 10616 + }, + { + "epoch": 0.8166294900392278, + "learning_rate": 0.00024208957237159057, + "loss": 1.1937, + "step": 10617 + }, + { + "epoch": 0.8167064071994462, + "learning_rate": 0.00024189216178287614, + "loss": 1.1317, + "step": 10618 + }, + { + "epoch": 0.8167833243596646, + "learning_rate": 0.00024169482465638954, + "loss": 1.4476, + "step": 10619 + }, + { + "epoch": 0.8168602415198831, + "learning_rate": 0.00024149756100365317, + "loss": 1.1497, + "step": 10620 + }, + { + "epoch": 0.8169371586801015, + "learning_rate": 0.0002413003708361865, + "loss": 1.1194, + "step": 10621 + }, + { + "epoch": 0.8170140758403199, + "learning_rate": 0.0002411032541655025, + "loss": 0.917, + "step": 10622 + }, + { + "epoch": 0.8170909930005384, + "learning_rate": 0.00024090621100311126, + "loss": 1.1918, + "step": 10623 + }, + { + "epoch": 0.8171679101607568, + "learning_rate": 0.0002407092413605188, + "loss": 1.0003, + "step": 10624 + }, + { + "epoch": 0.8172448273209754, + "learning_rate": 0.00024051234524922623, + "loss": 1.2944, + "step": 10625 + }, + { + "epoch": 0.8173217444811938, + "learning_rate": 0.0002403155226807305, + "loss": 1.279, + "step": 10626 + }, + { + "epoch": 0.8173986616414122, + "learning_rate": 0.00024011877366652407, + "loss": 1.1764, + "step": 10627 + }, + { + "epoch": 0.8174755788016307, + "learning_rate": 0.00023992209821809514, + "loss": 1.0114, + "step": 10628 + }, + { + "epoch": 0.8175524959618491, + "learning_rate": 0.00023972549634692847, + "loss": 0.8522, + "step": 10629 + }, + { + "epoch": 0.8176294131220675, + "learning_rate": 0.0002395289680645039, + "loss": 1.3193, + "step": 10630 + }, + { + "epoch": 0.817706330282286, + "learning_rate": 0.00023933251338229577, + "loss": 1.3139, + "step": 10631 + }, + { + "epoch": 0.8177832474425044, + "learning_rate": 0.0002391361323117764, + "loss": 1.1277, + "step": 10632 + }, + { + "epoch": 0.8178601646027228, + "learning_rate": 0.0002389398248644124, + "loss": 1.0314, + "step": 10633 + }, + { + "epoch": 0.8179370817629413, + "learning_rate": 0.00023874359105166616, + "loss": 0.8421, + "step": 10634 + }, + { + "epoch": 0.8180139989231597, + "learning_rate": 0.00023854743088499598, + "loss": 1.0115, + "step": 10635 + }, + { + "epoch": 0.8180909160833783, + "learning_rate": 0.0002383513443758557, + "loss": 1.1188, + "step": 10636 + }, + { + "epoch": 0.8181678332435967, + "learning_rate": 0.0002381553315356957, + "loss": 0.9807, + "step": 10637 + }, + { + "epoch": 0.8182447504038151, + "learning_rate": 0.00023795939237596092, + "loss": 1.0553, + "step": 10638 + }, + { + "epoch": 0.8183216675640336, + "learning_rate": 0.00023776352690809245, + "loss": 1.022, + "step": 10639 + }, + { + "epoch": 0.818398584724252, + "learning_rate": 0.00023756773514352708, + "loss": 1.2374, + "step": 10640 + }, + { + "epoch": 0.8184755018844704, + "learning_rate": 0.00023737201709369754, + "loss": 1.1, + "step": 10641 + }, + { + "epoch": 0.8185524190446889, + "learning_rate": 0.00023717637277003117, + "loss": 0.9657, + "step": 10642 + }, + { + "epoch": 0.8186293362049073, + "learning_rate": 0.00023698080218395335, + "loss": 1.0946, + "step": 10643 + }, + { + "epoch": 0.8187062533651258, + "learning_rate": 0.00023678530534688207, + "loss": 1.2417, + "step": 10644 + }, + { + "epoch": 0.8187831705253442, + "learning_rate": 0.0002365898822702336, + "loss": 0.8935, + "step": 10645 + }, + { + "epoch": 0.8188600876855626, + "learning_rate": 0.00023639453296541858, + "loss": 1.094, + "step": 10646 + }, + { + "epoch": 0.8189370048457811, + "learning_rate": 0.00023619925744384373, + "loss": 1.0916, + "step": 10647 + }, + { + "epoch": 0.8190139220059995, + "learning_rate": 0.0002360040557169113, + "loss": 1.0222, + "step": 10648 + }, + { + "epoch": 0.8190908391662179, + "learning_rate": 0.00023580892779601949, + "loss": 1.042, + "step": 10649 + }, + { + "epoch": 0.8191677563264365, + "learning_rate": 0.00023561387369256133, + "loss": 1.0733, + "step": 10650 + }, + { + "epoch": 0.8192446734866549, + "learning_rate": 0.00023541889341792722, + "loss": 1.036, + "step": 10651 + }, + { + "epoch": 0.8193215906468733, + "learning_rate": 0.0002352239869835024, + "loss": 1.2857, + "step": 10652 + }, + { + "epoch": 0.8193985078070918, + "learning_rate": 0.00023502915440066607, + "loss": 0.8074, + "step": 10653 + }, + { + "epoch": 0.8194754249673102, + "learning_rate": 0.00023483439568079635, + "loss": 1.1465, + "step": 10654 + }, + { + "epoch": 0.8195523421275287, + "learning_rate": 0.00023463971083526424, + "loss": 1.1189, + "step": 10655 + }, + { + "epoch": 0.8196292592877471, + "learning_rate": 0.00023444509987543887, + "loss": 1.1119, + "step": 10656 + }, + { + "epoch": 0.8197061764479655, + "learning_rate": 0.0002342505628126828, + "loss": 1.2435, + "step": 10657 + }, + { + "epoch": 0.819783093608184, + "learning_rate": 0.00023405609965835512, + "loss": 0.6836, + "step": 10658 + }, + { + "epoch": 0.8198600107684024, + "learning_rate": 0.00023386171042381155, + "loss": 1.1743, + "step": 10659 + }, + { + "epoch": 0.8199369279286208, + "learning_rate": 0.00023366739512040204, + "loss": 1.7128, + "step": 10660 + }, + { + "epoch": 0.8200138450888393, + "learning_rate": 0.00023347315375947314, + "loss": 1.4693, + "step": 10661 + }, + { + "epoch": 0.8200907622490577, + "learning_rate": 0.00023327898635236682, + "loss": 1.5164, + "step": 10662 + }, + { + "epoch": 0.8201676794092763, + "learning_rate": 0.00023308489291042035, + "loss": 1.0043, + "step": 10663 + }, + { + "epoch": 0.8202445965694947, + "learning_rate": 0.00023289087344496745, + "loss": 1.3352, + "step": 10664 + }, + { + "epoch": 0.8203215137297131, + "learning_rate": 0.00023269692796733742, + "loss": 0.928, + "step": 10665 + }, + { + "epoch": 0.8203984308899316, + "learning_rate": 0.00023250305648885388, + "loss": 1.1806, + "step": 10666 + }, + { + "epoch": 0.82047534805015, + "learning_rate": 0.00023230925902083827, + "loss": 1.5782, + "step": 10667 + }, + { + "epoch": 0.8205522652103684, + "learning_rate": 0.000232115535574606, + "loss": 1.1112, + "step": 10668 + }, + { + "epoch": 0.8206291823705869, + "learning_rate": 0.00023192188616146908, + "loss": 1.3983, + "step": 10669 + }, + { + "epoch": 0.8207060995308053, + "learning_rate": 0.00023172831079273476, + "loss": 0.8936, + "step": 10670 + }, + { + "epoch": 0.8207830166910237, + "learning_rate": 0.000231534809479706, + "loss": 1.1173, + "step": 10671 + }, + { + "epoch": 0.8208599338512422, + "learning_rate": 0.00023134138223368145, + "loss": 1.4427, + "step": 10672 + }, + { + "epoch": 0.8209368510114606, + "learning_rate": 0.00023114802906595622, + "loss": 0.9714, + "step": 10673 + }, + { + "epoch": 0.8210137681716791, + "learning_rate": 0.0002309547499878198, + "loss": 1.3299, + "step": 10674 + }, + { + "epoch": 0.8210906853318976, + "learning_rate": 0.00023076154501055812, + "loss": 1.1207, + "step": 10675 + }, + { + "epoch": 0.821167602492116, + "learning_rate": 0.00023056841414545253, + "loss": 1.7848, + "step": 10676 + }, + { + "epoch": 0.8212445196523345, + "learning_rate": 0.00023037535740377986, + "loss": 1.3192, + "step": 10677 + }, + { + "epoch": 0.8213214368125529, + "learning_rate": 0.00023018237479681403, + "loss": 1.3084, + "step": 10678 + }, + { + "epoch": 0.8213983539727713, + "learning_rate": 0.0002299894663358224, + "loss": 1.2936, + "step": 10679 + }, + { + "epoch": 0.8214752711329898, + "learning_rate": 0.00022979663203206895, + "loss": 1.2973, + "step": 10680 + }, + { + "epoch": 0.8215521882932082, + "learning_rate": 0.0002296038718968142, + "loss": 0.8722, + "step": 10681 + }, + { + "epoch": 0.8216291054534267, + "learning_rate": 0.0002294111859413136, + "loss": 1.0495, + "step": 10682 + }, + { + "epoch": 0.8217060226136451, + "learning_rate": 0.00022921857417681785, + "loss": 1.1018, + "step": 10683 + }, + { + "epoch": 0.8217829397738635, + "learning_rate": 0.00022902603661457393, + "loss": 1.0434, + "step": 10684 + }, + { + "epoch": 0.821859856934082, + "learning_rate": 0.00022883357326582403, + "loss": 1.5451, + "step": 10685 + }, + { + "epoch": 0.8219367740943004, + "learning_rate": 0.00022864118414180686, + "loss": 1.0836, + "step": 10686 + }, + { + "epoch": 0.8220136912545188, + "learning_rate": 0.00022844886925375618, + "loss": 0.9152, + "step": 10687 + }, + { + "epoch": 0.8220906084147374, + "learning_rate": 0.00022825662861290075, + "loss": 1.0145, + "step": 10688 + }, + { + "epoch": 0.8221675255749558, + "learning_rate": 0.0002280644622304664, + "loss": 1.1943, + "step": 10689 + }, + { + "epoch": 0.8222444427351742, + "learning_rate": 0.00022787237011767365, + "loss": 1.2422, + "step": 10690 + }, + { + "epoch": 0.8223213598953927, + "learning_rate": 0.00022768035228573884, + "loss": 1.1322, + "step": 10691 + }, + { + "epoch": 0.8223982770556111, + "learning_rate": 0.0002274884087458744, + "loss": 1.1814, + "step": 10692 + }, + { + "epoch": 0.8224751942158296, + "learning_rate": 0.00022729653950928763, + "loss": 0.9778, + "step": 10693 + }, + { + "epoch": 0.822552111376048, + "learning_rate": 0.0002271047445871826, + "loss": 0.9629, + "step": 10694 + }, + { + "epoch": 0.8226290285362664, + "learning_rate": 0.0002269130239907582, + "loss": 0.7977, + "step": 10695 + }, + { + "epoch": 0.8227059456964849, + "learning_rate": 0.00022672137773120893, + "loss": 1.2588, + "step": 10696 + }, + { + "epoch": 0.8227828628567033, + "learning_rate": 0.00022652980581972553, + "loss": 0.876, + "step": 10697 + }, + { + "epoch": 0.8228597800169217, + "learning_rate": 0.00022633830826749392, + "loss": 0.8721, + "step": 10698 + }, + { + "epoch": 0.8229366971771402, + "learning_rate": 0.00022614688508569558, + "loss": 1.2949, + "step": 10699 + }, + { + "epoch": 0.8230136143373586, + "learning_rate": 0.00022595553628550847, + "loss": 1.1846, + "step": 10700 + }, + { + "epoch": 0.8230905314975772, + "learning_rate": 0.0002257642618781053, + "loss": 0.7855, + "step": 10701 + }, + { + "epoch": 0.8231674486577956, + "learning_rate": 0.00022557306187465494, + "loss": 1.0839, + "step": 10702 + }, + { + "epoch": 0.823244365818014, + "learning_rate": 0.00022538193628632164, + "loss": 0.9837, + "step": 10703 + }, + { + "epoch": 0.8233212829782325, + "learning_rate": 0.0002251908851242655, + "loss": 1.0355, + "step": 10704 + }, + { + "epoch": 0.8233982001384509, + "learning_rate": 0.00022499990839964224, + "loss": 0.9371, + "step": 10705 + }, + { + "epoch": 0.8234751172986693, + "learning_rate": 0.00022480900612360295, + "loss": 1.0933, + "step": 10706 + }, + { + "epoch": 0.8235520344588878, + "learning_rate": 0.00022461817830729435, + "loss": 1.2113, + "step": 10707 + }, + { + "epoch": 0.8236289516191062, + "learning_rate": 0.00022442742496186008, + "loss": 1.0762, + "step": 10708 + }, + { + "epoch": 0.8237058687793246, + "learning_rate": 0.0002242367460984377, + "loss": 1.0536, + "step": 10709 + }, + { + "epoch": 0.8237827859395431, + "learning_rate": 0.0002240461417281613, + "loss": 1.0907, + "step": 10710 + }, + { + "epoch": 0.8238597030997615, + "learning_rate": 0.00022385561186216046, + "loss": 1.6937, + "step": 10711 + }, + { + "epoch": 0.82393662025998, + "learning_rate": 0.00022366515651155999, + "loss": 1.5094, + "step": 10712 + }, + { + "epoch": 0.8240135374201984, + "learning_rate": 0.00022347477568748165, + "loss": 1.0817, + "step": 10713 + }, + { + "epoch": 0.8240904545804169, + "learning_rate": 0.00022328446940104185, + "loss": 1.0013, + "step": 10714 + }, + { + "epoch": 0.8241673717406354, + "learning_rate": 0.00022309423766335158, + "loss": 0.9153, + "step": 10715 + }, + { + "epoch": 0.8242442889008538, + "learning_rate": 0.00022290408048552008, + "loss": 1.051, + "step": 10716 + }, + { + "epoch": 0.8243212060610722, + "learning_rate": 0.00022271399787865016, + "loss": 1.0477, + "step": 10717 + }, + { + "epoch": 0.8243981232212907, + "learning_rate": 0.0002225239898538411, + "loss": 1.0299, + "step": 10718 + }, + { + "epoch": 0.8244750403815091, + "learning_rate": 0.00022233405642218767, + "loss": 1.0845, + "step": 10719 + }, + { + "epoch": 0.8245519575417276, + "learning_rate": 0.00022214419759477989, + "loss": 1.2588, + "step": 10720 + }, + { + "epoch": 0.824628874701946, + "learning_rate": 0.0002219544133827045, + "loss": 1.2467, + "step": 10721 + }, + { + "epoch": 0.8247057918621644, + "learning_rate": 0.00022176470379704283, + "loss": 1.2005, + "step": 10722 + }, + { + "epoch": 0.8247827090223829, + "learning_rate": 0.0002215750688488722, + "loss": 0.8573, + "step": 10723 + }, + { + "epoch": 0.8248596261826013, + "learning_rate": 0.0002213855085492657, + "loss": 1.2205, + "step": 10724 + }, + { + "epoch": 0.8249365433428197, + "learning_rate": 0.00022119602290929187, + "loss": 1.3672, + "step": 10725 + }, + { + "epoch": 0.8250134605030383, + "learning_rate": 0.00022100661194001482, + "loss": 1.318, + "step": 10726 + }, + { + "epoch": 0.8250903776632567, + "learning_rate": 0.00022081727565249505, + "loss": 1.3921, + "step": 10727 + }, + { + "epoch": 0.8251672948234752, + "learning_rate": 0.00022062801405778722, + "loss": 1.3497, + "step": 10728 + }, + { + "epoch": 0.8252442119836936, + "learning_rate": 0.00022043882716694314, + "loss": 0.9178, + "step": 10729 + }, + { + "epoch": 0.825321129143912, + "learning_rate": 0.00022024971499100965, + "loss": 0.969, + "step": 10730 + }, + { + "epoch": 0.8253980463041305, + "learning_rate": 0.00022006067754102893, + "loss": 0.9037, + "step": 10731 + }, + { + "epoch": 0.8254749634643489, + "learning_rate": 0.00021987171482803914, + "loss": 1.1075, + "step": 10732 + }, + { + "epoch": 0.8255518806245673, + "learning_rate": 0.00021968282686307412, + "loss": 1.393, + "step": 10733 + }, + { + "epoch": 0.8256287977847858, + "learning_rate": 0.0002194940136571627, + "loss": 1.3812, + "step": 10734 + }, + { + "epoch": 0.8257057149450042, + "learning_rate": 0.00021930527522133077, + "loss": 0.9854, + "step": 10735 + }, + { + "epoch": 0.8257826321052226, + "learning_rate": 0.0002191166115665988, + "loss": 0.9693, + "step": 10736 + }, + { + "epoch": 0.8258595492654411, + "learning_rate": 0.00021892802270398216, + "loss": 1.1318, + "step": 10737 + }, + { + "epoch": 0.8259364664256595, + "learning_rate": 0.0002187395086444937, + "loss": 1.0286, + "step": 10738 + }, + { + "epoch": 0.8260133835858781, + "learning_rate": 0.0002185510693991403, + "loss": 1.1151, + "step": 10739 + }, + { + "epoch": 0.8260903007460965, + "learning_rate": 0.0002183627049789262, + "loss": 1.4246, + "step": 10740 + }, + { + "epoch": 0.8261672179063149, + "learning_rate": 0.000218174415394849, + "loss": 1.0629, + "step": 10741 + }, + { + "epoch": 0.8262441350665334, + "learning_rate": 0.0002179862006579033, + "loss": 1.3737, + "step": 10742 + }, + { + "epoch": 0.8263210522267518, + "learning_rate": 0.00021779806077907966, + "loss": 1.0107, + "step": 10743 + }, + { + "epoch": 0.8263979693869702, + "learning_rate": 0.00021760999576936357, + "loss": 1.1733, + "step": 10744 + }, + { + "epoch": 0.8264748865471887, + "learning_rate": 0.0002174220056397363, + "loss": 1.6141, + "step": 10745 + }, + { + "epoch": 0.8265518037074071, + "learning_rate": 0.00021723409040117487, + "loss": 1.1024, + "step": 10746 + }, + { + "epoch": 0.8266287208676256, + "learning_rate": 0.00021704625006465134, + "loss": 1.4375, + "step": 10747 + }, + { + "epoch": 0.826705638027844, + "learning_rate": 0.0002168584846411348, + "loss": 1.1996, + "step": 10748 + }, + { + "epoch": 0.8267825551880624, + "learning_rate": 0.00021667079414158875, + "loss": 1.2223, + "step": 10749 + }, + { + "epoch": 0.8268594723482809, + "learning_rate": 0.00021648317857697198, + "loss": 1.2452, + "step": 10750 + }, + { + "epoch": 0.8269363895084993, + "learning_rate": 0.00021629563795824035, + "loss": 1.4031, + "step": 10751 + }, + { + "epoch": 0.8270133066687178, + "learning_rate": 0.00021610817229634417, + "loss": 0.9523, + "step": 10752 + }, + { + "epoch": 0.8270902238289363, + "learning_rate": 0.00021592078160222994, + "loss": 0.6703, + "step": 10753 + }, + { + "epoch": 0.8271671409891547, + "learning_rate": 0.00021573346588683946, + "loss": 1.1687, + "step": 10754 + }, + { + "epoch": 0.8272440581493731, + "learning_rate": 0.00021554622516111043, + "loss": 1.5381, + "step": 10755 + }, + { + "epoch": 0.8273209753095916, + "learning_rate": 0.0002153590594359755, + "loss": 1.1097, + "step": 10756 + }, + { + "epoch": 0.82739789246981, + "learning_rate": 0.0002151719687223642, + "loss": 0.9915, + "step": 10757 + }, + { + "epoch": 0.8274748096300285, + "learning_rate": 0.00021498495303120085, + "loss": 1.5455, + "step": 10758 + }, + { + "epoch": 0.8275517267902469, + "learning_rate": 0.00021479801237340518, + "loss": 1.4729, + "step": 10759 + }, + { + "epoch": 0.8276286439504653, + "learning_rate": 0.00021461114675989297, + "loss": 1.0881, + "step": 10760 + }, + { + "epoch": 0.8277055611106838, + "learning_rate": 0.00021442435620157513, + "loss": 1.2372, + "step": 10761 + }, + { + "epoch": 0.8277824782709022, + "learning_rate": 0.00021423764070935963, + "loss": 1.3053, + "step": 10762 + }, + { + "epoch": 0.8278593954311206, + "learning_rate": 0.0002140510002941481, + "loss": 1.3, + "step": 10763 + }, + { + "epoch": 0.8279363125913392, + "learning_rate": 0.00021386443496683826, + "loss": 0.9578, + "step": 10764 + }, + { + "epoch": 0.8280132297515576, + "learning_rate": 0.00021367794473832496, + "loss": 1.2897, + "step": 10765 + }, + { + "epoch": 0.8280901469117761, + "learning_rate": 0.00021349152961949697, + "loss": 1.1766, + "step": 10766 + }, + { + "epoch": 0.8281670640719945, + "learning_rate": 0.0002133051896212394, + "loss": 1.1793, + "step": 10767 + }, + { + "epoch": 0.8282439812322129, + "learning_rate": 0.00021311892475443258, + "loss": 0.8118, + "step": 10768 + }, + { + "epoch": 0.8283208983924314, + "learning_rate": 0.00021293273502995274, + "loss": 1.0039, + "step": 10769 + }, + { + "epoch": 0.8283978155526498, + "learning_rate": 0.0002127466204586721, + "loss": 0.9122, + "step": 10770 + }, + { + "epoch": 0.8284747327128682, + "learning_rate": 0.00021256058105145826, + "loss": 1.2355, + "step": 10771 + }, + { + "epoch": 0.8285516498730867, + "learning_rate": 0.00021237461681917303, + "loss": 0.9744, + "step": 10772 + }, + { + "epoch": 0.8286285670333051, + "learning_rate": 0.00021218872777267623, + "loss": 0.9132, + "step": 10773 + }, + { + "epoch": 0.8287054841935235, + "learning_rate": 0.00021200291392282184, + "loss": 1.4135, + "step": 10774 + }, + { + "epoch": 0.828782401353742, + "learning_rate": 0.00021181717528045952, + "loss": 1.5483, + "step": 10775 + }, + { + "epoch": 0.8288593185139604, + "learning_rate": 0.00021163151185643492, + "loss": 1.1269, + "step": 10776 + }, + { + "epoch": 0.828936235674179, + "learning_rate": 0.00021144592366158872, + "loss": 1.0945, + "step": 10777 + }, + { + "epoch": 0.8290131528343974, + "learning_rate": 0.00021126041070675821, + "loss": 1.1136, + "step": 10778 + }, + { + "epoch": 0.8290900699946158, + "learning_rate": 0.00021107497300277544, + "loss": 0.9133, + "step": 10779 + }, + { + "epoch": 0.8291669871548343, + "learning_rate": 0.00021088961056046822, + "loss": 0.9289, + "step": 10780 + }, + { + "epoch": 0.8292439043150527, + "learning_rate": 0.00021070432339066026, + "loss": 1.1636, + "step": 10781 + }, + { + "epoch": 0.8293208214752711, + "learning_rate": 0.00021051911150417037, + "loss": 1.0531, + "step": 10782 + }, + { + "epoch": 0.8293977386354896, + "learning_rate": 0.00021033397491181322, + "loss": 1.2099, + "step": 10783 + }, + { + "epoch": 0.829474655795708, + "learning_rate": 0.00021014891362439986, + "loss": 1.0567, + "step": 10784 + }, + { + "epoch": 0.8295515729559265, + "learning_rate": 0.00020996392765273514, + "loss": 0.7669, + "step": 10785 + }, + { + "epoch": 0.8296284901161449, + "learning_rate": 0.0002097790170076214, + "loss": 1.1889, + "step": 10786 + }, + { + "epoch": 0.8297054072763633, + "learning_rate": 0.00020959418169985556, + "loss": 0.9922, + "step": 10787 + }, + { + "epoch": 0.8297823244365818, + "learning_rate": 0.0002094094217402301, + "loss": 1.3012, + "step": 10788 + }, + { + "epoch": 0.8298592415968002, + "learning_rate": 0.00020922473713953373, + "loss": 1.0292, + "step": 10789 + }, + { + "epoch": 0.8299361587570186, + "learning_rate": 0.00020904012790855014, + "loss": 1.3112, + "step": 10790 + }, + { + "epoch": 0.8300130759172372, + "learning_rate": 0.00020885559405805842, + "loss": 1.3005, + "step": 10791 + }, + { + "epoch": 0.8300899930774556, + "learning_rate": 0.0002086711355988346, + "loss": 1.5351, + "step": 10792 + }, + { + "epoch": 0.830166910237674, + "learning_rate": 0.00020848675254164922, + "loss": 1.3236, + "step": 10793 + }, + { + "epoch": 0.8302438273978925, + "learning_rate": 0.00020830244489726784, + "loss": 1.219, + "step": 10794 + }, + { + "epoch": 0.8303207445581109, + "learning_rate": 0.00020811821267645303, + "loss": 1.4288, + "step": 10795 + }, + { + "epoch": 0.8303976617183294, + "learning_rate": 0.00020793405588996202, + "loss": 1.5462, + "step": 10796 + }, + { + "epoch": 0.8304745788785478, + "learning_rate": 0.0002077499745485485, + "loss": 1.122, + "step": 10797 + }, + { + "epoch": 0.8305514960387662, + "learning_rate": 0.0002075659686629604, + "loss": 1.2538, + "step": 10798 + }, + { + "epoch": 0.8306284131989847, + "learning_rate": 0.00020738203824394213, + "loss": 1.1927, + "step": 10799 + }, + { + "epoch": 0.8307053303592031, + "learning_rate": 0.0002071981833022341, + "loss": 1.1776, + "step": 10800 + }, + { + "epoch": 0.8307822475194215, + "learning_rate": 0.00020701440384857139, + "loss": 1.1312, + "step": 10801 + }, + { + "epoch": 0.83085916467964, + "learning_rate": 0.00020683069989368524, + "loss": 1.0845, + "step": 10802 + }, + { + "epoch": 0.8309360818398585, + "learning_rate": 0.00020664707144830207, + "loss": 1.2887, + "step": 10803 + }, + { + "epoch": 0.831012999000077, + "learning_rate": 0.00020646351852314415, + "loss": 1.193, + "step": 10804 + }, + { + "epoch": 0.8310899161602954, + "learning_rate": 0.00020628004112892973, + "loss": 1.2198, + "step": 10805 + }, + { + "epoch": 0.8311668333205138, + "learning_rate": 0.0002060966392763724, + "loss": 1.1254, + "step": 10806 + }, + { + "epoch": 0.8312437504807323, + "learning_rate": 0.00020591331297617994, + "loss": 1.6282, + "step": 10807 + }, + { + "epoch": 0.8313206676409507, + "learning_rate": 0.00020573006223905828, + "loss": 1.0798, + "step": 10808 + }, + { + "epoch": 0.8313975848011691, + "learning_rate": 0.00020554688707570703, + "loss": 1.1077, + "step": 10809 + }, + { + "epoch": 0.8314745019613876, + "learning_rate": 0.00020536378749682227, + "loss": 1.4114, + "step": 10810 + }, + { + "epoch": 0.831551419121606, + "learning_rate": 0.00020518076351309502, + "loss": 1.2684, + "step": 10811 + }, + { + "epoch": 0.8316283362818244, + "learning_rate": 0.00020499781513521248, + "loss": 1.2452, + "step": 10812 + }, + { + "epoch": 0.8317052534420429, + "learning_rate": 0.00020481494237385684, + "loss": 1.1362, + "step": 10813 + }, + { + "epoch": 0.8317821706022613, + "learning_rate": 0.00020463214523970668, + "loss": 1.2861, + "step": 10814 + }, + { + "epoch": 0.8318590877624799, + "learning_rate": 0.00020444942374343584, + "loss": 0.899, + "step": 10815 + }, + { + "epoch": 0.8319360049226983, + "learning_rate": 0.00020426677789571308, + "loss": 1.0705, + "step": 10816 + }, + { + "epoch": 0.8320129220829167, + "learning_rate": 0.0002040842077072037, + "loss": 0.8464, + "step": 10817 + }, + { + "epoch": 0.8320898392431352, + "learning_rate": 0.0002039017131885677, + "loss": 1.2695, + "step": 10818 + }, + { + "epoch": 0.8321667564033536, + "learning_rate": 0.00020371929435046188, + "loss": 1.0746, + "step": 10819 + }, + { + "epoch": 0.832243673563572, + "learning_rate": 0.00020353695120353723, + "loss": 0.9288, + "step": 10820 + }, + { + "epoch": 0.8323205907237905, + "learning_rate": 0.00020335468375844086, + "loss": 1.1749, + "step": 10821 + }, + { + "epoch": 0.8323975078840089, + "learning_rate": 0.0002031724920258161, + "loss": 1.0873, + "step": 10822 + }, + { + "epoch": 0.8324744250442274, + "learning_rate": 0.000202990376016301, + "loss": 1.2259, + "step": 10823 + }, + { + "epoch": 0.8325513422044458, + "learning_rate": 0.00020280833574052966, + "loss": 1.2233, + "step": 10824 + }, + { + "epoch": 0.8326282593646642, + "learning_rate": 0.0002026263712091314, + "loss": 1.1299, + "step": 10825 + }, + { + "epoch": 0.8327051765248827, + "learning_rate": 0.0002024444824327311, + "loss": 1.2207, + "step": 10826 + }, + { + "epoch": 0.8327820936851011, + "learning_rate": 0.00020226266942195025, + "loss": 1.6364, + "step": 10827 + }, + { + "epoch": 0.8328590108453195, + "learning_rate": 0.0002020809321874047, + "loss": 0.9727, + "step": 10828 + }, + { + "epoch": 0.8329359280055381, + "learning_rate": 0.0002018992707397056, + "loss": 1.3102, + "step": 10829 + }, + { + "epoch": 0.8330128451657565, + "learning_rate": 0.0002017176850894613, + "loss": 1.2125, + "step": 10830 + }, + { + "epoch": 0.8330897623259749, + "learning_rate": 0.00020153617524727436, + "loss": 1.2022, + "step": 10831 + }, + { + "epoch": 0.8331666794861934, + "learning_rate": 0.00020135474122374343, + "loss": 0.9467, + "step": 10832 + }, + { + "epoch": 0.8332435966464118, + "learning_rate": 0.00020117338302946252, + "loss": 0.5734, + "step": 10833 + }, + { + "epoch": 0.8333205138066303, + "learning_rate": 0.00020099210067502104, + "loss": 1.1923, + "step": 10834 + }, + { + "epoch": 0.8333974309668487, + "learning_rate": 0.000200810894171005, + "loss": 0.9654, + "step": 10835 + }, + { + "epoch": 0.8334743481270671, + "learning_rate": 0.00020062976352799478, + "loss": 1.4249, + "step": 10836 + }, + { + "epoch": 0.8335512652872856, + "learning_rate": 0.0002004487087565669, + "loss": 1.0348, + "step": 10837 + }, + { + "epoch": 0.833628182447504, + "learning_rate": 0.00020026772986729324, + "loss": 0.9627, + "step": 10838 + }, + { + "epoch": 0.8337050996077224, + "learning_rate": 0.00020008682687074137, + "loss": 1.2621, + "step": 10839 + }, + { + "epoch": 0.833782016767941, + "learning_rate": 0.00019990599977747398, + "loss": 1.4122, + "step": 10840 + }, + { + "epoch": 0.8338589339281594, + "learning_rate": 0.00019972524859805096, + "loss": 0.7352, + "step": 10841 + }, + { + "epoch": 0.8339358510883779, + "learning_rate": 0.00019954457334302522, + "loss": 1.0593, + "step": 10842 + }, + { + "epoch": 0.8340127682485963, + "learning_rate": 0.00019936397402294732, + "loss": 0.8924, + "step": 10843 + }, + { + "epoch": 0.8340896854088147, + "learning_rate": 0.00019918345064836245, + "loss": 1.2799, + "step": 10844 + }, + { + "epoch": 0.8341666025690332, + "learning_rate": 0.00019900300322981174, + "loss": 1.2259, + "step": 10845 + }, + { + "epoch": 0.8342435197292516, + "learning_rate": 0.00019882263177783139, + "loss": 0.8012, + "step": 10846 + }, + { + "epoch": 0.83432043688947, + "learning_rate": 0.00019864233630295365, + "loss": 1.1015, + "step": 10847 + }, + { + "epoch": 0.8343973540496885, + "learning_rate": 0.00019846211681570592, + "loss": 1.0674, + "step": 10848 + }, + { + "epoch": 0.8344742712099069, + "learning_rate": 0.0001982819733266118, + "loss": 1.1149, + "step": 10849 + }, + { + "epoch": 0.8345511883701254, + "learning_rate": 0.00019810190584618987, + "loss": 1.0664, + "step": 10850 + }, + { + "epoch": 0.8346281055303438, + "learning_rate": 0.00019792191438495454, + "loss": 1.2631, + "step": 10851 + }, + { + "epoch": 0.8347050226905622, + "learning_rate": 0.0001977419989534156, + "loss": 0.9738, + "step": 10852 + }, + { + "epoch": 0.8347819398507808, + "learning_rate": 0.0001975621595620783, + "loss": 1.1769, + "step": 10853 + }, + { + "epoch": 0.8348588570109992, + "learning_rate": 0.0001973823962214444, + "loss": 1.1159, + "step": 10854 + }, + { + "epoch": 0.8349357741712176, + "learning_rate": 0.00019720270894200964, + "loss": 1.1199, + "step": 10855 + }, + { + "epoch": 0.8350126913314361, + "learning_rate": 0.00019702309773426618, + "loss": 1.0585, + "step": 10856 + }, + { + "epoch": 0.8350896084916545, + "learning_rate": 0.00019684356260870212, + "loss": 0.7886, + "step": 10857 + }, + { + "epoch": 0.8351665256518729, + "learning_rate": 0.00019666410357580073, + "loss": 1.6249, + "step": 10858 + }, + { + "epoch": 0.8352434428120914, + "learning_rate": 0.00019648472064604062, + "loss": 1.015, + "step": 10859 + }, + { + "epoch": 0.8353203599723098, + "learning_rate": 0.00019630541382989608, + "loss": 1.0349, + "step": 10860 + }, + { + "epoch": 0.8353972771325283, + "learning_rate": 0.00019612618313783693, + "loss": 1.2132, + "step": 10861 + }, + { + "epoch": 0.8354741942927467, + "learning_rate": 0.0001959470285803289, + "loss": 1.3673, + "step": 10862 + }, + { + "epoch": 0.8355511114529651, + "learning_rate": 0.00019576795016783304, + "loss": 0.7714, + "step": 10863 + }, + { + "epoch": 0.8356280286131836, + "learning_rate": 0.00019558894791080596, + "loss": 0.9195, + "step": 10864 + }, + { + "epoch": 0.835704945773402, + "learning_rate": 0.00019541002181969942, + "loss": 1.4968, + "step": 10865 + }, + { + "epoch": 0.8357818629336204, + "learning_rate": 0.00019523117190496143, + "loss": 0.7924, + "step": 10866 + }, + { + "epoch": 0.835858780093839, + "learning_rate": 0.00019505239817703513, + "loss": 1.1613, + "step": 10867 + }, + { + "epoch": 0.8359356972540574, + "learning_rate": 0.00019487370064635918, + "loss": 1.0735, + "step": 10868 + }, + { + "epoch": 0.8360126144142759, + "learning_rate": 0.00019469507932336772, + "loss": 1.4261, + "step": 10869 + }, + { + "epoch": 0.8360895315744943, + "learning_rate": 0.00019451653421849124, + "loss": 1.076, + "step": 10870 + }, + { + "epoch": 0.8361664487347127, + "learning_rate": 0.00019433806534215493, + "loss": 1.3461, + "step": 10871 + }, + { + "epoch": 0.8362433658949312, + "learning_rate": 0.00019415967270477975, + "loss": 0.9975, + "step": 10872 + }, + { + "epoch": 0.8363202830551496, + "learning_rate": 0.0001939813563167822, + "loss": 1.06, + "step": 10873 + }, + { + "epoch": 0.836397200215368, + "learning_rate": 0.00019380311618857432, + "loss": 1.4747, + "step": 10874 + }, + { + "epoch": 0.8364741173755865, + "learning_rate": 0.00019362495233056344, + "loss": 1.2087, + "step": 10875 + }, + { + "epoch": 0.8365510345358049, + "learning_rate": 0.0001934468647531532, + "loss": 1.3013, + "step": 10876 + }, + { + "epoch": 0.8366279516960233, + "learning_rate": 0.00019326885346674283, + "loss": 0.9965, + "step": 10877 + }, + { + "epoch": 0.8367048688562418, + "learning_rate": 0.00019309091848172517, + "loss": 1.1463, + "step": 10878 + }, + { + "epoch": 0.8367817860164602, + "learning_rate": 0.00019291305980849105, + "loss": 1.2645, + "step": 10879 + }, + { + "epoch": 0.8368587031766788, + "learning_rate": 0.00019273527745742537, + "loss": 1.1815, + "step": 10880 + }, + { + "epoch": 0.8369356203368972, + "learning_rate": 0.00019255757143890977, + "loss": 1.062, + "step": 10881 + }, + { + "epoch": 0.8370125374971156, + "learning_rate": 0.00019237994176331984, + "loss": 1.0678, + "step": 10882 + }, + { + "epoch": 0.8370894546573341, + "learning_rate": 0.00019220238844102762, + "loss": 1.1956, + "step": 10883 + }, + { + "epoch": 0.8371663718175525, + "learning_rate": 0.00019202491148240125, + "loss": 1.1834, + "step": 10884 + }, + { + "epoch": 0.8372432889777709, + "learning_rate": 0.0001918475108978035, + "loss": 1.1271, + "step": 10885 + }, + { + "epoch": 0.8373202061379894, + "learning_rate": 0.00019167018669759273, + "loss": 1.2681, + "step": 10886 + }, + { + "epoch": 0.8373971232982078, + "learning_rate": 0.00019149293889212334, + "loss": 1.3992, + "step": 10887 + }, + { + "epoch": 0.8374740404584263, + "learning_rate": 0.00019131576749174467, + "loss": 1.0591, + "step": 10888 + }, + { + "epoch": 0.8375509576186447, + "learning_rate": 0.0001911386725068025, + "loss": 1.1505, + "step": 10889 + }, + { + "epoch": 0.8376278747788631, + "learning_rate": 0.00019096165394763754, + "loss": 0.7873, + "step": 10890 + }, + { + "epoch": 0.8377047919390816, + "learning_rate": 0.00019078471182458535, + "loss": 1.3096, + "step": 10891 + }, + { + "epoch": 0.8377817090993, + "learning_rate": 0.00019060784614797848, + "loss": 1.1801, + "step": 10892 + }, + { + "epoch": 0.8378586262595185, + "learning_rate": 0.0001904310569281442, + "loss": 1.2087, + "step": 10893 + }, + { + "epoch": 0.837935543419737, + "learning_rate": 0.00019025434417540527, + "loss": 1.1831, + "step": 10894 + }, + { + "epoch": 0.8380124605799554, + "learning_rate": 0.00019007770790008006, + "loss": 1.3378, + "step": 10895 + }, + { + "epoch": 0.8380893777401738, + "learning_rate": 0.00018990114811248283, + "loss": 1.2176, + "step": 10896 + }, + { + "epoch": 0.8381662949003923, + "learning_rate": 0.0001897246648229225, + "loss": 0.8967, + "step": 10897 + }, + { + "epoch": 0.8382432120606107, + "learning_rate": 0.00018954825804170483, + "loss": 1.0882, + "step": 10898 + }, + { + "epoch": 0.8383201292208292, + "learning_rate": 0.0001893719277791301, + "loss": 1.1616, + "step": 10899 + }, + { + "epoch": 0.8383970463810476, + "learning_rate": 0.00018919567404549438, + "loss": 1.2519, + "step": 10900 + }, + { + "epoch": 0.838473963541266, + "learning_rate": 0.00018901949685108945, + "loss": 1.2597, + "step": 10901 + }, + { + "epoch": 0.8385508807014845, + "learning_rate": 0.00018884339620620205, + "loss": 1.0918, + "step": 10902 + }, + { + "epoch": 0.8386277978617029, + "learning_rate": 0.0001886673721211157, + "loss": 1.1742, + "step": 10903 + }, + { + "epoch": 0.8387047150219213, + "learning_rate": 0.00018849142460610792, + "loss": 1.1534, + "step": 10904 + }, + { + "epoch": 0.8387816321821399, + "learning_rate": 0.00018831555367145237, + "loss": 1.1128, + "step": 10905 + }, + { + "epoch": 0.8388585493423583, + "learning_rate": 0.00018813975932741882, + "loss": 1.1676, + "step": 10906 + }, + { + "epoch": 0.8389354665025768, + "learning_rate": 0.0001879640415842721, + "loss": 0.8937, + "step": 10907 + }, + { + "epoch": 0.8390123836627952, + "learning_rate": 0.00018778840045227213, + "loss": 1.0896, + "step": 10908 + }, + { + "epoch": 0.8390893008230136, + "learning_rate": 0.0001876128359416752, + "loss": 1.0042, + "step": 10909 + }, + { + "epoch": 0.8391662179832321, + "learning_rate": 0.00018743734806273216, + "loss": 1.146, + "step": 10910 + }, + { + "epoch": 0.8392431351434505, + "learning_rate": 0.0001872619368256906, + "loss": 1.3506, + "step": 10911 + }, + { + "epoch": 0.8393200523036689, + "learning_rate": 0.00018708660224079298, + "loss": 1.035, + "step": 10912 + }, + { + "epoch": 0.8393969694638874, + "learning_rate": 0.00018691134431827634, + "loss": 0.9361, + "step": 10913 + }, + { + "epoch": 0.8394738866241058, + "learning_rate": 0.00018673616306837492, + "loss": 1.183, + "step": 10914 + }, + { + "epoch": 0.8395508037843242, + "learning_rate": 0.00018656105850131777, + "loss": 0.7885, + "step": 10915 + }, + { + "epoch": 0.8396277209445427, + "learning_rate": 0.00018638603062732918, + "loss": 1.3637, + "step": 10916 + }, + { + "epoch": 0.8397046381047611, + "learning_rate": 0.0001862110794566293, + "loss": 1.2342, + "step": 10917 + }, + { + "epoch": 0.8397815552649797, + "learning_rate": 0.0001860362049994333, + "loss": 1.1524, + "step": 10918 + }, + { + "epoch": 0.8398584724251981, + "learning_rate": 0.0001858614072659532, + "loss": 1.0019, + "step": 10919 + }, + { + "epoch": 0.8399353895854165, + "learning_rate": 0.0001856866862663948, + "loss": 1.1074, + "step": 10920 + }, + { + "epoch": 0.840012306745635, + "learning_rate": 0.00018551204201096077, + "loss": 1.1043, + "step": 10921 + }, + { + "epoch": 0.8400892239058534, + "learning_rate": 0.00018533747450984834, + "loss": 1.2372, + "step": 10922 + }, + { + "epoch": 0.8401661410660718, + "learning_rate": 0.00018516298377325096, + "loss": 0.9992, + "step": 10923 + }, + { + "epoch": 0.8402430582262903, + "learning_rate": 0.00018498856981135703, + "loss": 1.3434, + "step": 10924 + }, + { + "epoch": 0.8403199753865087, + "learning_rate": 0.00018481423263435153, + "loss": 1.2279, + "step": 10925 + }, + { + "epoch": 0.8403968925467272, + "learning_rate": 0.00018463997225241302, + "loss": 0.9389, + "step": 10926 + }, + { + "epoch": 0.8404738097069456, + "learning_rate": 0.00018446578867571767, + "loss": 1.5451, + "step": 10927 + }, + { + "epoch": 0.840550726867164, + "learning_rate": 0.00018429168191443586, + "loss": 0.9528, + "step": 10928 + }, + { + "epoch": 0.8406276440273825, + "learning_rate": 0.0001841176519787341, + "loss": 1.0809, + "step": 10929 + }, + { + "epoch": 0.840704561187601, + "learning_rate": 0.00018394369887877394, + "loss": 1.2432, + "step": 10930 + }, + { + "epoch": 0.8407814783478194, + "learning_rate": 0.00018376982262471276, + "loss": 1.2846, + "step": 10931 + }, + { + "epoch": 0.8408583955080379, + "learning_rate": 0.0001835960232267031, + "loss": 1.5822, + "step": 10932 + }, + { + "epoch": 0.8409353126682563, + "learning_rate": 0.00018342230069489385, + "loss": 1.1378, + "step": 10933 + }, + { + "epoch": 0.8410122298284747, + "learning_rate": 0.0001832486550394289, + "loss": 1.1332, + "step": 10934 + }, + { + "epoch": 0.8410891469886932, + "learning_rate": 0.00018307508627044662, + "loss": 1.1435, + "step": 10935 + }, + { + "epoch": 0.8411660641489116, + "learning_rate": 0.00018290159439808274, + "loss": 1.3661, + "step": 10936 + }, + { + "epoch": 0.8412429813091301, + "learning_rate": 0.00018272817943246718, + "loss": 0.9954, + "step": 10937 + }, + { + "epoch": 0.8413198984693485, + "learning_rate": 0.00018255484138372663, + "loss": 1.2505, + "step": 10938 + }, + { + "epoch": 0.8413968156295669, + "learning_rate": 0.0001823815802619817, + "loss": 1.0661, + "step": 10939 + }, + { + "epoch": 0.8414737327897854, + "learning_rate": 0.00018220839607734912, + "loss": 0.9761, + "step": 10940 + }, + { + "epoch": 0.8415506499500038, + "learning_rate": 0.00018203528883994197, + "loss": 1.3057, + "step": 10941 + }, + { + "epoch": 0.8416275671102222, + "learning_rate": 0.00018186225855986783, + "loss": 1.1566, + "step": 10942 + }, + { + "epoch": 0.8417044842704408, + "learning_rate": 0.00018168930524723026, + "loss": 1.2366, + "step": 10943 + }, + { + "epoch": 0.8417814014306592, + "learning_rate": 0.00018151642891212788, + "loss": 1.225, + "step": 10944 + }, + { + "epoch": 0.8418583185908777, + "learning_rate": 0.00018134362956465512, + "loss": 1.2227, + "step": 10945 + }, + { + "epoch": 0.8419352357510961, + "learning_rate": 0.00018117090721490238, + "loss": 0.9353, + "step": 10946 + }, + { + "epoch": 0.8420121529113145, + "learning_rate": 0.0001809982618729551, + "loss": 1.1659, + "step": 10947 + }, + { + "epoch": 0.842089070071533, + "learning_rate": 0.00018082569354889338, + "loss": 1.3555, + "step": 10948 + }, + { + "epoch": 0.8421659872317514, + "learning_rate": 0.00018065320225279453, + "loss": 0.8292, + "step": 10949 + }, + { + "epoch": 0.8422429043919698, + "learning_rate": 0.0001804807879947301, + "loss": 1.7349, + "step": 10950 + }, + { + "epoch": 0.8423198215521883, + "learning_rate": 0.00018030845078476776, + "loss": 1.0104, + "step": 10951 + }, + { + "epoch": 0.8423967387124067, + "learning_rate": 0.00018013619063297028, + "loss": 1.321, + "step": 10952 + }, + { + "epoch": 0.8424736558726252, + "learning_rate": 0.00017996400754939575, + "loss": 1.0762, + "step": 10953 + }, + { + "epoch": 0.8425505730328436, + "learning_rate": 0.00017979190154409898, + "loss": 0.8755, + "step": 10954 + }, + { + "epoch": 0.842627490193062, + "learning_rate": 0.00017961987262712892, + "loss": 1.2364, + "step": 10955 + }, + { + "epoch": 0.8427044073532806, + "learning_rate": 0.00017944792080853056, + "loss": 0.9473, + "step": 10956 + }, + { + "epoch": 0.842781324513499, + "learning_rate": 0.00017927604609834435, + "loss": 1.2731, + "step": 10957 + }, + { + "epoch": 0.8428582416737174, + "learning_rate": 0.0001791042485066061, + "loss": 0.9911, + "step": 10958 + }, + { + "epoch": 0.8429351588339359, + "learning_rate": 0.0001789325280433473, + "loss": 1.261, + "step": 10959 + }, + { + "epoch": 0.8430120759941543, + "learning_rate": 0.00017876088471859542, + "loss": 1.2491, + "step": 10960 + }, + { + "epoch": 0.8430889931543727, + "learning_rate": 0.00017858931854237227, + "loss": 1.2727, + "step": 10961 + }, + { + "epoch": 0.8431659103145912, + "learning_rate": 0.00017841782952469565, + "loss": 1.2604, + "step": 10962 + }, + { + "epoch": 0.8432428274748096, + "learning_rate": 0.0001782464176755794, + "loss": 1.2202, + "step": 10963 + }, + { + "epoch": 0.8433197446350281, + "learning_rate": 0.00017807508300503257, + "loss": 0.8169, + "step": 10964 + }, + { + "epoch": 0.8433966617952465, + "learning_rate": 0.00017790382552305907, + "loss": 1.2243, + "step": 10965 + }, + { + "epoch": 0.8434735789554649, + "learning_rate": 0.00017773264523965925, + "loss": 1.3926, + "step": 10966 + }, + { + "epoch": 0.8435504961156834, + "learning_rate": 0.00017756154216482783, + "loss": 1.0529, + "step": 10967 + }, + { + "epoch": 0.8436274132759018, + "learning_rate": 0.00017739051630855663, + "loss": 1.1103, + "step": 10968 + }, + { + "epoch": 0.8437043304361203, + "learning_rate": 0.0001772195676808318, + "loss": 1.0959, + "step": 10969 + }, + { + "epoch": 0.8437812475963388, + "learning_rate": 0.0001770486962916344, + "loss": 0.7226, + "step": 10970 + }, + { + "epoch": 0.8438581647565572, + "learning_rate": 0.00017687790215094274, + "loss": 0.8492, + "step": 10971 + }, + { + "epoch": 0.8439350819167757, + "learning_rate": 0.00017670718526872887, + "loss": 1.2645, + "step": 10972 + }, + { + "epoch": 0.8440119990769941, + "learning_rate": 0.0001765365456549623, + "loss": 1.0794, + "step": 10973 + }, + { + "epoch": 0.8440889162372125, + "learning_rate": 0.00017636598331960556, + "loss": 1.0627, + "step": 10974 + }, + { + "epoch": 0.844165833397431, + "learning_rate": 0.00017619549827261837, + "loss": 1.168, + "step": 10975 + }, + { + "epoch": 0.8442427505576494, + "learning_rate": 0.0001760250905239561, + "loss": 0.9741, + "step": 10976 + }, + { + "epoch": 0.8443196677178678, + "learning_rate": 0.00017585476008356843, + "loss": 1.2224, + "step": 10977 + }, + { + "epoch": 0.8443965848780863, + "learning_rate": 0.00017568450696140141, + "loss": 1.1392, + "step": 10978 + }, + { + "epoch": 0.8444735020383047, + "learning_rate": 0.00017551433116739612, + "loss": 1.1975, + "step": 10979 + }, + { + "epoch": 0.8445504191985231, + "learning_rate": 0.00017534423271148942, + "loss": 0.9705, + "step": 10980 + }, + { + "epoch": 0.8446273363587417, + "learning_rate": 0.00017517421160361319, + "loss": 0.9703, + "step": 10981 + }, + { + "epoch": 0.8447042535189601, + "learning_rate": 0.000175004267853696, + "loss": 1.2962, + "step": 10982 + }, + { + "epoch": 0.8447811706791786, + "learning_rate": 0.0001748344014716599, + "loss": 1.2797, + "step": 10983 + }, + { + "epoch": 0.844858087839397, + "learning_rate": 0.00017466461246742448, + "loss": 1.0698, + "step": 10984 + }, + { + "epoch": 0.8449350049996154, + "learning_rate": 0.00017449490085090364, + "loss": 1.005, + "step": 10985 + }, + { + "epoch": 0.8450119221598339, + "learning_rate": 0.00017432526663200694, + "loss": 0.8879, + "step": 10986 + }, + { + "epoch": 0.8450888393200523, + "learning_rate": 0.00017415570982063944, + "loss": 1.1531, + "step": 10987 + }, + { + "epoch": 0.8451657564802707, + "learning_rate": 0.00017398623042670175, + "loss": 1.3468, + "step": 10988 + }, + { + "epoch": 0.8452426736404892, + "learning_rate": 0.0001738168284600898, + "loss": 1.2146, + "step": 10989 + }, + { + "epoch": 0.8453195908007076, + "learning_rate": 0.00017364750393069562, + "loss": 1.4544, + "step": 10990 + }, + { + "epoch": 0.8453965079609261, + "learning_rate": 0.000173478256848406, + "loss": 0.9959, + "step": 10991 + }, + { + "epoch": 0.8454734251211445, + "learning_rate": 0.0001733090872231034, + "loss": 1.2112, + "step": 10992 + }, + { + "epoch": 0.8455503422813629, + "learning_rate": 0.0001731399950646657, + "loss": 0.7243, + "step": 10993 + }, + { + "epoch": 0.8456272594415815, + "learning_rate": 0.00017297098038296637, + "loss": 1.3572, + "step": 10994 + }, + { + "epoch": 0.8457041766017999, + "learning_rate": 0.00017280204318787518, + "loss": 1.1198, + "step": 10995 + }, + { + "epoch": 0.8457810937620183, + "learning_rate": 0.0001726331834892554, + "loss": 1.2802, + "step": 10996 + }, + { + "epoch": 0.8458580109222368, + "learning_rate": 0.0001724644012969671, + "loss": 0.6415, + "step": 10997 + }, + { + "epoch": 0.8459349280824552, + "learning_rate": 0.00017229569662086631, + "loss": 1.3411, + "step": 10998 + }, + { + "epoch": 0.8460118452426736, + "learning_rate": 0.00017212706947080343, + "loss": 1.438, + "step": 10999 + }, + { + "epoch": 0.8460887624028921, + "learning_rate": 0.00017195851985662492, + "loss": 1.42, + "step": 11000 + }, + { + "epoch": 0.8461656795631105, + "learning_rate": 0.00017179004778817243, + "loss": 1.189, + "step": 11001 + }, + { + "epoch": 0.846242596723329, + "learning_rate": 0.00017162165327528307, + "loss": 1.1512, + "step": 11002 + }, + { + "epoch": 0.8463195138835474, + "learning_rate": 0.00017145333632778999, + "loss": 1.4032, + "step": 11003 + }, + { + "epoch": 0.8463964310437658, + "learning_rate": 0.00017128509695552114, + "loss": 1.7623, + "step": 11004 + }, + { + "epoch": 0.8464733482039843, + "learning_rate": 0.00017111693516830019, + "loss": 0.8291, + "step": 11005 + }, + { + "epoch": 0.8465502653642027, + "learning_rate": 0.00017094885097594642, + "loss": 1.1812, + "step": 11006 + }, + { + "epoch": 0.8466271825244212, + "learning_rate": 0.00017078084438827435, + "loss": 1.0123, + "step": 11007 + }, + { + "epoch": 0.8467040996846397, + "learning_rate": 0.0001706129154150941, + "loss": 1.2354, + "step": 11008 + }, + { + "epoch": 0.8467810168448581, + "learning_rate": 0.00017044506406621102, + "loss": 1.1638, + "step": 11009 + }, + { + "epoch": 0.8468579340050766, + "learning_rate": 0.00017027729035142607, + "loss": 1.0892, + "step": 11010 + }, + { + "epoch": 0.846934851165295, + "learning_rate": 0.0001701095942805363, + "loss": 1.315, + "step": 11011 + }, + { + "epoch": 0.8470117683255134, + "learning_rate": 0.00016994197586333315, + "loss": 1.2207, + "step": 11012 + }, + { + "epoch": 0.8470886854857319, + "learning_rate": 0.00016977443510960433, + "loss": 1.2968, + "step": 11013 + }, + { + "epoch": 0.8471656026459503, + "learning_rate": 0.0001696069720291325, + "loss": 0.9474, + "step": 11014 + }, + { + "epoch": 0.8472425198061687, + "learning_rate": 0.000169439586631696, + "loss": 1.1572, + "step": 11015 + }, + { + "epoch": 0.8473194369663872, + "learning_rate": 0.00016927227892706852, + "loss": 1.0062, + "step": 11016 + }, + { + "epoch": 0.8473963541266056, + "learning_rate": 0.00016910504892501992, + "loss": 1.2504, + "step": 11017 + }, + { + "epoch": 0.847473271286824, + "learning_rate": 0.00016893789663531471, + "loss": 1.1055, + "step": 11018 + }, + { + "epoch": 0.8475501884470426, + "learning_rate": 0.00016877082206771227, + "loss": 1.0222, + "step": 11019 + }, + { + "epoch": 0.847627105607261, + "learning_rate": 0.00016860382523196926, + "loss": 1.2512, + "step": 11020 + }, + { + "epoch": 0.8477040227674795, + "learning_rate": 0.00016843690613783642, + "loss": 0.9854, + "step": 11021 + }, + { + "epoch": 0.8477809399276979, + "learning_rate": 0.00016827006479506042, + "loss": 1.1565, + "step": 11022 + }, + { + "epoch": 0.8478578570879163, + "learning_rate": 0.00016810330121338313, + "loss": 1.1712, + "step": 11023 + }, + { + "epoch": 0.8479347742481348, + "learning_rate": 0.00016793661540254195, + "loss": 1.1295, + "step": 11024 + }, + { + "epoch": 0.8480116914083532, + "learning_rate": 0.00016777000737227028, + "loss": 0.8593, + "step": 11025 + }, + { + "epoch": 0.8480886085685716, + "learning_rate": 0.0001676034771322963, + "loss": 1.0704, + "step": 11026 + }, + { + "epoch": 0.8481655257287901, + "learning_rate": 0.0001674370246923439, + "loss": 1.2174, + "step": 11027 + }, + { + "epoch": 0.8482424428890085, + "learning_rate": 0.0001672706500621325, + "loss": 1.0962, + "step": 11028 + }, + { + "epoch": 0.848319360049227, + "learning_rate": 0.0001671043532513763, + "loss": 0.9685, + "step": 11029 + }, + { + "epoch": 0.8483962772094454, + "learning_rate": 0.0001669381342697864, + "loss": 1.2179, + "step": 11030 + }, + { + "epoch": 0.8484731943696638, + "learning_rate": 0.00016677199312706848, + "loss": 1.2541, + "step": 11031 + }, + { + "epoch": 0.8485501115298824, + "learning_rate": 0.00016660592983292266, + "loss": 1.2204, + "step": 11032 + }, + { + "epoch": 0.8486270286901008, + "learning_rate": 0.0001664399443970465, + "loss": 1.0257, + "step": 11033 + }, + { + "epoch": 0.8487039458503192, + "learning_rate": 0.00016627403682913179, + "loss": 1.3559, + "step": 11034 + }, + { + "epoch": 0.8487808630105377, + "learning_rate": 0.00016610820713886605, + "loss": 0.806, + "step": 11035 + }, + { + "epoch": 0.8488577801707561, + "learning_rate": 0.00016594245533593238, + "loss": 0.9572, + "step": 11036 + }, + { + "epoch": 0.8489346973309745, + "learning_rate": 0.0001657767814300089, + "loss": 0.7416, + "step": 11037 + }, + { + "epoch": 0.849011614491193, + "learning_rate": 0.00016561118543076937, + "loss": 1.2735, + "step": 11038 + }, + { + "epoch": 0.8490885316514114, + "learning_rate": 0.00016544566734788386, + "loss": 0.7056, + "step": 11039 + }, + { + "epoch": 0.8491654488116299, + "learning_rate": 0.00016528022719101648, + "loss": 1.0106, + "step": 11040 + }, + { + "epoch": 0.8492423659718483, + "learning_rate": 0.0001651148649698277, + "loss": 0.8831, + "step": 11041 + }, + { + "epoch": 0.8493192831320667, + "learning_rate": 0.00016494958069397309, + "loss": 1.3064, + "step": 11042 + }, + { + "epoch": 0.8493962002922852, + "learning_rate": 0.00016478437437310363, + "loss": 1.1316, + "step": 11043 + }, + { + "epoch": 0.8494731174525036, + "learning_rate": 0.00016461924601686656, + "loss": 1.179, + "step": 11044 + }, + { + "epoch": 0.849550034612722, + "learning_rate": 0.0001644541956349032, + "loss": 1.3716, + "step": 11045 + }, + { + "epoch": 0.8496269517729406, + "learning_rate": 0.00016428922323685101, + "loss": 0.8783, + "step": 11046 + }, + { + "epoch": 0.849703868933159, + "learning_rate": 0.0001641243288323433, + "loss": 1.3848, + "step": 11047 + }, + { + "epoch": 0.8497807860933775, + "learning_rate": 0.0001639595124310083, + "loss": 0.7817, + "step": 11048 + }, + { + "epoch": 0.8498577032535959, + "learning_rate": 0.00016379477404246973, + "loss": 0.8854, + "step": 11049 + }, + { + "epoch": 0.8499346204138143, + "learning_rate": 0.00016363011367634685, + "loss": 1.3142, + "step": 11050 + }, + { + "epoch": 0.8500115375740328, + "learning_rate": 0.00016346553134225395, + "loss": 1.1499, + "step": 11051 + }, + { + "epoch": 0.8500884547342512, + "learning_rate": 0.00016330102704980204, + "loss": 0.986, + "step": 11052 + }, + { + "epoch": 0.8501653718944696, + "learning_rate": 0.00016313660080859643, + "loss": 0.9656, + "step": 11053 + }, + { + "epoch": 0.8502422890546881, + "learning_rate": 0.00016297225262823745, + "loss": 1.2177, + "step": 11054 + }, + { + "epoch": 0.8503192062149065, + "learning_rate": 0.0001628079825183224, + "loss": 0.8681, + "step": 11055 + }, + { + "epoch": 0.8503961233751249, + "learning_rate": 0.00016264379048844263, + "loss": 1.1863, + "step": 11056 + }, + { + "epoch": 0.8504730405353434, + "learning_rate": 0.00016247967654818595, + "loss": 1.0493, + "step": 11057 + }, + { + "epoch": 0.8505499576955619, + "learning_rate": 0.00016231564070713473, + "loss": 1.2471, + "step": 11058 + }, + { + "epoch": 0.8506268748557804, + "learning_rate": 0.0001621516829748671, + "loss": 1.0571, + "step": 11059 + }, + { + "epoch": 0.8507037920159988, + "learning_rate": 0.00016198780336095726, + "loss": 1.2245, + "step": 11060 + }, + { + "epoch": 0.8507807091762172, + "learning_rate": 0.00016182400187497386, + "loss": 1.179, + "step": 11061 + }, + { + "epoch": 0.8508576263364357, + "learning_rate": 0.0001616602785264818, + "loss": 1.7001, + "step": 11062 + }, + { + "epoch": 0.8509345434966541, + "learning_rate": 0.0001614966333250409, + "loss": 0.9951, + "step": 11063 + }, + { + "epoch": 0.8510114606568725, + "learning_rate": 0.00016133306628020633, + "loss": 1.1903, + "step": 11064 + }, + { + "epoch": 0.851088377817091, + "learning_rate": 0.00016116957740152898, + "loss": 1.2345, + "step": 11065 + }, + { + "epoch": 0.8511652949773094, + "learning_rate": 0.0001610061666985559, + "loss": 1.3622, + "step": 11066 + }, + { + "epoch": 0.8512422121375279, + "learning_rate": 0.00016084283418082757, + "loss": 0.9611, + "step": 11067 + }, + { + "epoch": 0.8513191292977463, + "learning_rate": 0.00016067957985788222, + "loss": 1.3876, + "step": 11068 + }, + { + "epoch": 0.8513960464579647, + "learning_rate": 0.00016051640373925192, + "loss": 1.1933, + "step": 11069 + }, + { + "epoch": 0.8514729636181833, + "learning_rate": 0.00016035330583446467, + "loss": 0.8288, + "step": 11070 + }, + { + "epoch": 0.8515498807784017, + "learning_rate": 0.0001601902861530442, + "loss": 1.235, + "step": 11071 + }, + { + "epoch": 0.8516267979386201, + "learning_rate": 0.00016002734470450908, + "loss": 0.9444, + "step": 11072 + }, + { + "epoch": 0.8517037150988386, + "learning_rate": 0.00015986448149837368, + "loss": 1.292, + "step": 11073 + }, + { + "epoch": 0.851780632259057, + "learning_rate": 0.00015970169654414807, + "loss": 1.448, + "step": 11074 + }, + { + "epoch": 0.8518575494192755, + "learning_rate": 0.00015953898985133713, + "loss": 1.6343, + "step": 11075 + }, + { + "epoch": 0.8519344665794939, + "learning_rate": 0.0001593763614294416, + "loss": 0.9717, + "step": 11076 + }, + { + "epoch": 0.8520113837397123, + "learning_rate": 0.00015921381128795759, + "loss": 1.1789, + "step": 11077 + }, + { + "epoch": 0.8520883008999308, + "learning_rate": 0.00015905133943637612, + "loss": 1.2542, + "step": 11078 + }, + { + "epoch": 0.8521652180601492, + "learning_rate": 0.00015888894588418512, + "loss": 1.2433, + "step": 11079 + }, + { + "epoch": 0.8522421352203676, + "learning_rate": 0.00015872663064086585, + "loss": 1.2715, + "step": 11080 + }, + { + "epoch": 0.8523190523805861, + "learning_rate": 0.0001585643937158962, + "loss": 1.5103, + "step": 11081 + }, + { + "epoch": 0.8523959695408045, + "learning_rate": 0.00015840223511874997, + "loss": 1.3577, + "step": 11082 + }, + { + "epoch": 0.852472886701023, + "learning_rate": 0.00015824015485889521, + "loss": 1.3692, + "step": 11083 + }, + { + "epoch": 0.8525498038612415, + "learning_rate": 0.00015807815294579635, + "loss": 1.246, + "step": 11084 + }, + { + "epoch": 0.8526267210214599, + "learning_rate": 0.00015791622938891254, + "loss": 1.1344, + "step": 11085 + }, + { + "epoch": 0.8527036381816784, + "learning_rate": 0.0001577543841976985, + "loss": 1.0745, + "step": 11086 + }, + { + "epoch": 0.8527805553418968, + "learning_rate": 0.0001575926173816052, + "loss": 1.0883, + "step": 11087 + }, + { + "epoch": 0.8528574725021152, + "learning_rate": 0.0001574309289500782, + "loss": 0.9141, + "step": 11088 + }, + { + "epoch": 0.8529343896623337, + "learning_rate": 0.000157269318912558, + "loss": 1.255, + "step": 11089 + }, + { + "epoch": 0.8530113068225521, + "learning_rate": 0.00015710778727848174, + "loss": 1.2988, + "step": 11090 + }, + { + "epoch": 0.8530882239827705, + "learning_rate": 0.00015694633405728147, + "loss": 1.0386, + "step": 11091 + }, + { + "epoch": 0.853165141142989, + "learning_rate": 0.00015678495925838438, + "loss": 1.1427, + "step": 11092 + }, + { + "epoch": 0.8532420583032074, + "learning_rate": 0.00015662366289121338, + "loss": 1.1574, + "step": 11093 + }, + { + "epoch": 0.8533189754634259, + "learning_rate": 0.00015646244496518648, + "loss": 1.198, + "step": 11094 + }, + { + "epoch": 0.8533958926236443, + "learning_rate": 0.00015630130548971777, + "loss": 1.2668, + "step": 11095 + }, + { + "epoch": 0.8534728097838628, + "learning_rate": 0.00015614024447421627, + "loss": 1.2075, + "step": 11096 + }, + { + "epoch": 0.8535497269440813, + "learning_rate": 0.00015597926192808658, + "loss": 1.3358, + "step": 11097 + }, + { + "epoch": 0.8536266441042997, + "learning_rate": 0.0001558183578607284, + "loss": 1.3366, + "step": 11098 + }, + { + "epoch": 0.8537035612645181, + "learning_rate": 0.00015565753228153717, + "loss": 1.1686, + "step": 11099 + }, + { + "epoch": 0.8537804784247366, + "learning_rate": 0.00015549678519990323, + "loss": 1.1919, + "step": 11100 + }, + { + "epoch": 0.853857395584955, + "learning_rate": 0.00015533611662521408, + "loss": 0.9732, + "step": 11101 + }, + { + "epoch": 0.8539343127451734, + "learning_rate": 0.00015517552656684985, + "loss": 1.2336, + "step": 11102 + }, + { + "epoch": 0.8540112299053919, + "learning_rate": 0.00015501501503418808, + "loss": 1.2513, + "step": 11103 + }, + { + "epoch": 0.8540881470656103, + "learning_rate": 0.00015485458203660156, + "loss": 1.1273, + "step": 11104 + }, + { + "epoch": 0.8541650642258288, + "learning_rate": 0.00015469422758345786, + "loss": 1.333, + "step": 11105 + }, + { + "epoch": 0.8542419813860472, + "learning_rate": 0.00015453395168412026, + "loss": 0.917, + "step": 11106 + }, + { + "epoch": 0.8543188985462656, + "learning_rate": 0.0001543737543479475, + "loss": 1.2608, + "step": 11107 + }, + { + "epoch": 0.8543958157064842, + "learning_rate": 0.0001542136355842932, + "loss": 1.4238, + "step": 11108 + }, + { + "epoch": 0.8544727328667026, + "learning_rate": 0.00015405359540250757, + "loss": 0.9659, + "step": 11109 + }, + { + "epoch": 0.854549650026921, + "learning_rate": 0.0001538936338119355, + "loss": 1.1839, + "step": 11110 + }, + { + "epoch": 0.8546265671871395, + "learning_rate": 0.0001537337508219165, + "loss": 1.2577, + "step": 11111 + }, + { + "epoch": 0.8547034843473579, + "learning_rate": 0.00015357394644178706, + "loss": 0.7599, + "step": 11112 + }, + { + "epoch": 0.8547804015075764, + "learning_rate": 0.00015341422068087778, + "loss": 1.1779, + "step": 11113 + }, + { + "epoch": 0.8548573186677948, + "learning_rate": 0.00015325457354851618, + "loss": 1.4574, + "step": 11114 + }, + { + "epoch": 0.8549342358280132, + "learning_rate": 0.0001530950050540233, + "loss": 0.805, + "step": 11115 + }, + { + "epoch": 0.8550111529882317, + "learning_rate": 0.0001529355152067165, + "loss": 0.7373, + "step": 11116 + }, + { + "epoch": 0.8550880701484501, + "learning_rate": 0.0001527761040159092, + "loss": 1.3628, + "step": 11117 + }, + { + "epoch": 0.8551649873086685, + "learning_rate": 0.00015261677149090908, + "loss": 1.4228, + "step": 11118 + }, + { + "epoch": 0.855241904468887, + "learning_rate": 0.00015245751764101989, + "loss": 1.4497, + "step": 11119 + }, + { + "epoch": 0.8553188216291054, + "learning_rate": 0.0001522983424755407, + "loss": 1.2204, + "step": 11120 + }, + { + "epoch": 0.8553957387893238, + "learning_rate": 0.0001521392460037659, + "loss": 1.2439, + "step": 11121 + }, + { + "epoch": 0.8554726559495424, + "learning_rate": 0.00015198022823498486, + "loss": 1.0578, + "step": 11122 + }, + { + "epoch": 0.8555495731097608, + "learning_rate": 0.0001518212891784837, + "loss": 1.174, + "step": 11123 + }, + { + "epoch": 0.8556264902699793, + "learning_rate": 0.00015166242884354198, + "loss": 1.1494, + "step": 11124 + }, + { + "epoch": 0.8557034074301977, + "learning_rate": 0.00015150364723943643, + "loss": 1.0044, + "step": 11125 + }, + { + "epoch": 0.8557803245904161, + "learning_rate": 0.00015134494437543834, + "loss": 1.2293, + "step": 11126 + }, + { + "epoch": 0.8558572417506346, + "learning_rate": 0.00015118632026081457, + "loss": 1.3714, + "step": 11127 + }, + { + "epoch": 0.855934158910853, + "learning_rate": 0.00015102777490482727, + "loss": 1.324, + "step": 11128 + }, + { + "epoch": 0.8560110760710714, + "learning_rate": 0.00015086930831673402, + "loss": 1.3615, + "step": 11129 + }, + { + "epoch": 0.8560879932312899, + "learning_rate": 0.00015071092050578756, + "loss": 1.0112, + "step": 11130 + }, + { + "epoch": 0.8561649103915083, + "learning_rate": 0.00015055261148123682, + "loss": 1.5117, + "step": 11131 + }, + { + "epoch": 0.8562418275517268, + "learning_rate": 0.00015039438125232562, + "loss": 0.9133, + "step": 11132 + }, + { + "epoch": 0.8563187447119452, + "learning_rate": 0.00015023622982829305, + "loss": 0.9773, + "step": 11133 + }, + { + "epoch": 0.8563956618721636, + "learning_rate": 0.00015007815721837354, + "loss": 1.2601, + "step": 11134 + }, + { + "epoch": 0.8564725790323822, + "learning_rate": 0.0001499201634317972, + "loss": 1.1649, + "step": 11135 + }, + { + "epoch": 0.8565494961926006, + "learning_rate": 0.00014976224847779003, + "loss": 1.318, + "step": 11136 + }, + { + "epoch": 0.856626413352819, + "learning_rate": 0.00014960441236557193, + "loss": 1.1077, + "step": 11137 + }, + { + "epoch": 0.8567033305130375, + "learning_rate": 0.0001494466551043594, + "loss": 1.1102, + "step": 11138 + }, + { + "epoch": 0.8567802476732559, + "learning_rate": 0.00014928897670336456, + "loss": 1.279, + "step": 11139 + }, + { + "epoch": 0.8568571648334743, + "learning_rate": 0.00014913137717179387, + "loss": 1.0541, + "step": 11140 + }, + { + "epoch": 0.8569340819936928, + "learning_rate": 0.00014897385651884998, + "loss": 1.0891, + "step": 11141 + }, + { + "epoch": 0.8570109991539112, + "learning_rate": 0.00014881641475373054, + "loss": 1.2631, + "step": 11142 + }, + { + "epoch": 0.8570879163141297, + "learning_rate": 0.00014865905188562867, + "loss": 1.088, + "step": 11143 + }, + { + "epoch": 0.8571648334743481, + "learning_rate": 0.00014850176792373338, + "loss": 1.1013, + "step": 11144 + }, + { + "epoch": 0.8572417506345665, + "learning_rate": 0.00014834456287722848, + "loss": 1.132, + "step": 11145 + }, + { + "epoch": 0.857318667794785, + "learning_rate": 0.00014818743675529278, + "loss": 1.3541, + "step": 11146 + }, + { + "epoch": 0.8573955849550035, + "learning_rate": 0.00014803038956710162, + "loss": 1.2979, + "step": 11147 + }, + { + "epoch": 0.8574725021152219, + "learning_rate": 0.0001478734213218252, + "loss": 0.6047, + "step": 11148 + }, + { + "epoch": 0.8575494192754404, + "learning_rate": 0.00014771653202862877, + "loss": 1.199, + "step": 11149 + }, + { + "epoch": 0.8576263364356588, + "learning_rate": 0.00014755972169667325, + "loss": 1.1863, + "step": 11150 + }, + { + "epoch": 0.8577032535958773, + "learning_rate": 0.00014740299033511495, + "loss": 1.0097, + "step": 11151 + }, + { + "epoch": 0.8577801707560957, + "learning_rate": 0.00014724633795310587, + "loss": 1.2456, + "step": 11152 + }, + { + "epoch": 0.8578570879163141, + "learning_rate": 0.0001470897645597929, + "loss": 0.9478, + "step": 11153 + }, + { + "epoch": 0.8579340050765326, + "learning_rate": 0.00014693327016431868, + "loss": 1.3032, + "step": 11154 + }, + { + "epoch": 0.858010922236751, + "learning_rate": 0.00014677685477582075, + "loss": 1.2877, + "step": 11155 + }, + { + "epoch": 0.8580878393969694, + "learning_rate": 0.00014662051840343265, + "loss": 1.2861, + "step": 11156 + }, + { + "epoch": 0.8581647565571879, + "learning_rate": 0.00014646426105628275, + "loss": 1.3966, + "step": 11157 + }, + { + "epoch": 0.8582416737174063, + "learning_rate": 0.00014630808274349573, + "loss": 1.1975, + "step": 11158 + }, + { + "epoch": 0.8583185908776247, + "learning_rate": 0.0001461519834741903, + "loss": 1.2797, + "step": 11159 + }, + { + "epoch": 0.8583955080378433, + "learning_rate": 0.00014599596325748116, + "loss": 0.9818, + "step": 11160 + }, + { + "epoch": 0.8584724251980617, + "learning_rate": 0.00014584002210247922, + "loss": 1.2435, + "step": 11161 + }, + { + "epoch": 0.8585493423582802, + "learning_rate": 0.00014568416001828964, + "loss": 1.1772, + "step": 11162 + }, + { + "epoch": 0.8586262595184986, + "learning_rate": 0.0001455283770140134, + "loss": 0.9049, + "step": 11163 + }, + { + "epoch": 0.858703176678717, + "learning_rate": 0.00014537267309874664, + "loss": 1.3119, + "step": 11164 + }, + { + "epoch": 0.8587800938389355, + "learning_rate": 0.0001452170482815811, + "loss": 0.9825, + "step": 11165 + }, + { + "epoch": 0.8588570109991539, + "learning_rate": 0.00014506150257160438, + "loss": 1.3915, + "step": 11166 + }, + { + "epoch": 0.8589339281593723, + "learning_rate": 0.00014490603597789854, + "loss": 1.8288, + "step": 11167 + }, + { + "epoch": 0.8590108453195908, + "learning_rate": 0.0001447506485095416, + "loss": 1.139, + "step": 11168 + }, + { + "epoch": 0.8590877624798092, + "learning_rate": 0.00014459534017560654, + "loss": 1.1336, + "step": 11169 + }, + { + "epoch": 0.8591646796400277, + "learning_rate": 0.00014444011098516192, + "loss": 1.4344, + "step": 11170 + }, + { + "epoch": 0.8592415968002461, + "learning_rate": 0.00014428496094727228, + "loss": 1.0494, + "step": 11171 + }, + { + "epoch": 0.8593185139604645, + "learning_rate": 0.00014412989007099692, + "loss": 1.2346, + "step": 11172 + }, + { + "epoch": 0.8593954311206831, + "learning_rate": 0.00014397489836538973, + "loss": 1.1324, + "step": 11173 + }, + { + "epoch": 0.8594723482809015, + "learning_rate": 0.00014381998583950186, + "loss": 1.3367, + "step": 11174 + }, + { + "epoch": 0.8595492654411199, + "learning_rate": 0.00014366515250237817, + "loss": 1.2786, + "step": 11175 + }, + { + "epoch": 0.8596261826013384, + "learning_rate": 0.00014351039836305995, + "loss": 0.9845, + "step": 11176 + }, + { + "epoch": 0.8597030997615568, + "learning_rate": 0.00014335572343058316, + "loss": 1.1422, + "step": 11177 + }, + { + "epoch": 0.8597800169217753, + "learning_rate": 0.00014320112771397926, + "loss": 1.3816, + "step": 11178 + }, + { + "epoch": 0.8598569340819937, + "learning_rate": 0.00014304661122227597, + "loss": 1.5254, + "step": 11179 + }, + { + "epoch": 0.8599338512422121, + "learning_rate": 0.0001428921739644951, + "loss": 1.329, + "step": 11180 + }, + { + "epoch": 0.8600107684024306, + "learning_rate": 0.00014273781594965445, + "loss": 0.9533, + "step": 11181 + }, + { + "epoch": 0.860087685562649, + "learning_rate": 0.00014258353718676748, + "loss": 0.9062, + "step": 11182 + }, + { + "epoch": 0.8601646027228674, + "learning_rate": 0.0001424293376848423, + "loss": 0.987, + "step": 11183 + }, + { + "epoch": 0.860241519883086, + "learning_rate": 0.00014227521745288253, + "loss": 1.0842, + "step": 11184 + }, + { + "epoch": 0.8603184370433044, + "learning_rate": 0.00014212117649988848, + "loss": 1.2322, + "step": 11185 + }, + { + "epoch": 0.8603953542035228, + "learning_rate": 0.0001419672148348538, + "loss": 0.941, + "step": 11186 + }, + { + "epoch": 0.8604722713637413, + "learning_rate": 0.00014181333246676846, + "loss": 1.1537, + "step": 11187 + }, + { + "epoch": 0.8605491885239597, + "learning_rate": 0.0001416595294046183, + "loss": 1.3088, + "step": 11188 + }, + { + "epoch": 0.8606261056841782, + "learning_rate": 0.0001415058056573839, + "loss": 1.2471, + "step": 11189 + }, + { + "epoch": 0.8607030228443966, + "learning_rate": 0.0001413521612340415, + "loss": 1.3289, + "step": 11190 + }, + { + "epoch": 0.860779940004615, + "learning_rate": 0.00014119859614356218, + "loss": 1.1891, + "step": 11191 + }, + { + "epoch": 0.8608568571648335, + "learning_rate": 0.00014104511039491263, + "loss": 1.2171, + "step": 11192 + }, + { + "epoch": 0.8609337743250519, + "learning_rate": 0.00014089170399705553, + "loss": 0.9272, + "step": 11193 + }, + { + "epoch": 0.8610106914852703, + "learning_rate": 0.0001407383769589487, + "loss": 1.0157, + "step": 11194 + }, + { + "epoch": 0.8610876086454888, + "learning_rate": 0.00014058512928954414, + "loss": 1.2028, + "step": 11195 + }, + { + "epoch": 0.8611645258057072, + "learning_rate": 0.0001404319609977907, + "loss": 1.1606, + "step": 11196 + }, + { + "epoch": 0.8612414429659258, + "learning_rate": 0.00014027887209263192, + "loss": 0.9443, + "step": 11197 + }, + { + "epoch": 0.8613183601261442, + "learning_rate": 0.00014012586258300725, + "loss": 0.9861, + "step": 11198 + }, + { + "epoch": 0.8613952772863626, + "learning_rate": 0.00013997293247785042, + "loss": 0.9838, + "step": 11199 + }, + { + "epoch": 0.8614721944465811, + "learning_rate": 0.0001398200817860913, + "loss": 1.4646, + "step": 11200 + }, + { + "epoch": 0.8615491116067995, + "learning_rate": 0.0001396673105166552, + "loss": 0.9187, + "step": 11201 + }, + { + "epoch": 0.8616260287670179, + "learning_rate": 0.00013951461867846271, + "loss": 1.0571, + "step": 11202 + }, + { + "epoch": 0.8617029459272364, + "learning_rate": 0.0001393620062804295, + "loss": 0.979, + "step": 11203 + }, + { + "epoch": 0.8617798630874548, + "learning_rate": 0.00013920947333146677, + "loss": 1.1068, + "step": 11204 + }, + { + "epoch": 0.8618567802476732, + "learning_rate": 0.00013905701984048109, + "loss": 1.1729, + "step": 11205 + }, + { + "epoch": 0.8619336974078917, + "learning_rate": 0.00013890464581637395, + "loss": 1.2125, + "step": 11206 + }, + { + "epoch": 0.8620106145681101, + "learning_rate": 0.0001387523512680438, + "loss": 1.3013, + "step": 11207 + }, + { + "epoch": 0.8620875317283286, + "learning_rate": 0.00013860013620438182, + "loss": 1.1944, + "step": 11208 + }, + { + "epoch": 0.862164448888547, + "learning_rate": 0.00013844800063427692, + "loss": 1.2554, + "step": 11209 + }, + { + "epoch": 0.8622413660487654, + "learning_rate": 0.00013829594456661232, + "loss": 1.2009, + "step": 11210 + }, + { + "epoch": 0.862318283208984, + "learning_rate": 0.00013814396801026675, + "loss": 1.027, + "step": 11211 + }, + { + "epoch": 0.8623952003692024, + "learning_rate": 0.0001379920709741141, + "loss": 0.7939, + "step": 11212 + }, + { + "epoch": 0.8624721175294208, + "learning_rate": 0.00013784025346702379, + "loss": 1.0146, + "step": 11213 + }, + { + "epoch": 0.8625490346896393, + "learning_rate": 0.00013768851549786056, + "loss": 1.5782, + "step": 11214 + }, + { + "epoch": 0.8626259518498577, + "learning_rate": 0.0001375368570754848, + "loss": 1.0979, + "step": 11215 + }, + { + "epoch": 0.8627028690100762, + "learning_rate": 0.00013738527820875198, + "loss": 1.163, + "step": 11216 + }, + { + "epoch": 0.8627797861702946, + "learning_rate": 0.00013723377890651278, + "loss": 0.9345, + "step": 11217 + }, + { + "epoch": 0.862856703330513, + "learning_rate": 0.00013708235917761353, + "loss": 1.1542, + "step": 11218 + }, + { + "epoch": 0.8629336204907315, + "learning_rate": 0.00013693101903089545, + "loss": 1.1283, + "step": 11219 + }, + { + "epoch": 0.8630105376509499, + "learning_rate": 0.0001367797584751963, + "loss": 0.9504, + "step": 11220 + }, + { + "epoch": 0.8630874548111683, + "learning_rate": 0.00013662857751934755, + "loss": 0.9247, + "step": 11221 + }, + { + "epoch": 0.8631643719713868, + "learning_rate": 0.0001364774761721766, + "loss": 1.3241, + "step": 11222 + }, + { + "epoch": 0.8632412891316052, + "learning_rate": 0.00013632645444250723, + "loss": 1.2632, + "step": 11223 + }, + { + "epoch": 0.8633182062918237, + "learning_rate": 0.00013617551233915725, + "loss": 0.9872, + "step": 11224 + }, + { + "epoch": 0.8633951234520422, + "learning_rate": 0.0001360246498709406, + "loss": 1.1072, + "step": 11225 + }, + { + "epoch": 0.8634720406122606, + "learning_rate": 0.00013587386704666605, + "loss": 1.4621, + "step": 11226 + }, + { + "epoch": 0.8635489577724791, + "learning_rate": 0.00013572316387513772, + "loss": 1.1754, + "step": 11227 + }, + { + "epoch": 0.8636258749326975, + "learning_rate": 0.00013557254036515609, + "loss": 1.1917, + "step": 11228 + }, + { + "epoch": 0.8637027920929159, + "learning_rate": 0.00013542199652551596, + "loss": 1.0406, + "step": 11229 + }, + { + "epoch": 0.8637797092531344, + "learning_rate": 0.0001352715323650071, + "loss": 1.1595, + "step": 11230 + }, + { + "epoch": 0.8638566264133528, + "learning_rate": 0.00013512114789241604, + "loss": 1.4831, + "step": 11231 + }, + { + "epoch": 0.8639335435735712, + "learning_rate": 0.00013497084311652353, + "loss": 1.4832, + "step": 11232 + }, + { + "epoch": 0.8640104607337897, + "learning_rate": 0.00013482061804610622, + "loss": 1.3306, + "step": 11233 + }, + { + "epoch": 0.8640873778940081, + "learning_rate": 0.00013467047268993565, + "loss": 1.2282, + "step": 11234 + }, + { + "epoch": 0.8641642950542266, + "learning_rate": 0.00013452040705677875, + "loss": 1.1446, + "step": 11235 + }, + { + "epoch": 0.864241212214445, + "learning_rate": 0.0001343704211553989, + "loss": 1.041, + "step": 11236 + }, + { + "epoch": 0.8643181293746635, + "learning_rate": 0.00013422051499455319, + "loss": 1.1237, + "step": 11237 + }, + { + "epoch": 0.864395046534882, + "learning_rate": 0.00013407068858299498, + "loss": 0.8111, + "step": 11238 + }, + { + "epoch": 0.8644719636951004, + "learning_rate": 0.0001339209419294728, + "loss": 1.0908, + "step": 11239 + }, + { + "epoch": 0.8645488808553188, + "learning_rate": 0.00013377127504273062, + "loss": 1.1345, + "step": 11240 + }, + { + "epoch": 0.8646257980155373, + "learning_rate": 0.00013362168793150714, + "loss": 1.269, + "step": 11241 + }, + { + "epoch": 0.8647027151757557, + "learning_rate": 0.00013347218060453787, + "loss": 1.4384, + "step": 11242 + }, + { + "epoch": 0.8647796323359741, + "learning_rate": 0.00013332275307055197, + "loss": 1.1863, + "step": 11243 + }, + { + "epoch": 0.8648565494961926, + "learning_rate": 0.0001331734053382745, + "loss": 1.0937, + "step": 11244 + }, + { + "epoch": 0.864933466656411, + "learning_rate": 0.0001330241374164266, + "loss": 1.4534, + "step": 11245 + }, + { + "epoch": 0.8650103838166295, + "learning_rate": 0.00013287494931372401, + "loss": 1.105, + "step": 11246 + }, + { + "epoch": 0.8650873009768479, + "learning_rate": 0.00013272584103887774, + "loss": 1.1691, + "step": 11247 + }, + { + "epoch": 0.8651642181370663, + "learning_rate": 0.00013257681260059483, + "loss": 1.1537, + "step": 11248 + }, + { + "epoch": 0.8652411352972849, + "learning_rate": 0.00013242786400757651, + "loss": 1.097, + "step": 11249 + }, + { + "epoch": 0.8653180524575033, + "learning_rate": 0.00013227899526852062, + "loss": 0.9412, + "step": 11250 + }, + { + "epoch": 0.8653949696177217, + "learning_rate": 0.00013213020639212003, + "loss": 1.1851, + "step": 11251 + }, + { + "epoch": 0.8654718867779402, + "learning_rate": 0.00013198149738706167, + "loss": 1.1797, + "step": 11252 + }, + { + "epoch": 0.8655488039381586, + "learning_rate": 0.00013183286826202973, + "loss": 1.1292, + "step": 11253 + }, + { + "epoch": 0.8656257210983771, + "learning_rate": 0.0001316843190257021, + "loss": 1.3159, + "step": 11254 + }, + { + "epoch": 0.8657026382585955, + "learning_rate": 0.0001315358496867537, + "loss": 0.8295, + "step": 11255 + }, + { + "epoch": 0.8657795554188139, + "learning_rate": 0.0001313874602538529, + "loss": 1.0835, + "step": 11256 + }, + { + "epoch": 0.8658564725790324, + "learning_rate": 0.00013123915073566427, + "loss": 1.1864, + "step": 11257 + }, + { + "epoch": 0.8659333897392508, + "learning_rate": 0.0001310909211408486, + "loss": 1.0999, + "step": 11258 + }, + { + "epoch": 0.8660103068994692, + "learning_rate": 0.00013094277147806055, + "loss": 1.0201, + "step": 11259 + }, + { + "epoch": 0.8660872240596877, + "learning_rate": 0.00013079470175595092, + "loss": 1.3484, + "step": 11260 + }, + { + "epoch": 0.8661641412199061, + "learning_rate": 0.00013064671198316564, + "loss": 1.1714, + "step": 11261 + }, + { + "epoch": 0.8662410583801246, + "learning_rate": 0.00013049880216834593, + "loss": 1.1316, + "step": 11262 + }, + { + "epoch": 0.8663179755403431, + "learning_rate": 0.00013035097232012805, + "loss": 1.0454, + "step": 11263 + }, + { + "epoch": 0.8663948927005615, + "learning_rate": 0.00013020322244714495, + "loss": 1.5866, + "step": 11264 + }, + { + "epoch": 0.86647180986078, + "learning_rate": 0.0001300555525580227, + "loss": 0.8978, + "step": 11265 + }, + { + "epoch": 0.8665487270209984, + "learning_rate": 0.00012990796266138476, + "loss": 1.2143, + "step": 11266 + }, + { + "epoch": 0.8666256441812168, + "learning_rate": 0.00012976045276584874, + "loss": 1.3761, + "step": 11267 + }, + { + "epoch": 0.8667025613414353, + "learning_rate": 0.00012961302288002792, + "loss": 0.7814, + "step": 11268 + }, + { + "epoch": 0.8667794785016537, + "learning_rate": 0.00012946567301253087, + "loss": 1.6697, + "step": 11269 + }, + { + "epoch": 0.8668563956618721, + "learning_rate": 0.00012931840317196143, + "loss": 0.7685, + "step": 11270 + }, + { + "epoch": 0.8669333128220906, + "learning_rate": 0.00012917121336691883, + "loss": 0.9561, + "step": 11271 + }, + { + "epoch": 0.867010229982309, + "learning_rate": 0.00012902410360599777, + "loss": 1.2734, + "step": 11272 + }, + { + "epoch": 0.8670871471425275, + "learning_rate": 0.0001288770738977883, + "loss": 1.241, + "step": 11273 + }, + { + "epoch": 0.867164064302746, + "learning_rate": 0.00012873012425087526, + "loss": 0.8868, + "step": 11274 + }, + { + "epoch": 0.8672409814629644, + "learning_rate": 0.00012858325467383946, + "loss": 1.0719, + "step": 11275 + }, + { + "epoch": 0.8673178986231829, + "learning_rate": 0.00012843646517525636, + "loss": 1.1167, + "step": 11276 + }, + { + "epoch": 0.8673948157834013, + "learning_rate": 0.00012828975576369812, + "loss": 0.6641, + "step": 11277 + }, + { + "epoch": 0.8674717329436197, + "learning_rate": 0.00012814312644773019, + "loss": 0.9545, + "step": 11278 + }, + { + "epoch": 0.8675486501038382, + "learning_rate": 0.00012799657723591473, + "loss": 1.1492, + "step": 11279 + }, + { + "epoch": 0.8676255672640566, + "learning_rate": 0.0001278501081368091, + "loss": 1.3582, + "step": 11280 + }, + { + "epoch": 0.8677024844242751, + "learning_rate": 0.00012770371915896573, + "loss": 1.2481, + "step": 11281 + }, + { + "epoch": 0.8677794015844935, + "learning_rate": 0.00012755741031093232, + "loss": 1.2525, + "step": 11282 + }, + { + "epoch": 0.8678563187447119, + "learning_rate": 0.0001274111816012522, + "loss": 1.2596, + "step": 11283 + }, + { + "epoch": 0.8679332359049304, + "learning_rate": 0.0001272650330384632, + "loss": 0.8427, + "step": 11284 + }, + { + "epoch": 0.8680101530651488, + "learning_rate": 0.0001271189646311, + "loss": 1.0808, + "step": 11285 + }, + { + "epoch": 0.8680870702253672, + "learning_rate": 0.00012697297638769145, + "loss": 1.0008, + "step": 11286 + }, + { + "epoch": 0.8681639873855858, + "learning_rate": 0.00012682706831676106, + "loss": 1.4559, + "step": 11287 + }, + { + "epoch": 0.8682409045458042, + "learning_rate": 0.00012668124042682967, + "loss": 1.5305, + "step": 11288 + }, + { + "epoch": 0.8683178217060226, + "learning_rate": 0.0001265354927264118, + "loss": 1.0053, + "step": 11289 + }, + { + "epoch": 0.8683947388662411, + "learning_rate": 0.000126389825224018, + "loss": 1.183, + "step": 11290 + }, + { + "epoch": 0.8684716560264595, + "learning_rate": 0.00012624423792815376, + "loss": 1.0886, + "step": 11291 + }, + { + "epoch": 0.868548573186678, + "learning_rate": 0.00012609873084732, + "loss": 1.2597, + "step": 11292 + }, + { + "epoch": 0.8686254903468964, + "learning_rate": 0.00012595330399001336, + "loss": 0.8537, + "step": 11293 + }, + { + "epoch": 0.8687024075071148, + "learning_rate": 0.00012580795736472528, + "loss": 0.8841, + "step": 11294 + }, + { + "epoch": 0.8687793246673333, + "learning_rate": 0.00012566269097994277, + "loss": 1.1812, + "step": 11295 + }, + { + "epoch": 0.8688562418275517, + "learning_rate": 0.00012551750484414802, + "loss": 0.7635, + "step": 11296 + }, + { + "epoch": 0.8689331589877701, + "learning_rate": 0.00012537239896581864, + "loss": 0.9661, + "step": 11297 + }, + { + "epoch": 0.8690100761479886, + "learning_rate": 0.00012522737335342717, + "loss": 1.2354, + "step": 11298 + }, + { + "epoch": 0.869086993308207, + "learning_rate": 0.0001250824280154426, + "loss": 1.0257, + "step": 11299 + }, + { + "epoch": 0.8691639104684256, + "learning_rate": 0.0001249375629603276, + "loss": 1.1796, + "step": 11300 + }, + { + "epoch": 0.869240827628644, + "learning_rate": 0.0001247927781965415, + "loss": 1.1895, + "step": 11301 + }, + { + "epoch": 0.8693177447888624, + "learning_rate": 0.0001246480737325382, + "loss": 1.1967, + "step": 11302 + }, + { + "epoch": 0.8693946619490809, + "learning_rate": 0.0001245034495767675, + "loss": 1.1796, + "step": 11303 + }, + { + "epoch": 0.8694715791092993, + "learning_rate": 0.00012435890573767361, + "loss": 1.1002, + "step": 11304 + }, + { + "epoch": 0.8695484962695177, + "learning_rate": 0.0001242144422236971, + "loss": 1.2278, + "step": 11305 + }, + { + "epoch": 0.8696254134297362, + "learning_rate": 0.00012407005904327274, + "loss": 0.9246, + "step": 11306 + }, + { + "epoch": 0.8697023305899546, + "learning_rate": 0.00012392575620483192, + "loss": 0.9358, + "step": 11307 + }, + { + "epoch": 0.869779247750173, + "learning_rate": 0.0001237815337168001, + "loss": 1.2992, + "step": 11308 + }, + { + "epoch": 0.8698561649103915, + "learning_rate": 0.00012363739158759908, + "loss": 1.2913, + "step": 11309 + }, + { + "epoch": 0.8699330820706099, + "learning_rate": 0.00012349332982564503, + "loss": 1.1436, + "step": 11310 + }, + { + "epoch": 0.8700099992308284, + "learning_rate": 0.0001233493484393498, + "loss": 0.8496, + "step": 11311 + }, + { + "epoch": 0.8700869163910468, + "learning_rate": 0.0001232054474371213, + "loss": 1.3898, + "step": 11312 + }, + { + "epoch": 0.8701638335512653, + "learning_rate": 0.00012306162682736133, + "loss": 1.2717, + "step": 11313 + }, + { + "epoch": 0.8702407507114838, + "learning_rate": 0.00012291788661846764, + "loss": 1.4172, + "step": 11314 + }, + { + "epoch": 0.8703176678717022, + "learning_rate": 0.00012277422681883403, + "loss": 0.8232, + "step": 11315 + }, + { + "epoch": 0.8703945850319206, + "learning_rate": 0.00012263064743684866, + "loss": 1.4226, + "step": 11316 + }, + { + "epoch": 0.8704715021921391, + "learning_rate": 0.00012248714848089526, + "loss": 1.2452, + "step": 11317 + }, + { + "epoch": 0.8705484193523575, + "learning_rate": 0.00012234372995935283, + "loss": 1.2566, + "step": 11318 + }, + { + "epoch": 0.870625336512576, + "learning_rate": 0.0001222003918805955, + "loss": 0.9356, + "step": 11319 + }, + { + "epoch": 0.8707022536727944, + "learning_rate": 0.00012205713425299358, + "loss": 1.2432, + "step": 11320 + }, + { + "epoch": 0.8707791708330128, + "learning_rate": 0.00012191395708491154, + "loss": 1.3988, + "step": 11321 + }, + { + "epoch": 0.8708560879932313, + "learning_rate": 0.0001217708603847097, + "loss": 1.4929, + "step": 11322 + }, + { + "epoch": 0.8709330051534497, + "learning_rate": 0.00012162784416074369, + "loss": 1.0432, + "step": 11323 + }, + { + "epoch": 0.8710099223136681, + "learning_rate": 0.0001214849084213645, + "loss": 1.2808, + "step": 11324 + }, + { + "epoch": 0.8710868394738867, + "learning_rate": 0.00012134205317491798, + "loss": 1.1407, + "step": 11325 + }, + { + "epoch": 0.8711637566341051, + "learning_rate": 0.00012119927842974592, + "loss": 1.1389, + "step": 11326 + }, + { + "epoch": 0.8712406737943235, + "learning_rate": 0.00012105658419418486, + "loss": 1.197, + "step": 11327 + }, + { + "epoch": 0.871317590954542, + "learning_rate": 0.00012091397047656677, + "loss": 1.1079, + "step": 11328 + }, + { + "epoch": 0.8713945081147604, + "learning_rate": 0.00012077143728521933, + "loss": 1.0327, + "step": 11329 + }, + { + "epoch": 0.8714714252749789, + "learning_rate": 0.00012062898462846523, + "loss": 1.0879, + "step": 11330 + }, + { + "epoch": 0.8715483424351973, + "learning_rate": 0.00012048661251462212, + "loss": 0.8193, + "step": 11331 + }, + { + "epoch": 0.8716252595954157, + "learning_rate": 0.00012034432095200354, + "loss": 1.2071, + "step": 11332 + }, + { + "epoch": 0.8717021767556342, + "learning_rate": 0.00012020210994891751, + "loss": 1.6704, + "step": 11333 + }, + { + "epoch": 0.8717790939158526, + "learning_rate": 0.00012005997951366854, + "loss": 1.1841, + "step": 11334 + }, + { + "epoch": 0.871856011076071, + "learning_rate": 0.00011991792965455584, + "loss": 1.469, + "step": 11335 + }, + { + "epoch": 0.8719329282362895, + "learning_rate": 0.00011977596037987293, + "loss": 1.0647, + "step": 11336 + }, + { + "epoch": 0.8720098453965079, + "learning_rate": 0.00011963407169791052, + "loss": 1.1584, + "step": 11337 + }, + { + "epoch": 0.8720867625567265, + "learning_rate": 0.00011949226361695297, + "loss": 1.1649, + "step": 11338 + }, + { + "epoch": 0.8721636797169449, + "learning_rate": 0.00011935053614528118, + "loss": 1.0068, + "step": 11339 + }, + { + "epoch": 0.8722405968771633, + "learning_rate": 0.00011920888929117035, + "loss": 1.155, + "step": 11340 + }, + { + "epoch": 0.8723175140373818, + "learning_rate": 0.0001190673230628912, + "loss": 0.931, + "step": 11341 + }, + { + "epoch": 0.8723944311976002, + "learning_rate": 0.00011892583746871045, + "loss": 0.879, + "step": 11342 + }, + { + "epoch": 0.8724713483578186, + "learning_rate": 0.00011878443251688947, + "loss": 0.7196, + "step": 11343 + }, + { + "epoch": 0.8725482655180371, + "learning_rate": 0.00011864310821568486, + "loss": 1.265, + "step": 11344 + }, + { + "epoch": 0.8726251826782555, + "learning_rate": 0.00011850186457334882, + "loss": 1.2083, + "step": 11345 + }, + { + "epoch": 0.8727020998384739, + "learning_rate": 0.00011836070159812878, + "loss": 1.1793, + "step": 11346 + }, + { + "epoch": 0.8727790169986924, + "learning_rate": 0.0001182196192982668, + "loss": 1.3855, + "step": 11347 + }, + { + "epoch": 0.8728559341589108, + "learning_rate": 0.00011807861768200196, + "loss": 1.7267, + "step": 11348 + }, + { + "epoch": 0.8729328513191293, + "learning_rate": 0.00011793769675756632, + "loss": 1.0257, + "step": 11349 + }, + { + "epoch": 0.8730097684793477, + "learning_rate": 0.00011779685653318916, + "loss": 1.1233, + "step": 11350 + }, + { + "epoch": 0.8730866856395662, + "learning_rate": 0.00011765609701709407, + "loss": 0.7911, + "step": 11351 + }, + { + "epoch": 0.8731636027997847, + "learning_rate": 0.00011751541821750012, + "loss": 1.1124, + "step": 11352 + }, + { + "epoch": 0.8732405199600031, + "learning_rate": 0.00011737482014262157, + "loss": 0.952, + "step": 11353 + }, + { + "epoch": 0.8733174371202215, + "learning_rate": 0.00011723430280066838, + "loss": 1.2934, + "step": 11354 + }, + { + "epoch": 0.87339435428044, + "learning_rate": 0.00011709386619984496, + "loss": 1.0059, + "step": 11355 + }, + { + "epoch": 0.8734712714406584, + "learning_rate": 0.00011695351034835227, + "loss": 1.4836, + "step": 11356 + }, + { + "epoch": 0.8735481886008769, + "learning_rate": 0.00011681323525438542, + "loss": 1.098, + "step": 11357 + }, + { + "epoch": 0.8736251057610953, + "learning_rate": 0.00011667304092613534, + "loss": 0.8327, + "step": 11358 + }, + { + "epoch": 0.8737020229213137, + "learning_rate": 0.000116532927371788, + "loss": 0.9604, + "step": 11359 + }, + { + "epoch": 0.8737789400815322, + "learning_rate": 0.00011639289459952467, + "loss": 1.1297, + "step": 11360 + }, + { + "epoch": 0.8738558572417506, + "learning_rate": 0.00011625294261752268, + "loss": 1.4265, + "step": 11361 + }, + { + "epoch": 0.873932774401969, + "learning_rate": 0.0001161130714339531, + "loss": 1.3408, + "step": 11362 + }, + { + "epoch": 0.8740096915621876, + "learning_rate": 0.00011597328105698329, + "loss": 1.0941, + "step": 11363 + }, + { + "epoch": 0.874086608722406, + "learning_rate": 0.00011583357149477636, + "loss": 0.908, + "step": 11364 + }, + { + "epoch": 0.8741635258826244, + "learning_rate": 0.00011569394275548978, + "loss": 1.6905, + "step": 11365 + }, + { + "epoch": 0.8742404430428429, + "learning_rate": 0.00011555439484727654, + "loss": 0.9637, + "step": 11366 + }, + { + "epoch": 0.8743173602030613, + "learning_rate": 0.00011541492777828493, + "loss": 1.2012, + "step": 11367 + }, + { + "epoch": 0.8743942773632798, + "learning_rate": 0.00011527554155665848, + "loss": 0.9998, + "step": 11368 + }, + { + "epoch": 0.8744711945234982, + "learning_rate": 0.00011513623619053648, + "loss": 1.4104, + "step": 11369 + }, + { + "epoch": 0.8745481116837166, + "learning_rate": 0.00011499701168805326, + "loss": 1.2468, + "step": 11370 + }, + { + "epoch": 0.8746250288439351, + "learning_rate": 0.00011485786805733733, + "loss": 0.9052, + "step": 11371 + }, + { + "epoch": 0.8747019460041535, + "learning_rate": 0.00011471880530651435, + "loss": 1.192, + "step": 11372 + }, + { + "epoch": 0.8747788631643719, + "learning_rate": 0.00011457982344370399, + "loss": 1.2228, + "step": 11373 + }, + { + "epoch": 0.8748557803245904, + "learning_rate": 0.00011444092247702159, + "loss": 1.0131, + "step": 11374 + }, + { + "epoch": 0.8749326974848088, + "learning_rate": 0.00011430210241457784, + "loss": 1.2772, + "step": 11375 + }, + { + "epoch": 0.8750096146450274, + "learning_rate": 0.00011416336326447807, + "loss": 1.3133, + "step": 11376 + }, + { + "epoch": 0.8750865318052458, + "learning_rate": 0.00011402470503482415, + "loss": 1.3592, + "step": 11377 + }, + { + "epoch": 0.8751634489654642, + "learning_rate": 0.0001138861277337121, + "loss": 1.0208, + "step": 11378 + }, + { + "epoch": 0.8752403661256827, + "learning_rate": 0.00011374763136923377, + "loss": 1.2365, + "step": 11379 + }, + { + "epoch": 0.8753172832859011, + "learning_rate": 0.00011360921594947571, + "loss": 1.454, + "step": 11380 + }, + { + "epoch": 0.8753942004461195, + "learning_rate": 0.0001134708814825206, + "loss": 1.2529, + "step": 11381 + }, + { + "epoch": 0.875471117606338, + "learning_rate": 0.00011333262797644533, + "loss": 1.3665, + "step": 11382 + }, + { + "epoch": 0.8755480347665564, + "learning_rate": 0.00011319445543932361, + "loss": 1.1139, + "step": 11383 + }, + { + "epoch": 0.8756249519267748, + "learning_rate": 0.00011305636387922263, + "loss": 1.2085, + "step": 11384 + }, + { + "epoch": 0.8757018690869933, + "learning_rate": 0.00011291835330420596, + "loss": 1.2508, + "step": 11385 + }, + { + "epoch": 0.8757787862472117, + "learning_rate": 0.00011278042372233233, + "loss": 1.2565, + "step": 11386 + }, + { + "epoch": 0.8758557034074302, + "learning_rate": 0.00011264257514165543, + "loss": 1.2357, + "step": 11387 + }, + { + "epoch": 0.8759326205676486, + "learning_rate": 0.00011250480757022452, + "loss": 1.2803, + "step": 11388 + }, + { + "epoch": 0.876009537727867, + "learning_rate": 0.00011236712101608382, + "loss": 1.3607, + "step": 11389 + }, + { + "epoch": 0.8760864548880856, + "learning_rate": 0.00011222951548727289, + "loss": 0.8425, + "step": 11390 + }, + { + "epoch": 0.876163372048304, + "learning_rate": 0.00011209199099182715, + "loss": 1.1194, + "step": 11391 + }, + { + "epoch": 0.8762402892085224, + "learning_rate": 0.00011195454753777668, + "loss": 1.0913, + "step": 11392 + }, + { + "epoch": 0.8763172063687409, + "learning_rate": 0.0001118171851331462, + "loss": 1.1493, + "step": 11393 + }, + { + "epoch": 0.8763941235289593, + "learning_rate": 0.00011167990378595733, + "loss": 1.5724, + "step": 11394 + }, + { + "epoch": 0.8764710406891778, + "learning_rate": 0.00011154270350422547, + "loss": 1.3541, + "step": 11395 + }, + { + "epoch": 0.8765479578493962, + "learning_rate": 0.00011140558429596271, + "loss": 1.1884, + "step": 11396 + }, + { + "epoch": 0.8766248750096146, + "learning_rate": 0.00011126854616917481, + "loss": 1.22, + "step": 11397 + }, + { + "epoch": 0.8767017921698331, + "learning_rate": 0.00011113158913186355, + "loss": 1.2094, + "step": 11398 + }, + { + "epoch": 0.8767787093300515, + "learning_rate": 0.00011099471319202653, + "loss": 1.1837, + "step": 11399 + }, + { + "epoch": 0.8768556264902699, + "learning_rate": 0.00011085791835765568, + "loss": 1.4606, + "step": 11400 + }, + { + "epoch": 0.8769325436504884, + "learning_rate": 0.00011072120463673879, + "loss": 0.9544, + "step": 11401 + }, + { + "epoch": 0.8770094608107069, + "learning_rate": 0.00011058457203725864, + "loss": 0.907, + "step": 11402 + }, + { + "epoch": 0.8770863779709254, + "learning_rate": 0.00011044802056719299, + "loss": 1.0953, + "step": 11403 + }, + { + "epoch": 0.8771632951311438, + "learning_rate": 0.00011031155023451584, + "loss": 1.5112, + "step": 11404 + }, + { + "epoch": 0.8772402122913622, + "learning_rate": 0.00011017516104719593, + "loss": 1.0694, + "step": 11405 + }, + { + "epoch": 0.8773171294515807, + "learning_rate": 0.00011003885301319627, + "loss": 0.8511, + "step": 11406 + }, + { + "epoch": 0.8773940466117991, + "learning_rate": 0.00010990262614047681, + "loss": 0.9416, + "step": 11407 + }, + { + "epoch": 0.8774709637720175, + "learning_rate": 0.00010976648043699167, + "loss": 1.1338, + "step": 11408 + }, + { + "epoch": 0.877547880932236, + "learning_rate": 0.00010963041591069067, + "loss": 0.8955, + "step": 11409 + }, + { + "epoch": 0.8776247980924544, + "learning_rate": 0.00010949443256951864, + "loss": 1.1223, + "step": 11410 + }, + { + "epoch": 0.8777017152526728, + "learning_rate": 0.00010935853042141586, + "loss": 0.9867, + "step": 11411 + }, + { + "epoch": 0.8777786324128913, + "learning_rate": 0.00010922270947431767, + "loss": 0.6568, + "step": 11412 + }, + { + "epoch": 0.8778555495731097, + "learning_rate": 0.00010908696973615506, + "loss": 1.3113, + "step": 11413 + }, + { + "epoch": 0.8779324667333283, + "learning_rate": 0.00010895131121485369, + "loss": 1.2506, + "step": 11414 + }, + { + "epoch": 0.8780093838935467, + "learning_rate": 0.00010881573391833521, + "loss": 1.4292, + "step": 11415 + }, + { + "epoch": 0.8780863010537651, + "learning_rate": 0.00010868023785451581, + "loss": 0.8695, + "step": 11416 + }, + { + "epoch": 0.8781632182139836, + "learning_rate": 0.00010854482303130698, + "loss": 1.0954, + "step": 11417 + }, + { + "epoch": 0.878240135374202, + "learning_rate": 0.00010840948945661671, + "loss": 1.2705, + "step": 11418 + }, + { + "epoch": 0.8783170525344204, + "learning_rate": 0.00010827423713834638, + "loss": 1.3253, + "step": 11419 + }, + { + "epoch": 0.8783939696946389, + "learning_rate": 0.00010813906608439333, + "loss": 0.9382, + "step": 11420 + }, + { + "epoch": 0.8784708868548573, + "learning_rate": 0.00010800397630265125, + "loss": 1.2581, + "step": 11421 + }, + { + "epoch": 0.8785478040150758, + "learning_rate": 0.00010786896780100763, + "loss": 1.7604, + "step": 11422 + }, + { + "epoch": 0.8786247211752942, + "learning_rate": 0.00010773404058734571, + "loss": 1.1944, + "step": 11423 + }, + { + "epoch": 0.8787016383355126, + "learning_rate": 0.00010759919466954432, + "loss": 1.0712, + "step": 11424 + }, + { + "epoch": 0.8787785554957311, + "learning_rate": 0.00010746443005547668, + "loss": 1.0611, + "step": 11425 + }, + { + "epoch": 0.8788554726559495, + "learning_rate": 0.0001073297467530125, + "loss": 1.0331, + "step": 11426 + }, + { + "epoch": 0.878932389816168, + "learning_rate": 0.00010719514477001613, + "loss": 0.9628, + "step": 11427 + }, + { + "epoch": 0.8790093069763865, + "learning_rate": 0.00010706062411434631, + "loss": 1.0546, + "step": 11428 + }, + { + "epoch": 0.8790862241366049, + "learning_rate": 0.00010692618479385857, + "loss": 1.1125, + "step": 11429 + }, + { + "epoch": 0.8791631412968233, + "learning_rate": 0.00010679182681640264, + "loss": 1.4027, + "step": 11430 + }, + { + "epoch": 0.8792400584570418, + "learning_rate": 0.00010665755018982392, + "loss": 1.3157, + "step": 11431 + }, + { + "epoch": 0.8793169756172602, + "learning_rate": 0.00010652335492196314, + "loss": 0.9309, + "step": 11432 + }, + { + "epoch": 0.8793938927774787, + "learning_rate": 0.0001063892410206555, + "loss": 1.253, + "step": 11433 + }, + { + "epoch": 0.8794708099376971, + "learning_rate": 0.00010625520849373277, + "loss": 0.9708, + "step": 11434 + }, + { + "epoch": 0.8795477270979155, + "learning_rate": 0.00010612125734902083, + "loss": 1.188, + "step": 11435 + }, + { + "epoch": 0.879624644258134, + "learning_rate": 0.00010598738759434146, + "loss": 1.3302, + "step": 11436 + }, + { + "epoch": 0.8797015614183524, + "learning_rate": 0.00010585359923751137, + "loss": 1.0724, + "step": 11437 + }, + { + "epoch": 0.8797784785785708, + "learning_rate": 0.00010571989228634249, + "loss": 1.0742, + "step": 11438 + }, + { + "epoch": 0.8798553957387893, + "learning_rate": 0.00010558626674864191, + "loss": 0.8101, + "step": 11439 + }, + { + "epoch": 0.8799323128990078, + "learning_rate": 0.00010545272263221289, + "loss": 0.9729, + "step": 11440 + }, + { + "epoch": 0.8800092300592263, + "learning_rate": 0.00010531925994485236, + "loss": 1.0727, + "step": 11441 + }, + { + "epoch": 0.8800861472194447, + "learning_rate": 0.00010518587869435392, + "loss": 1.1956, + "step": 11442 + }, + { + "epoch": 0.8801630643796631, + "learning_rate": 0.00010505257888850584, + "loss": 1.2086, + "step": 11443 + }, + { + "epoch": 0.8802399815398816, + "learning_rate": 0.00010491936053509122, + "loss": 1.3478, + "step": 11444 + }, + { + "epoch": 0.8803168987001, + "learning_rate": 0.00010478622364188917, + "loss": 1.0753, + "step": 11445 + }, + { + "epoch": 0.8803938158603184, + "learning_rate": 0.00010465316821667364, + "loss": 1.152, + "step": 11446 + }, + { + "epoch": 0.8804707330205369, + "learning_rate": 0.00010452019426721344, + "loss": 1.4289, + "step": 11447 + }, + { + "epoch": 0.8805476501807553, + "learning_rate": 0.00010438730180127382, + "loss": 1.1304, + "step": 11448 + }, + { + "epoch": 0.8806245673409737, + "learning_rate": 0.00010425449082661425, + "loss": 1.1484, + "step": 11449 + }, + { + "epoch": 0.8807014845011922, + "learning_rate": 0.00010412176135098905, + "loss": 1.1992, + "step": 11450 + }, + { + "epoch": 0.8807784016614106, + "learning_rate": 0.00010398911338214934, + "loss": 0.9903, + "step": 11451 + }, + { + "epoch": 0.8808553188216292, + "learning_rate": 0.00010385654692783991, + "loss": 1.2172, + "step": 11452 + }, + { + "epoch": 0.8809322359818476, + "learning_rate": 0.00010372406199580209, + "loss": 0.9563, + "step": 11453 + }, + { + "epoch": 0.881009153142066, + "learning_rate": 0.00010359165859377135, + "loss": 1.0242, + "step": 11454 + }, + { + "epoch": 0.8810860703022845, + "learning_rate": 0.00010345933672947866, + "loss": 1.1362, + "step": 11455 + }, + { + "epoch": 0.8811629874625029, + "learning_rate": 0.00010332709641065102, + "loss": 1.1323, + "step": 11456 + }, + { + "epoch": 0.8812399046227213, + "learning_rate": 0.00010319493764500976, + "loss": 1.2717, + "step": 11457 + }, + { + "epoch": 0.8813168217829398, + "learning_rate": 0.00010306286044027168, + "loss": 1.0881, + "step": 11458 + }, + { + "epoch": 0.8813937389431582, + "learning_rate": 0.00010293086480414932, + "loss": 0.8135, + "step": 11459 + }, + { + "epoch": 0.8814706561033767, + "learning_rate": 0.00010279895074434931, + "loss": 1.1207, + "step": 11460 + }, + { + "epoch": 0.8815475732635951, + "learning_rate": 0.000102667118268575, + "loss": 1.0214, + "step": 11461 + }, + { + "epoch": 0.8816244904238135, + "learning_rate": 0.0001025353673845239, + "loss": 1.2395, + "step": 11462 + }, + { + "epoch": 0.881701407584032, + "learning_rate": 0.00010240369809988903, + "loss": 1.1762, + "step": 11463 + }, + { + "epoch": 0.8817783247442504, + "learning_rate": 0.0001022721104223589, + "loss": 0.9741, + "step": 11464 + }, + { + "epoch": 0.8818552419044688, + "learning_rate": 0.00010214060435961686, + "loss": 1.0985, + "step": 11465 + }, + { + "epoch": 0.8819321590646874, + "learning_rate": 0.00010200917991934177, + "loss": 1.1406, + "step": 11466 + }, + { + "epoch": 0.8820090762249058, + "learning_rate": 0.0001018778371092075, + "loss": 1.1153, + "step": 11467 + }, + { + "epoch": 0.8820859933851242, + "learning_rate": 0.00010174657593688358, + "loss": 0.7835, + "step": 11468 + }, + { + "epoch": 0.8821629105453427, + "learning_rate": 0.00010161539641003404, + "loss": 1.6216, + "step": 11469 + }, + { + "epoch": 0.8822398277055611, + "learning_rate": 0.00010148429853631907, + "loss": 1.2589, + "step": 11470 + }, + { + "epoch": 0.8823167448657796, + "learning_rate": 0.00010135328232339341, + "loss": 1.2109, + "step": 11471 + }, + { + "epoch": 0.882393662025998, + "learning_rate": 0.0001012223477789071, + "loss": 0.8953, + "step": 11472 + }, + { + "epoch": 0.8824705791862164, + "learning_rate": 0.00010109149491050585, + "loss": 1.0965, + "step": 11473 + }, + { + "epoch": 0.8825474963464349, + "learning_rate": 0.00010096072372582987, + "loss": 0.8941, + "step": 11474 + }, + { + "epoch": 0.8826244135066533, + "learning_rate": 0.0001008300342325154, + "loss": 0.9229, + "step": 11475 + }, + { + "epoch": 0.8827013306668717, + "learning_rate": 0.00010069942643819368, + "loss": 1.3066, + "step": 11476 + }, + { + "epoch": 0.8827782478270902, + "learning_rate": 0.00010056890035049043, + "loss": 1.253, + "step": 11477 + }, + { + "epoch": 0.8828551649873086, + "learning_rate": 0.00010043845597702756, + "loss": 1.2333, + "step": 11478 + }, + { + "epoch": 0.8829320821475272, + "learning_rate": 0.00010030809332542195, + "loss": 1.166, + "step": 11479 + }, + { + "epoch": 0.8830089993077456, + "learning_rate": 0.00010017781240328555, + "loss": 1.142, + "step": 11480 + }, + { + "epoch": 0.883085916467964, + "learning_rate": 0.00010004761321822559, + "loss": 1.2751, + "step": 11481 + }, + { + "epoch": 0.8831628336281825, + "learning_rate": 9.991749577784415e-05, + "loss": 0.9672, + "step": 11482 + }, + { + "epoch": 0.8832397507884009, + "learning_rate": 9.978746008973949e-05, + "loss": 0.7696, + "step": 11483 + }, + { + "epoch": 0.8833166679486193, + "learning_rate": 9.96575061615042e-05, + "loss": 0.9291, + "step": 11484 + }, + { + "epoch": 0.8833935851088378, + "learning_rate": 9.952763400072673e-05, + "loss": 1.2045, + "step": 11485 + }, + { + "epoch": 0.8834705022690562, + "learning_rate": 9.939784361499016e-05, + "loss": 1.2746, + "step": 11486 + }, + { + "epoch": 0.8835474194292746, + "learning_rate": 9.92681350118731e-05, + "loss": 0.9339, + "step": 11487 + }, + { + "epoch": 0.8836243365894931, + "learning_rate": 9.913850819894931e-05, + "loss": 0.9674, + "step": 11488 + }, + { + "epoch": 0.8837012537497115, + "learning_rate": 9.900896318378844e-05, + "loss": 1.3145, + "step": 11489 + }, + { + "epoch": 0.88377817090993, + "learning_rate": 9.887949997395373e-05, + "loss": 1.014, + "step": 11490 + }, + { + "epoch": 0.8838550880701485, + "learning_rate": 9.87501185770055e-05, + "loss": 0.8902, + "step": 11491 + }, + { + "epoch": 0.8839320052303669, + "learning_rate": 9.862081900049819e-05, + "loss": 0.862, + "step": 11492 + }, + { + "epoch": 0.8840089223905854, + "learning_rate": 9.849160125198192e-05, + "loss": 1.4363, + "step": 11493 + }, + { + "epoch": 0.8840858395508038, + "learning_rate": 9.836246533900151e-05, + "loss": 1.3161, + "step": 11494 + }, + { + "epoch": 0.8841627567110222, + "learning_rate": 9.823341126909741e-05, + "loss": 1.1359, + "step": 11495 + }, + { + "epoch": 0.8842396738712407, + "learning_rate": 9.810443904980525e-05, + "loss": 1.0161, + "step": 11496 + }, + { + "epoch": 0.8843165910314591, + "learning_rate": 9.797554868865604e-05, + "loss": 1.4108, + "step": 11497 + }, + { + "epoch": 0.8843935081916776, + "learning_rate": 9.784674019317574e-05, + "loss": 1.1903, + "step": 11498 + }, + { + "epoch": 0.884470425351896, + "learning_rate": 9.77180135708855e-05, + "loss": 1.2703, + "step": 11499 + }, + { + "epoch": 0.8845473425121144, + "learning_rate": 9.758936882930197e-05, + "loss": 1.2174, + "step": 11500 + }, + { + "epoch": 0.8846242596723329, + "learning_rate": 9.746080597593648e-05, + "loss": 1.3184, + "step": 11501 + }, + { + "epoch": 0.8847011768325513, + "learning_rate": 9.73323250182967e-05, + "loss": 1.0426, + "step": 11502 + }, + { + "epoch": 0.8847780939927697, + "learning_rate": 9.720392596388395e-05, + "loss": 1.0595, + "step": 11503 + }, + { + "epoch": 0.8848550111529883, + "learning_rate": 9.707560882019589e-05, + "loss": 1.3072, + "step": 11504 + }, + { + "epoch": 0.8849319283132067, + "learning_rate": 9.694737359472522e-05, + "loss": 1.1604, + "step": 11505 + }, + { + "epoch": 0.8850088454734252, + "learning_rate": 9.68192202949596e-05, + "loss": 1.2572, + "step": 11506 + }, + { + "epoch": 0.8850857626336436, + "learning_rate": 9.669114892838221e-05, + "loss": 1.3691, + "step": 11507 + }, + { + "epoch": 0.885162679793862, + "learning_rate": 9.656315950247108e-05, + "loss": 1.0542, + "step": 11508 + }, + { + "epoch": 0.8852395969540805, + "learning_rate": 9.643525202469938e-05, + "loss": 1.2763, + "step": 11509 + }, + { + "epoch": 0.8853165141142989, + "learning_rate": 9.630742650253632e-05, + "loss": 1.1088, + "step": 11510 + }, + { + "epoch": 0.8853934312745173, + "learning_rate": 9.617968294344576e-05, + "loss": 1.1865, + "step": 11511 + }, + { + "epoch": 0.8854703484347358, + "learning_rate": 9.605202135488605e-05, + "loss": 1.0677, + "step": 11512 + }, + { + "epoch": 0.8855472655949542, + "learning_rate": 9.592444174431226e-05, + "loss": 1.1806, + "step": 11513 + }, + { + "epoch": 0.8856241827551726, + "learning_rate": 9.579694411917373e-05, + "loss": 0.933, + "step": 11514 + }, + { + "epoch": 0.8857010999153911, + "learning_rate": 9.566952848691484e-05, + "loss": 1.099, + "step": 11515 + }, + { + "epoch": 0.8857780170756095, + "learning_rate": 9.554219485497584e-05, + "loss": 1.0346, + "step": 11516 + }, + { + "epoch": 0.8858549342358281, + "learning_rate": 9.541494323079159e-05, + "loss": 1.0292, + "step": 11517 + }, + { + "epoch": 0.8859318513960465, + "learning_rate": 9.52877736217928e-05, + "loss": 0.9372, + "step": 11518 + }, + { + "epoch": 0.8860087685562649, + "learning_rate": 9.516068603540473e-05, + "loss": 1.1542, + "step": 11519 + }, + { + "epoch": 0.8860856857164834, + "learning_rate": 9.503368047904842e-05, + "loss": 1.0551, + "step": 11520 + }, + { + "epoch": 0.8861626028767018, + "learning_rate": 9.490675696013978e-05, + "loss": 0.8332, + "step": 11521 + }, + { + "epoch": 0.8862395200369202, + "learning_rate": 9.477991548608989e-05, + "loss": 1.1166, + "step": 11522 + }, + { + "epoch": 0.8863164371971387, + "learning_rate": 9.465315606430513e-05, + "loss": 0.8214, + "step": 11523 + }, + { + "epoch": 0.8863933543573571, + "learning_rate": 9.452647870218745e-05, + "loss": 1.0597, + "step": 11524 + }, + { + "epoch": 0.8864702715175756, + "learning_rate": 9.439988340713306e-05, + "loss": 1.2656, + "step": 11525 + }, + { + "epoch": 0.886547188677794, + "learning_rate": 9.42733701865346e-05, + "loss": 1.2766, + "step": 11526 + }, + { + "epoch": 0.8866241058380124, + "learning_rate": 9.41469390477791e-05, + "loss": 1.1914, + "step": 11527 + }, + { + "epoch": 0.886701022998231, + "learning_rate": 9.402058999824886e-05, + "loss": 1.3831, + "step": 11528 + }, + { + "epoch": 0.8867779401584494, + "learning_rate": 9.389432304532181e-05, + "loss": 1.2206, + "step": 11529 + }, + { + "epoch": 0.8868548573186678, + "learning_rate": 9.376813819637054e-05, + "loss": 1.1589, + "step": 11530 + }, + { + "epoch": 0.8869317744788863, + "learning_rate": 9.364203545876298e-05, + "loss": 1.1878, + "step": 11531 + }, + { + "epoch": 0.8870086916391047, + "learning_rate": 9.351601483986294e-05, + "loss": 1.0108, + "step": 11532 + }, + { + "epoch": 0.8870856087993231, + "learning_rate": 9.339007634702867e-05, + "loss": 0.888, + "step": 11533 + }, + { + "epoch": 0.8871625259595416, + "learning_rate": 9.326421998761347e-05, + "loss": 1.2003, + "step": 11534 + }, + { + "epoch": 0.88723944311976, + "learning_rate": 9.313844576896662e-05, + "loss": 1.4812, + "step": 11535 + }, + { + "epoch": 0.8873163602799785, + "learning_rate": 9.30127536984321e-05, + "loss": 0.8595, + "step": 11536 + }, + { + "epoch": 0.8873932774401969, + "learning_rate": 9.28871437833495e-05, + "loss": 1.125, + "step": 11537 + }, + { + "epoch": 0.8874701946004153, + "learning_rate": 9.276161603105282e-05, + "loss": 1.3038, + "step": 11538 + }, + { + "epoch": 0.8875471117606338, + "learning_rate": 9.263617044887185e-05, + "loss": 1.1381, + "step": 11539 + }, + { + "epoch": 0.8876240289208522, + "learning_rate": 9.251080704413173e-05, + "loss": 0.8655, + "step": 11540 + }, + { + "epoch": 0.8877009460810706, + "learning_rate": 9.23855258241526e-05, + "loss": 1.2277, + "step": 11541 + }, + { + "epoch": 0.8877778632412892, + "learning_rate": 9.226032679624946e-05, + "loss": 0.9171, + "step": 11542 + }, + { + "epoch": 0.8878547804015076, + "learning_rate": 9.213520996773295e-05, + "loss": 1.4198, + "step": 11543 + }, + { + "epoch": 0.8879316975617261, + "learning_rate": 9.201017534590855e-05, + "loss": 1.0412, + "step": 11544 + }, + { + "epoch": 0.8880086147219445, + "learning_rate": 9.18852229380776e-05, + "loss": 1.0927, + "step": 11545 + }, + { + "epoch": 0.8880855318821629, + "learning_rate": 9.176035275153626e-05, + "loss": 1.1522, + "step": 11546 + }, + { + "epoch": 0.8881624490423814, + "learning_rate": 9.163556479357504e-05, + "loss": 1.1021, + "step": 11547 + }, + { + "epoch": 0.8882393662025998, + "learning_rate": 9.151085907148127e-05, + "loss": 1.2182, + "step": 11548 + }, + { + "epoch": 0.8883162833628182, + "learning_rate": 9.138623559253612e-05, + "loss": 0.8963, + "step": 11549 + }, + { + "epoch": 0.8883932005230367, + "learning_rate": 9.126169436401693e-05, + "loss": 0.8112, + "step": 11550 + }, + { + "epoch": 0.8884701176832551, + "learning_rate": 9.11372353931954e-05, + "loss": 0.9991, + "step": 11551 + }, + { + "epoch": 0.8885470348434735, + "learning_rate": 9.101285868733905e-05, + "loss": 1.1995, + "step": 11552 + }, + { + "epoch": 0.888623952003692, + "learning_rate": 9.088856425371005e-05, + "loss": 1.1579, + "step": 11553 + }, + { + "epoch": 0.8887008691639104, + "learning_rate": 9.076435209956663e-05, + "loss": 0.8164, + "step": 11554 + }, + { + "epoch": 0.888777786324129, + "learning_rate": 9.064022223216129e-05, + "loss": 1.2062, + "step": 11555 + }, + { + "epoch": 0.8888547034843474, + "learning_rate": 9.051617465874223e-05, + "loss": 1.3546, + "step": 11556 + }, + { + "epoch": 0.8889316206445658, + "learning_rate": 9.039220938655268e-05, + "loss": 1.2306, + "step": 11557 + }, + { + "epoch": 0.8890085378047843, + "learning_rate": 9.026832642283067e-05, + "loss": 0.76, + "step": 11558 + }, + { + "epoch": 0.8890854549650027, + "learning_rate": 9.014452577481092e-05, + "loss": 1.4175, + "step": 11559 + }, + { + "epoch": 0.8891623721252211, + "learning_rate": 9.00208074497213e-05, + "loss": 1.4658, + "step": 11560 + }, + { + "epoch": 0.8892392892854396, + "learning_rate": 8.989717145478587e-05, + "loss": 1.3656, + "step": 11561 + }, + { + "epoch": 0.889316206445658, + "learning_rate": 8.977361779722453e-05, + "loss": 1.3049, + "step": 11562 + }, + { + "epoch": 0.8893931236058765, + "learning_rate": 8.965014648425135e-05, + "loss": 1.3541, + "step": 11563 + }, + { + "epoch": 0.8894700407660949, + "learning_rate": 8.952675752307588e-05, + "loss": 1.1139, + "step": 11564 + }, + { + "epoch": 0.8895469579263133, + "learning_rate": 8.940345092090318e-05, + "loss": 1.5293, + "step": 11565 + }, + { + "epoch": 0.8896238750865318, + "learning_rate": 8.928022668493269e-05, + "loss": 1.3007, + "step": 11566 + }, + { + "epoch": 0.8897007922467502, + "learning_rate": 8.915708482236013e-05, + "loss": 0.9564, + "step": 11567 + }, + { + "epoch": 0.8897777094069687, + "learning_rate": 8.903402534037608e-05, + "loss": 1.2317, + "step": 11568 + }, + { + "epoch": 0.8898546265671872, + "learning_rate": 8.89110482461653e-05, + "loss": 1.1855, + "step": 11569 + }, + { + "epoch": 0.8899315437274056, + "learning_rate": 8.878815354690923e-05, + "loss": 1.3336, + "step": 11570 + }, + { + "epoch": 0.890008460887624, + "learning_rate": 8.866534124978359e-05, + "loss": 1.3226, + "step": 11571 + }, + { + "epoch": 0.8900853780478425, + "learning_rate": 8.854261136195952e-05, + "loss": 1.1299, + "step": 11572 + }, + { + "epoch": 0.8901622952080609, + "learning_rate": 8.841996389060325e-05, + "loss": 0.9352, + "step": 11573 + }, + { + "epoch": 0.8902392123682794, + "learning_rate": 8.829739884287641e-05, + "loss": 0.9524, + "step": 11574 + }, + { + "epoch": 0.8903161295284978, + "learning_rate": 8.817491622593576e-05, + "loss": 0.7863, + "step": 11575 + }, + { + "epoch": 0.8903930466887162, + "learning_rate": 8.805251604693326e-05, + "loss": 1.3396, + "step": 11576 + }, + { + "epoch": 0.8904699638489347, + "learning_rate": 8.793019831301569e-05, + "loss": 1.0705, + "step": 11577 + }, + { + "epoch": 0.8905468810091531, + "learning_rate": 8.780796303132548e-05, + "loss": 0.8763, + "step": 11578 + }, + { + "epoch": 0.8906237981693715, + "learning_rate": 8.768581020900012e-05, + "loss": 1.0995, + "step": 11579 + }, + { + "epoch": 0.89070071532959, + "learning_rate": 8.756373985317189e-05, + "loss": 1.2277, + "step": 11580 + }, + { + "epoch": 0.8907776324898085, + "learning_rate": 8.744175197096943e-05, + "loss": 1.2304, + "step": 11581 + }, + { + "epoch": 0.890854549650027, + "learning_rate": 8.73198465695147e-05, + "loss": 1.4348, + "step": 11582 + }, + { + "epoch": 0.8909314668102454, + "learning_rate": 8.719802365592667e-05, + "loss": 1.0804, + "step": 11583 + }, + { + "epoch": 0.8910083839704638, + "learning_rate": 8.707628323731848e-05, + "loss": 1.4491, + "step": 11584 + }, + { + "epoch": 0.8910853011306823, + "learning_rate": 8.695462532079846e-05, + "loss": 1.4276, + "step": 11585 + }, + { + "epoch": 0.8911622182909007, + "learning_rate": 8.683304991347057e-05, + "loss": 1.1899, + "step": 11586 + }, + { + "epoch": 0.8912391354511191, + "learning_rate": 8.671155702243366e-05, + "loss": 0.8688, + "step": 11587 + }, + { + "epoch": 0.8913160526113376, + "learning_rate": 8.659014665478171e-05, + "loss": 1.2816, + "step": 11588 + }, + { + "epoch": 0.891392969771556, + "learning_rate": 8.646881881760421e-05, + "loss": 0.819, + "step": 11589 + }, + { + "epoch": 0.8914698869317744, + "learning_rate": 8.634757351798583e-05, + "loss": 1.4842, + "step": 11590 + }, + { + "epoch": 0.8915468040919929, + "learning_rate": 8.622641076300541e-05, + "loss": 1.0664, + "step": 11591 + }, + { + "epoch": 0.8916237212522113, + "learning_rate": 8.610533055973847e-05, + "loss": 0.8228, + "step": 11592 + }, + { + "epoch": 0.8917006384124299, + "learning_rate": 8.598433291525464e-05, + "loss": 0.9508, + "step": 11593 + }, + { + "epoch": 0.8917775555726483, + "learning_rate": 8.586341783661966e-05, + "loss": 0.8014, + "step": 11594 + }, + { + "epoch": 0.8918544727328667, + "learning_rate": 8.574258533089335e-05, + "loss": 1.1235, + "step": 11595 + }, + { + "epoch": 0.8919313898930852, + "learning_rate": 8.562183540513107e-05, + "loss": 1.075, + "step": 11596 + }, + { + "epoch": 0.8920083070533036, + "learning_rate": 8.550116806638403e-05, + "loss": 1.0194, + "step": 11597 + }, + { + "epoch": 0.892085224213522, + "learning_rate": 8.538058332169807e-05, + "loss": 1.0437, + "step": 11598 + }, + { + "epoch": 0.8921621413737405, + "learning_rate": 8.526008117811407e-05, + "loss": 1.1943, + "step": 11599 + }, + { + "epoch": 0.8922390585339589, + "learning_rate": 8.51396616426684e-05, + "loss": 0.6849, + "step": 11600 + }, + { + "epoch": 0.8923159756941774, + "learning_rate": 8.50193247223921e-05, + "loss": 1.1121, + "step": 11601 + }, + { + "epoch": 0.8923928928543958, + "learning_rate": 8.489907042431238e-05, + "loss": 1.0803, + "step": 11602 + }, + { + "epoch": 0.8924698100146142, + "learning_rate": 8.47788987554508e-05, + "loss": 1.3998, + "step": 11603 + }, + { + "epoch": 0.8925467271748327, + "learning_rate": 8.465880972282391e-05, + "loss": 0.9228, + "step": 11604 + }, + { + "epoch": 0.8926236443350511, + "learning_rate": 8.453880333344444e-05, + "loss": 1.2237, + "step": 11605 + }, + { + "epoch": 0.8927005614952696, + "learning_rate": 8.441887959431926e-05, + "loss": 1.1221, + "step": 11606 + }, + { + "epoch": 0.8927774786554881, + "learning_rate": 8.429903851245114e-05, + "loss": 1.1364, + "step": 11607 + }, + { + "epoch": 0.8928543958157065, + "learning_rate": 8.417928009483744e-05, + "loss": 1.4891, + "step": 11608 + }, + { + "epoch": 0.8929313129759249, + "learning_rate": 8.405960434847109e-05, + "loss": 1.404, + "step": 11609 + }, + { + "epoch": 0.8930082301361434, + "learning_rate": 8.394001128034001e-05, + "loss": 1.1013, + "step": 11610 + }, + { + "epoch": 0.8930851472963618, + "learning_rate": 8.38205008974276e-05, + "loss": 0.9517, + "step": 11611 + }, + { + "epoch": 0.8931620644565803, + "learning_rate": 8.370107320671194e-05, + "loss": 0.8864, + "step": 11612 + }, + { + "epoch": 0.8932389816167987, + "learning_rate": 8.35817282151668e-05, + "loss": 0.9827, + "step": 11613 + }, + { + "epoch": 0.8933158987770171, + "learning_rate": 8.346246592976075e-05, + "loss": 1.2023, + "step": 11614 + }, + { + "epoch": 0.8933928159372356, + "learning_rate": 8.334328635745724e-05, + "loss": 0.9427, + "step": 11615 + }, + { + "epoch": 0.893469733097454, + "learning_rate": 8.322418950521605e-05, + "loss": 1.4364, + "step": 11616 + }, + { + "epoch": 0.8935466502576724, + "learning_rate": 8.310517537999091e-05, + "loss": 1.4936, + "step": 11617 + }, + { + "epoch": 0.893623567417891, + "learning_rate": 8.298624398873095e-05, + "loss": 1.0511, + "step": 11618 + }, + { + "epoch": 0.8937004845781094, + "learning_rate": 8.286739533838111e-05, + "loss": 1.1745, + "step": 11619 + }, + { + "epoch": 0.8937774017383279, + "learning_rate": 8.274862943588102e-05, + "loss": 1.005, + "step": 11620 + }, + { + "epoch": 0.8938543188985463, + "learning_rate": 8.262994628816545e-05, + "loss": 1.0272, + "step": 11621 + }, + { + "epoch": 0.8939312360587647, + "learning_rate": 8.251134590216452e-05, + "loss": 0.984, + "step": 11622 + }, + { + "epoch": 0.8940081532189832, + "learning_rate": 8.23928282848032e-05, + "loss": 1.2201, + "step": 11623 + }, + { + "epoch": 0.8940850703792016, + "learning_rate": 8.227439344300214e-05, + "loss": 1.3198, + "step": 11624 + }, + { + "epoch": 0.89416198753942, + "learning_rate": 8.215604138367678e-05, + "loss": 0.9837, + "step": 11625 + }, + { + "epoch": 0.8942389046996385, + "learning_rate": 8.203777211373792e-05, + "loss": 1.4332, + "step": 11626 + }, + { + "epoch": 0.8943158218598569, + "learning_rate": 8.191958564009122e-05, + "loss": 1.2763, + "step": 11627 + }, + { + "epoch": 0.8943927390200754, + "learning_rate": 8.18014819696375e-05, + "loss": 0.8833, + "step": 11628 + }, + { + "epoch": 0.8944696561802938, + "learning_rate": 8.168346110927388e-05, + "loss": 1.1425, + "step": 11629 + }, + { + "epoch": 0.8945465733405122, + "learning_rate": 8.156552306589071e-05, + "loss": 1.0878, + "step": 11630 + }, + { + "epoch": 0.8946234905007308, + "learning_rate": 8.144766784637481e-05, + "loss": 0.8387, + "step": 11631 + }, + { + "epoch": 0.8947004076609492, + "learning_rate": 8.132989545760816e-05, + "loss": 1.0278, + "step": 11632 + }, + { + "epoch": 0.8947773248211676, + "learning_rate": 8.121220590646727e-05, + "loss": 1.01, + "step": 11633 + }, + { + "epoch": 0.8948542419813861, + "learning_rate": 8.109459919982448e-05, + "loss": 1.0479, + "step": 11634 + }, + { + "epoch": 0.8949311591416045, + "learning_rate": 8.097707534454678e-05, + "loss": 1.2038, + "step": 11635 + }, + { + "epoch": 0.8950080763018229, + "learning_rate": 8.085963434749639e-05, + "loss": 0.6738, + "step": 11636 + }, + { + "epoch": 0.8950849934620414, + "learning_rate": 8.074227621553076e-05, + "loss": 0.9903, + "step": 11637 + }, + { + "epoch": 0.8951619106222598, + "learning_rate": 8.062500095550313e-05, + "loss": 1.1692, + "step": 11638 + }, + { + "epoch": 0.8952388277824783, + "learning_rate": 8.050780857426082e-05, + "loss": 1.1571, + "step": 11639 + }, + { + "epoch": 0.8953157449426967, + "learning_rate": 8.039069907864688e-05, + "loss": 1.0482, + "step": 11640 + }, + { + "epoch": 0.8953926621029151, + "learning_rate": 8.027367247549949e-05, + "loss": 1.3399, + "step": 11641 + }, + { + "epoch": 0.8954695792631336, + "learning_rate": 8.015672877165187e-05, + "loss": 1.1453, + "step": 11642 + }, + { + "epoch": 0.895546496423352, + "learning_rate": 8.003986797393286e-05, + "loss": 1.1411, + "step": 11643 + }, + { + "epoch": 0.8956234135835704, + "learning_rate": 7.99230900891657e-05, + "loss": 1.2096, + "step": 11644 + }, + { + "epoch": 0.895700330743789, + "learning_rate": 7.980639512416893e-05, + "loss": 0.8934, + "step": 11645 + }, + { + "epoch": 0.8957772479040074, + "learning_rate": 7.968978308575725e-05, + "loss": 0.8719, + "step": 11646 + }, + { + "epoch": 0.8958541650642259, + "learning_rate": 7.957325398073922e-05, + "loss": 1.1567, + "step": 11647 + }, + { + "epoch": 0.8959310822244443, + "learning_rate": 7.945680781591924e-05, + "loss": 1.2281, + "step": 11648 + }, + { + "epoch": 0.8960079993846627, + "learning_rate": 7.93404445980967e-05, + "loss": 0.8987, + "step": 11649 + }, + { + "epoch": 0.8960849165448812, + "learning_rate": 7.922416433406598e-05, + "loss": 1.1464, + "step": 11650 + }, + { + "epoch": 0.8961618337050996, + "learning_rate": 7.910796703061734e-05, + "loss": 1.4386, + "step": 11651 + }, + { + "epoch": 0.896238750865318, + "learning_rate": 7.899185269453534e-05, + "loss": 1.4999, + "step": 11652 + }, + { + "epoch": 0.8963156680255365, + "learning_rate": 7.887582133259974e-05, + "loss": 1.1821, + "step": 11653 + }, + { + "epoch": 0.8963925851857549, + "learning_rate": 7.87598729515861e-05, + "loss": 1.2184, + "step": 11654 + }, + { + "epoch": 0.8964695023459733, + "learning_rate": 7.864400755826467e-05, + "loss": 1.1649, + "step": 11655 + }, + { + "epoch": 0.8965464195061918, + "learning_rate": 7.85282251594009e-05, + "loss": 1.0316, + "step": 11656 + }, + { + "epoch": 0.8966233366664103, + "learning_rate": 7.841252576175551e-05, + "loss": 1.9651, + "step": 11657 + }, + { + "epoch": 0.8967002538266288, + "learning_rate": 7.829690937208412e-05, + "loss": 1.2488, + "step": 11658 + }, + { + "epoch": 0.8967771709868472, + "learning_rate": 7.818137599713798e-05, + "loss": 0.976, + "step": 11659 + }, + { + "epoch": 0.8968540881470656, + "learning_rate": 7.806592564366322e-05, + "loss": 0.83, + "step": 11660 + }, + { + "epoch": 0.8969310053072841, + "learning_rate": 7.795055831840092e-05, + "loss": 1.1996, + "step": 11661 + }, + { + "epoch": 0.8970079224675025, + "learning_rate": 7.783527402808737e-05, + "loss": 1.0164, + "step": 11662 + }, + { + "epoch": 0.8970848396277209, + "learning_rate": 7.772007277945454e-05, + "loss": 1.0251, + "step": 11663 + }, + { + "epoch": 0.8971617567879394, + "learning_rate": 7.760495457922868e-05, + "loss": 1.5952, + "step": 11664 + }, + { + "epoch": 0.8972386739481578, + "learning_rate": 7.748991943413225e-05, + "loss": 0.7782, + "step": 11665 + }, + { + "epoch": 0.8973155911083763, + "learning_rate": 7.737496735088156e-05, + "loss": 1.1199, + "step": 11666 + }, + { + "epoch": 0.8973925082685947, + "learning_rate": 7.726009833618925e-05, + "loss": 1.1331, + "step": 11667 + }, + { + "epoch": 0.8974694254288131, + "learning_rate": 7.714531239676259e-05, + "loss": 1.1998, + "step": 11668 + }, + { + "epoch": 0.8975463425890317, + "learning_rate": 7.703060953930408e-05, + "loss": 1.2644, + "step": 11669 + }, + { + "epoch": 0.8976232597492501, + "learning_rate": 7.691598977051117e-05, + "loss": 1.1396, + "step": 11670 + }, + { + "epoch": 0.8977001769094685, + "learning_rate": 7.680145309707687e-05, + "loss": 1.3195, + "step": 11671 + }, + { + "epoch": 0.897777094069687, + "learning_rate": 7.668699952568863e-05, + "loss": 1.2649, + "step": 11672 + }, + { + "epoch": 0.8978540112299054, + "learning_rate": 7.657262906302998e-05, + "loss": 0.6796, + "step": 11673 + }, + { + "epoch": 0.8979309283901238, + "learning_rate": 7.645834171577936e-05, + "loss": 1.3104, + "step": 11674 + }, + { + "epoch": 0.8980078455503423, + "learning_rate": 7.634413749060915e-05, + "loss": 1.3297, + "step": 11675 + }, + { + "epoch": 0.8980847627105607, + "learning_rate": 7.62300163941888e-05, + "loss": 1.1521, + "step": 11676 + }, + { + "epoch": 0.8981616798707792, + "learning_rate": 7.611597843318118e-05, + "loss": 1.1726, + "step": 11677 + }, + { + "epoch": 0.8982385970309976, + "learning_rate": 7.600202361424613e-05, + "loss": 1.2127, + "step": 11678 + }, + { + "epoch": 0.898315514191216, + "learning_rate": 7.588815194403681e-05, + "loss": 0.9094, + "step": 11679 + }, + { + "epoch": 0.8983924313514345, + "learning_rate": 7.577436342920208e-05, + "loss": 1.1092, + "step": 11680 + }, + { + "epoch": 0.8984693485116529, + "learning_rate": 7.566065807638694e-05, + "loss": 1.2423, + "step": 11681 + }, + { + "epoch": 0.8985462656718713, + "learning_rate": 7.554703589223028e-05, + "loss": 1.2819, + "step": 11682 + }, + { + "epoch": 0.8986231828320899, + "learning_rate": 7.543349688336676e-05, + "loss": 0.9137, + "step": 11683 + }, + { + "epoch": 0.8987000999923083, + "learning_rate": 7.532004105642609e-05, + "loss": 0.9771, + "step": 11684 + }, + { + "epoch": 0.8987770171525268, + "learning_rate": 7.520666841803265e-05, + "loss": 1.3371, + "step": 11685 + }, + { + "epoch": 0.8988539343127452, + "learning_rate": 7.509337897480695e-05, + "loss": 1.3606, + "step": 11686 + }, + { + "epoch": 0.8989308514729636, + "learning_rate": 7.498017273336405e-05, + "loss": 1.3945, + "step": 11687 + }, + { + "epoch": 0.8990077686331821, + "learning_rate": 7.486704970031366e-05, + "loss": 0.8943, + "step": 11688 + }, + { + "epoch": 0.8990846857934005, + "learning_rate": 7.475400988226166e-05, + "loss": 1.0878, + "step": 11689 + }, + { + "epoch": 0.8991616029536189, + "learning_rate": 7.464105328580828e-05, + "loss": 1.2263, + "step": 11690 + }, + { + "epoch": 0.8992385201138374, + "learning_rate": 7.45281799175494e-05, + "loss": 1.3543, + "step": 11691 + }, + { + "epoch": 0.8993154372740558, + "learning_rate": 7.441538978407558e-05, + "loss": 0.594, + "step": 11692 + }, + { + "epoch": 0.8993923544342742, + "learning_rate": 7.430268289197289e-05, + "loss": 1.0057, + "step": 11693 + }, + { + "epoch": 0.8994692715944927, + "learning_rate": 7.419005924782224e-05, + "loss": 1.1454, + "step": 11694 + }, + { + "epoch": 0.8995461887547112, + "learning_rate": 7.407751885820002e-05, + "loss": 1.2633, + "step": 11695 + }, + { + "epoch": 0.8996231059149297, + "learning_rate": 7.396506172967766e-05, + "loss": 1.1363, + "step": 11696 + }, + { + "epoch": 0.8997000230751481, + "learning_rate": 7.385268786882154e-05, + "loss": 1.1124, + "step": 11697 + }, + { + "epoch": 0.8997769402353665, + "learning_rate": 7.374039728219329e-05, + "loss": 1.1851, + "step": 11698 + }, + { + "epoch": 0.899853857395585, + "learning_rate": 7.362818997634929e-05, + "loss": 1.4856, + "step": 11699 + }, + { + "epoch": 0.8999307745558034, + "learning_rate": 7.351606595784233e-05, + "loss": 0.9974, + "step": 11700 + }, + { + "epoch": 0.9000076917160218, + "learning_rate": 7.340402523321865e-05, + "loss": 1.0596, + "step": 11701 + }, + { + "epoch": 0.9000846088762403, + "learning_rate": 7.329206780902054e-05, + "loss": 1.3564, + "step": 11702 + }, + { + "epoch": 0.9001615260364587, + "learning_rate": 7.318019369178574e-05, + "loss": 1.1683, + "step": 11703 + }, + { + "epoch": 0.9002384431966772, + "learning_rate": 7.30684028880464e-05, + "loss": 1.2756, + "step": 11704 + }, + { + "epoch": 0.9003153603568956, + "learning_rate": 7.295669540433009e-05, + "loss": 1.3974, + "step": 11705 + }, + { + "epoch": 0.900392277517114, + "learning_rate": 7.28450712471596e-05, + "loss": 1.5267, + "step": 11706 + }, + { + "epoch": 0.9004691946773326, + "learning_rate": 7.273353042305241e-05, + "loss": 1.3843, + "step": 11707 + }, + { + "epoch": 0.900546111837551, + "learning_rate": 7.262207293852214e-05, + "loss": 0.9572, + "step": 11708 + }, + { + "epoch": 0.9006230289977694, + "learning_rate": 7.251069880007688e-05, + "loss": 1.2914, + "step": 11709 + }, + { + "epoch": 0.9006999461579879, + "learning_rate": 7.239940801421913e-05, + "loss": 1.0633, + "step": 11710 + }, + { + "epoch": 0.9007768633182063, + "learning_rate": 7.2288200587448e-05, + "loss": 0.968, + "step": 11711 + }, + { + "epoch": 0.9008537804784247, + "learning_rate": 7.217707652625666e-05, + "loss": 1.3555, + "step": 11712 + }, + { + "epoch": 0.9009306976386432, + "learning_rate": 7.206603583713405e-05, + "loss": 0.9122, + "step": 11713 + }, + { + "epoch": 0.9010076147988616, + "learning_rate": 7.19550785265637e-05, + "loss": 1.3127, + "step": 11714 + }, + { + "epoch": 0.9010845319590801, + "learning_rate": 7.18442046010242e-05, + "loss": 1.1678, + "step": 11715 + }, + { + "epoch": 0.9011614491192985, + "learning_rate": 7.173341406699041e-05, + "loss": 1.2193, + "step": 11716 + }, + { + "epoch": 0.9012383662795169, + "learning_rate": 7.162270693093098e-05, + "loss": 1.2674, + "step": 11717 + }, + { + "epoch": 0.9013152834397354, + "learning_rate": 7.151208319931024e-05, + "loss": 1.3066, + "step": 11718 + }, + { + "epoch": 0.9013922005999538, + "learning_rate": 7.140154287858786e-05, + "loss": 1.194, + "step": 11719 + }, + { + "epoch": 0.9014691177601722, + "learning_rate": 7.129108597521816e-05, + "loss": 1.1326, + "step": 11720 + }, + { + "epoch": 0.9015460349203908, + "learning_rate": 7.118071249565067e-05, + "loss": 1.4602, + "step": 11721 + }, + { + "epoch": 0.9016229520806092, + "learning_rate": 7.107042244633088e-05, + "loss": 1.2625, + "step": 11722 + }, + { + "epoch": 0.9016998692408277, + "learning_rate": 7.0960215833698e-05, + "loss": 1.1909, + "step": 11723 + }, + { + "epoch": 0.9017767864010461, + "learning_rate": 7.085009266418752e-05, + "loss": 0.9271, + "step": 11724 + }, + { + "epoch": 0.9018537035612645, + "learning_rate": 7.074005294422964e-05, + "loss": 0.9104, + "step": 11725 + }, + { + "epoch": 0.901930620721483, + "learning_rate": 7.063009668024956e-05, + "loss": 1.2579, + "step": 11726 + }, + { + "epoch": 0.9020075378817014, + "learning_rate": 7.05202238786678e-05, + "loss": 0.9298, + "step": 11727 + }, + { + "epoch": 0.9020844550419198, + "learning_rate": 7.04104345458999e-05, + "loss": 1.3275, + "step": 11728 + }, + { + "epoch": 0.9021613722021383, + "learning_rate": 7.030072868835641e-05, + "loss": 1.109, + "step": 11729 + }, + { + "epoch": 0.9022382893623567, + "learning_rate": 7.019110631244369e-05, + "loss": 1.2055, + "step": 11730 + }, + { + "epoch": 0.9023152065225752, + "learning_rate": 7.008156742456229e-05, + "loss": 0.8952, + "step": 11731 + }, + { + "epoch": 0.9023921236827936, + "learning_rate": 6.997211203110825e-05, + "loss": 1.0172, + "step": 11732 + }, + { + "epoch": 0.902469040843012, + "learning_rate": 6.986274013847316e-05, + "loss": 0.9172, + "step": 11733 + }, + { + "epoch": 0.9025459580032306, + "learning_rate": 6.975345175304287e-05, + "loss": 1.2287, + "step": 11734 + }, + { + "epoch": 0.902622875163449, + "learning_rate": 6.964424688119947e-05, + "loss": 1.2064, + "step": 11735 + }, + { + "epoch": 0.9026997923236674, + "learning_rate": 6.953512552931901e-05, + "loss": 1.1301, + "step": 11736 + }, + { + "epoch": 0.9027767094838859, + "learning_rate": 6.942608770377307e-05, + "loss": 1.4835, + "step": 11737 + }, + { + "epoch": 0.9028536266441043, + "learning_rate": 6.931713341092905e-05, + "loss": 1.1486, + "step": 11738 + }, + { + "epoch": 0.9029305438043227, + "learning_rate": 6.920826265714852e-05, + "loss": 1.1691, + "step": 11739 + }, + { + "epoch": 0.9030074609645412, + "learning_rate": 6.909947544878875e-05, + "loss": 1.2209, + "step": 11740 + }, + { + "epoch": 0.9030843781247596, + "learning_rate": 6.899077179220198e-05, + "loss": 1.6455, + "step": 11741 + }, + { + "epoch": 0.9031612952849781, + "learning_rate": 6.888215169373512e-05, + "loss": 1.069, + "step": 11742 + }, + { + "epoch": 0.9032382124451965, + "learning_rate": 6.877361515973096e-05, + "loss": 1.3897, + "step": 11743 + }, + { + "epoch": 0.9033151296054149, + "learning_rate": 6.86651621965274e-05, + "loss": 1.392, + "step": 11744 + }, + { + "epoch": 0.9033920467656334, + "learning_rate": 6.855679281045624e-05, + "loss": 1.4806, + "step": 11745 + }, + { + "epoch": 0.9034689639258519, + "learning_rate": 6.844850700784588e-05, + "loss": 1.0319, + "step": 11746 + }, + { + "epoch": 0.9035458810860703, + "learning_rate": 6.834030479501929e-05, + "loss": 1.1348, + "step": 11747 + }, + { + "epoch": 0.9036227982462888, + "learning_rate": 6.823218617829424e-05, + "loss": 1.0737, + "step": 11748 + }, + { + "epoch": 0.9036997154065072, + "learning_rate": 6.812415116398401e-05, + "loss": 1.2461, + "step": 11749 + }, + { + "epoch": 0.9037766325667257, + "learning_rate": 6.801619975839673e-05, + "loss": 0.7658, + "step": 11750 + }, + { + "epoch": 0.9038535497269441, + "learning_rate": 6.790833196783602e-05, + "loss": 1.1775, + "step": 11751 + }, + { + "epoch": 0.9039304668871625, + "learning_rate": 6.780054779860034e-05, + "loss": 1.1126, + "step": 11752 + }, + { + "epoch": 0.904007384047381, + "learning_rate": 6.769284725698316e-05, + "loss": 1.1478, + "step": 11753 + }, + { + "epoch": 0.9040843012075994, + "learning_rate": 6.758523034927344e-05, + "loss": 1.1646, + "step": 11754 + }, + { + "epoch": 0.9041612183678178, + "learning_rate": 6.7477697081755e-05, + "loss": 1.466, + "step": 11755 + }, + { + "epoch": 0.9042381355280363, + "learning_rate": 6.737024746070647e-05, + "loss": 1.5361, + "step": 11756 + }, + { + "epoch": 0.9043150526882547, + "learning_rate": 6.726288149240284e-05, + "loss": 0.8641, + "step": 11757 + }, + { + "epoch": 0.9043919698484731, + "learning_rate": 6.715559918311226e-05, + "loss": 1.3044, + "step": 11758 + }, + { + "epoch": 0.9044688870086917, + "learning_rate": 6.704840053909955e-05, + "loss": 1.1893, + "step": 11759 + }, + { + "epoch": 0.9045458041689101, + "learning_rate": 6.694128556662422e-05, + "loss": 0.8801, + "step": 11760 + }, + { + "epoch": 0.9046227213291286, + "learning_rate": 6.683425427194057e-05, + "loss": 1.366, + "step": 11761 + }, + { + "epoch": 0.904699638489347, + "learning_rate": 6.672730666129861e-05, + "loss": 1.1066, + "step": 11762 + }, + { + "epoch": 0.9047765556495654, + "learning_rate": 6.662044274094286e-05, + "loss": 1.1511, + "step": 11763 + }, + { + "epoch": 0.9048534728097839, + "learning_rate": 6.651366251711316e-05, + "loss": 1.4392, + "step": 11764 + }, + { + "epoch": 0.9049303899700023, + "learning_rate": 6.640696599604485e-05, + "loss": 1.2444, + "step": 11765 + }, + { + "epoch": 0.9050073071302207, + "learning_rate": 6.630035318396776e-05, + "loss": 1.4095, + "step": 11766 + }, + { + "epoch": 0.9050842242904392, + "learning_rate": 6.619382408710728e-05, + "loss": 0.9544, + "step": 11767 + }, + { + "epoch": 0.9051611414506576, + "learning_rate": 6.608737871168357e-05, + "loss": 1.0499, + "step": 11768 + }, + { + "epoch": 0.9052380586108761, + "learning_rate": 6.598101706391219e-05, + "loss": 1.0724, + "step": 11769 + }, + { + "epoch": 0.9053149757710945, + "learning_rate": 6.587473915000397e-05, + "loss": 1.0502, + "step": 11770 + }, + { + "epoch": 0.905391892931313, + "learning_rate": 6.57685449761643e-05, + "loss": 0.8725, + "step": 11771 + }, + { + "epoch": 0.9054688100915315, + "learning_rate": 6.566243454859372e-05, + "loss": 1.0236, + "step": 11772 + }, + { + "epoch": 0.9055457272517499, + "learning_rate": 6.55564078734886e-05, + "loss": 0.8038, + "step": 11773 + }, + { + "epoch": 0.9056226444119683, + "learning_rate": 6.545046495703966e-05, + "loss": 1.3477, + "step": 11774 + }, + { + "epoch": 0.9056995615721868, + "learning_rate": 6.534460580543327e-05, + "loss": 1.1728, + "step": 11775 + }, + { + "epoch": 0.9057764787324052, + "learning_rate": 6.523883042485052e-05, + "loss": 1.3801, + "step": 11776 + }, + { + "epoch": 0.9058533958926236, + "learning_rate": 6.51331388214676e-05, + "loss": 1.0779, + "step": 11777 + }, + { + "epoch": 0.9059303130528421, + "learning_rate": 6.502753100145609e-05, + "loss": 1.2253, + "step": 11778 + }, + { + "epoch": 0.9060072302130605, + "learning_rate": 6.492200697098255e-05, + "loss": 0.7492, + "step": 11779 + }, + { + "epoch": 0.906084147373279, + "learning_rate": 6.481656673620873e-05, + "loss": 1.4086, + "step": 11780 + }, + { + "epoch": 0.9061610645334974, + "learning_rate": 6.471121030329119e-05, + "loss": 1.1758, + "step": 11781 + }, + { + "epoch": 0.9062379816937158, + "learning_rate": 6.460593767838203e-05, + "loss": 1.1186, + "step": 11782 + }, + { + "epoch": 0.9063148988539343, + "learning_rate": 6.450074886762813e-05, + "loss": 1.094, + "step": 11783 + }, + { + "epoch": 0.9063918160141528, + "learning_rate": 6.439564387717145e-05, + "loss": 1.0366, + "step": 11784 + }, + { + "epoch": 0.9064687331743712, + "learning_rate": 6.429062271314923e-05, + "loss": 1.0036, + "step": 11785 + }, + { + "epoch": 0.9065456503345897, + "learning_rate": 6.418568538169372e-05, + "loss": 1.1917, + "step": 11786 + }, + { + "epoch": 0.9066225674948081, + "learning_rate": 6.408083188893255e-05, + "loss": 1.2543, + "step": 11787 + }, + { + "epoch": 0.9066994846550266, + "learning_rate": 6.397606224098812e-05, + "loss": 1.7143, + "step": 11788 + }, + { + "epoch": 0.906776401815245, + "learning_rate": 6.387137644397789e-05, + "loss": 1.221, + "step": 11789 + }, + { + "epoch": 0.9068533189754634, + "learning_rate": 6.376677450401481e-05, + "loss": 1.5333, + "step": 11790 + }, + { + "epoch": 0.9069302361356819, + "learning_rate": 6.366225642720629e-05, + "loss": 1.2567, + "step": 11791 + }, + { + "epoch": 0.9070071532959003, + "learning_rate": 6.355782221965583e-05, + "loss": 1.2335, + "step": 11792 + }, + { + "epoch": 0.9070840704561187, + "learning_rate": 6.34534718874612e-05, + "loss": 1.305, + "step": 11793 + }, + { + "epoch": 0.9071609876163372, + "learning_rate": 6.334920543671519e-05, + "loss": 0.6841, + "step": 11794 + }, + { + "epoch": 0.9072379047765556, + "learning_rate": 6.324502287350659e-05, + "loss": 1.4732, + "step": 11795 + }, + { + "epoch": 0.907314821936774, + "learning_rate": 6.314092420391837e-05, + "loss": 1.0183, + "step": 11796 + }, + { + "epoch": 0.9073917390969926, + "learning_rate": 6.303690943402902e-05, + "loss": 1.175, + "step": 11797 + }, + { + "epoch": 0.907468656257211, + "learning_rate": 6.293297856991214e-05, + "loss": 1.0436, + "step": 11798 + }, + { + "epoch": 0.9075455734174295, + "learning_rate": 6.282913161763609e-05, + "loss": 0.7594, + "step": 11799 + }, + { + "epoch": 0.9076224905776479, + "learning_rate": 6.272536858326499e-05, + "loss": 0.7875, + "step": 11800 + }, + { + "epoch": 0.9076994077378663, + "learning_rate": 6.262168947285768e-05, + "loss": 1.0691, + "step": 11801 + }, + { + "epoch": 0.9077763248980848, + "learning_rate": 6.25180942924678e-05, + "loss": 1.4212, + "step": 11802 + }, + { + "epoch": 0.9078532420583032, + "learning_rate": 6.241458304814434e-05, + "loss": 0.9927, + "step": 11803 + }, + { + "epoch": 0.9079301592185216, + "learning_rate": 6.231115574593182e-05, + "loss": 1.0773, + "step": 11804 + }, + { + "epoch": 0.9080070763787401, + "learning_rate": 6.220781239186906e-05, + "loss": 1.1155, + "step": 11805 + }, + { + "epoch": 0.9080839935389585, + "learning_rate": 6.210455299199075e-05, + "loss": 1.5424, + "step": 11806 + }, + { + "epoch": 0.908160910699177, + "learning_rate": 6.200137755232587e-05, + "loss": 1.1167, + "step": 11807 + }, + { + "epoch": 0.9082378278593954, + "learning_rate": 6.189828607889947e-05, + "loss": 1.1444, + "step": 11808 + }, + { + "epoch": 0.9083147450196138, + "learning_rate": 6.179527857773071e-05, + "loss": 0.7555, + "step": 11809 + }, + { + "epoch": 0.9083916621798324, + "learning_rate": 6.169235505483462e-05, + "loss": 1.2785, + "step": 11810 + }, + { + "epoch": 0.9084685793400508, + "learning_rate": 6.158951551622088e-05, + "loss": 1.3092, + "step": 11811 + }, + { + "epoch": 0.9085454965002692, + "learning_rate": 6.148675996789437e-05, + "loss": 1.1576, + "step": 11812 + }, + { + "epoch": 0.9086224136604877, + "learning_rate": 6.138408841585497e-05, + "loss": 1.1901, + "step": 11813 + }, + { + "epoch": 0.9086993308207061, + "learning_rate": 6.128150086609819e-05, + "loss": 1.3538, + "step": 11814 + }, + { + "epoch": 0.9087762479809245, + "learning_rate": 6.117899732461408e-05, + "loss": 0.8202, + "step": 11815 + }, + { + "epoch": 0.908853165141143, + "learning_rate": 6.107657779738735e-05, + "loss": 0.7978, + "step": 11816 + }, + { + "epoch": 0.9089300823013614, + "learning_rate": 6.097424229039922e-05, + "loss": 1.192, + "step": 11817 + }, + { + "epoch": 0.9090069994615799, + "learning_rate": 6.087199080962458e-05, + "loss": 1.1199, + "step": 11818 + }, + { + "epoch": 0.9090839166217983, + "learning_rate": 6.0769823361034484e-05, + "loss": 1.6453, + "step": 11819 + }, + { + "epoch": 0.9091608337820167, + "learning_rate": 6.066773995059416e-05, + "loss": 1.581, + "step": 11820 + }, + { + "epoch": 0.9092377509422352, + "learning_rate": 6.056574058426451e-05, + "loss": 1.4239, + "step": 11821 + }, + { + "epoch": 0.9093146681024536, + "learning_rate": 6.0463825268001425e-05, + "loss": 1.0648, + "step": 11822 + }, + { + "epoch": 0.909391585262672, + "learning_rate": 6.0361994007755826e-05, + "loss": 1.4888, + "step": 11823 + }, + { + "epoch": 0.9094685024228906, + "learning_rate": 6.026024680947379e-05, + "loss": 0.6561, + "step": 11824 + }, + { + "epoch": 0.909545419583109, + "learning_rate": 6.0158583679096547e-05, + "loss": 1.6584, + "step": 11825 + }, + { + "epoch": 0.9096223367433275, + "learning_rate": 6.005700462255986e-05, + "loss": 1.0234, + "step": 11826 + }, + { + "epoch": 0.9096992539035459, + "learning_rate": 5.995550964579549e-05, + "loss": 0.9035, + "step": 11827 + }, + { + "epoch": 0.9097761710637643, + "learning_rate": 5.985409875472986e-05, + "loss": 1.1757, + "step": 11828 + }, + { + "epoch": 0.9098530882239828, + "learning_rate": 5.975277195528406e-05, + "loss": 1.0951, + "step": 11829 + }, + { + "epoch": 0.9099300053842012, + "learning_rate": 5.965152925337503e-05, + "loss": 1.3039, + "step": 11830 + }, + { + "epoch": 0.9100069225444196, + "learning_rate": 5.9550370654914195e-05, + "loss": 1.0342, + "step": 11831 + }, + { + "epoch": 0.9100838397046381, + "learning_rate": 5.944929616580852e-05, + "loss": 1.0479, + "step": 11832 + }, + { + "epoch": 0.9101607568648565, + "learning_rate": 5.934830579195977e-05, + "loss": 0.9659, + "step": 11833 + }, + { + "epoch": 0.9102376740250749, + "learning_rate": 5.924739953926472e-05, + "loss": 1.0638, + "step": 11834 + }, + { + "epoch": 0.9103145911852935, + "learning_rate": 5.914657741361551e-05, + "loss": 1.254, + "step": 11835 + }, + { + "epoch": 0.9103915083455119, + "learning_rate": 5.904583942089925e-05, + "loss": 1.3537, + "step": 11836 + }, + { + "epoch": 0.9104684255057304, + "learning_rate": 5.8945185566998405e-05, + "loss": 1.1721, + "step": 11837 + }, + { + "epoch": 0.9105453426659488, + "learning_rate": 5.8844615857789774e-05, + "loss": 1.312, + "step": 11838 + }, + { + "epoch": 0.9106222598261672, + "learning_rate": 5.874413029914616e-05, + "loss": 0.9252, + "step": 11839 + }, + { + "epoch": 0.9106991769863857, + "learning_rate": 5.864372889693454e-05, + "loss": 1.2886, + "step": 11840 + }, + { + "epoch": 0.9107760941466041, + "learning_rate": 5.854341165701804e-05, + "loss": 1.248, + "step": 11841 + }, + { + "epoch": 0.9108530113068225, + "learning_rate": 5.844317858525383e-05, + "loss": 1.3042, + "step": 11842 + }, + { + "epoch": 0.910929928467041, + "learning_rate": 5.834302968749472e-05, + "loss": 1.1956, + "step": 11843 + }, + { + "epoch": 0.9110068456272594, + "learning_rate": 5.824296496958853e-05, + "loss": 1.3257, + "step": 11844 + }, + { + "epoch": 0.9110837627874779, + "learning_rate": 5.8142984437378424e-05, + "loss": 1.1969, + "step": 11845 + }, + { + "epoch": 0.9111606799476963, + "learning_rate": 5.804308809670189e-05, + "loss": 1.0234, + "step": 11846 + }, + { + "epoch": 0.9112375971079147, + "learning_rate": 5.794327595339227e-05, + "loss": 1.3848, + "step": 11847 + }, + { + "epoch": 0.9113145142681333, + "learning_rate": 5.784354801327757e-05, + "loss": 1.2557, + "step": 11848 + }, + { + "epoch": 0.9113914314283517, + "learning_rate": 5.7743904282181293e-05, + "loss": 1.4429, + "step": 11849 + }, + { + "epoch": 0.9114683485885701, + "learning_rate": 5.764434476592162e-05, + "loss": 1.1427, + "step": 11850 + }, + { + "epoch": 0.9115452657487886, + "learning_rate": 5.75448694703114e-05, + "loss": 0.9271, + "step": 11851 + }, + { + "epoch": 0.911622182909007, + "learning_rate": 5.744547840115982e-05, + "loss": 1.1305, + "step": 11852 + }, + { + "epoch": 0.9116991000692255, + "learning_rate": 5.7346171564269914e-05, + "loss": 1.236, + "step": 11853 + }, + { + "epoch": 0.9117760172294439, + "learning_rate": 5.7246948965441025e-05, + "loss": 1.1155, + "step": 11854 + }, + { + "epoch": 0.9118529343896623, + "learning_rate": 5.714781061046603e-05, + "loss": 1.3916, + "step": 11855 + }, + { + "epoch": 0.9119298515498808, + "learning_rate": 5.704875650513414e-05, + "loss": 1.1753, + "step": 11856 + }, + { + "epoch": 0.9120067687100992, + "learning_rate": 5.694978665522921e-05, + "loss": 1.5762, + "step": 11857 + }, + { + "epoch": 0.9120836858703176, + "learning_rate": 5.6850901066530126e-05, + "loss": 1.3602, + "step": 11858 + }, + { + "epoch": 0.9121606030305361, + "learning_rate": 5.6752099744810937e-05, + "loss": 1.1882, + "step": 11859 + }, + { + "epoch": 0.9122375201907545, + "learning_rate": 5.6653382695840696e-05, + "loss": 1.444, + "step": 11860 + }, + { + "epoch": 0.912314437350973, + "learning_rate": 5.655474992538378e-05, + "loss": 0.9654, + "step": 11861 + }, + { + "epoch": 0.9123913545111915, + "learning_rate": 5.6456201439199105e-05, + "loss": 1.1835, + "step": 11862 + }, + { + "epoch": 0.9124682716714099, + "learning_rate": 5.635773724304172e-05, + "loss": 1.0858, + "step": 11863 + }, + { + "epoch": 0.9125451888316284, + "learning_rate": 5.6259357342660024e-05, + "loss": 0.9491, + "step": 11864 + }, + { + "epoch": 0.9126221059918468, + "learning_rate": 5.6161061743799424e-05, + "loss": 1.1276, + "step": 11865 + }, + { + "epoch": 0.9126990231520652, + "learning_rate": 5.606285045219917e-05, + "loss": 1.0894, + "step": 11866 + }, + { + "epoch": 0.9127759403122837, + "learning_rate": 5.5964723473593826e-05, + "loss": 1.0319, + "step": 11867 + }, + { + "epoch": 0.9128528574725021, + "learning_rate": 5.5866680813713324e-05, + "loss": 1.1197, + "step": 11868 + }, + { + "epoch": 0.9129297746327205, + "learning_rate": 5.576872247828224e-05, + "loss": 1.1269, + "step": 11869 + }, + { + "epoch": 0.913006691792939, + "learning_rate": 5.567084847302051e-05, + "loss": 1.0615, + "step": 11870 + }, + { + "epoch": 0.9130836089531574, + "learning_rate": 5.557305880364322e-05, + "loss": 1.0588, + "step": 11871 + }, + { + "epoch": 0.913160526113376, + "learning_rate": 5.5475353475860466e-05, + "loss": 1.1178, + "step": 11872 + }, + { + "epoch": 0.9132374432735944, + "learning_rate": 5.5377732495377364e-05, + "loss": 1.5532, + "step": 11873 + }, + { + "epoch": 0.9133143604338128, + "learning_rate": 5.5280195867893843e-05, + "loss": 1.2373, + "step": 11874 + }, + { + "epoch": 0.9133912775940313, + "learning_rate": 5.518274359910519e-05, + "loss": 1.2619, + "step": 11875 + }, + { + "epoch": 0.9134681947542497, + "learning_rate": 5.508537569470234e-05, + "loss": 1.2821, + "step": 11876 + }, + { + "epoch": 0.9135451119144681, + "learning_rate": 5.49880921603701e-05, + "loss": 1.1162, + "step": 11877 + }, + { + "epoch": 0.9136220290746866, + "learning_rate": 5.489089300178873e-05, + "loss": 0.8705, + "step": 11878 + }, + { + "epoch": 0.913698946234905, + "learning_rate": 5.479377822463455e-05, + "loss": 1.3552, + "step": 11879 + }, + { + "epoch": 0.9137758633951234, + "learning_rate": 5.4696747834577846e-05, + "loss": 0.9738, + "step": 11880 + }, + { + "epoch": 0.9138527805553419, + "learning_rate": 5.459980183728408e-05, + "loss": 1.042, + "step": 11881 + }, + { + "epoch": 0.9139296977155603, + "learning_rate": 5.45029402384144e-05, + "loss": 1.2623, + "step": 11882 + }, + { + "epoch": 0.9140066148757788, + "learning_rate": 5.440616304362428e-05, + "loss": 1.2096, + "step": 11883 + }, + { + "epoch": 0.9140835320359972, + "learning_rate": 5.430947025856503e-05, + "loss": 1.2831, + "step": 11884 + }, + { + "epoch": 0.9141604491962156, + "learning_rate": 5.42128618888828e-05, + "loss": 1.2196, + "step": 11885 + }, + { + "epoch": 0.9142373663564342, + "learning_rate": 5.4116337940217756e-05, + "loss": 0.8838, + "step": 11886 + }, + { + "epoch": 0.9143142835166526, + "learning_rate": 5.401989841820687e-05, + "loss": 1.0429, + "step": 11887 + }, + { + "epoch": 0.914391200676871, + "learning_rate": 5.392354332848115e-05, + "loss": 1.365, + "step": 11888 + }, + { + "epoch": 0.9144681178370895, + "learning_rate": 5.382727267666676e-05, + "loss": 0.9185, + "step": 11889 + }, + { + "epoch": 0.9145450349973079, + "learning_rate": 5.3731086468385035e-05, + "loss": 1.2586, + "step": 11890 + }, + { + "epoch": 0.9146219521575264, + "learning_rate": 5.3634984709252156e-05, + "loss": 1.1925, + "step": 11891 + }, + { + "epoch": 0.9146988693177448, + "learning_rate": 5.353896740488029e-05, + "loss": 1.3063, + "step": 11892 + }, + { + "epoch": 0.9147757864779632, + "learning_rate": 5.344303456087546e-05, + "loss": 1.0249, + "step": 11893 + }, + { + "epoch": 0.9148527036381817, + "learning_rate": 5.3347186182839344e-05, + "loss": 1.1116, + "step": 11894 + }, + { + "epoch": 0.9149296207984001, + "learning_rate": 5.325142227636881e-05, + "loss": 1.2574, + "step": 11895 + }, + { + "epoch": 0.9150065379586185, + "learning_rate": 5.3155742847055544e-05, + "loss": 0.8567, + "step": 11896 + }, + { + "epoch": 0.915083455118837, + "learning_rate": 5.306014790048608e-05, + "loss": 1.3659, + "step": 11897 + }, + { + "epoch": 0.9151603722790554, + "learning_rate": 5.296463744224278e-05, + "loss": 1.1399, + "step": 11898 + }, + { + "epoch": 0.9152372894392738, + "learning_rate": 5.2869211477902354e-05, + "loss": 1.3039, + "step": 11899 + }, + { + "epoch": 0.9153142065994924, + "learning_rate": 5.277387001303652e-05, + "loss": 0.8714, + "step": 11900 + }, + { + "epoch": 0.9153911237597108, + "learning_rate": 5.267861305321298e-05, + "loss": 1.0065, + "step": 11901 + }, + { + "epoch": 0.9154680409199293, + "learning_rate": 5.258344060399345e-05, + "loss": 1.0591, + "step": 11902 + }, + { + "epoch": 0.9155449580801477, + "learning_rate": 5.2488352670935325e-05, + "loss": 1.2966, + "step": 11903 + }, + { + "epoch": 0.9156218752403661, + "learning_rate": 5.239334925959099e-05, + "loss": 1.0625, + "step": 11904 + }, + { + "epoch": 0.9156987924005846, + "learning_rate": 5.2298430375507354e-05, + "loss": 1.1937, + "step": 11905 + }, + { + "epoch": 0.915775709560803, + "learning_rate": 5.220359602422731e-05, + "loss": 0.8867, + "step": 11906 + }, + { + "epoch": 0.9158526267210214, + "learning_rate": 5.210884621128825e-05, + "loss": 1.236, + "step": 11907 + }, + { + "epoch": 0.9159295438812399, + "learning_rate": 5.2014180942222435e-05, + "loss": 1.1658, + "step": 11908 + }, + { + "epoch": 0.9160064610414583, + "learning_rate": 5.191960022255776e-05, + "loss": 1.125, + "step": 11909 + }, + { + "epoch": 0.9160833782016768, + "learning_rate": 5.182510405781649e-05, + "loss": 1.4948, + "step": 11910 + }, + { + "epoch": 0.9161602953618952, + "learning_rate": 5.173069245351719e-05, + "loss": 1.0294, + "step": 11911 + }, + { + "epoch": 0.9162372125221137, + "learning_rate": 5.163636541517181e-05, + "loss": 1.3724, + "step": 11912 + }, + { + "epoch": 0.9163141296823322, + "learning_rate": 5.154212294828825e-05, + "loss": 0.9865, + "step": 11913 + }, + { + "epoch": 0.9163910468425506, + "learning_rate": 5.144796505836996e-05, + "loss": 1.3347, + "step": 11914 + }, + { + "epoch": 0.916467964002769, + "learning_rate": 5.13538917509147e-05, + "loss": 1.3131, + "step": 11915 + }, + { + "epoch": 0.9165448811629875, + "learning_rate": 5.1259903031415254e-05, + "loss": 1.2446, + "step": 11916 + }, + { + "epoch": 0.9166217983232059, + "learning_rate": 5.1165998905360055e-05, + "loss": 1.2039, + "step": 11917 + }, + { + "epoch": 0.9166987154834243, + "learning_rate": 5.107217937823205e-05, + "loss": 1.0967, + "step": 11918 + }, + { + "epoch": 0.9167756326436428, + "learning_rate": 5.097844445550953e-05, + "loss": 1.3681, + "step": 11919 + }, + { + "epoch": 0.9168525498038612, + "learning_rate": 5.0884794142665944e-05, + "loss": 0.7595, + "step": 11920 + }, + { + "epoch": 0.9169294669640797, + "learning_rate": 5.079122844516909e-05, + "loss": 1.2948, + "step": 11921 + }, + { + "epoch": 0.9170063841242981, + "learning_rate": 5.069774736848309e-05, + "loss": 1.2925, + "step": 11922 + }, + { + "epoch": 0.9170833012845165, + "learning_rate": 5.060435091806592e-05, + "loss": 0.9102, + "step": 11923 + }, + { + "epoch": 0.917160218444735, + "learning_rate": 5.051103909937121e-05, + "loss": 1.0557, + "step": 11924 + }, + { + "epoch": 0.9172371356049535, + "learning_rate": 5.041781191784761e-05, + "loss": 1.4557, + "step": 11925 + }, + { + "epoch": 0.9173140527651719, + "learning_rate": 5.032466937893859e-05, + "loss": 0.9919, + "step": 11926 + }, + { + "epoch": 0.9173909699253904, + "learning_rate": 5.0231611488082806e-05, + "loss": 1.1075, + "step": 11927 + }, + { + "epoch": 0.9174678870856088, + "learning_rate": 5.0138638250714406e-05, + "loss": 1.0142, + "step": 11928 + }, + { + "epoch": 0.9175448042458273, + "learning_rate": 5.004574967226172e-05, + "loss": 0.9889, + "step": 11929 + }, + { + "epoch": 0.9176217214060457, + "learning_rate": 4.995294575814907e-05, + "loss": 0.9621, + "step": 11930 + }, + { + "epoch": 0.9176986385662641, + "learning_rate": 4.986022651379496e-05, + "loss": 1.2588, + "step": 11931 + }, + { + "epoch": 0.9177755557264826, + "learning_rate": 4.976759194461339e-05, + "loss": 0.7398, + "step": 11932 + }, + { + "epoch": 0.917852472886701, + "learning_rate": 4.9675042056013695e-05, + "loss": 1.2111, + "step": 11933 + }, + { + "epoch": 0.9179293900469194, + "learning_rate": 4.9582576853399894e-05, + "loss": 1.2539, + "step": 11934 + }, + { + "epoch": 0.9180063072071379, + "learning_rate": 4.9490196342170836e-05, + "loss": 1.0612, + "step": 11935 + }, + { + "epoch": 0.9180832243673563, + "learning_rate": 4.939790052772103e-05, + "loss": 1.5774, + "step": 11936 + }, + { + "epoch": 0.9181601415275747, + "learning_rate": 4.93056894154395e-05, + "loss": 0.9043, + "step": 11937 + }, + { + "epoch": 0.9182370586877933, + "learning_rate": 4.9213563010710614e-05, + "loss": 1.3238, + "step": 11938 + }, + { + "epoch": 0.9183139758480117, + "learning_rate": 4.91215213189139e-05, + "loss": 1.0755, + "step": 11939 + }, + { + "epoch": 0.9183908930082302, + "learning_rate": 4.902956434542338e-05, + "loss": 1.2284, + "step": 11940 + }, + { + "epoch": 0.9184678101684486, + "learning_rate": 4.893769209560894e-05, + "loss": 0.9964, + "step": 11941 + }, + { + "epoch": 0.918544727328667, + "learning_rate": 4.8845904574834956e-05, + "loss": 1.2363, + "step": 11942 + }, + { + "epoch": 0.9186216444888855, + "learning_rate": 4.875420178846079e-05, + "loss": 1.29, + "step": 11943 + }, + { + "epoch": 0.9186985616491039, + "learning_rate": 4.8662583741841515e-05, + "loss": 0.6637, + "step": 11944 + }, + { + "epoch": 0.9187754788093223, + "learning_rate": 4.8571050440326333e-05, + "loss": 0.9039, + "step": 11945 + }, + { + "epoch": 0.9188523959695408, + "learning_rate": 4.8479601889259984e-05, + "loss": 1.3058, + "step": 11946 + }, + { + "epoch": 0.9189293131297592, + "learning_rate": 4.838823809398285e-05, + "loss": 1.5151, + "step": 11947 + }, + { + "epoch": 0.9190062302899777, + "learning_rate": 4.829695905982884e-05, + "loss": 0.7713, + "step": 11948 + }, + { + "epoch": 0.9190831474501961, + "learning_rate": 4.820576479212868e-05, + "loss": 0.9511, + "step": 11949 + }, + { + "epoch": 0.9191600646104146, + "learning_rate": 4.811465529620679e-05, + "loss": 1.5938, + "step": 11950 + }, + { + "epoch": 0.9192369817706331, + "learning_rate": 4.8023630577383404e-05, + "loss": 0.6984, + "step": 11951 + }, + { + "epoch": 0.9193138989308515, + "learning_rate": 4.793269064097344e-05, + "loss": 1.1737, + "step": 11952 + }, + { + "epoch": 0.9193908160910699, + "learning_rate": 4.7841835492286823e-05, + "loss": 1.3361, + "step": 11953 + }, + { + "epoch": 0.9194677332512884, + "learning_rate": 4.77510651366288e-05, + "loss": 1.1423, + "step": 11954 + }, + { + "epoch": 0.9195446504115068, + "learning_rate": 4.76603795792998e-05, + "loss": 1.0514, + "step": 11955 + }, + { + "epoch": 0.9196215675717253, + "learning_rate": 4.7569778825594604e-05, + "loss": 0.7971, + "step": 11956 + }, + { + "epoch": 0.9196984847319437, + "learning_rate": 4.7479262880803955e-05, + "loss": 1.0356, + "step": 11957 + }, + { + "epoch": 0.9197754018921621, + "learning_rate": 4.73888317502128e-05, + "loss": 1.0134, + "step": 11958 + }, + { + "epoch": 0.9198523190523806, + "learning_rate": 4.7298485439101426e-05, + "loss": 1.2458, + "step": 11959 + }, + { + "epoch": 0.919929236212599, + "learning_rate": 4.7208223952745935e-05, + "loss": 1.1579, + "step": 11960 + }, + { + "epoch": 0.9200061533728174, + "learning_rate": 4.711804729641611e-05, + "loss": 1.464, + "step": 11961 + }, + { + "epoch": 0.920083070533036, + "learning_rate": 4.7027955475377416e-05, + "loss": 1.223, + "step": 11962 + }, + { + "epoch": 0.9201599876932544, + "learning_rate": 4.6937948494890805e-05, + "loss": 1.3812, + "step": 11963 + }, + { + "epoch": 0.9202369048534728, + "learning_rate": 4.684802636021174e-05, + "loss": 1.089, + "step": 11964 + }, + { + "epoch": 0.9203138220136913, + "learning_rate": 4.6758189076591015e-05, + "loss": 1.0512, + "step": 11965 + }, + { + "epoch": 0.9203907391739097, + "learning_rate": 4.666843664927395e-05, + "loss": 0.7737, + "step": 11966 + }, + { + "epoch": 0.9204676563341282, + "learning_rate": 4.65787690835015e-05, + "loss": 0.8502, + "step": 11967 + }, + { + "epoch": 0.9205445734943466, + "learning_rate": 4.64891863845095e-05, + "loss": 1.0651, + "step": 11968 + }, + { + "epoch": 0.920621490654565, + "learning_rate": 4.6399688557528907e-05, + "loss": 1.2638, + "step": 11969 + }, + { + "epoch": 0.9206984078147835, + "learning_rate": 4.631027560778506e-05, + "loss": 1.1519, + "step": 11970 + }, + { + "epoch": 0.9207753249750019, + "learning_rate": 4.6220947540499433e-05, + "loss": 1.3492, + "step": 11971 + }, + { + "epoch": 0.9208522421352203, + "learning_rate": 4.6131704360887706e-05, + "loss": 1.1754, + "step": 11972 + }, + { + "epoch": 0.9209291592954388, + "learning_rate": 4.604254607416103e-05, + "loss": 1.3276, + "step": 11973 + }, + { + "epoch": 0.9210060764556572, + "learning_rate": 4.595347268552524e-05, + "loss": 1.0822, + "step": 11974 + }, + { + "epoch": 0.9210829936158758, + "learning_rate": 4.586448420018152e-05, + "loss": 1.2038, + "step": 11975 + }, + { + "epoch": 0.9211599107760942, + "learning_rate": 4.577558062332604e-05, + "loss": 1.0735, + "step": 11976 + }, + { + "epoch": 0.9212368279363126, + "learning_rate": 4.568676196015015e-05, + "loss": 0.9305, + "step": 11977 + }, + { + "epoch": 0.9213137450965311, + "learning_rate": 4.559802821583969e-05, + "loss": 1.2265, + "step": 11978 + }, + { + "epoch": 0.9213906622567495, + "learning_rate": 4.55093793955762e-05, + "loss": 1.3947, + "step": 11979 + }, + { + "epoch": 0.9214675794169679, + "learning_rate": 4.542081550453603e-05, + "loss": 1.1411, + "step": 11980 + }, + { + "epoch": 0.9215444965771864, + "learning_rate": 4.533233654789004e-05, + "loss": 1.0006, + "step": 11981 + }, + { + "epoch": 0.9216214137374048, + "learning_rate": 4.524394253080527e-05, + "loss": 0.8626, + "step": 11982 + }, + { + "epoch": 0.9216983308976232, + "learning_rate": 4.515563345844259e-05, + "loss": 0.9308, + "step": 11983 + }, + { + "epoch": 0.9217752480578417, + "learning_rate": 4.506740933595854e-05, + "loss": 0.9866, + "step": 11984 + }, + { + "epoch": 0.9218521652180601, + "learning_rate": 4.4979270168505004e-05, + "loss": 1.2695, + "step": 11985 + }, + { + "epoch": 0.9219290823782786, + "learning_rate": 4.489121596122819e-05, + "loss": 1.0283, + "step": 11986 + }, + { + "epoch": 0.922005999538497, + "learning_rate": 4.480324671926983e-05, + "loss": 1.3385, + "step": 11987 + }, + { + "epoch": 0.9220829166987154, + "learning_rate": 4.471536244776647e-05, + "loss": 0.9881, + "step": 11988 + }, + { + "epoch": 0.922159833858934, + "learning_rate": 4.462756315184951e-05, + "loss": 1.2224, + "step": 11989 + }, + { + "epoch": 0.9222367510191524, + "learning_rate": 4.453984883664603e-05, + "loss": 1.3394, + "step": 11990 + }, + { + "epoch": 0.9223136681793708, + "learning_rate": 4.44522195072779e-05, + "loss": 1.0757, + "step": 11991 + }, + { + "epoch": 0.9223905853395893, + "learning_rate": 4.4364675168861226e-05, + "loss": 1.2237, + "step": 11992 + }, + { + "epoch": 0.9224675024998077, + "learning_rate": 4.42772158265084e-05, + "loss": 0.7563, + "step": 11993 + }, + { + "epoch": 0.9225444196600262, + "learning_rate": 4.4189841485325844e-05, + "loss": 1.928, + "step": 11994 + }, + { + "epoch": 0.9226213368202446, + "learning_rate": 4.410255215041598e-05, + "loss": 1.156, + "step": 11995 + }, + { + "epoch": 0.922698253980463, + "learning_rate": 4.401534782687522e-05, + "loss": 1.1637, + "step": 11996 + }, + { + "epoch": 0.9227751711406815, + "learning_rate": 4.39282285197955e-05, + "loss": 0.9752, + "step": 11997 + }, + { + "epoch": 0.9228520883008999, + "learning_rate": 4.384119423426425e-05, + "loss": 1.1092, + "step": 11998 + }, + { + "epoch": 0.9229290054611183, + "learning_rate": 4.375424497536323e-05, + "loss": 1.0258, + "step": 11999 + }, + { + "epoch": 0.9230059226213368, + "learning_rate": 4.3667380748169396e-05, + "loss": 1.3356, + "step": 12000 + }, + { + "epoch": 0.9230828397815553, + "learning_rate": 4.3580601557755004e-05, + "loss": 0.8829, + "step": 12001 + }, + { + "epoch": 0.9231597569417737, + "learning_rate": 4.349390740918718e-05, + "loss": 1.0759, + "step": 12002 + }, + { + "epoch": 0.9232366741019922, + "learning_rate": 4.3407298307527876e-05, + "loss": 1.1812, + "step": 12003 + }, + { + "epoch": 0.9233135912622106, + "learning_rate": 4.3320774257834704e-05, + "loss": 1.2493, + "step": 12004 + }, + { + "epoch": 0.9233905084224291, + "learning_rate": 4.3234335265159466e-05, + "loss": 0.9774, + "step": 12005 + }, + { + "epoch": 0.9234674255826475, + "learning_rate": 4.314798133454961e-05, + "loss": 0.9717, + "step": 12006 + }, + { + "epoch": 0.9235443427428659, + "learning_rate": 4.306171247104746e-05, + "loss": 1.1435, + "step": 12007 + }, + { + "epoch": 0.9236212599030844, + "learning_rate": 4.297552867969029e-05, + "loss": 1.2968, + "step": 12008 + }, + { + "epoch": 0.9236981770633028, + "learning_rate": 4.28894299655106e-05, + "loss": 1.1827, + "step": 12009 + }, + { + "epoch": 0.9237750942235212, + "learning_rate": 4.2803416333535514e-05, + "loss": 1.2274, + "step": 12010 + }, + { + "epoch": 0.9238520113837397, + "learning_rate": 4.2717487788787536e-05, + "loss": 1.2271, + "step": 12011 + }, + { + "epoch": 0.9239289285439581, + "learning_rate": 4.2631644336284136e-05, + "loss": 1.1826, + "step": 12012 + }, + { + "epoch": 0.9240058457041767, + "learning_rate": 4.254588598103815e-05, + "loss": 1.0768, + "step": 12013 + }, + { + "epoch": 0.9240827628643951, + "learning_rate": 4.2460212728056555e-05, + "loss": 0.9962, + "step": 12014 + }, + { + "epoch": 0.9241596800246135, + "learning_rate": 4.2374624582342366e-05, + "loss": 1.0629, + "step": 12015 + }, + { + "epoch": 0.924236597184832, + "learning_rate": 4.228912154889258e-05, + "loss": 0.6099, + "step": 12016 + }, + { + "epoch": 0.9243135143450504, + "learning_rate": 4.2203703632700705e-05, + "loss": 0.9834, + "step": 12017 + }, + { + "epoch": 0.9243904315052688, + "learning_rate": 4.2118370838753584e-05, + "loss": 0.7054, + "step": 12018 + }, + { + "epoch": 0.9244673486654873, + "learning_rate": 4.2033123172033895e-05, + "loss": 1.1603, + "step": 12019 + }, + { + "epoch": 0.9245442658257057, + "learning_rate": 4.194796063751999e-05, + "loss": 0.8002, + "step": 12020 + }, + { + "epoch": 0.9246211829859241, + "learning_rate": 4.1862883240184055e-05, + "loss": 1.2631, + "step": 12021 + }, + { + "epoch": 0.9246981001461426, + "learning_rate": 4.177789098499396e-05, + "loss": 1.0248, + "step": 12022 + }, + { + "epoch": 0.924775017306361, + "learning_rate": 4.169298387691273e-05, + "loss": 0.9804, + "step": 12023 + }, + { + "epoch": 0.9248519344665795, + "learning_rate": 4.160816192089772e-05, + "loss": 1.2527, + "step": 12024 + }, + { + "epoch": 0.9249288516267979, + "learning_rate": 4.152342512190216e-05, + "loss": 0.9706, + "step": 12025 + }, + { + "epoch": 0.9250057687870163, + "learning_rate": 4.143877348487391e-05, + "loss": 1.1578, + "step": 12026 + }, + { + "epoch": 0.9250826859472349, + "learning_rate": 4.135420701475551e-05, + "loss": 0.9385, + "step": 12027 + }, + { + "epoch": 0.9251596031074533, + "learning_rate": 4.1269725716485184e-05, + "loss": 1.1869, + "step": 12028 + }, + { + "epoch": 0.9252365202676717, + "learning_rate": 4.118532959499599e-05, + "loss": 1.2315, + "step": 12029 + }, + { + "epoch": 0.9253134374278902, + "learning_rate": 4.110101865521565e-05, + "loss": 1.4298, + "step": 12030 + }, + { + "epoch": 0.9253903545881086, + "learning_rate": 4.101679290206722e-05, + "loss": 1.627, + "step": 12031 + }, + { + "epoch": 0.9254672717483271, + "learning_rate": 4.093265234046878e-05, + "loss": 0.9546, + "step": 12032 + }, + { + "epoch": 0.9255441889085455, + "learning_rate": 4.084859697533338e-05, + "loss": 1.1561, + "step": 12033 + }, + { + "epoch": 0.9256211060687639, + "learning_rate": 4.076462681156912e-05, + "loss": 0.9854, + "step": 12034 + }, + { + "epoch": 0.9256980232289824, + "learning_rate": 4.068074185407905e-05, + "loss": 0.8717, + "step": 12035 + }, + { + "epoch": 0.9257749403892008, + "learning_rate": 4.0596942107761435e-05, + "loss": 1.3097, + "step": 12036 + }, + { + "epoch": 0.9258518575494192, + "learning_rate": 4.0513227577509195e-05, + "loss": 1.1436, + "step": 12037 + }, + { + "epoch": 0.9259287747096377, + "learning_rate": 4.0429598268210575e-05, + "loss": 1.1507, + "step": 12038 + }, + { + "epoch": 0.9260056918698562, + "learning_rate": 4.0346054184749016e-05, + "loss": 1.0428, + "step": 12039 + }, + { + "epoch": 0.9260826090300746, + "learning_rate": 4.02625953320026e-05, + "loss": 1.1619, + "step": 12040 + }, + { + "epoch": 0.9261595261902931, + "learning_rate": 4.017922171484428e-05, + "loss": 1.3877, + "step": 12041 + }, + { + "epoch": 0.9262364433505115, + "learning_rate": 4.009593333814282e-05, + "loss": 1.326, + "step": 12042 + }, + { + "epoch": 0.92631336051073, + "learning_rate": 4.001273020676116e-05, + "loss": 1.2347, + "step": 12043 + }, + { + "epoch": 0.9263902776709484, + "learning_rate": 3.992961232555792e-05, + "loss": 1.2249, + "step": 12044 + }, + { + "epoch": 0.9264671948311668, + "learning_rate": 3.9846579699386046e-05, + "loss": 1.1475, + "step": 12045 + }, + { + "epoch": 0.9265441119913853, + "learning_rate": 3.9763632333094156e-05, + "loss": 1.2705, + "step": 12046 + }, + { + "epoch": 0.9266210291516037, + "learning_rate": 3.968077023152572e-05, + "loss": 1.0513, + "step": 12047 + }, + { + "epoch": 0.9266979463118221, + "learning_rate": 3.959799339951903e-05, + "loss": 1.1733, + "step": 12048 + }, + { + "epoch": 0.9267748634720406, + "learning_rate": 3.951530184190738e-05, + "loss": 1.2158, + "step": 12049 + }, + { + "epoch": 0.926851780632259, + "learning_rate": 3.943269556351942e-05, + "loss": 1.0553, + "step": 12050 + }, + { + "epoch": 0.9269286977924776, + "learning_rate": 3.9350174569178273e-05, + "loss": 1.2771, + "step": 12051 + }, + { + "epoch": 0.927005614952696, + "learning_rate": 3.9267738863703105e-05, + "loss": 1.1422, + "step": 12052 + }, + { + "epoch": 0.9270825321129144, + "learning_rate": 3.918538845190689e-05, + "loss": 0.8328, + "step": 12053 + }, + { + "epoch": 0.9271594492731329, + "learning_rate": 3.910312333859794e-05, + "loss": 1.0746, + "step": 12054 + }, + { + "epoch": 0.9272363664333513, + "learning_rate": 3.902094352858043e-05, + "loss": 1.0455, + "step": 12055 + }, + { + "epoch": 0.9273132835935697, + "learning_rate": 3.893884902665268e-05, + "loss": 1.1229, + "step": 12056 + }, + { + "epoch": 0.9273902007537882, + "learning_rate": 3.8856839837608015e-05, + "loss": 1.0004, + "step": 12057 + }, + { + "epoch": 0.9274671179140066, + "learning_rate": 3.877491596623545e-05, + "loss": 1.0069, + "step": 12058 + }, + { + "epoch": 0.9275440350742251, + "learning_rate": 3.8693077417318145e-05, + "loss": 0.9788, + "step": 12059 + }, + { + "epoch": 0.9276209522344435, + "learning_rate": 3.861132419563512e-05, + "loss": 1.388, + "step": 12060 + }, + { + "epoch": 0.9276978693946619, + "learning_rate": 3.852965630596006e-05, + "loss": 1.3543, + "step": 12061 + }, + { + "epoch": 0.9277747865548804, + "learning_rate": 3.844807375306114e-05, + "loss": 1.283, + "step": 12062 + }, + { + "epoch": 0.9278517037150988, + "learning_rate": 3.836657654170256e-05, + "loss": 1.1871, + "step": 12063 + }, + { + "epoch": 0.9279286208753172, + "learning_rate": 3.8285164676642836e-05, + "loss": 0.9242, + "step": 12064 + }, + { + "epoch": 0.9280055380355358, + "learning_rate": 3.820383816263584e-05, + "loss": 0.9492, + "step": 12065 + }, + { + "epoch": 0.9280824551957542, + "learning_rate": 3.81225970044301e-05, + "loss": 1.2078, + "step": 12066 + }, + { + "epoch": 0.9281593723559726, + "learning_rate": 3.804144120676933e-05, + "loss": 1.0509, + "step": 12067 + }, + { + "epoch": 0.9282362895161911, + "learning_rate": 3.7960370774392406e-05, + "loss": 0.9027, + "step": 12068 + }, + { + "epoch": 0.9283132066764095, + "learning_rate": 3.7879385712033196e-05, + "loss": 0.9869, + "step": 12069 + }, + { + "epoch": 0.928390123836628, + "learning_rate": 3.77984860244206e-05, + "loss": 0.9502, + "step": 12070 + }, + { + "epoch": 0.9284670409968464, + "learning_rate": 3.771767171627816e-05, + "loss": 1.239, + "step": 12071 + }, + { + "epoch": 0.9285439581570648, + "learning_rate": 3.763694279232477e-05, + "loss": 1.1828, + "step": 12072 + }, + { + "epoch": 0.9286208753172833, + "learning_rate": 3.755629925727433e-05, + "loss": 0.8113, + "step": 12073 + }, + { + "epoch": 0.9286977924775017, + "learning_rate": 3.7475741115835906e-05, + "loss": 1.0285, + "step": 12074 + }, + { + "epoch": 0.9287747096377201, + "learning_rate": 3.739526837271307e-05, + "loss": 1.148, + "step": 12075 + }, + { + "epoch": 0.9288516267979386, + "learning_rate": 3.7314881032604553e-05, + "loss": 1.3272, + "step": 12076 + }, + { + "epoch": 0.928928543958157, + "learning_rate": 3.723457910020478e-05, + "loss": 1.1356, + "step": 12077 + }, + { + "epoch": 0.9290054611183756, + "learning_rate": 3.715436258020233e-05, + "loss": 1.0096, + "step": 12078 + }, + { + "epoch": 0.929082378278594, + "learning_rate": 3.707423147728112e-05, + "loss": 1.2124, + "step": 12079 + }, + { + "epoch": 0.9291592954388124, + "learning_rate": 3.699418579612024e-05, + "loss": 0.8465, + "step": 12080 + }, + { + "epoch": 0.9292362125990309, + "learning_rate": 3.6914225541393456e-05, + "loss": 1.2877, + "step": 12081 + }, + { + "epoch": 0.9293131297592493, + "learning_rate": 3.683435071776986e-05, + "loss": 0.923, + "step": 12082 + }, + { + "epoch": 0.9293900469194677, + "learning_rate": 3.675456132991339e-05, + "loss": 1.2139, + "step": 12083 + }, + { + "epoch": 0.9294669640796862, + "learning_rate": 3.667485738248316e-05, + "loss": 1.0502, + "step": 12084 + }, + { + "epoch": 0.9295438812399046, + "learning_rate": 3.6595238880132944e-05, + "loss": 1.1556, + "step": 12085 + }, + { + "epoch": 0.929620798400123, + "learning_rate": 3.6515705827512016e-05, + "loss": 1.2585, + "step": 12086 + }, + { + "epoch": 0.9296977155603415, + "learning_rate": 3.6436258229263996e-05, + "loss": 0.7302, + "step": 12087 + }, + { + "epoch": 0.9297746327205599, + "learning_rate": 3.635689609002835e-05, + "loss": 1.4461, + "step": 12088 + }, + { + "epoch": 0.9298515498807784, + "learning_rate": 3.627761941443869e-05, + "loss": 1.3529, + "step": 12089 + }, + { + "epoch": 0.9299284670409969, + "learning_rate": 3.619842820712449e-05, + "loss": 1.0811, + "step": 12090 + }, + { + "epoch": 0.9300053842012153, + "learning_rate": 3.611932247270955e-05, + "loss": 1.2806, + "step": 12091 + }, + { + "epoch": 0.9300823013614338, + "learning_rate": 3.6040302215813004e-05, + "loss": 1.1022, + "step": 12092 + }, + { + "epoch": 0.9301592185216522, + "learning_rate": 3.596136744104883e-05, + "loss": 1.0145, + "step": 12093 + }, + { + "epoch": 0.9302361356818706, + "learning_rate": 3.5882518153026344e-05, + "loss": 1.3617, + "step": 12094 + }, + { + "epoch": 0.9303130528420891, + "learning_rate": 3.580375435634919e-05, + "loss": 1.2117, + "step": 12095 + }, + { + "epoch": 0.9303899700023075, + "learning_rate": 3.572507605561703e-05, + "loss": 0.9597, + "step": 12096 + }, + { + "epoch": 0.930466887162526, + "learning_rate": 3.564648325542369e-05, + "loss": 0.8889, + "step": 12097 + }, + { + "epoch": 0.9305438043227444, + "learning_rate": 3.5567975960358326e-05, + "loss": 1.2961, + "step": 12098 + }, + { + "epoch": 0.9306207214829628, + "learning_rate": 3.5489554175004936e-05, + "loss": 1.1012, + "step": 12099 + }, + { + "epoch": 0.9306976386431813, + "learning_rate": 3.5411217903942696e-05, + "loss": 1.0175, + "step": 12100 + }, + { + "epoch": 0.9307745558033997, + "learning_rate": 3.533296715174611e-05, + "loss": 1.2612, + "step": 12101 + }, + { + "epoch": 0.9308514729636181, + "learning_rate": 3.525480192298369e-05, + "loss": 1.1085, + "step": 12102 + }, + { + "epoch": 0.9309283901238367, + "learning_rate": 3.517672222221996e-05, + "loss": 0.8248, + "step": 12103 + }, + { + "epoch": 0.9310053072840551, + "learning_rate": 3.50987280540141e-05, + "loss": 1.152, + "step": 12104 + }, + { + "epoch": 0.9310822244442735, + "learning_rate": 3.502081942292012e-05, + "loss": 1.1902, + "step": 12105 + }, + { + "epoch": 0.931159141604492, + "learning_rate": 3.494299633348741e-05, + "loss": 1.0832, + "step": 12106 + }, + { + "epoch": 0.9312360587647104, + "learning_rate": 3.486525879025998e-05, + "loss": 1.0917, + "step": 12107 + }, + { + "epoch": 0.9313129759249289, + "learning_rate": 3.47876067977767e-05, + "loss": 1.348, + "step": 12108 + }, + { + "epoch": 0.9313898930851473, + "learning_rate": 3.471004036057229e-05, + "loss": 1.3364, + "step": 12109 + }, + { + "epoch": 0.9314668102453657, + "learning_rate": 3.463255948317595e-05, + "loss": 1.5034, + "step": 12110 + }, + { + "epoch": 0.9315437274055842, + "learning_rate": 3.455516417011123e-05, + "loss": 1.4563, + "step": 12111 + }, + { + "epoch": 0.9316206445658026, + "learning_rate": 3.447785442589802e-05, + "loss": 1.2429, + "step": 12112 + }, + { + "epoch": 0.931697561726021, + "learning_rate": 3.440063025505002e-05, + "loss": 1.0759, + "step": 12113 + }, + { + "epoch": 0.9317744788862395, + "learning_rate": 3.4323491662076825e-05, + "loss": 0.9891, + "step": 12114 + }, + { + "epoch": 0.931851396046458, + "learning_rate": 3.42464386514823e-05, + "loss": 1.1364, + "step": 12115 + }, + { + "epoch": 0.9319283132066765, + "learning_rate": 3.41694712277657e-05, + "loss": 0.8841, + "step": 12116 + }, + { + "epoch": 0.9320052303668949, + "learning_rate": 3.4092589395421424e-05, + "loss": 0.8101, + "step": 12117 + }, + { + "epoch": 0.9320821475271133, + "learning_rate": 3.4015793158938544e-05, + "loss": 1.2486, + "step": 12118 + }, + { + "epoch": 0.9321590646873318, + "learning_rate": 3.393908252280148e-05, + "loss": 1.1403, + "step": 12119 + }, + { + "epoch": 0.9322359818475502, + "learning_rate": 3.386245749148914e-05, + "loss": 0.958, + "step": 12120 + }, + { + "epoch": 0.9323128990077686, + "learning_rate": 3.3785918069475953e-05, + "loss": 1.4061, + "step": 12121 + }, + { + "epoch": 0.9323898161679871, + "learning_rate": 3.370946426123101e-05, + "loss": 1.1242, + "step": 12122 + }, + { + "epoch": 0.9324667333282055, + "learning_rate": 3.3633096071218724e-05, + "loss": 1.0106, + "step": 12123 + }, + { + "epoch": 0.9325436504884239, + "learning_rate": 3.3556813503898056e-05, + "loss": 0.9884, + "step": 12124 + }, + { + "epoch": 0.9326205676486424, + "learning_rate": 3.348061656372309e-05, + "loss": 0.9873, + "step": 12125 + }, + { + "epoch": 0.9326974848088608, + "learning_rate": 3.3404505255143605e-05, + "loss": 1.1096, + "step": 12126 + }, + { + "epoch": 0.9327744019690793, + "learning_rate": 3.3328479582603376e-05, + "loss": 1.4038, + "step": 12127 + }, + { + "epoch": 0.9328513191292978, + "learning_rate": 3.325253955054186e-05, + "loss": 1.0945, + "step": 12128 + }, + { + "epoch": 0.9329282362895162, + "learning_rate": 3.317668516339317e-05, + "loss": 1.1887, + "step": 12129 + }, + { + "epoch": 0.9330051534497347, + "learning_rate": 3.310091642558627e-05, + "loss": 1.9605, + "step": 12130 + }, + { + "epoch": 0.9330820706099531, + "learning_rate": 3.3025233341545945e-05, + "loss": 1.4124, + "step": 12131 + }, + { + "epoch": 0.9331589877701715, + "learning_rate": 3.2949635915691e-05, + "loss": 1.2597, + "step": 12132 + }, + { + "epoch": 0.93323590493039, + "learning_rate": 3.287412415243557e-05, + "loss": 0.9069, + "step": 12133 + }, + { + "epoch": 0.9333128220906084, + "learning_rate": 3.279869805618929e-05, + "loss": 1.1168, + "step": 12134 + }, + { + "epoch": 0.9333897392508269, + "learning_rate": 3.272335763135581e-05, + "loss": 0.9069, + "step": 12135 + }, + { + "epoch": 0.9334666564110453, + "learning_rate": 3.264810288233494e-05, + "loss": 1.2822, + "step": 12136 + }, + { + "epoch": 0.9335435735712637, + "learning_rate": 3.25729338135205e-05, + "loss": 1.0019, + "step": 12137 + }, + { + "epoch": 0.9336204907314822, + "learning_rate": 3.2497850429301643e-05, + "loss": 1.2547, + "step": 12138 + }, + { + "epoch": 0.9336974078917006, + "learning_rate": 3.24228527340627e-05, + "loss": 1.4441, + "step": 12139 + }, + { + "epoch": 0.933774325051919, + "learning_rate": 3.2347940732183e-05, + "loss": 1.1251, + "step": 12140 + }, + { + "epoch": 0.9338512422121376, + "learning_rate": 3.2273114428036546e-05, + "loss": 1.2194, + "step": 12141 + }, + { + "epoch": 0.933928159372356, + "learning_rate": 3.21983738259925e-05, + "loss": 1.4413, + "step": 12142 + }, + { + "epoch": 0.9340050765325744, + "learning_rate": 3.212371893041521e-05, + "loss": 1.2906, + "step": 12143 + }, + { + "epoch": 0.9340819936927929, + "learning_rate": 3.204914974566353e-05, + "loss": 1.2077, + "step": 12144 + }, + { + "epoch": 0.9341589108530113, + "learning_rate": 3.1974666276092135e-05, + "loss": 1.3196, + "step": 12145 + }, + { + "epoch": 0.9342358280132298, + "learning_rate": 3.190026852604971e-05, + "loss": 1.1493, + "step": 12146 + }, + { + "epoch": 0.9343127451734482, + "learning_rate": 3.182595649988079e-05, + "loss": 1.0146, + "step": 12147 + }, + { + "epoch": 0.9343896623336666, + "learning_rate": 3.175173020192424e-05, + "loss": 1.3035, + "step": 12148 + }, + { + "epoch": 0.9344665794938851, + "learning_rate": 3.16775896365144e-05, + "loss": 1.1528, + "step": 12149 + }, + { + "epoch": 0.9345434966541035, + "learning_rate": 3.16035348079805e-05, + "loss": 1.3201, + "step": 12150 + }, + { + "epoch": 0.9346204138143219, + "learning_rate": 3.152956572064625e-05, + "loss": 1.0597, + "step": 12151 + }, + { + "epoch": 0.9346973309745404, + "learning_rate": 3.145568237883117e-05, + "loss": 1.223, + "step": 12152 + }, + { + "epoch": 0.9347742481347588, + "learning_rate": 3.138188478684917e-05, + "loss": 1.2984, + "step": 12153 + }, + { + "epoch": 0.9348511652949774, + "learning_rate": 3.130817294900962e-05, + "loss": 1.4366, + "step": 12154 + }, + { + "epoch": 0.9349280824551958, + "learning_rate": 3.123454686961657e-05, + "loss": 1.1027, + "step": 12155 + }, + { + "epoch": 0.9350049996154142, + "learning_rate": 3.116100655296894e-05, + "loss": 1.3012, + "step": 12156 + }, + { + "epoch": 0.9350819167756327, + "learning_rate": 3.1087552003360775e-05, + "loss": 1.1954, + "step": 12157 + }, + { + "epoch": 0.9351588339358511, + "learning_rate": 3.101418322508165e-05, + "loss": 1.4599, + "step": 12158 + }, + { + "epoch": 0.9352357510960695, + "learning_rate": 3.094090022241497e-05, + "loss": 0.9842, + "step": 12159 + }, + { + "epoch": 0.935312668256288, + "learning_rate": 3.086770299964031e-05, + "loss": 1.1034, + "step": 12160 + }, + { + "epoch": 0.9353895854165064, + "learning_rate": 3.0794591561031426e-05, + "loss": 1.6368, + "step": 12161 + }, + { + "epoch": 0.9354665025767248, + "learning_rate": 3.072156591085773e-05, + "loss": 1.0797, + "step": 12162 + }, + { + "epoch": 0.9355434197369433, + "learning_rate": 3.064862605338281e-05, + "loss": 1.1395, + "step": 12163 + }, + { + "epoch": 0.9356203368971617, + "learning_rate": 3.0575771992866095e-05, + "loss": 1.2444, + "step": 12164 + }, + { + "epoch": 0.9356972540573802, + "learning_rate": 3.050300373356135e-05, + "loss": 1.0648, + "step": 12165 + }, + { + "epoch": 0.9357741712175986, + "learning_rate": 3.043032127971784e-05, + "loss": 1.4334, + "step": 12166 + }, + { + "epoch": 0.935851088377817, + "learning_rate": 3.0357724635579342e-05, + "loss": 1.1942, + "step": 12167 + }, + { + "epoch": 0.9359280055380356, + "learning_rate": 3.0285213805384793e-05, + "loss": 1.1077, + "step": 12168 + }, + { + "epoch": 0.936004922698254, + "learning_rate": 3.0212788793368474e-05, + "loss": 0.946, + "step": 12169 + }, + { + "epoch": 0.9360818398584724, + "learning_rate": 3.0140449603759167e-05, + "loss": 1.0302, + "step": 12170 + }, + { + "epoch": 0.9361587570186909, + "learning_rate": 3.006819624078083e-05, + "loss": 1.0565, + "step": 12171 + }, + { + "epoch": 0.9362356741789093, + "learning_rate": 2.999602870865259e-05, + "loss": 0.9015, + "step": 12172 + }, + { + "epoch": 0.9363125913391278, + "learning_rate": 2.9923947011587904e-05, + "loss": 1.0329, + "step": 12173 + }, + { + "epoch": 0.9363895084993462, + "learning_rate": 2.985195115379624e-05, + "loss": 1.229, + "step": 12174 + }, + { + "epoch": 0.9364664256595646, + "learning_rate": 2.9780041139481405e-05, + "loss": 1.1141, + "step": 12175 + }, + { + "epoch": 0.9365433428197831, + "learning_rate": 2.9708216972842038e-05, + "loss": 1.3029, + "step": 12176 + }, + { + "epoch": 0.9366202599800015, + "learning_rate": 2.9636478658072287e-05, + "loss": 1.1784, + "step": 12177 + }, + { + "epoch": 0.9366971771402199, + "learning_rate": 2.9564826199360795e-05, + "loss": 1.3269, + "step": 12178 + }, + { + "epoch": 0.9367740943004385, + "learning_rate": 2.9493259600891555e-05, + "loss": 1.0073, + "step": 12179 + }, + { + "epoch": 0.9368510114606569, + "learning_rate": 2.942177886684355e-05, + "loss": 1.264, + "step": 12180 + }, + { + "epoch": 0.9369279286208754, + "learning_rate": 2.9350384001390285e-05, + "loss": 0.7756, + "step": 12181 + }, + { + "epoch": 0.9370048457810938, + "learning_rate": 2.9279075008700916e-05, + "loss": 0.9171, + "step": 12182 + }, + { + "epoch": 0.9370817629413122, + "learning_rate": 2.920785189293912e-05, + "loss": 1.203, + "step": 12183 + }, + { + "epoch": 0.9371586801015307, + "learning_rate": 2.9136714658263564e-05, + "loss": 1.1214, + "step": 12184 + }, + { + "epoch": 0.9372355972617491, + "learning_rate": 2.906566330882826e-05, + "loss": 1.3194, + "step": 12185 + }, + { + "epoch": 0.9373125144219675, + "learning_rate": 2.8994697848781725e-05, + "loss": 1.1999, + "step": 12186 + }, + { + "epoch": 0.937389431582186, + "learning_rate": 2.8923818282267812e-05, + "loss": 0.9972, + "step": 12187 + }, + { + "epoch": 0.9374663487424044, + "learning_rate": 2.8853024613425372e-05, + "loss": 1.1405, + "step": 12188 + }, + { + "epoch": 0.9375432659026228, + "learning_rate": 2.8782316846388102e-05, + "loss": 1.2609, + "step": 12189 + }, + { + "epoch": 0.9376201830628413, + "learning_rate": 2.8711694985284366e-05, + "loss": 0.7842, + "step": 12190 + }, + { + "epoch": 0.9376971002230597, + "learning_rate": 2.8641159034238363e-05, + "loss": 1.2734, + "step": 12191 + }, + { + "epoch": 0.9377740173832783, + "learning_rate": 2.85707089973683e-05, + "loss": 1.3886, + "step": 12192 + }, + { + "epoch": 0.9378509345434967, + "learning_rate": 2.8500344878788222e-05, + "loss": 1.0924, + "step": 12193 + }, + { + "epoch": 0.9379278517037151, + "learning_rate": 2.8430066682606337e-05, + "loss": 1.0277, + "step": 12194 + }, + { + "epoch": 0.9380047688639336, + "learning_rate": 2.8359874412926534e-05, + "loss": 1.2736, + "step": 12195 + }, + { + "epoch": 0.938081686024152, + "learning_rate": 2.828976807384753e-05, + "loss": 1.0204, + "step": 12196 + }, + { + "epoch": 0.9381586031843704, + "learning_rate": 2.8219747669462558e-05, + "loss": 1.2011, + "step": 12197 + }, + { + "epoch": 0.9382355203445889, + "learning_rate": 2.8149813203860507e-05, + "loss": 1.3007, + "step": 12198 + }, + { + "epoch": 0.9383124375048073, + "learning_rate": 2.807996468112478e-05, + "loss": 1.2816, + "step": 12199 + }, + { + "epoch": 0.9383893546650258, + "learning_rate": 2.8010202105333782e-05, + "loss": 0.9556, + "step": 12200 + }, + { + "epoch": 0.9384662718252442, + "learning_rate": 2.7940525480561252e-05, + "loss": 0.8778, + "step": 12201 + }, + { + "epoch": 0.9385431889854626, + "learning_rate": 2.787093481087577e-05, + "loss": 0.861, + "step": 12202 + }, + { + "epoch": 0.9386201061456811, + "learning_rate": 2.780143010034025e-05, + "loss": 1.3557, + "step": 12203 + }, + { + "epoch": 0.9386970233058995, + "learning_rate": 2.7732011353013786e-05, + "loss": 1.0216, + "step": 12204 + }, + { + "epoch": 0.938773940466118, + "learning_rate": 2.7662678572949462e-05, + "loss": 1.2426, + "step": 12205 + }, + { + "epoch": 0.9388508576263365, + "learning_rate": 2.7593431764195707e-05, + "loss": 0.9678, + "step": 12206 + }, + { + "epoch": 0.9389277747865549, + "learning_rate": 2.7524270930796125e-05, + "loss": 1.3174, + "step": 12207 + }, + { + "epoch": 0.9390046919467733, + "learning_rate": 2.7455196076788812e-05, + "loss": 0.9357, + "step": 12208 + }, + { + "epoch": 0.9390816091069918, + "learning_rate": 2.7386207206207214e-05, + "loss": 1.1123, + "step": 12209 + }, + { + "epoch": 0.9391585262672102, + "learning_rate": 2.7317304323079773e-05, + "loss": 1.0222, + "step": 12210 + }, + { + "epoch": 0.9392354434274287, + "learning_rate": 2.724848743142977e-05, + "loss": 1.2877, + "step": 12211 + }, + { + "epoch": 0.9393123605876471, + "learning_rate": 2.7179756535275325e-05, + "loss": 1.5601, + "step": 12212 + }, + { + "epoch": 0.9393892777478655, + "learning_rate": 2.711111163863006e-05, + "loss": 1.0236, + "step": 12213 + }, + { + "epoch": 0.939466194908084, + "learning_rate": 2.7042552745501605e-05, + "loss": 1.4442, + "step": 12214 + }, + { + "epoch": 0.9395431120683024, + "learning_rate": 2.6974079859894084e-05, + "loss": 0.8903, + "step": 12215 + }, + { + "epoch": 0.9396200292285208, + "learning_rate": 2.690569298580481e-05, + "loss": 0.8188, + "step": 12216 + }, + { + "epoch": 0.9396969463887394, + "learning_rate": 2.683739212722741e-05, + "loss": 0.9695, + "step": 12217 + }, + { + "epoch": 0.9397738635489578, + "learning_rate": 2.6769177288150036e-05, + "loss": 1.0914, + "step": 12218 + }, + { + "epoch": 0.9398507807091763, + "learning_rate": 2.6701048472555834e-05, + "loss": 1.0811, + "step": 12219 + }, + { + "epoch": 0.9399276978693947, + "learning_rate": 2.6633005684422785e-05, + "loss": 1.045, + "step": 12220 + }, + { + "epoch": 0.9400046150296131, + "learning_rate": 2.6565048927724045e-05, + "loss": 0.731, + "step": 12221 + }, + { + "epoch": 0.9400815321898316, + "learning_rate": 2.6497178206427443e-05, + "loss": 1.2308, + "step": 12222 + }, + { + "epoch": 0.94015844935005, + "learning_rate": 2.642939352449647e-05, + "loss": 0.9671, + "step": 12223 + }, + { + "epoch": 0.9402353665102684, + "learning_rate": 2.6361694885888967e-05, + "loss": 1.104, + "step": 12224 + }, + { + "epoch": 0.9403122836704869, + "learning_rate": 2.6294082294557764e-05, + "loss": 1.1925, + "step": 12225 + }, + { + "epoch": 0.9403892008307053, + "learning_rate": 2.622655575445104e-05, + "loss": 1.2977, + "step": 12226 + }, + { + "epoch": 0.9404661179909237, + "learning_rate": 2.6159115269511812e-05, + "loss": 1.1346, + "step": 12227 + }, + { + "epoch": 0.9405430351511422, + "learning_rate": 2.6091760843677593e-05, + "loss": 0.9481, + "step": 12228 + }, + { + "epoch": 0.9406199523113606, + "learning_rate": 2.6024492480881735e-05, + "loss": 0.8181, + "step": 12229 + }, + { + "epoch": 0.9406968694715792, + "learning_rate": 2.5957310185051597e-05, + "loss": 0.9931, + "step": 12230 + }, + { + "epoch": 0.9407737866317976, + "learning_rate": 2.5890213960110542e-05, + "loss": 1.4884, + "step": 12231 + }, + { + "epoch": 0.940850703792016, + "learning_rate": 2.5823203809976105e-05, + "loss": 1.0598, + "step": 12232 + }, + { + "epoch": 0.9409276209522345, + "learning_rate": 2.5756279738561317e-05, + "loss": 1.04, + "step": 12233 + }, + { + "epoch": 0.9410045381124529, + "learning_rate": 2.5689441749773555e-05, + "loss": 1.4693, + "step": 12234 + }, + { + "epoch": 0.9410814552726713, + "learning_rate": 2.562268984751587e-05, + "loss": 1.1455, + "step": 12235 + }, + { + "epoch": 0.9411583724328898, + "learning_rate": 2.55560240356858e-05, + "loss": 0.9774, + "step": 12236 + }, + { + "epoch": 0.9412352895931082, + "learning_rate": 2.5489444318176235e-05, + "loss": 1.2256, + "step": 12237 + }, + { + "epoch": 0.9413122067533267, + "learning_rate": 2.5422950698874735e-05, + "loss": 1.2458, + "step": 12238 + }, + { + "epoch": 0.9413891239135451, + "learning_rate": 2.535654318166386e-05, + "loss": 1.0736, + "step": 12239 + }, + { + "epoch": 0.9414660410737635, + "learning_rate": 2.5290221770421173e-05, + "loss": 1.4101, + "step": 12240 + }, + { + "epoch": 0.941542958233982, + "learning_rate": 2.522398646901941e-05, + "loss": 0.8386, + "step": 12241 + }, + { + "epoch": 0.9416198753942004, + "learning_rate": 2.5157837281326147e-05, + "loss": 0.9314, + "step": 12242 + }, + { + "epoch": 0.9416967925544188, + "learning_rate": 2.5091774211203787e-05, + "loss": 1.0781, + "step": 12243 + }, + { + "epoch": 0.9417737097146374, + "learning_rate": 2.5025797262509587e-05, + "loss": 1.1547, + "step": 12244 + }, + { + "epoch": 0.9418506268748558, + "learning_rate": 2.4959906439096626e-05, + "loss": 0.8614, + "step": 12245 + }, + { + "epoch": 0.9419275440350742, + "learning_rate": 2.4894101744811824e-05, + "loss": 1.0271, + "step": 12246 + }, + { + "epoch": 0.9420044611952927, + "learning_rate": 2.4828383183497617e-05, + "loss": 0.7347, + "step": 12247 + }, + { + "epoch": 0.9420813783555111, + "learning_rate": 2.476275075899176e-05, + "loss": 1.1688, + "step": 12248 + }, + { + "epoch": 0.9421582955157296, + "learning_rate": 2.4697204475126024e-05, + "loss": 1.2839, + "step": 12249 + }, + { + "epoch": 0.942235212675948, + "learning_rate": 2.463174433572818e-05, + "loss": 0.9427, + "step": 12250 + }, + { + "epoch": 0.9423121298361664, + "learning_rate": 2.456637034462067e-05, + "loss": 1.2026, + "step": 12251 + }, + { + "epoch": 0.9423890469963849, + "learning_rate": 2.4501082505620108e-05, + "loss": 0.7465, + "step": 12252 + }, + { + "epoch": 0.9424659641566033, + "learning_rate": 2.4435880822539116e-05, + "loss": 1.1824, + "step": 12253 + }, + { + "epoch": 0.9425428813168217, + "learning_rate": 2.4370765299184805e-05, + "loss": 1.3512, + "step": 12254 + }, + { + "epoch": 0.9426197984770402, + "learning_rate": 2.4305735939359473e-05, + "loss": 1.1434, + "step": 12255 + }, + { + "epoch": 0.9426967156372587, + "learning_rate": 2.424079274686025e-05, + "loss": 1.169, + "step": 12256 + }, + { + "epoch": 0.9427736327974772, + "learning_rate": 2.4175935725478938e-05, + "loss": 1.2582, + "step": 12257 + }, + { + "epoch": 0.9428505499576956, + "learning_rate": 2.4111164879002832e-05, + "loss": 1.2718, + "step": 12258 + }, + { + "epoch": 0.942927467117914, + "learning_rate": 2.4046480211214082e-05, + "loss": 0.9659, + "step": 12259 + }, + { + "epoch": 0.9430043842781325, + "learning_rate": 2.3981881725889332e-05, + "loss": 1.3946, + "step": 12260 + }, + { + "epoch": 0.9430813014383509, + "learning_rate": 2.391736942680106e-05, + "loss": 1.2674, + "step": 12261 + }, + { + "epoch": 0.9431582185985693, + "learning_rate": 2.3852943317715757e-05, + "loss": 1.0531, + "step": 12262 + }, + { + "epoch": 0.9432351357587878, + "learning_rate": 2.3788603402395415e-05, + "loss": 0.9711, + "step": 12263 + }, + { + "epoch": 0.9433120529190062, + "learning_rate": 2.372434968459719e-05, + "loss": 1.3543, + "step": 12264 + }, + { + "epoch": 0.9433889700792246, + "learning_rate": 2.3660182168072585e-05, + "loss": 0.9723, + "step": 12265 + }, + { + "epoch": 0.9434658872394431, + "learning_rate": 2.3596100856568603e-05, + "loss": 0.8284, + "step": 12266 + }, + { + "epoch": 0.9435428043996615, + "learning_rate": 2.3532105753826917e-05, + "loss": 1.4464, + "step": 12267 + }, + { + "epoch": 0.94361972155988, + "learning_rate": 2.3468196863584367e-05, + "loss": 0.8575, + "step": 12268 + }, + { + "epoch": 0.9436966387200985, + "learning_rate": 2.3404374189572643e-05, + "loss": 1.0575, + "step": 12269 + }, + { + "epoch": 0.9437735558803169, + "learning_rate": 2.3340637735518254e-05, + "loss": 0.663, + "step": 12270 + }, + { + "epoch": 0.9438504730405354, + "learning_rate": 2.32769875051429e-05, + "loss": 0.9797, + "step": 12271 + }, + { + "epoch": 0.9439273902007538, + "learning_rate": 2.3213423502163432e-05, + "loss": 0.8766, + "step": 12272 + }, + { + "epoch": 0.9440043073609722, + "learning_rate": 2.314994573029139e-05, + "loss": 0.9139, + "step": 12273 + }, + { + "epoch": 0.9440812245211907, + "learning_rate": 2.30865541932328e-05, + "loss": 1.2268, + "step": 12274 + }, + { + "epoch": 0.9441581416814091, + "learning_rate": 2.302324889468971e-05, + "loss": 1.0848, + "step": 12275 + }, + { + "epoch": 0.9442350588416276, + "learning_rate": 2.2960029838358154e-05, + "loss": 1.016, + "step": 12276 + }, + { + "epoch": 0.944311976001846, + "learning_rate": 2.2896897027930186e-05, + "loss": 1.0007, + "step": 12277 + }, + { + "epoch": 0.9443888931620644, + "learning_rate": 2.2833850467091523e-05, + "loss": 1.1234, + "step": 12278 + }, + { + "epoch": 0.9444658103222829, + "learning_rate": 2.2770890159523882e-05, + "loss": 1.1582, + "step": 12279 + }, + { + "epoch": 0.9445427274825013, + "learning_rate": 2.2708016108903494e-05, + "loss": 1.3735, + "step": 12280 + }, + { + "epoch": 0.9446196446427197, + "learning_rate": 2.264522831890159e-05, + "loss": 1.0893, + "step": 12281 + }, + { + "epoch": 0.9446965618029383, + "learning_rate": 2.2582526793184566e-05, + "loss": 1.427, + "step": 12282 + }, + { + "epoch": 0.9447734789631567, + "learning_rate": 2.251991153541333e-05, + "loss": 1.382, + "step": 12283 + }, + { + "epoch": 0.9448503961233752, + "learning_rate": 2.2457382549244288e-05, + "loss": 1.1619, + "step": 12284 + }, + { + "epoch": 0.9449273132835936, + "learning_rate": 2.239493983832869e-05, + "loss": 1.1343, + "step": 12285 + }, + { + "epoch": 0.945004230443812, + "learning_rate": 2.233258340631261e-05, + "loss": 0.9671, + "step": 12286 + }, + { + "epoch": 0.9450811476040305, + "learning_rate": 2.227031325683665e-05, + "loss": 0.9059, + "step": 12287 + }, + { + "epoch": 0.9451580647642489, + "learning_rate": 2.220812939353739e-05, + "loss": 1.4422, + "step": 12288 + }, + { + "epoch": 0.9452349819244673, + "learning_rate": 2.214603182004543e-05, + "loss": 1.3011, + "step": 12289 + }, + { + "epoch": 0.9453118990846858, + "learning_rate": 2.208402053998687e-05, + "loss": 1.0817, + "step": 12290 + }, + { + "epoch": 0.9453888162449042, + "learning_rate": 2.2022095556982645e-05, + "loss": 1.4671, + "step": 12291 + }, + { + "epoch": 0.9454657334051226, + "learning_rate": 2.1960256874648532e-05, + "loss": 0.9501, + "step": 12292 + }, + { + "epoch": 0.9455426505653411, + "learning_rate": 2.1898504496595306e-05, + "loss": 1.4825, + "step": 12293 + }, + { + "epoch": 0.9456195677255596, + "learning_rate": 2.1836838426429086e-05, + "loss": 1.1331, + "step": 12294 + }, + { + "epoch": 0.9456964848857781, + "learning_rate": 2.1775258667750153e-05, + "loss": 1.0971, + "step": 12295 + }, + { + "epoch": 0.9457734020459965, + "learning_rate": 2.1713765224154635e-05, + "loss": 1.2781, + "step": 12296 + }, + { + "epoch": 0.9458503192062149, + "learning_rate": 2.165235809923283e-05, + "loss": 1.1165, + "step": 12297 + }, + { + "epoch": 0.9459272363664334, + "learning_rate": 2.1591037296570526e-05, + "loss": 0.9566, + "step": 12298 + }, + { + "epoch": 0.9460041535266518, + "learning_rate": 2.1529802819748533e-05, + "loss": 1.3316, + "step": 12299 + }, + { + "epoch": 0.9460810706868702, + "learning_rate": 2.146865467234199e-05, + "loss": 1.2123, + "step": 12300 + }, + { + "epoch": 0.9461579878470887, + "learning_rate": 2.1407592857921542e-05, + "loss": 1.2362, + "step": 12301 + }, + { + "epoch": 0.9462349050073071, + "learning_rate": 2.1346617380052834e-05, + "loss": 1.0055, + "step": 12302 + }, + { + "epoch": 0.9463118221675256, + "learning_rate": 2.128572824229602e-05, + "loss": 1.2387, + "step": 12303 + }, + { + "epoch": 0.946388739327744, + "learning_rate": 2.1224925448206755e-05, + "loss": 1.1984, + "step": 12304 + }, + { + "epoch": 0.9464656564879624, + "learning_rate": 2.11642090013352e-05, + "loss": 1.3221, + "step": 12305 + }, + { + "epoch": 0.946542573648181, + "learning_rate": 2.110357890522652e-05, + "loss": 0.9596, + "step": 12306 + }, + { + "epoch": 0.9466194908083994, + "learning_rate": 2.1043035163421374e-05, + "loss": 1.3926, + "step": 12307 + }, + { + "epoch": 0.9466964079686178, + "learning_rate": 2.098257777945478e-05, + "loss": 1.5158, + "step": 12308 + }, + { + "epoch": 0.9467733251288363, + "learning_rate": 2.0922206756856577e-05, + "loss": 1.2633, + "step": 12309 + }, + { + "epoch": 0.9468502422890547, + "learning_rate": 2.0861922099152275e-05, + "loss": 1.2649, + "step": 12310 + }, + { + "epoch": 0.9469271594492731, + "learning_rate": 2.0801723809862062e-05, + "loss": 1.0729, + "step": 12311 + }, + { + "epoch": 0.9470040766094916, + "learning_rate": 2.0741611892500624e-05, + "loss": 0.7765, + "step": 12312 + }, + { + "epoch": 0.94708099376971, + "learning_rate": 2.0681586350578152e-05, + "loss": 1.2071, + "step": 12313 + }, + { + "epoch": 0.9471579109299285, + "learning_rate": 2.0621647187599347e-05, + "loss": 0.839, + "step": 12314 + }, + { + "epoch": 0.9472348280901469, + "learning_rate": 2.056179440706457e-05, + "loss": 1.1766, + "step": 12315 + }, + { + "epoch": 0.9473117452503653, + "learning_rate": 2.0502028012468367e-05, + "loss": 1.2616, + "step": 12316 + }, + { + "epoch": 0.9473886624105838, + "learning_rate": 2.044234800730077e-05, + "loss": 1.3573, + "step": 12317 + }, + { + "epoch": 0.9474655795708022, + "learning_rate": 2.0382754395046332e-05, + "loss": 1.0533, + "step": 12318 + }, + { + "epoch": 0.9475424967310206, + "learning_rate": 2.0323247179184935e-05, + "loss": 1.2565, + "step": 12319 + }, + { + "epoch": 0.9476194138912392, + "learning_rate": 2.0263826363190962e-05, + "loss": 1.0266, + "step": 12320 + }, + { + "epoch": 0.9476963310514576, + "learning_rate": 2.0204491950534642e-05, + "loss": 0.9547, + "step": 12321 + }, + { + "epoch": 0.9477732482116761, + "learning_rate": 2.0145243944680036e-05, + "loss": 1.3065, + "step": 12322 + }, + { + "epoch": 0.9478501653718945, + "learning_rate": 2.0086082349087043e-05, + "loss": 1.349, + "step": 12323 + }, + { + "epoch": 0.9479270825321129, + "learning_rate": 2.0027007167210066e-05, + "loss": 1.2137, + "step": 12324 + }, + { + "epoch": 0.9480039996923314, + "learning_rate": 1.9968018402498346e-05, + "loss": 1.0298, + "step": 12325 + }, + { + "epoch": 0.9480809168525498, + "learning_rate": 1.9909116058396626e-05, + "loss": 1.2058, + "step": 12326 + }, + { + "epoch": 0.9481578340127682, + "learning_rate": 1.985030013834416e-05, + "loss": 1.0441, + "step": 12327 + }, + { + "epoch": 0.9482347511729867, + "learning_rate": 1.9791570645775025e-05, + "loss": 0.878, + "step": 12328 + }, + { + "epoch": 0.9483116683332051, + "learning_rate": 1.973292758411882e-05, + "loss": 1.1512, + "step": 12329 + }, + { + "epoch": 0.9483885854934235, + "learning_rate": 1.96743709567998e-05, + "loss": 1.0311, + "step": 12330 + }, + { + "epoch": 0.948465502653642, + "learning_rate": 1.9615900767236904e-05, + "loss": 1.0456, + "step": 12331 + }, + { + "epoch": 0.9485424198138604, + "learning_rate": 1.9557517018844396e-05, + "loss": 1.007, + "step": 12332 + }, + { + "epoch": 0.948619336974079, + "learning_rate": 1.949921971503138e-05, + "loss": 1.0526, + "step": 12333 + }, + { + "epoch": 0.9486962541342974, + "learning_rate": 1.944100885920197e-05, + "loss": 1.2026, + "step": 12334 + }, + { + "epoch": 0.9487731712945158, + "learning_rate": 1.9382884454754947e-05, + "loss": 1.0834, + "step": 12335 + }, + { + "epoch": 0.9488500884547343, + "learning_rate": 1.932484650508426e-05, + "loss": 1.0423, + "step": 12336 + }, + { + "epoch": 0.9489270056149527, + "learning_rate": 1.9266895013578866e-05, + "loss": 1.0567, + "step": 12337 + }, + { + "epoch": 0.9490039227751711, + "learning_rate": 1.9209029983622893e-05, + "loss": 0.9646, + "step": 12338 + }, + { + "epoch": 0.9490808399353896, + "learning_rate": 1.9151251418594806e-05, + "loss": 1.1774, + "step": 12339 + }, + { + "epoch": 0.949157757095608, + "learning_rate": 1.9093559321868404e-05, + "loss": 0.8241, + "step": 12340 + }, + { + "epoch": 0.9492346742558265, + "learning_rate": 1.903595369681249e-05, + "loss": 1.0441, + "step": 12341 + }, + { + "epoch": 0.9493115914160449, + "learning_rate": 1.8978434546790546e-05, + "loss": 1.4644, + "step": 12342 + }, + { + "epoch": 0.9493885085762633, + "learning_rate": 1.8921001875161546e-05, + "loss": 1.3055, + "step": 12343 + }, + { + "epoch": 0.9494654257364818, + "learning_rate": 1.886365568527848e-05, + "loss": 1.0192, + "step": 12344 + }, + { + "epoch": 0.9495423428967003, + "learning_rate": 1.88063959804905e-05, + "loss": 1.0575, + "step": 12345 + }, + { + "epoch": 0.9496192600569187, + "learning_rate": 1.8749222764140426e-05, + "loss": 0.9611, + "step": 12346 + }, + { + "epoch": 0.9496961772171372, + "learning_rate": 1.86921360395671e-05, + "loss": 1.447, + "step": 12347 + }, + { + "epoch": 0.9497730943773556, + "learning_rate": 1.8635135810103676e-05, + "loss": 1.3901, + "step": 12348 + }, + { + "epoch": 0.949850011537574, + "learning_rate": 1.8578222079078665e-05, + "loss": 0.9232, + "step": 12349 + }, + { + "epoch": 0.9499269286977925, + "learning_rate": 1.852139484981491e-05, + "loss": 1.4881, + "step": 12350 + }, + { + "epoch": 0.9500038458580109, + "learning_rate": 1.8464654125631087e-05, + "loss": 1.4199, + "step": 12351 + }, + { + "epoch": 0.9500807630182294, + "learning_rate": 1.8407999909840045e-05, + "loss": 1.2796, + "step": 12352 + }, + { + "epoch": 0.9501576801784478, + "learning_rate": 1.8351432205749974e-05, + "loss": 1.4685, + "step": 12353 + }, + { + "epoch": 0.9502345973386662, + "learning_rate": 1.8294951016663895e-05, + "loss": 0.8389, + "step": 12354 + }, + { + "epoch": 0.9503115144988847, + "learning_rate": 1.8238556345879675e-05, + "loss": 1.0521, + "step": 12355 + }, + { + "epoch": 0.9503884316591031, + "learning_rate": 1.8182248196690678e-05, + "loss": 1.272, + "step": 12356 + }, + { + "epoch": 0.9504653488193215, + "learning_rate": 1.8126026572384435e-05, + "loss": 1.0702, + "step": 12357 + }, + { + "epoch": 0.9505422659795401, + "learning_rate": 1.8069891476243827e-05, + "loss": 0.8724, + "step": 12358 + }, + { + "epoch": 0.9506191831397585, + "learning_rate": 1.801384291154673e-05, + "loss": 1.3949, + "step": 12359 + }, + { + "epoch": 0.950696100299977, + "learning_rate": 1.795788088156586e-05, + "loss": 1.252, + "step": 12360 + }, + { + "epoch": 0.9507730174601954, + "learning_rate": 1.7902005389568765e-05, + "loss": 1.1387, + "step": 12361 + }, + { + "epoch": 0.9508499346204138, + "learning_rate": 1.784621643881834e-05, + "loss": 1.2472, + "step": 12362 + }, + { + "epoch": 0.9509268517806323, + "learning_rate": 1.7790514032571815e-05, + "loss": 1.1041, + "step": 12363 + }, + { + "epoch": 0.9510037689408507, + "learning_rate": 1.7734898174081916e-05, + "loss": 0.9482, + "step": 12364 + }, + { + "epoch": 0.9510806861010691, + "learning_rate": 1.767936886659621e-05, + "loss": 1.2077, + "step": 12365 + }, + { + "epoch": 0.9511576032612876, + "learning_rate": 1.7623926113356946e-05, + "loss": 1.1982, + "step": 12366 + }, + { + "epoch": 0.951234520421506, + "learning_rate": 1.756856991760153e-05, + "loss": 1.4091, + "step": 12367 + }, + { + "epoch": 0.9513114375817244, + "learning_rate": 1.7513300282562206e-05, + "loss": 1.1945, + "step": 12368 + }, + { + "epoch": 0.9513883547419429, + "learning_rate": 1.7458117211466395e-05, + "loss": 1.4266, + "step": 12369 + }, + { + "epoch": 0.9514652719021613, + "learning_rate": 1.740302070753619e-05, + "loss": 1.3746, + "step": 12370 + }, + { + "epoch": 0.9515421890623799, + "learning_rate": 1.7348010773988676e-05, + "loss": 1.2154, + "step": 12371 + }, + { + "epoch": 0.9516191062225983, + "learning_rate": 1.7293087414035956e-05, + "loss": 1.1474, + "step": 12372 + }, + { + "epoch": 0.9516960233828167, + "learning_rate": 1.7238250630885132e-05, + "loss": 0.9347, + "step": 12373 + }, + { + "epoch": 0.9517729405430352, + "learning_rate": 1.7183500427738142e-05, + "loss": 0.8565, + "step": 12374 + }, + { + "epoch": 0.9518498577032536, + "learning_rate": 1.7128836807791926e-05, + "loss": 1.2474, + "step": 12375 + }, + { + "epoch": 0.951926774863472, + "learning_rate": 1.7074259774238432e-05, + "loss": 1.1588, + "step": 12376 + }, + { + "epoch": 0.9520036920236905, + "learning_rate": 1.701976933026428e-05, + "loss": 0.9822, + "step": 12377 + }, + { + "epoch": 0.9520806091839089, + "learning_rate": 1.696536547905142e-05, + "loss": 1.2391, + "step": 12378 + }, + { + "epoch": 0.9521575263441274, + "learning_rate": 1.6911048223776316e-05, + "loss": 0.9747, + "step": 12379 + }, + { + "epoch": 0.9522344435043458, + "learning_rate": 1.685681756761076e-05, + "loss": 1.4819, + "step": 12380 + }, + { + "epoch": 0.9523113606645642, + "learning_rate": 1.680267351372139e-05, + "loss": 1.3563, + "step": 12381 + }, + { + "epoch": 0.9523882778247827, + "learning_rate": 1.6748616065269673e-05, + "loss": 1.059, + "step": 12382 + }, + { + "epoch": 0.9524651949850012, + "learning_rate": 1.6694645225412087e-05, + "loss": 1.4287, + "step": 12383 + }, + { + "epoch": 0.9525421121452196, + "learning_rate": 1.6640760997299942e-05, + "loss": 1.0282, + "step": 12384 + }, + { + "epoch": 0.9526190293054381, + "learning_rate": 1.6586963384079722e-05, + "loss": 1.0989, + "step": 12385 + }, + { + "epoch": 0.9526959464656565, + "learning_rate": 1.6533252388892584e-05, + "loss": 1.3305, + "step": 12386 + }, + { + "epoch": 0.9527728636258749, + "learning_rate": 1.6479628014875014e-05, + "loss": 1.2509, + "step": 12387 + }, + { + "epoch": 0.9528497807860934, + "learning_rate": 1.642609026515801e-05, + "loss": 1.2583, + "step": 12388 + }, + { + "epoch": 0.9529266979463118, + "learning_rate": 1.637263914286774e-05, + "loss": 0.9596, + "step": 12389 + }, + { + "epoch": 0.9530036151065303, + "learning_rate": 1.6319274651125204e-05, + "loss": 0.7987, + "step": 12390 + }, + { + "epoch": 0.9530805322667487, + "learning_rate": 1.626599679304658e-05, + "loss": 1.459, + "step": 12391 + }, + { + "epoch": 0.9531574494269671, + "learning_rate": 1.621280557174287e-05, + "loss": 0.8505, + "step": 12392 + }, + { + "epoch": 0.9532343665871856, + "learning_rate": 1.6159700990319436e-05, + "loss": 1.2911, + "step": 12393 + }, + { + "epoch": 0.953311283747404, + "learning_rate": 1.6106683051877623e-05, + "loss": 0.6452, + "step": 12394 + }, + { + "epoch": 0.9533882009076224, + "learning_rate": 1.6053751759513124e-05, + "loss": 1.2833, + "step": 12395 + }, + { + "epoch": 0.953465118067841, + "learning_rate": 1.600090711631663e-05, + "loss": 1.3633, + "step": 12396 + }, + { + "epoch": 0.9535420352280594, + "learning_rate": 1.5948149125373678e-05, + "loss": 0.9192, + "step": 12397 + }, + { + "epoch": 0.9536189523882779, + "learning_rate": 1.5895477789764966e-05, + "loss": 1.0924, + "step": 12398 + }, + { + "epoch": 0.9536958695484963, + "learning_rate": 1.584289311256587e-05, + "loss": 1.0657, + "step": 12399 + }, + { + "epoch": 0.9537727867087147, + "learning_rate": 1.5790395096847264e-05, + "loss": 1.6161, + "step": 12400 + }, + { + "epoch": 0.9538497038689332, + "learning_rate": 1.5737983745674035e-05, + "loss": 1.2523, + "step": 12401 + }, + { + "epoch": 0.9539266210291516, + "learning_rate": 1.568565906210706e-05, + "loss": 1.4484, + "step": 12402 + }, + { + "epoch": 0.95400353818937, + "learning_rate": 1.5633421049201236e-05, + "loss": 1.0825, + "step": 12403 + }, + { + "epoch": 0.9540804553495885, + "learning_rate": 1.5581269710006785e-05, + "loss": 1.1632, + "step": 12404 + }, + { + "epoch": 0.9541573725098069, + "learning_rate": 1.5529205047569273e-05, + "loss": 0.9711, + "step": 12405 + }, + { + "epoch": 0.9542342896700254, + "learning_rate": 1.547722706492827e-05, + "loss": 0.9484, + "step": 12406 + }, + { + "epoch": 0.9543112068302438, + "learning_rate": 1.542533576511934e-05, + "loss": 1.3789, + "step": 12407 + }, + { + "epoch": 0.9543881239904622, + "learning_rate": 1.5373531151172236e-05, + "loss": 0.9206, + "step": 12408 + }, + { + "epoch": 0.9544650411506808, + "learning_rate": 1.5321813226111703e-05, + "loss": 1.1812, + "step": 12409 + }, + { + "epoch": 0.9545419583108992, + "learning_rate": 1.5270181992957987e-05, + "loss": 0.9874, + "step": 12410 + }, + { + "epoch": 0.9546188754711176, + "learning_rate": 1.5218637454725682e-05, + "loss": 1.3513, + "step": 12411 + }, + { + "epoch": 0.9546957926313361, + "learning_rate": 1.5167179614424376e-05, + "loss": 1.4242, + "step": 12412 + }, + { + "epoch": 0.9547727097915545, + "learning_rate": 1.5115808475059001e-05, + "loss": 1.3357, + "step": 12413 + }, + { + "epoch": 0.9548496269517729, + "learning_rate": 1.5064524039629157e-05, + "loss": 1.1638, + "step": 12414 + }, + { + "epoch": 0.9549265441119914, + "learning_rate": 1.5013326311129117e-05, + "loss": 1.4263, + "step": 12415 + }, + { + "epoch": 0.9550034612722098, + "learning_rate": 1.496221529254882e-05, + "loss": 1.3826, + "step": 12416 + }, + { + "epoch": 0.9550803784324283, + "learning_rate": 1.4911190986872213e-05, + "loss": 1.1397, + "step": 12417 + }, + { + "epoch": 0.9551572955926467, + "learning_rate": 1.4860253397079082e-05, + "loss": 1.0621, + "step": 12418 + }, + { + "epoch": 0.9552342127528651, + "learning_rate": 1.4809402526143378e-05, + "loss": 1.0134, + "step": 12419 + }, + { + "epoch": 0.9553111299130836, + "learning_rate": 1.4758638377034562e-05, + "loss": 1.1756, + "step": 12420 + }, + { + "epoch": 0.955388047073302, + "learning_rate": 1.470796095271676e-05, + "loss": 1.0331, + "step": 12421 + }, + { + "epoch": 0.9554649642335205, + "learning_rate": 1.4657370256148939e-05, + "loss": 0.7614, + "step": 12422 + }, + { + "epoch": 0.955541881393739, + "learning_rate": 1.4606866290285404e-05, + "loss": 1.2887, + "step": 12423 + }, + { + "epoch": 0.9556187985539574, + "learning_rate": 1.4556449058074794e-05, + "loss": 1.156, + "step": 12424 + }, + { + "epoch": 0.9556957157141759, + "learning_rate": 1.4506118562461257e-05, + "loss": 0.9547, + "step": 12425 + }, + { + "epoch": 0.9557726328743943, + "learning_rate": 1.4455874806383772e-05, + "loss": 0.8543, + "step": 12426 + }, + { + "epoch": 0.9558495500346127, + "learning_rate": 1.4405717792775996e-05, + "loss": 1.0461, + "step": 12427 + }, + { + "epoch": 0.9559264671948312, + "learning_rate": 1.4355647524566417e-05, + "loss": 0.9796, + "step": 12428 + }, + { + "epoch": 0.9560033843550496, + "learning_rate": 1.4305664004678865e-05, + "loss": 1.3172, + "step": 12429 + }, + { + "epoch": 0.956080301515268, + "learning_rate": 1.4255767236032002e-05, + "loss": 1.2666, + "step": 12430 + }, + { + "epoch": 0.9561572186754865, + "learning_rate": 1.4205957221539334e-05, + "loss": 1.1766, + "step": 12431 + }, + { + "epoch": 0.9562341358357049, + "learning_rate": 1.4156233964109366e-05, + "loss": 1.1884, + "step": 12432 + }, + { + "epoch": 0.9563110529959233, + "learning_rate": 1.4106597466645277e-05, + "loss": 0.9319, + "step": 12433 + }, + { + "epoch": 0.9563879701561419, + "learning_rate": 1.4057047732045414e-05, + "loss": 1.0771, + "step": 12434 + }, + { + "epoch": 0.9564648873163603, + "learning_rate": 1.4007584763203297e-05, + "loss": 1.1815, + "step": 12435 + }, + { + "epoch": 0.9565418044765788, + "learning_rate": 1.3958208563006947e-05, + "loss": 1.0333, + "step": 12436 + }, + { + "epoch": 0.9566187216367972, + "learning_rate": 1.390891913433956e-05, + "loss": 1.2589, + "step": 12437 + }, + { + "epoch": 0.9566956387970156, + "learning_rate": 1.3859716480079165e-05, + "loss": 1.2224, + "step": 12438 + }, + { + "epoch": 0.9567725559572341, + "learning_rate": 1.3810600603098466e-05, + "loss": 1.0726, + "step": 12439 + }, + { + "epoch": 0.9568494731174525, + "learning_rate": 1.3761571506266002e-05, + "loss": 0.8625, + "step": 12440 + }, + { + "epoch": 0.9569263902776709, + "learning_rate": 1.3712629192444314e-05, + "loss": 1.1425, + "step": 12441 + }, + { + "epoch": 0.9570033074378894, + "learning_rate": 1.3663773664490953e-05, + "loss": 1.7033, + "step": 12442 + }, + { + "epoch": 0.9570802245981078, + "learning_rate": 1.3615004925258968e-05, + "loss": 1.3728, + "step": 12443 + }, + { + "epoch": 0.9571571417583263, + "learning_rate": 1.3566322977595913e-05, + "loss": 1.5058, + "step": 12444 + }, + { + "epoch": 0.9572340589185447, + "learning_rate": 1.351772782434435e-05, + "loss": 1.179, + "step": 12445 + }, + { + "epoch": 0.9573109760787631, + "learning_rate": 1.3469219468341842e-05, + "loss": 1.303, + "step": 12446 + }, + { + "epoch": 0.9573878932389817, + "learning_rate": 1.342079791242079e-05, + "loss": 1.1784, + "step": 12447 + }, + { + "epoch": 0.9574648103992001, + "learning_rate": 1.3372463159408765e-05, + "loss": 1.2931, + "step": 12448 + }, + { + "epoch": 0.9575417275594185, + "learning_rate": 1.332421521212801e-05, + "loss": 1.0764, + "step": 12449 + }, + { + "epoch": 0.957618644719637, + "learning_rate": 1.3276054073395438e-05, + "loss": 0.9208, + "step": 12450 + }, + { + "epoch": 0.9576955618798554, + "learning_rate": 1.3227979746023633e-05, + "loss": 1.3162, + "step": 12451 + }, + { + "epoch": 0.9577724790400738, + "learning_rate": 1.3179992232819515e-05, + "loss": 1.0227, + "step": 12452 + }, + { + "epoch": 0.9578493962002923, + "learning_rate": 1.3132091536585344e-05, + "loss": 1.5541, + "step": 12453 + }, + { + "epoch": 0.9579263133605107, + "learning_rate": 1.3084277660117716e-05, + "loss": 1.3618, + "step": 12454 + }, + { + "epoch": 0.9580032305207292, + "learning_rate": 1.303655060620873e-05, + "loss": 1.1838, + "step": 12455 + }, + { + "epoch": 0.9580801476809476, + "learning_rate": 1.298891037764549e-05, + "loss": 1.0831, + "step": 12456 + }, + { + "epoch": 0.958157064841166, + "learning_rate": 1.2941356977209273e-05, + "loss": 1.1509, + "step": 12457 + }, + { + "epoch": 0.9582339820013845, + "learning_rate": 1.2893890407677023e-05, + "loss": 0.9196, + "step": 12458 + }, + { + "epoch": 0.958310899161603, + "learning_rate": 1.2846510671820356e-05, + "loss": 1.3795, + "step": 12459 + }, + { + "epoch": 0.9583878163218214, + "learning_rate": 1.2799217772405891e-05, + "loss": 1.2916, + "step": 12460 + }, + { + "epoch": 0.9584647334820399, + "learning_rate": 1.2752011712194922e-05, + "loss": 1.0964, + "step": 12461 + }, + { + "epoch": 0.9585416506422583, + "learning_rate": 1.2704892493944076e-05, + "loss": 1.1146, + "step": 12462 + }, + { + "epoch": 0.9586185678024768, + "learning_rate": 1.2657860120404485e-05, + "loss": 0.8127, + "step": 12463 + }, + { + "epoch": 0.9586954849626952, + "learning_rate": 1.2610914594322453e-05, + "loss": 1.0171, + "step": 12464 + }, + { + "epoch": 0.9587724021229136, + "learning_rate": 1.2564055918439287e-05, + "loss": 1.2993, + "step": 12465 + }, + { + "epoch": 0.9588493192831321, + "learning_rate": 1.2517284095491133e-05, + "loss": 0.9041, + "step": 12466 + }, + { + "epoch": 0.9589262364433505, + "learning_rate": 1.247059912820897e-05, + "loss": 1.2157, + "step": 12467 + }, + { + "epoch": 0.9590031536035689, + "learning_rate": 1.2424001019318786e-05, + "loss": 1.3155, + "step": 12468 + }, + { + "epoch": 0.9590800707637874, + "learning_rate": 1.2377489771541406e-05, + "loss": 0.9835, + "step": 12469 + }, + { + "epoch": 0.9591569879240058, + "learning_rate": 1.2331065387592821e-05, + "loss": 0.9406, + "step": 12470 + }, + { + "epoch": 0.9592339050842242, + "learning_rate": 1.2284727870183865e-05, + "loss": 0.9119, + "step": 12471 + }, + { + "epoch": 0.9593108222444428, + "learning_rate": 1.2238477222019872e-05, + "loss": 1.0957, + "step": 12472 + }, + { + "epoch": 0.9593877394046612, + "learning_rate": 1.219231344580185e-05, + "loss": 1.2068, + "step": 12473 + }, + { + "epoch": 0.9594646565648797, + "learning_rate": 1.2146236544224975e-05, + "loss": 1.003, + "step": 12474 + }, + { + "epoch": 0.9595415737250981, + "learning_rate": 1.2100246519980262e-05, + "loss": 1.226, + "step": 12475 + }, + { + "epoch": 0.9596184908853165, + "learning_rate": 1.205434337575273e-05, + "loss": 1.0051, + "step": 12476 + }, + { + "epoch": 0.959695408045535, + "learning_rate": 1.2008527114222734e-05, + "loss": 1.2554, + "step": 12477 + }, + { + "epoch": 0.9597723252057534, + "learning_rate": 1.1962797738065633e-05, + "loss": 0.9832, + "step": 12478 + }, + { + "epoch": 0.9598492423659718, + "learning_rate": 1.1917155249951461e-05, + "loss": 1.0128, + "step": 12479 + }, + { + "epoch": 0.9599261595261903, + "learning_rate": 1.1871599652545584e-05, + "loss": 1.3611, + "step": 12480 + }, + { + "epoch": 0.9600030766864087, + "learning_rate": 1.1826130948507707e-05, + "loss": 1.4286, + "step": 12481 + }, + { + "epoch": 0.9600799938466272, + "learning_rate": 1.1780749140493041e-05, + "loss": 1.0974, + "step": 12482 + }, + { + "epoch": 0.9601569110068456, + "learning_rate": 1.1735454231151465e-05, + "loss": 1.0853, + "step": 12483 + }, + { + "epoch": 0.960233828167064, + "learning_rate": 1.1690246223127865e-05, + "loss": 1.3005, + "step": 12484 + }, + { + "epoch": 0.9603107453272826, + "learning_rate": 1.1645125119061795e-05, + "loss": 0.9851, + "step": 12485 + }, + { + "epoch": 0.960387662487501, + "learning_rate": 1.1600090921587813e-05, + "loss": 1.1901, + "step": 12486 + }, + { + "epoch": 0.9604645796477194, + "learning_rate": 1.1555143633335986e-05, + "loss": 1.0553, + "step": 12487 + }, + { + "epoch": 0.9605414968079379, + "learning_rate": 1.1510283256930376e-05, + "loss": 0.9551, + "step": 12488 + }, + { + "epoch": 0.9606184139681563, + "learning_rate": 1.1465509794990559e-05, + "loss": 1.1452, + "step": 12489 + }, + { + "epoch": 0.9606953311283747, + "learning_rate": 1.1420823250131107e-05, + "loss": 1.1404, + "step": 12490 + }, + { + "epoch": 0.9607722482885932, + "learning_rate": 1.1376223624960935e-05, + "loss": 1.4506, + "step": 12491 + }, + { + "epoch": 0.9608491654488116, + "learning_rate": 1.133171092208446e-05, + "loss": 1.1455, + "step": 12492 + }, + { + "epoch": 0.9609260826090301, + "learning_rate": 1.1287285144100933e-05, + "loss": 1.0521, + "step": 12493 + }, + { + "epoch": 0.9610029997692485, + "learning_rate": 1.124294629360445e-05, + "loss": 1.0328, + "step": 12494 + }, + { + "epoch": 0.9610799169294669, + "learning_rate": 1.119869437318377e-05, + "loss": 1.0154, + "step": 12495 + }, + { + "epoch": 0.9611568340896854, + "learning_rate": 1.115452938542283e-05, + "loss": 0.9358, + "step": 12496 + }, + { + "epoch": 0.9612337512499038, + "learning_rate": 1.1110451332900729e-05, + "loss": 1.2245, + "step": 12497 + }, + { + "epoch": 0.9613106684101222, + "learning_rate": 1.106646021819091e-05, + "loss": 1.2063, + "step": 12498 + }, + { + "epoch": 0.9613875855703408, + "learning_rate": 1.102255604386232e-05, + "loss": 1.1398, + "step": 12499 + }, + { + "epoch": 0.9614645027305592, + "learning_rate": 1.0978738812478405e-05, + "loss": 1.2032, + "step": 12500 + }, + { + "epoch": 0.9615414198907777, + "learning_rate": 1.0935008526597623e-05, + "loss": 1.3763, + "step": 12501 + }, + { + "epoch": 0.9616183370509961, + "learning_rate": 1.089136518877376e-05, + "loss": 1.1248, + "step": 12502 + }, + { + "epoch": 0.9616952542112145, + "learning_rate": 1.0847808801554949e-05, + "loss": 1.2183, + "step": 12503 + }, + { + "epoch": 0.961772171371433, + "learning_rate": 1.0804339367484484e-05, + "loss": 1.0694, + "step": 12504 + }, + { + "epoch": 0.9618490885316514, + "learning_rate": 1.076095688910067e-05, + "loss": 1.0625, + "step": 12505 + }, + { + "epoch": 0.9619260056918698, + "learning_rate": 1.071766136893665e-05, + "loss": 1.0587, + "step": 12506 + }, + { + "epoch": 0.9620029228520883, + "learning_rate": 1.0674452809520397e-05, + "loss": 0.867, + "step": 12507 + }, + { + "epoch": 0.9620798400123067, + "learning_rate": 1.0631331213375061e-05, + "loss": 1.0942, + "step": 12508 + }, + { + "epoch": 0.9621567571725252, + "learning_rate": 1.0588296583018464e-05, + "loss": 1.0028, + "step": 12509 + }, + { + "epoch": 0.9622336743327436, + "learning_rate": 1.0545348920963426e-05, + "loss": 1.3383, + "step": 12510 + }, + { + "epoch": 0.962310591492962, + "learning_rate": 1.0502488229717777e-05, + "loss": 1.1583, + "step": 12511 + }, + { + "epoch": 0.9623875086531806, + "learning_rate": 1.045971451178418e-05, + "loss": 0.9778, + "step": 12512 + }, + { + "epoch": 0.962464425813399, + "learning_rate": 1.0417027769660137e-05, + "loss": 1.3648, + "step": 12513 + }, + { + "epoch": 0.9625413429736174, + "learning_rate": 1.037442800583832e-05, + "loss": 1.3772, + "step": 12514 + }, + { + "epoch": 0.9626182601338359, + "learning_rate": 1.0331915222806077e-05, + "loss": 1.37, + "step": 12515 + }, + { + "epoch": 0.9626951772940543, + "learning_rate": 1.028948942304575e-05, + "loss": 1.0969, + "step": 12516 + }, + { + "epoch": 0.9627720944542727, + "learning_rate": 1.024715060903486e-05, + "loss": 1.1798, + "step": 12517 + }, + { + "epoch": 0.9628490116144912, + "learning_rate": 1.0204898783245097e-05, + "loss": 0.8726, + "step": 12518 + }, + { + "epoch": 0.9629259287747096, + "learning_rate": 1.0162733948144321e-05, + "loss": 1.2034, + "step": 12519 + }, + { + "epoch": 0.9630028459349281, + "learning_rate": 1.0120656106193894e-05, + "loss": 1.0849, + "step": 12520 + }, + { + "epoch": 0.9630797630951465, + "learning_rate": 1.0078665259851183e-05, + "loss": 0.9967, + "step": 12521 + }, + { + "epoch": 0.9631566802553649, + "learning_rate": 1.0036761411567897e-05, + "loss": 0.8526, + "step": 12522 + }, + { + "epoch": 0.9632335974155835, + "learning_rate": 9.994944563790908e-06, + "loss": 1.4954, + "step": 12523 + }, + { + "epoch": 0.9633105145758019, + "learning_rate": 9.9532147189621e-06, + "loss": 1.2077, + "step": 12524 + }, + { + "epoch": 0.9633874317360203, + "learning_rate": 9.911571879517856e-06, + "loss": 1.0329, + "step": 12525 + }, + { + "epoch": 0.9634643488962388, + "learning_rate": 9.87001604788973e-06, + "loss": 1.0367, + "step": 12526 + }, + { + "epoch": 0.9635412660564572, + "learning_rate": 9.828547226504614e-06, + "loss": 0.7542, + "step": 12527 + }, + { + "epoch": 0.9636181832166757, + "learning_rate": 9.787165417783405e-06, + "loss": 1.1541, + "step": 12528 + }, + { + "epoch": 0.9636951003768941, + "learning_rate": 9.745870624142838e-06, + "loss": 1.4798, + "step": 12529 + }, + { + "epoch": 0.9637720175371125, + "learning_rate": 9.704662847993984e-06, + "loss": 1.0329, + "step": 12530 + }, + { + "epoch": 0.963848934697331, + "learning_rate": 9.663542091742916e-06, + "loss": 1.0714, + "step": 12531 + }, + { + "epoch": 0.9639258518575494, + "learning_rate": 9.622508357790883e-06, + "loss": 1.1419, + "step": 12532 + }, + { + "epoch": 0.9640027690177678, + "learning_rate": 9.581561648533798e-06, + "loss": 1.1956, + "step": 12533 + }, + { + "epoch": 0.9640796861779863, + "learning_rate": 9.540701966362586e-06, + "loss": 1.0352, + "step": 12534 + }, + { + "epoch": 0.9641566033382047, + "learning_rate": 9.49992931366317e-06, + "loss": 1.0903, + "step": 12535 + }, + { + "epoch": 0.9642335204984231, + "learning_rate": 9.45924369281631e-06, + "loss": 1.1682, + "step": 12536 + }, + { + "epoch": 0.9643104376586417, + "learning_rate": 9.418645106197443e-06, + "loss": 1.2397, + "step": 12537 + }, + { + "epoch": 0.9643873548188601, + "learning_rate": 9.378133556177504e-06, + "loss": 1.0185, + "step": 12538 + }, + { + "epoch": 0.9644642719790786, + "learning_rate": 9.337709045121601e-06, + "loss": 0.9796, + "step": 12539 + }, + { + "epoch": 0.964541189139297, + "learning_rate": 9.297371575390512e-06, + "loss": 1.3726, + "step": 12540 + }, + { + "epoch": 0.9646181062995154, + "learning_rate": 9.25712114933952e-06, + "loss": 0.9965, + "step": 12541 + }, + { + "epoch": 0.9646950234597339, + "learning_rate": 9.216957769318912e-06, + "loss": 1.0806, + "step": 12542 + }, + { + "epoch": 0.9647719406199523, + "learning_rate": 9.176881437673645e-06, + "loss": 1.4456, + "step": 12543 + }, + { + "epoch": 0.9648488577801707, + "learning_rate": 9.136892156744015e-06, + "loss": 1.9841, + "step": 12544 + }, + { + "epoch": 0.9649257749403892, + "learning_rate": 9.096989928864984e-06, + "loss": 1.0315, + "step": 12545 + }, + { + "epoch": 0.9650026921006076, + "learning_rate": 9.057174756366526e-06, + "loss": 1.4377, + "step": 12546 + }, + { + "epoch": 0.9650796092608261, + "learning_rate": 9.017446641573446e-06, + "loss": 1.2672, + "step": 12547 + }, + { + "epoch": 0.9651565264210445, + "learning_rate": 8.977805586805555e-06, + "loss": 1.2858, + "step": 12548 + }, + { + "epoch": 0.965233443581263, + "learning_rate": 8.938251594377334e-06, + "loss": 1.4411, + "step": 12549 + }, + { + "epoch": 0.9653103607414815, + "learning_rate": 8.89878466659877e-06, + "loss": 1.3483, + "step": 12550 + }, + { + "epoch": 0.9653872779016999, + "learning_rate": 8.859404805774018e-06, + "loss": 1.1452, + "step": 12551 + }, + { + "epoch": 0.9654641950619183, + "learning_rate": 8.820112014202742e-06, + "loss": 1.4143, + "step": 12552 + }, + { + "epoch": 0.9655411122221368, + "learning_rate": 8.780906294179104e-06, + "loss": 1.2448, + "step": 12553 + }, + { + "epoch": 0.9656180293823552, + "learning_rate": 8.741787647992439e-06, + "loss": 1.277, + "step": 12554 + }, + { + "epoch": 0.9656949465425736, + "learning_rate": 8.702756077927087e-06, + "loss": 1.1117, + "step": 12555 + }, + { + "epoch": 0.9657718637027921, + "learning_rate": 8.663811586261894e-06, + "loss": 1.2591, + "step": 12556 + }, + { + "epoch": 0.9658487808630105, + "learning_rate": 8.624954175271038e-06, + "loss": 1.0719, + "step": 12557 + }, + { + "epoch": 0.965925698023229, + "learning_rate": 8.586183847223371e-06, + "loss": 1.2018, + "step": 12558 + }, + { + "epoch": 0.9660026151834474, + "learning_rate": 8.547500604382752e-06, + "loss": 1.4612, + "step": 12559 + }, + { + "epoch": 0.9660795323436658, + "learning_rate": 8.508904449007871e-06, + "loss": 1.2367, + "step": 12560 + }, + { + "epoch": 0.9661564495038844, + "learning_rate": 8.470395383352425e-06, + "loss": 1.4056, + "step": 12561 + }, + { + "epoch": 0.9662333666641028, + "learning_rate": 8.431973409665116e-06, + "loss": 1.1051, + "step": 12562 + }, + { + "epoch": 0.9663102838243212, + "learning_rate": 8.393638530189319e-06, + "loss": 1.113, + "step": 12563 + }, + { + "epoch": 0.9663872009845397, + "learning_rate": 8.355390747163405e-06, + "loss": 1.101, + "step": 12564 + }, + { + "epoch": 0.9664641181447581, + "learning_rate": 8.317230062820925e-06, + "loss": 1.1721, + "step": 12565 + }, + { + "epoch": 0.9665410353049766, + "learning_rate": 8.279156479389926e-06, + "loss": 1.3101, + "step": 12566 + }, + { + "epoch": 0.966617952465195, + "learning_rate": 8.241169999093468e-06, + "loss": 1.8327, + "step": 12567 + }, + { + "epoch": 0.9666948696254134, + "learning_rate": 8.203270624150106e-06, + "loss": 1.2764, + "step": 12568 + }, + { + "epoch": 0.9667717867856319, + "learning_rate": 8.165458356772238e-06, + "loss": 0.9915, + "step": 12569 + }, + { + "epoch": 0.9668487039458503, + "learning_rate": 8.127733199167931e-06, + "loss": 1.4001, + "step": 12570 + }, + { + "epoch": 0.9669256211060687, + "learning_rate": 8.090095153540255e-06, + "loss": 1.1301, + "step": 12571 + }, + { + "epoch": 0.9670025382662872, + "learning_rate": 8.052544222086621e-06, + "loss": 1.5621, + "step": 12572 + }, + { + "epoch": 0.9670794554265056, + "learning_rate": 8.015080406999942e-06, + "loss": 1.0522, + "step": 12573 + }, + { + "epoch": 0.967156372586724, + "learning_rate": 7.977703710467632e-06, + "loss": 1.1197, + "step": 12574 + }, + { + "epoch": 0.9672332897469426, + "learning_rate": 7.940414134671948e-06, + "loss": 0.9458, + "step": 12575 + }, + { + "epoch": 0.967310206907161, + "learning_rate": 7.903211681790645e-06, + "loss": 1.316, + "step": 12576 + }, + { + "epoch": 0.9673871240673795, + "learning_rate": 7.866096353995823e-06, + "loss": 1.4618, + "step": 12577 + }, + { + "epoch": 0.9674640412275979, + "learning_rate": 7.829068153454577e-06, + "loss": 1.0523, + "step": 12578 + }, + { + "epoch": 0.9675409583878163, + "learning_rate": 7.792127082329348e-06, + "loss": 1.1657, + "step": 12579 + }, + { + "epoch": 0.9676178755480348, + "learning_rate": 7.755273142776742e-06, + "loss": 1.1556, + "step": 12580 + }, + { + "epoch": 0.9676947927082532, + "learning_rate": 7.718506336949038e-06, + "loss": 1.3058, + "step": 12581 + }, + { + "epoch": 0.9677717098684716, + "learning_rate": 7.681826666993019e-06, + "loss": 1.0612, + "step": 12582 + }, + { + "epoch": 0.9678486270286901, + "learning_rate": 7.645234135050305e-06, + "loss": 1.1075, + "step": 12583 + }, + { + "epoch": 0.9679255441889085, + "learning_rate": 7.608728743257687e-06, + "loss": 1.2045, + "step": 12584 + }, + { + "epoch": 0.968002461349127, + "learning_rate": 7.572310493746793e-06, + "loss": 1.2167, + "step": 12585 + }, + { + "epoch": 0.9680793785093454, + "learning_rate": 7.535979388643921e-06, + "loss": 1.2357, + "step": 12586 + }, + { + "epoch": 0.9681562956695638, + "learning_rate": 7.499735430070709e-06, + "loss": 1.1877, + "step": 12587 + }, + { + "epoch": 0.9682332128297824, + "learning_rate": 7.463578620143297e-06, + "loss": 0.8865, + "step": 12588 + }, + { + "epoch": 0.9683101299900008, + "learning_rate": 7.427508960973162e-06, + "loss": 0.8846, + "step": 12589 + }, + { + "epoch": 0.9683870471502192, + "learning_rate": 7.391526454666286e-06, + "loss": 1.0404, + "step": 12590 + }, + { + "epoch": 0.9684639643104377, + "learning_rate": 7.355631103323657e-06, + "loss": 0.8971, + "step": 12591 + }, + { + "epoch": 0.9685408814706561, + "learning_rate": 7.319822909041429e-06, + "loss": 1.1847, + "step": 12592 + }, + { + "epoch": 0.9686177986308745, + "learning_rate": 7.2841018739102646e-06, + "loss": 0.8368, + "step": 12593 + }, + { + "epoch": 0.968694715791093, + "learning_rate": 7.2484680000161615e-06, + "loss": 1.0476, + "step": 12594 + }, + { + "epoch": 0.9687716329513114, + "learning_rate": 7.2129212894396224e-06, + "loss": 1.1231, + "step": 12595 + }, + { + "epoch": 0.9688485501115299, + "learning_rate": 7.177461744256319e-06, + "loss": 1.1153, + "step": 12596 + }, + { + "epoch": 0.9689254672717483, + "learning_rate": 7.14208936653693e-06, + "loss": 0.981, + "step": 12597 + }, + { + "epoch": 0.9690023844319667, + "learning_rate": 7.106804158346636e-06, + "loss": 1.3899, + "step": 12598 + }, + { + "epoch": 0.9690793015921852, + "learning_rate": 7.071606121745955e-06, + "loss": 0.9187, + "step": 12599 + }, + { + "epoch": 0.9691562187524037, + "learning_rate": 7.0364952587902435e-06, + "loss": 1.1633, + "step": 12600 + }, + { + "epoch": 0.9692331359126221, + "learning_rate": 7.001471571529361e-06, + "loss": 1.2645, + "step": 12601 + }, + { + "epoch": 0.9693100530728406, + "learning_rate": 6.9665350620085055e-06, + "loss": 1.2485, + "step": 12602 + }, + { + "epoch": 0.969386970233059, + "learning_rate": 6.931685732267712e-06, + "loss": 1.1676, + "step": 12603 + }, + { + "epoch": 0.9694638873932775, + "learning_rate": 6.8969235843416855e-06, + "loss": 1.0614, + "step": 12604 + }, + { + "epoch": 0.9695408045534959, + "learning_rate": 6.862248620260636e-06, + "loss": 1.1671, + "step": 12605 + }, + { + "epoch": 0.9696177217137143, + "learning_rate": 6.827660842048777e-06, + "loss": 1.6335, + "step": 12606 + }, + { + "epoch": 0.9696946388739328, + "learning_rate": 6.793160251725994e-06, + "loss": 1.5102, + "step": 12607 + }, + { + "epoch": 0.9697715560341512, + "learning_rate": 6.75874685130684e-06, + "loss": 1.2357, + "step": 12608 + }, + { + "epoch": 0.9698484731943696, + "learning_rate": 6.72442064280071e-06, + "loss": 0.9505, + "step": 12609 + }, + { + "epoch": 0.9699253903545881, + "learning_rate": 6.6901816282116665e-06, + "loss": 0.8884, + "step": 12610 + }, + { + "epoch": 0.9700023075148065, + "learning_rate": 6.6560298095394434e-06, + "loss": 1.1603, + "step": 12611 + }, + { + "epoch": 0.9700792246750249, + "learning_rate": 6.621965188777945e-06, + "loss": 0.9623, + "step": 12612 + }, + { + "epoch": 0.9701561418352435, + "learning_rate": 6.587987767916248e-06, + "loss": 1.0489, + "step": 12613 + }, + { + "epoch": 0.9702330589954619, + "learning_rate": 6.554097548938265e-06, + "loss": 1.1253, + "step": 12614 + }, + { + "epoch": 0.9703099761556804, + "learning_rate": 6.5202945338229124e-06, + "loss": 1.2029, + "step": 12615 + }, + { + "epoch": 0.9703868933158988, + "learning_rate": 6.486578724544279e-06, + "loss": 1.2329, + "step": 12616 + }, + { + "epoch": 0.9704638104761172, + "learning_rate": 6.4529501230706225e-06, + "loss": 1.3183, + "step": 12617 + }, + { + "epoch": 0.9705407276363357, + "learning_rate": 6.419408731365706e-06, + "loss": 0.7932, + "step": 12618 + }, + { + "epoch": 0.9706176447965541, + "learning_rate": 6.385954551388129e-06, + "loss": 1.4808, + "step": 12619 + }, + { + "epoch": 0.9706945619567725, + "learning_rate": 6.35258758509133e-06, + "loss": 1.3753, + "step": 12620 + }, + { + "epoch": 0.970771479116991, + "learning_rate": 6.319307834423582e-06, + "loss": 1.1444, + "step": 12621 + }, + { + "epoch": 0.9708483962772094, + "learning_rate": 6.286115301328166e-06, + "loss": 1.3147, + "step": 12622 + }, + { + "epoch": 0.9709253134374279, + "learning_rate": 6.25300998774303e-06, + "loss": 1.4053, + "step": 12623 + }, + { + "epoch": 0.9710022305976463, + "learning_rate": 6.2199918956014625e-06, + "loss": 1.0932, + "step": 12624 + }, + { + "epoch": 0.9710791477578647, + "learning_rate": 6.18706102683142e-06, + "loss": 1.3351, + "step": 12625 + }, + { + "epoch": 0.9711560649180833, + "learning_rate": 6.154217383355698e-06, + "loss": 1.2115, + "step": 12626 + }, + { + "epoch": 0.9712329820783017, + "learning_rate": 6.1214609670919295e-06, + "loss": 1.1545, + "step": 12627 + }, + { + "epoch": 0.9713098992385201, + "learning_rate": 6.088791779953085e-06, + "loss": 1.0827, + "step": 12628 + }, + { + "epoch": 0.9713868163987386, + "learning_rate": 6.056209823846637e-06, + "loss": 1.1002, + "step": 12629 + }, + { + "epoch": 0.971463733558957, + "learning_rate": 6.023715100674898e-06, + "loss": 0.9341, + "step": 12630 + }, + { + "epoch": 0.9715406507191755, + "learning_rate": 5.991307612335517e-06, + "loss": 1.0185, + "step": 12631 + }, + { + "epoch": 0.9716175678793939, + "learning_rate": 5.958987360720647e-06, + "loss": 1.1625, + "step": 12632 + }, + { + "epoch": 0.9716944850396123, + "learning_rate": 5.926754347717611e-06, + "loss": 1.3745, + "step": 12633 + }, + { + "epoch": 0.9717714021998308, + "learning_rate": 5.8946085752084046e-06, + "loss": 1.1692, + "step": 12634 + }, + { + "epoch": 0.9718483193600492, + "learning_rate": 5.8625500450701914e-06, + "loss": 1.182, + "step": 12635 + }, + { + "epoch": 0.9719252365202676, + "learning_rate": 5.830578759174809e-06, + "loss": 1.4195, + "step": 12636 + }, + { + "epoch": 0.9720021536804861, + "learning_rate": 5.7986947193889305e-06, + "loss": 1.0184, + "step": 12637 + }, + { + "epoch": 0.9720790708407046, + "learning_rate": 5.766897927574733e-06, + "loss": 1.1037, + "step": 12638 + }, + { + "epoch": 0.972155988000923, + "learning_rate": 5.735188385588564e-06, + "loss": 1.1467, + "step": 12639 + }, + { + "epoch": 0.9722329051611415, + "learning_rate": 5.703566095282109e-06, + "loss": 1.176, + "step": 12640 + }, + { + "epoch": 0.9723098223213599, + "learning_rate": 5.6720310585015586e-06, + "loss": 1.0016, + "step": 12641 + }, + { + "epoch": 0.9723867394815784, + "learning_rate": 5.6405832770886065e-06, + "loss": 1.064, + "step": 12642 + }, + { + "epoch": 0.9724636566417968, + "learning_rate": 5.609222752879284e-06, + "loss": 1.0039, + "step": 12643 + }, + { + "epoch": 0.9725405738020152, + "learning_rate": 5.5779494877049585e-06, + "loss": 1.4048, + "step": 12644 + }, + { + "epoch": 0.9726174909622337, + "learning_rate": 5.5467634833915035e-06, + "loss": 1.0396, + "step": 12645 + }, + { + "epoch": 0.9726944081224521, + "learning_rate": 5.515664741760129e-06, + "loss": 1.2299, + "step": 12646 + }, + { + "epoch": 0.9727713252826705, + "learning_rate": 5.4846532646267155e-06, + "loss": 1.2927, + "step": 12647 + }, + { + "epoch": 0.972848242442889, + "learning_rate": 5.453729053801648e-06, + "loss": 1.4449, + "step": 12648 + }, + { + "epoch": 0.9729251596031074, + "learning_rate": 5.422892111091149e-06, + "loss": 1.2064, + "step": 12649 + }, + { + "epoch": 0.973002076763326, + "learning_rate": 5.392142438295445e-06, + "loss": 1.3436, + "step": 12650 + }, + { + "epoch": 0.9730789939235444, + "learning_rate": 5.361480037210265e-06, + "loss": 1.117, + "step": 12651 + }, + { + "epoch": 0.9731559110837628, + "learning_rate": 5.330904909625845e-06, + "loss": 1.1208, + "step": 12652 + }, + { + "epoch": 0.9732328282439813, + "learning_rate": 5.300417057327589e-06, + "loss": 1.3434, + "step": 12653 + }, + { + "epoch": 0.9733097454041997, + "learning_rate": 5.270016482095908e-06, + "loss": 1.2208, + "step": 12654 + }, + { + "epoch": 0.9733866625644181, + "learning_rate": 5.239703185705546e-06, + "loss": 1.2378, + "step": 12655 + }, + { + "epoch": 0.9734635797246366, + "learning_rate": 5.209477169926758e-06, + "loss": 1.2216, + "step": 12656 + }, + { + "epoch": 0.973540496884855, + "learning_rate": 5.1793384365246274e-06, + "loss": 1.2038, + "step": 12657 + }, + { + "epoch": 0.9736174140450734, + "learning_rate": 5.149286987258583e-06, + "loss": 1.2282, + "step": 12658 + }, + { + "epoch": 0.9736943312052919, + "learning_rate": 5.119322823883554e-06, + "loss": 1.1535, + "step": 12659 + }, + { + "epoch": 0.9737712483655103, + "learning_rate": 5.0894459481493075e-06, + "loss": 1.2682, + "step": 12660 + }, + { + "epoch": 0.9738481655257288, + "learning_rate": 5.059656361800114e-06, + "loss": 1.9436, + "step": 12661 + }, + { + "epoch": 0.9739250826859472, + "learning_rate": 5.029954066575748e-06, + "loss": 1.5157, + "step": 12662 + }, + { + "epoch": 0.9740019998461656, + "learning_rate": 5.000339064210324e-06, + "loss": 1.2199, + "step": 12663 + }, + { + "epoch": 0.9740789170063842, + "learning_rate": 4.970811356433124e-06, + "loss": 0.8043, + "step": 12664 + }, + { + "epoch": 0.9741558341666026, + "learning_rate": 4.941370944968271e-06, + "loss": 0.9901, + "step": 12665 + }, + { + "epoch": 0.974232751326821, + "learning_rate": 4.9120178315350535e-06, + "loss": 0.8299, + "step": 12666 + }, + { + "epoch": 0.9743096684870395, + "learning_rate": 4.882752017847103e-06, + "loss": 1.0728, + "step": 12667 + }, + { + "epoch": 0.9743865856472579, + "learning_rate": 4.853573505613551e-06, + "loss": 0.8153, + "step": 12668 + }, + { + "epoch": 0.9744635028074764, + "learning_rate": 4.824482296537869e-06, + "loss": 0.9062, + "step": 12669 + }, + { + "epoch": 0.9745404199676948, + "learning_rate": 4.79547839231903e-06, + "loss": 1.367, + "step": 12670 + }, + { + "epoch": 0.9746173371279132, + "learning_rate": 4.766561794650515e-06, + "loss": 0.9308, + "step": 12671 + }, + { + "epoch": 0.9746942542881317, + "learning_rate": 4.737732505220638e-06, + "loss": 0.8003, + "step": 12672 + }, + { + "epoch": 0.9747711714483501, + "learning_rate": 4.708990525713053e-06, + "loss": 1.1944, + "step": 12673 + }, + { + "epoch": 0.9748480886085685, + "learning_rate": 4.680335857805751e-06, + "loss": 1.4942, + "step": 12674 + }, + { + "epoch": 0.974925005768787, + "learning_rate": 4.65176850317206e-06, + "loss": 1.3383, + "step": 12675 + }, + { + "epoch": 0.9750019229290054, + "learning_rate": 4.62328846347998e-06, + "loss": 1.0785, + "step": 12676 + }, + { + "epoch": 0.9750788400892239, + "learning_rate": 4.5948957403925126e-06, + "loss": 1.28, + "step": 12677 + }, + { + "epoch": 0.9751557572494424, + "learning_rate": 4.566590335567666e-06, + "loss": 1.4324, + "step": 12678 + }, + { + "epoch": 0.9752326744096608, + "learning_rate": 4.538372250657952e-06, + "loss": 1.1537, + "step": 12679 + }, + { + "epoch": 0.9753095915698793, + "learning_rate": 4.5102414873112176e-06, + "loss": 1.1609, + "step": 12680 + }, + { + "epoch": 0.9753865087300977, + "learning_rate": 4.48219804717015e-06, + "loss": 1.3438, + "step": 12681 + }, + { + "epoch": 0.9754634258903161, + "learning_rate": 4.454241931872105e-06, + "loss": 1.1787, + "step": 12682 + }, + { + "epoch": 0.9755403430505346, + "learning_rate": 4.426373143049611e-06, + "loss": 1.0784, + "step": 12683 + }, + { + "epoch": 0.975617260210753, + "learning_rate": 4.3985916823297e-06, + "loss": 0.9945, + "step": 12684 + }, + { + "epoch": 0.9756941773709714, + "learning_rate": 4.370897551334741e-06, + "loss": 1.4363, + "step": 12685 + }, + { + "epoch": 0.9757710945311899, + "learning_rate": 4.343290751681772e-06, + "loss": 1.2473, + "step": 12686 + }, + { + "epoch": 0.9758480116914083, + "learning_rate": 4.315771284982672e-06, + "loss": 0.9725, + "step": 12687 + }, + { + "epoch": 0.9759249288516268, + "learning_rate": 4.288339152844656e-06, + "loss": 1.2346, + "step": 12688 + }, + { + "epoch": 0.9760018460118453, + "learning_rate": 4.260994356869108e-06, + "loss": 1.38, + "step": 12689 + }, + { + "epoch": 0.9760787631720637, + "learning_rate": 4.233736898653084e-06, + "loss": 1.3028, + "step": 12690 + }, + { + "epoch": 0.9761556803322822, + "learning_rate": 4.206566779787979e-06, + "loss": 1.2133, + "step": 12691 + }, + { + "epoch": 0.9762325974925006, + "learning_rate": 4.179484001860356e-06, + "loss": 1.0899, + "step": 12692 + }, + { + "epoch": 0.976309514652719, + "learning_rate": 4.1524885664514515e-06, + "loss": 0.9205, + "step": 12693 + }, + { + "epoch": 0.9763864318129375, + "learning_rate": 4.1255804751376695e-06, + "loss": 1.0226, + "step": 12694 + }, + { + "epoch": 0.9764633489731559, + "learning_rate": 4.098759729490254e-06, + "loss": 1.0614, + "step": 12695 + }, + { + "epoch": 0.9765402661333743, + "learning_rate": 4.072026331075284e-06, + "loss": 1.2945, + "step": 12696 + }, + { + "epoch": 0.9766171832935928, + "learning_rate": 4.045380281453681e-06, + "loss": 1.066, + "step": 12697 + }, + { + "epoch": 0.9766941004538112, + "learning_rate": 4.018821582181364e-06, + "loss": 1.1933, + "step": 12698 + }, + { + "epoch": 0.9767710176140297, + "learning_rate": 3.992350234809261e-06, + "loss": 0.955, + "step": 12699 + }, + { + "epoch": 0.9768479347742481, + "learning_rate": 3.965966240882802e-06, + "loss": 0.9804, + "step": 12700 + }, + { + "epoch": 0.9769248519344665, + "learning_rate": 3.939669601942753e-06, + "loss": 1.519, + "step": 12701 + }, + { + "epoch": 0.9770017690946851, + "learning_rate": 3.913460319524387e-06, + "loss": 0.7899, + "step": 12702 + }, + { + "epoch": 0.9770786862549035, + "learning_rate": 3.8873383951584795e-06, + "loss": 1.2498, + "step": 12703 + }, + { + "epoch": 0.9771556034151219, + "learning_rate": 3.8613038303699775e-06, + "loss": 1.149, + "step": 12704 + }, + { + "epoch": 0.9772325205753404, + "learning_rate": 3.835356626679165e-06, + "loss": 1.2062, + "step": 12705 + }, + { + "epoch": 0.9773094377355588, + "learning_rate": 3.8094967856013297e-06, + "loss": 1.079, + "step": 12706 + }, + { + "epoch": 0.9773863548957773, + "learning_rate": 3.7837243086459303e-06, + "loss": 1.4197, + "step": 12707 + }, + { + "epoch": 0.9774632720559957, + "learning_rate": 3.7580391973184302e-06, + "loss": 1.7759, + "step": 12708 + }, + { + "epoch": 0.9775401892162141, + "learning_rate": 3.7324414531184626e-06, + "loss": 1.1377, + "step": 12709 + }, + { + "epoch": 0.9776171063764326, + "learning_rate": 3.706931077540332e-06, + "loss": 1.1648, + "step": 12710 + }, + { + "epoch": 0.977694023536651, + "learning_rate": 3.6815080720740136e-06, + "loss": 1.1485, + "step": 12711 + }, + { + "epoch": 0.9777709406968694, + "learning_rate": 3.6561724382039863e-06, + "loss": 1.1104, + "step": 12712 + }, + { + "epoch": 0.9778478578570879, + "learning_rate": 3.6309241774094004e-06, + "loss": 1.0424, + "step": 12713 + }, + { + "epoch": 0.9779247750173063, + "learning_rate": 3.6057632911645766e-06, + "loss": 1.3053, + "step": 12714 + }, + { + "epoch": 0.9780016921775248, + "learning_rate": 3.5806897809388395e-06, + "loss": 1.2633, + "step": 12715 + }, + { + "epoch": 0.9780786093377433, + "learning_rate": 3.555703648196018e-06, + "loss": 1.0491, + "step": 12716 + }, + { + "epoch": 0.9781555264979617, + "learning_rate": 3.5308048943954453e-06, + "loss": 1.366, + "step": 12717 + }, + { + "epoch": 0.9782324436581802, + "learning_rate": 3.5059935209906247e-06, + "loss": 1.0456, + "step": 12718 + }, + { + "epoch": 0.9783093608183986, + "learning_rate": 3.4812695294303977e-06, + "loss": 1.1009, + "step": 12719 + }, + { + "epoch": 0.978386277978617, + "learning_rate": 3.456632921158609e-06, + "loss": 1.1536, + "step": 12720 + }, + { + "epoch": 0.9784631951388355, + "learning_rate": 3.4320836976134418e-06, + "loss": 1.8793, + "step": 12721 + }, + { + "epoch": 0.9785401122990539, + "learning_rate": 3.4076218602289156e-06, + "loss": 1.0634, + "step": 12722 + }, + { + "epoch": 0.9786170294592723, + "learning_rate": 3.383247410432888e-06, + "loss": 1.1016, + "step": 12723 + }, + { + "epoch": 0.9786939466194908, + "learning_rate": 3.358960349648721e-06, + "loss": 1.2406, + "step": 12724 + }, + { + "epoch": 0.9787708637797092, + "learning_rate": 3.3347606792947793e-06, + "loss": 1.2966, + "step": 12725 + }, + { + "epoch": 0.9788477809399277, + "learning_rate": 3.3106484007837667e-06, + "loss": 0.7612, + "step": 12726 + }, + { + "epoch": 0.9789246981001462, + "learning_rate": 3.2866235155238897e-06, + "loss": 1.1185, + "step": 12727 + }, + { + "epoch": 0.9790016152603646, + "learning_rate": 3.262686024918027e-06, + "loss": 1.002, + "step": 12728 + }, + { + "epoch": 0.9790785324205831, + "learning_rate": 3.2388359303635596e-06, + "loss": 1.3115, + "step": 12729 + }, + { + "epoch": 0.9791554495808015, + "learning_rate": 3.2150732332535404e-06, + "loss": 1.2793, + "step": 12730 + }, + { + "epoch": 0.9792323667410199, + "learning_rate": 3.1913979349751932e-06, + "loss": 1.0122, + "step": 12731 + }, + { + "epoch": 0.9793092839012384, + "learning_rate": 3.1678100369110783e-06, + "loss": 1.0464, + "step": 12732 + }, + { + "epoch": 0.9793862010614568, + "learning_rate": 3.1443095404385945e-06, + "loss": 0.8579, + "step": 12733 + }, + { + "epoch": 0.9794631182216753, + "learning_rate": 3.120896446929644e-06, + "loss": 1.3535, + "step": 12734 + }, + { + "epoch": 0.9795400353818937, + "learning_rate": 3.0975707577517997e-06, + "loss": 0.8382, + "step": 12735 + }, + { + "epoch": 0.9796169525421121, + "learning_rate": 3.0743324742668054e-06, + "loss": 1.0523, + "step": 12736 + }, + { + "epoch": 0.9796938697023306, + "learning_rate": 3.0511815978314096e-06, + "loss": 1.3411, + "step": 12737 + }, + { + "epoch": 0.979770786862549, + "learning_rate": 3.0281181297976965e-06, + "loss": 1.0219, + "step": 12738 + }, + { + "epoch": 0.9798477040227674, + "learning_rate": 3.005142071512257e-06, + "loss": 0.8128, + "step": 12739 + }, + { + "epoch": 0.979924621182986, + "learning_rate": 2.98225342431685e-06, + "loss": 1.1951, + "step": 12740 + }, + { + "epoch": 0.9800015383432044, + "learning_rate": 2.95945218954774e-06, + "loss": 0.8736, + "step": 12741 + }, + { + "epoch": 0.9800784555034228, + "learning_rate": 2.9367383685363625e-06, + "loss": 1.0182, + "step": 12742 + }, + { + "epoch": 0.9801553726636413, + "learning_rate": 2.9141119626089898e-06, + "loss": 1.1931, + "step": 12743 + }, + { + "epoch": 0.9802322898238597, + "learning_rate": 2.8915729730868976e-06, + "loss": 1.1738, + "step": 12744 + }, + { + "epoch": 0.9803092069840782, + "learning_rate": 2.8691214012860343e-06, + "loss": 1.3247, + "step": 12745 + }, + { + "epoch": 0.9803861241442966, + "learning_rate": 2.8467572485175176e-06, + "loss": 1.1912, + "step": 12746 + }, + { + "epoch": 0.980463041304515, + "learning_rate": 2.824480516087136e-06, + "loss": 1.1418, + "step": 12747 + }, + { + "epoch": 0.9805399584647335, + "learning_rate": 2.802291205295682e-06, + "loss": 1.3108, + "step": 12748 + }, + { + "epoch": 0.9806168756249519, + "learning_rate": 2.780189317438786e-06, + "loss": 1.0052, + "step": 12749 + }, + { + "epoch": 0.9806937927851703, + "learning_rate": 2.758174853806916e-06, + "loss": 1.3417, + "step": 12750 + }, + { + "epoch": 0.9807707099453888, + "learning_rate": 2.7362478156857105e-06, + "loss": 1.1238, + "step": 12751 + }, + { + "epoch": 0.9808476271056072, + "learning_rate": 2.714408204355312e-06, + "loss": 1.1047, + "step": 12752 + }, + { + "epoch": 0.9809245442658258, + "learning_rate": 2.6926560210911997e-06, + "loss": 1.1214, + "step": 12753 + }, + { + "epoch": 0.9810014614260442, + "learning_rate": 2.670991267163192e-06, + "loss": 1.2733, + "step": 12754 + }, + { + "epoch": 0.9810783785862626, + "learning_rate": 2.6494139438366092e-06, + "loss": 1.2716, + "step": 12755 + }, + { + "epoch": 0.9811552957464811, + "learning_rate": 2.6279240523711113e-06, + "loss": 1.1705, + "step": 12756 + }, + { + "epoch": 0.9812322129066995, + "learning_rate": 2.6065215940216935e-06, + "loss": 0.8963, + "step": 12757 + }, + { + "epoch": 0.9813091300669179, + "learning_rate": 2.58520657003819e-06, + "loss": 0.8876, + "step": 12758 + }, + { + "epoch": 0.9813860472271364, + "learning_rate": 2.563978981664772e-06, + "loss": 1.2798, + "step": 12759 + }, + { + "epoch": 0.9814629643873548, + "learning_rate": 2.5428388301414474e-06, + "loss": 0.9096, + "step": 12760 + }, + { + "epoch": 0.9815398815475732, + "learning_rate": 2.5217861167022295e-06, + "loss": 0.9567, + "step": 12761 + }, + { + "epoch": 0.9816167987077917, + "learning_rate": 2.5008208425766345e-06, + "loss": 0.9947, + "step": 12762 + }, + { + "epoch": 0.9816937158680101, + "learning_rate": 2.479943008988683e-06, + "loss": 1.5009, + "step": 12763 + }, + { + "epoch": 0.9817706330282286, + "learning_rate": 2.4591526171575673e-06, + "loss": 1.259, + "step": 12764 + }, + { + "epoch": 0.981847550188447, + "learning_rate": 2.4384496682973158e-06, + "loss": 1.125, + "step": 12765 + }, + { + "epoch": 0.9819244673486655, + "learning_rate": 2.417834163616628e-06, + "loss": 1.0267, + "step": 12766 + }, + { + "epoch": 0.982001384508884, + "learning_rate": 2.3973061043193746e-06, + "loss": 1.3648, + "step": 12767 + }, + { + "epoch": 0.9820783016691024, + "learning_rate": 2.376865491604263e-06, + "loss": 1.1793, + "step": 12768 + }, + { + "epoch": 0.9821552188293208, + "learning_rate": 2.3565123266646725e-06, + "loss": 1.2886, + "step": 12769 + }, + { + "epoch": 0.9822321359895393, + "learning_rate": 2.3362466106891524e-06, + "loss": 1.2724, + "step": 12770 + }, + { + "epoch": 0.9823090531497577, + "learning_rate": 2.3160683448609223e-06, + "loss": 1.0261, + "step": 12771 + }, + { + "epoch": 0.9823859703099762, + "learning_rate": 2.295977530358373e-06, + "loss": 0.873, + "step": 12772 + }, + { + "epoch": 0.9824628874701946, + "learning_rate": 2.2759741683545665e-06, + "loss": 1.3855, + "step": 12773 + }, + { + "epoch": 0.982539804630413, + "learning_rate": 2.2560582600175684e-06, + "loss": 1.3186, + "step": 12774 + }, + { + "epoch": 0.9826167217906315, + "learning_rate": 2.2362298065101148e-06, + "loss": 1.333, + "step": 12775 + }, + { + "epoch": 0.9826936389508499, + "learning_rate": 2.21648880899028e-06, + "loss": 1.3524, + "step": 12776 + }, + { + "epoch": 0.9827705561110683, + "learning_rate": 2.196835268610642e-06, + "loss": 1.0582, + "step": 12777 + }, + { + "epoch": 0.9828474732712869, + "learning_rate": 2.1772691865186155e-06, + "loss": 1.4895, + "step": 12778 + }, + { + "epoch": 0.9829243904315053, + "learning_rate": 2.157790563856954e-06, + "loss": 1.1967, + "step": 12779 + }, + { + "epoch": 0.9830013075917237, + "learning_rate": 2.1383994017629137e-06, + "loss": 1.2466, + "step": 12780 + }, + { + "epoch": 0.9830782247519422, + "learning_rate": 2.1190957013685895e-06, + "loss": 1.1307, + "step": 12781 + }, + { + "epoch": 0.9831551419121606, + "learning_rate": 2.0998794638014128e-06, + "loss": 0.887, + "step": 12782 + }, + { + "epoch": 0.9832320590723791, + "learning_rate": 2.080750690183486e-06, + "loss": 1.3147, + "step": 12783 + }, + { + "epoch": 0.9833089762325975, + "learning_rate": 2.061709381631582e-06, + "loss": 0.7593, + "step": 12784 + }, + { + "epoch": 0.9833858933928159, + "learning_rate": 2.0427555392574793e-06, + "loss": 1.0692, + "step": 12785 + }, + { + "epoch": 0.9834628105530344, + "learning_rate": 2.023889164167958e-06, + "loss": 0.9097, + "step": 12786 + }, + { + "epoch": 0.9835397277132528, + "learning_rate": 2.0051102574648037e-06, + "loss": 1.2492, + "step": 12787 + }, + { + "epoch": 0.9836166448734712, + "learning_rate": 1.9864188202444734e-06, + "loss": 1.042, + "step": 12788 + }, + { + "epoch": 0.9836935620336897, + "learning_rate": 1.96781485359826e-06, + "loss": 0.9379, + "step": 12789 + }, + { + "epoch": 0.9837704791939081, + "learning_rate": 1.949298358612628e-06, + "loss": 1.2095, + "step": 12790 + }, + { + "epoch": 0.9838473963541267, + "learning_rate": 1.930869336368546e-06, + "loss": 1.1779, + "step": 12791 + }, + { + "epoch": 0.9839243135143451, + "learning_rate": 1.9125277879424864e-06, + "loss": 0.7955, + "step": 12792 + }, + { + "epoch": 0.9840012306745635, + "learning_rate": 1.8942737144050926e-06, + "loss": 1.16, + "step": 12793 + }, + { + "epoch": 0.984078147834782, + "learning_rate": 1.8761071168223454e-06, + "loss": 0.9418, + "step": 12794 + }, + { + "epoch": 0.9841550649950004, + "learning_rate": 1.8580279962548963e-06, + "loss": 1.0033, + "step": 12795 + }, + { + "epoch": 0.9842319821552188, + "learning_rate": 1.8400363537585674e-06, + "loss": 1.109, + "step": 12796 + }, + { + "epoch": 0.9843088993154373, + "learning_rate": 1.8221321903840182e-06, + "loss": 1.2688, + "step": 12797 + }, + { + "epoch": 0.9843858164756557, + "learning_rate": 1.8043155071764129e-06, + "loss": 1.2503, + "step": 12798 + }, + { + "epoch": 0.9844627336358741, + "learning_rate": 1.786586305176252e-06, + "loss": 0.8924, + "step": 12799 + }, + { + "epoch": 0.9845396507960926, + "learning_rate": 1.7689445854185416e-06, + "loss": 1.4539, + "step": 12800 + }, + { + "epoch": 0.984616567956311, + "learning_rate": 1.75139034893379e-06, + "loss": 1.037, + "step": 12801 + }, + { + "epoch": 0.9846934851165295, + "learning_rate": 1.7339235967466783e-06, + "loss": 1.0452, + "step": 12802 + }, + { + "epoch": 0.984770402276748, + "learning_rate": 1.7165443298772233e-06, + "loss": 0.9585, + "step": 12803 + }, + { + "epoch": 0.9848473194369664, + "learning_rate": 1.69925254934028e-06, + "loss": 1.0599, + "step": 12804 + }, + { + "epoch": 0.9849242365971849, + "learning_rate": 1.6820482561455409e-06, + "loss": 0.885, + "step": 12805 + }, + { + "epoch": 0.9850011537574033, + "learning_rate": 1.664931451297369e-06, + "loss": 1.106, + "step": 12806 + }, + { + "epoch": 0.9850780709176217, + "learning_rate": 1.6479021357954649e-06, + "loss": 1.3251, + "step": 12807 + }, + { + "epoch": 0.9851549880778402, + "learning_rate": 1.6309603106340331e-06, + "loss": 1.2526, + "step": 12808 + }, + { + "epoch": 0.9852319052380586, + "learning_rate": 1.6141059768024492e-06, + "loss": 1.0131, + "step": 12809 + }, + { + "epoch": 0.9853088223982771, + "learning_rate": 1.5973391352847588e-06, + "loss": 1.4628, + "step": 12810 + }, + { + "epoch": 0.9853857395584955, + "learning_rate": 1.5806597870600126e-06, + "loss": 1.3573, + "step": 12811 + }, + { + "epoch": 0.9854626567187139, + "learning_rate": 1.5640679331022644e-06, + "loss": 1.0887, + "step": 12812 + }, + { + "epoch": 0.9855395738789324, + "learning_rate": 1.547563574380073e-06, + "loss": 1.1108, + "step": 12813 + }, + { + "epoch": 0.9856164910391508, + "learning_rate": 1.5311467118575006e-06, + "loss": 1.5457, + "step": 12814 + }, + { + "epoch": 0.9856934081993692, + "learning_rate": 1.5148173464927806e-06, + "loss": 1.2486, + "step": 12815 + }, + { + "epoch": 0.9857703253595878, + "learning_rate": 1.4985754792394833e-06, + "loss": 1.2806, + "step": 12816 + }, + { + "epoch": 0.9858472425198062, + "learning_rate": 1.4824211110461839e-06, + "loss": 1.2897, + "step": 12817 + }, + { + "epoch": 0.9859241596800246, + "learning_rate": 1.4663542428561272e-06, + "loss": 1.2803, + "step": 12818 + }, + { + "epoch": 0.9860010768402431, + "learning_rate": 1.45037487560723e-06, + "loss": 1.1977, + "step": 12819 + }, + { + "epoch": 0.9860779940004615, + "learning_rate": 1.4344830102327456e-06, + "loss": 1.0733, + "step": 12820 + }, + { + "epoch": 0.98615491116068, + "learning_rate": 1.4186786476604318e-06, + "loss": 0.967, + "step": 12821 + }, + { + "epoch": 0.9862318283208984, + "learning_rate": 1.402961788813384e-06, + "loss": 0.895, + "step": 12822 + }, + { + "epoch": 0.9863087454811168, + "learning_rate": 1.3873324346093673e-06, + "loss": 1.3207, + "step": 12823 + }, + { + "epoch": 0.9863856626413353, + "learning_rate": 1.3717905859604862e-06, + "loss": 1.4154, + "step": 12824 + }, + { + "epoch": 0.9864625798015537, + "learning_rate": 1.3563362437748472e-06, + "loss": 1.1526, + "step": 12825 + }, + { + "epoch": 0.9865394969617721, + "learning_rate": 1.3409694089543956e-06, + "loss": 1.3104, + "step": 12826 + }, + { + "epoch": 0.9866164141219906, + "learning_rate": 1.3256900823967465e-06, + "loss": 0.8908, + "step": 12827 + }, + { + "epoch": 0.986693331282209, + "learning_rate": 1.3104982649938534e-06, + "loss": 1.2367, + "step": 12828 + }, + { + "epoch": 0.9867702484424276, + "learning_rate": 1.2953939576330065e-06, + "loss": 1.1903, + "step": 12829 + }, + { + "epoch": 0.986847165602646, + "learning_rate": 1.2803771611958337e-06, + "loss": 1.1108, + "step": 12830 + }, + { + "epoch": 0.9869240827628644, + "learning_rate": 1.2654478765594668e-06, + "loss": 1.2312, + "step": 12831 + }, + { + "epoch": 0.9870009999230829, + "learning_rate": 1.250606104595542e-06, + "loss": 1.1009, + "step": 12832 + }, + { + "epoch": 0.9870779170833013, + "learning_rate": 1.235851846170699e-06, + "loss": 0.9126, + "step": 12833 + }, + { + "epoch": 0.9871548342435197, + "learning_rate": 1.221185102146416e-06, + "loss": 1.1545, + "step": 12834 + }, + { + "epoch": 0.9872317514037382, + "learning_rate": 1.2066058733790075e-06, + "loss": 1.1847, + "step": 12835 + }, + { + "epoch": 0.9873086685639566, + "learning_rate": 1.1921141607201257e-06, + "loss": 1.2383, + "step": 12836 + }, + { + "epoch": 0.9873855857241751, + "learning_rate": 1.1777099650154277e-06, + "loss": 1.4643, + "step": 12837 + }, + { + "epoch": 0.9874625028843935, + "learning_rate": 1.1633932871065733e-06, + "loss": 1.1681, + "step": 12838 + }, + { + "epoch": 0.9875394200446119, + "learning_rate": 1.149164127829061e-06, + "loss": 1.3279, + "step": 12839 + }, + { + "epoch": 0.9876163372048304, + "learning_rate": 1.1350224880138926e-06, + "loss": 1.1594, + "step": 12840 + }, + { + "epoch": 0.9876932543650488, + "learning_rate": 1.1209683684869075e-06, + "loss": 1.2254, + "step": 12841 + }, + { + "epoch": 0.9877701715252672, + "learning_rate": 1.107001770068783e-06, + "loss": 1.0006, + "step": 12842 + }, + { + "epoch": 0.9878470886854858, + "learning_rate": 1.0931226935746996e-06, + "loss": 1.1089, + "step": 12843 + }, + { + "epoch": 0.9879240058457042, + "learning_rate": 1.0793311398155093e-06, + "loss": 1.3209, + "step": 12844 + }, + { + "epoch": 0.9880009230059226, + "learning_rate": 1.0656271095962345e-06, + "loss": 0.9682, + "step": 12845 + }, + { + "epoch": 0.9880778401661411, + "learning_rate": 1.0520106037170685e-06, + "loss": 1.1248, + "step": 12846 + }, + { + "epoch": 0.9881547573263595, + "learning_rate": 1.0384816229732086e-06, + "loss": 1.2041, + "step": 12847 + }, + { + "epoch": 0.988231674486578, + "learning_rate": 1.0250401681546896e-06, + "loss": 1.0256, + "step": 12848 + }, + { + "epoch": 0.9883085916467964, + "learning_rate": 1.0116862400462167e-06, + "loss": 0.9725, + "step": 12849 + }, + { + "epoch": 0.9883855088070148, + "learning_rate": 9.984198394275002e-07, + "loss": 0.7537, + "step": 12850 + }, + { + "epoch": 0.9884624259672333, + "learning_rate": 9.852409670734197e-07, + "loss": 1.165, + "step": 12851 + }, + { + "epoch": 0.9885393431274517, + "learning_rate": 9.721496237531934e-07, + "loss": 1.3722, + "step": 12852 + }, + { + "epoch": 0.9886162602876701, + "learning_rate": 9.59145810231543e-07, + "loss": 0.8697, + "step": 12853 + }, + { + "epoch": 0.9886931774478886, + "learning_rate": 9.46229527267528e-07, + "loss": 1.3825, + "step": 12854 + }, + { + "epoch": 0.988770094608107, + "learning_rate": 9.33400775615545e-07, + "loss": 1.2608, + "step": 12855 + }, + { + "epoch": 0.9888470117683256, + "learning_rate": 9.206595560244946e-07, + "loss": 0.8974, + "step": 12856 + }, + { + "epoch": 0.988923928928544, + "learning_rate": 9.080058692384485e-07, + "loss": 0.872, + "step": 12857 + }, + { + "epoch": 0.9890008460887624, + "learning_rate": 8.954397159963157e-07, + "loss": 1.4197, + "step": 12858 + }, + { + "epoch": 0.9890777632489809, + "learning_rate": 8.829610970318425e-07, + "loss": 1.2919, + "step": 12859 + }, + { + "epoch": 0.9891546804091993, + "learning_rate": 8.705700130734462e-07, + "loss": 1.3476, + "step": 12860 + }, + { + "epoch": 0.9892315975694177, + "learning_rate": 8.582664648448813e-07, + "loss": 1.3725, + "step": 12861 + }, + { + "epoch": 0.9893085147296362, + "learning_rate": 8.460504530645729e-07, + "loss": 0.9024, + "step": 12862 + }, + { + "epoch": 0.9893854318898546, + "learning_rate": 8.339219784457841e-07, + "loss": 0.93, + "step": 12863 + }, + { + "epoch": 0.989462349050073, + "learning_rate": 8.218810416966149e-07, + "loss": 1.0988, + "step": 12864 + }, + { + "epoch": 0.9895392662102915, + "learning_rate": 8.099276435201697e-07, + "loss": 1.0911, + "step": 12865 + }, + { + "epoch": 0.9896161833705099, + "learning_rate": 7.980617846145566e-07, + "loss": 0.839, + "step": 12866 + }, + { + "epoch": 0.9896931005307285, + "learning_rate": 7.862834656723882e-07, + "loss": 1.2614, + "step": 12867 + }, + { + "epoch": 0.9897700176909469, + "learning_rate": 7.74592687381781e-07, + "loss": 1.2021, + "step": 12868 + }, + { + "epoch": 0.9898469348511653, + "learning_rate": 7.629894504250223e-07, + "loss": 0.8893, + "step": 12869 + }, + { + "epoch": 0.9899238520113838, + "learning_rate": 7.514737554797368e-07, + "loss": 1.1556, + "step": 12870 + }, + { + "epoch": 0.9900007691716022, + "learning_rate": 7.400456032185532e-07, + "loss": 0.9451, + "step": 12871 + }, + { + "epoch": 0.9900776863318206, + "learning_rate": 7.287049943084379e-07, + "loss": 1.2126, + "step": 12872 + }, + { + "epoch": 0.9901546034920391, + "learning_rate": 7.174519294118609e-07, + "loss": 0.9153, + "step": 12873 + }, + { + "epoch": 0.9902315206522575, + "learning_rate": 7.062864091857967e-07, + "loss": 1.0964, + "step": 12874 + }, + { + "epoch": 0.990308437812476, + "learning_rate": 6.952084342820574e-07, + "loss": 1.3543, + "step": 12875 + }, + { + "epoch": 0.9903853549726944, + "learning_rate": 6.84218005347792e-07, + "loss": 1.212, + "step": 12876 + }, + { + "epoch": 0.9904622721329128, + "learning_rate": 6.733151230244872e-07, + "loss": 0.9412, + "step": 12877 + }, + { + "epoch": 0.9905391892931313, + "learning_rate": 6.624997879488004e-07, + "loss": 1.377, + "step": 12878 + }, + { + "epoch": 0.9906161064533497, + "learning_rate": 6.51772000752393e-07, + "loss": 0.9764, + "step": 12879 + }, + { + "epoch": 0.9906930236135681, + "learning_rate": 6.411317620615975e-07, + "loss": 1.0797, + "step": 12880 + }, + { + "epoch": 0.9907699407737867, + "learning_rate": 6.305790724977501e-07, + "loss": 0.9857, + "step": 12881 + }, + { + "epoch": 0.9908468579340051, + "learning_rate": 6.201139326770245e-07, + "loss": 0.8654, + "step": 12882 + }, + { + "epoch": 0.9909237750942235, + "learning_rate": 6.09736343210432e-07, + "loss": 1.3121, + "step": 12883 + }, + { + "epoch": 0.991000692254442, + "learning_rate": 5.994463047039878e-07, + "loss": 1.084, + "step": 12884 + }, + { + "epoch": 0.9910776094146604, + "learning_rate": 5.892438177583781e-07, + "loss": 1.2455, + "step": 12885 + }, + { + "epoch": 0.9911545265748789, + "learning_rate": 5.791288829696262e-07, + "loss": 1.2246, + "step": 12886 + }, + { + "epoch": 0.9912314437350973, + "learning_rate": 5.691015009280931e-07, + "loss": 1.2855, + "step": 12887 + }, + { + "epoch": 0.9913083608953157, + "learning_rate": 5.59161672219477e-07, + "loss": 1.3215, + "step": 12888 + }, + { + "epoch": 0.9913852780555342, + "learning_rate": 5.493093974239804e-07, + "loss": 1.4003, + "step": 12889 + }, + { + "epoch": 0.9914621952157526, + "learning_rate": 5.39544677117143e-07, + "loss": 1.0938, + "step": 12890 + }, + { + "epoch": 0.991539112375971, + "learning_rate": 5.298675118688423e-07, + "loss": 1.211, + "step": 12891 + }, + { + "epoch": 0.9916160295361895, + "learning_rate": 5.202779022444592e-07, + "loss": 1.3325, + "step": 12892 + }, + { + "epoch": 0.991692946696408, + "learning_rate": 5.107758488035464e-07, + "loss": 1.0643, + "step": 12893 + }, + { + "epoch": 0.9917698638566265, + "learning_rate": 5.013613521013261e-07, + "loss": 1.3499, + "step": 12894 + }, + { + "epoch": 0.9918467810168449, + "learning_rate": 4.920344126873588e-07, + "loss": 1.4176, + "step": 12895 + }, + { + "epoch": 0.9919236981770633, + "learning_rate": 4.827950311062091e-07, + "loss": 0.9883, + "step": 12896 + }, + { + "epoch": 0.9920006153372818, + "learning_rate": 4.736432078974451e-07, + "loss": 1.3637, + "step": 12897 + }, + { + "epoch": 0.9920775324975002, + "learning_rate": 4.6457894359530627e-07, + "loss": 1.0593, + "step": 12898 + }, + { + "epoch": 0.9921544496577186, + "learning_rate": 4.55602238729369e-07, + "loss": 1.0567, + "step": 12899 + }, + { + "epoch": 0.9922313668179371, + "learning_rate": 4.46713093823381e-07, + "loss": 1.1161, + "step": 12900 + }, + { + "epoch": 0.9923082839781555, + "learning_rate": 4.3791150939676003e-07, + "loss": 1.0579, + "step": 12901 + }, + { + "epoch": 0.9923852011383739, + "learning_rate": 4.291974859632619e-07, + "loss": 1.2888, + "step": 12902 + }, + { + "epoch": 0.9924621182985924, + "learning_rate": 4.205710240318128e-07, + "loss": 1.2811, + "step": 12903 + }, + { + "epoch": 0.9925390354588108, + "learning_rate": 4.1203212410600987e-07, + "loss": 1.1893, + "step": 12904 + }, + { + "epoch": 0.9926159526190294, + "learning_rate": 4.035807866844543e-07, + "loss": 1.189, + "step": 12905 + }, + { + "epoch": 0.9926928697792478, + "learning_rate": 3.952170122605847e-07, + "loss": 1.0412, + "step": 12906 + }, + { + "epoch": 0.9927697869394662, + "learning_rate": 3.8694080132301023e-07, + "loss": 0.9806, + "step": 12907 + }, + { + "epoch": 0.9928467040996847, + "learning_rate": 3.787521543546779e-07, + "loss": 0.9748, + "step": 12908 + }, + { + "epoch": 0.9929236212599031, + "learning_rate": 3.706510718338718e-07, + "loss": 1.4697, + "step": 12909 + }, + { + "epoch": 0.9930005384201215, + "learning_rate": 3.626375542337135e-07, + "loss": 1.1323, + "step": 12910 + }, + { + "epoch": 0.99307745558034, + "learning_rate": 3.5471160202199535e-07, + "loss": 1.2568, + "step": 12911 + }, + { + "epoch": 0.9931543727405584, + "learning_rate": 3.46873215661514e-07, + "loss": 1.2582, + "step": 12912 + }, + { + "epoch": 0.9932312899007769, + "learning_rate": 3.3912239561006975e-07, + "loss": 1.1581, + "step": 12913 + }, + { + "epoch": 0.9933082070609953, + "learning_rate": 3.314591423201341e-07, + "loss": 1.2994, + "step": 12914 + }, + { + "epoch": 0.9933851242212137, + "learning_rate": 3.238834562393489e-07, + "loss": 1.1417, + "step": 12915 + }, + { + "epoch": 0.9934620413814322, + "learning_rate": 3.163953378098605e-07, + "loss": 1.2175, + "step": 12916 + }, + { + "epoch": 0.9935389585416506, + "learning_rate": 3.0899478746881925e-07, + "loss": 1.0339, + "step": 12917 + }, + { + "epoch": 0.993615875701869, + "learning_rate": 3.016818056487125e-07, + "loss": 1.2117, + "step": 12918 + }, + { + "epoch": 0.9936927928620876, + "learning_rate": 2.944563927763655e-07, + "loss": 1.1594, + "step": 12919 + }, + { + "epoch": 0.993769710022306, + "learning_rate": 2.873185492734409e-07, + "loss": 1.1495, + "step": 12920 + }, + { + "epoch": 0.9938466271825244, + "learning_rate": 2.802682755571051e-07, + "loss": 1.4044, + "step": 12921 + }, + { + "epoch": 0.9939235443427429, + "learning_rate": 2.7330557203886217e-07, + "loss": 1.4963, + "step": 12922 + }, + { + "epoch": 0.9940004615029613, + "learning_rate": 2.664304391253869e-07, + "loss": 0.805, + "step": 12923 + }, + { + "epoch": 0.9940773786631798, + "learning_rate": 2.5964287721785827e-07, + "loss": 1.0027, + "step": 12924 + }, + { + "epoch": 0.9941542958233982, + "learning_rate": 2.5294288671279255e-07, + "loss": 0.9452, + "step": 12925 + }, + { + "epoch": 0.9942312129836166, + "learning_rate": 2.4633046800154325e-07, + "loss": 0.9951, + "step": 12926 + }, + { + "epoch": 0.9943081301438351, + "learning_rate": 2.398056214699684e-07, + "loss": 1.0305, + "step": 12927 + }, + { + "epoch": 0.9943850473040535, + "learning_rate": 2.3336834749909663e-07, + "loss": 1.3752, + "step": 12928 + }, + { + "epoch": 0.9944619644642719, + "learning_rate": 2.270186464649604e-07, + "loss": 1.1421, + "step": 12929 + }, + { + "epoch": 0.9945388816244904, + "learning_rate": 2.2075651873809667e-07, + "loss": 1.1417, + "step": 12930 + }, + { + "epoch": 0.9946157987847088, + "learning_rate": 2.1458196468437942e-07, + "loss": 1.2761, + "step": 12931 + }, + { + "epoch": 0.9946927159449274, + "learning_rate": 2.0849498466435357e-07, + "loss": 1.4439, + "step": 12932 + }, + { + "epoch": 0.9947696331051458, + "learning_rate": 2.02495579033235e-07, + "loss": 1.2078, + "step": 12933 + }, + { + "epoch": 0.9948465502653642, + "learning_rate": 1.965837481414101e-07, + "loss": 1.1953, + "step": 12934 + }, + { + "epoch": 0.9949234674255827, + "learning_rate": 1.9075949233426925e-07, + "loss": 1.4175, + "step": 12935 + }, + { + "epoch": 0.9950003845858011, + "learning_rate": 1.850228119515407e-07, + "loss": 1.3363, + "step": 12936 + }, + { + "epoch": 0.9950773017460195, + "learning_rate": 1.793737073286228e-07, + "loss": 0.9397, + "step": 12937 + }, + { + "epoch": 0.995154218906238, + "learning_rate": 1.7381217879508527e-07, + "loss": 1.2953, + "step": 12938 + }, + { + "epoch": 0.9952311360664564, + "learning_rate": 1.6833822667566834e-07, + "loss": 1.5886, + "step": 12939 + }, + { + "epoch": 0.9953080532266748, + "learning_rate": 1.6295185129011625e-07, + "loss": 1.1365, + "step": 12940 + }, + { + "epoch": 0.9953849703868933, + "learning_rate": 1.5765305295284416e-07, + "loss": 1.2621, + "step": 12941 + }, + { + "epoch": 0.9954618875471117, + "learning_rate": 1.5244183197327122e-07, + "loss": 1.1917, + "step": 12942 + }, + { + "epoch": 0.9955388047073302, + "learning_rate": 1.473181886558206e-07, + "loss": 1.0094, + "step": 12943 + }, + { + "epoch": 0.9956157218675487, + "learning_rate": 1.4228212329958634e-07, + "loss": 1.0105, + "step": 12944 + }, + { + "epoch": 0.9956926390277671, + "learning_rate": 1.3733363619850003e-07, + "loss": 1.1703, + "step": 12945 + }, + { + "epoch": 0.9957695561879856, + "learning_rate": 1.3247272764166373e-07, + "loss": 0.9972, + "step": 12946 + }, + { + "epoch": 0.995846473348204, + "learning_rate": 1.27699397913017e-07, + "loss": 1.2288, + "step": 12947 + }, + { + "epoch": 0.9959233905084224, + "learning_rate": 1.2301364729100374e-07, + "loss": 1.133, + "step": 12948 + }, + { + "epoch": 0.9960003076686409, + "learning_rate": 1.1841547604940494e-07, + "loss": 1.0169, + "step": 12949 + }, + { + "epoch": 0.9960772248288593, + "learning_rate": 1.1390488445650604e-07, + "loss": 1.1699, + "step": 12950 + }, + { + "epoch": 0.9961541419890778, + "learning_rate": 1.09481872776096e-07, + "loss": 1.3666, + "step": 12951 + }, + { + "epoch": 0.9962310591492962, + "learning_rate": 1.0514644126596862e-07, + "loss": 1.343, + "step": 12952 + }, + { + "epoch": 0.9963079763095146, + "learning_rate": 1.008985901795878e-07, + "loss": 1.0168, + "step": 12953 + }, + { + "epoch": 0.9963848934697331, + "learning_rate": 9.673831976492187e-08, + "loss": 0.9452, + "step": 12954 + }, + { + "epoch": 0.9964618106299515, + "learning_rate": 9.26656302647766e-08, + "loss": 1.0083, + "step": 12955 + }, + { + "epoch": 0.9965387277901699, + "learning_rate": 8.868052191712828e-08, + "loss": 1.3389, + "step": 12956 + }, + { + "epoch": 0.9966156449503885, + "learning_rate": 8.478299495462416e-08, + "loss": 1.0279, + "step": 12957 + }, + { + "epoch": 0.9966925621106069, + "learning_rate": 8.097304960474894e-08, + "loss": 0.9121, + "step": 12958 + }, + { + "epoch": 0.9967694792708254, + "learning_rate": 7.72506860899913e-08, + "loss": 1.2067, + "step": 12959 + }, + { + "epoch": 0.9968463964310438, + "learning_rate": 7.361590462784395e-08, + "loss": 1.0315, + "step": 12960 + }, + { + "epoch": 0.9969233135912622, + "learning_rate": 7.00687054304705e-08, + "loss": 1.1543, + "step": 12961 + }, + { + "epoch": 0.9970002307514807, + "learning_rate": 6.660908870487203e-08, + "loss": 1.5518, + "step": 12962 + }, + { + "epoch": 0.9970771479116991, + "learning_rate": 6.323705465322016e-08, + "loss": 1.4173, + "step": 12963 + }, + { + "epoch": 0.9971540650719175, + "learning_rate": 5.995260347235742e-08, + "loss": 1.4398, + "step": 12964 + }, + { + "epoch": 0.997230982232136, + "learning_rate": 5.675573535413037e-08, + "loss": 1.5524, + "step": 12965 + }, + { + "epoch": 0.9973078993923544, + "learning_rate": 5.3646450485056455e-08, + "loss": 1.7572, + "step": 12966 + }, + { + "epoch": 0.9973848165525728, + "learning_rate": 5.062474904665715e-08, + "loss": 1.2099, + "step": 12967 + }, + { + "epoch": 0.9974617337127913, + "learning_rate": 4.7690631215624444e-08, + "loss": 1.2903, + "step": 12968 + }, + { + "epoch": 0.9975386508730097, + "learning_rate": 4.484409716315474e-08, + "loss": 0.8052, + "step": 12969 + }, + { + "epoch": 0.9976155680332283, + "learning_rate": 4.208514705544841e-08, + "loss": 1.1351, + "step": 12970 + }, + { + "epoch": 0.9976924851934467, + "learning_rate": 3.941378105354332e-08, + "loss": 1.1921, + "step": 12971 + }, + { + "epoch": 0.9977694023536651, + "learning_rate": 3.682999931364783e-08, + "loss": 1.4235, + "step": 12972 + }, + { + "epoch": 0.9978463195138836, + "learning_rate": 3.43338019863082e-08, + "loss": 1.1581, + "step": 12973 + }, + { + "epoch": 0.997923236674102, + "learning_rate": 3.192518921740772e-08, + "loss": 0.9665, + "step": 12974 + }, + { + "epoch": 0.9980001538343204, + "learning_rate": 2.960416114783371e-08, + "loss": 1.4974, + "step": 12975 + }, + { + "epoch": 0.9980770709945389, + "learning_rate": 2.7370717912811316e-08, + "loss": 1.2694, + "step": 12976 + }, + { + "epoch": 0.9981539881547573, + "learning_rate": 2.522485964273624e-08, + "loss": 1.3921, + "step": 12977 + }, + { + "epoch": 0.9982309053149758, + "learning_rate": 2.316658646317471e-08, + "loss": 1.1122, + "step": 12978 + }, + { + "epoch": 0.9983078224751942, + "learning_rate": 2.119589849403081e-08, + "loss": 1.2248, + "step": 12979 + }, + { + "epoch": 0.9983847396354126, + "learning_rate": 1.9312795850545682e-08, + "loss": 1.2476, + "step": 12980 + }, + { + "epoch": 0.9984616567956311, + "learning_rate": 1.7517278642631416e-08, + "loss": 0.8254, + "step": 12981 + }, + { + "epoch": 0.9985385739558496, + "learning_rate": 1.5809346975204087e-08, + "loss": 0.9032, + "step": 12982 + }, + { + "epoch": 0.998615491116068, + "learning_rate": 1.418900094768416e-08, + "loss": 1.2986, + "step": 12983 + }, + { + "epoch": 0.9986924082762865, + "learning_rate": 1.2656240655162243e-08, + "loss": 1.372, + "step": 12984 + }, + { + "epoch": 0.9987693254365049, + "learning_rate": 1.1211066186733732e-08, + "loss": 0.941, + "step": 12985 + }, + { + "epoch": 0.9988462425967233, + "learning_rate": 9.85347762699762e-09, + "loss": 1.1172, + "step": 12986 + }, + { + "epoch": 0.9989231597569418, + "learning_rate": 8.583475055223832e-09, + "loss": 1.1312, + "step": 12987 + }, + { + "epoch": 0.9990000769171602, + "learning_rate": 7.4010585455197566e-09, + "loss": 1.2059, + "step": 12988 + }, + { + "epoch": 0.9990769940773787, + "learning_rate": 6.306228166830241e-09, + "loss": 1.466, + "step": 12989 + }, + { + "epoch": 0.9991539112375971, + "learning_rate": 5.2989839832706665e-09, + "loss": 1.1152, + "step": 12990 + }, + { + "epoch": 0.9992308283978155, + "learning_rate": 4.379326053460808e-09, + "loss": 1.209, + "step": 12991 + }, + { + "epoch": 0.999307745558034, + "learning_rate": 3.547254431357505e-09, + "loss": 1.1204, + "step": 12992 + }, + { + "epoch": 0.9993846627182524, + "learning_rate": 2.8027691652554587e-09, + "loss": 1.1816, + "step": 12993 + }, + { + "epoch": 0.9994615798784708, + "learning_rate": 2.145870298786434e-09, + "loss": 1.1467, + "step": 12994 + }, + { + "epoch": 0.9995384970386894, + "learning_rate": 1.5765578702531258e-09, + "loss": 1.1449, + "step": 12995 + }, + { + "epoch": 0.9996154141989078, + "learning_rate": 1.0948319129622242e-09, + "loss": 1.2414, + "step": 12996 + }, + { + "epoch": 0.9996923313591263, + "learning_rate": 7.006924548913496e-10, + "loss": 0.8868, + "step": 12997 + }, + { + "epoch": 0.9997692485193447, + "learning_rate": 3.9413951935518555e-10, + "loss": 1.1303, + "step": 12998 + }, + { + "epoch": 0.9998461656795631, + "learning_rate": 1.7517312400627817e-10, + "loss": 1.3999, + "step": 12999 + }, + { + "epoch": 0.9999230828397816, + "learning_rate": 4.3793281667703356e-11, + "loss": 1.2557, + "step": 13000 + } + ], + "logging_steps": 1, + "max_steps": 13001, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 100, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 3.419903693543424e+16, + "train_batch_size": 4, + "trial_name": null, + "trial_params": null +}