diff --git "a/trainer_state.json" "b/trainer_state.json" --- "a/trainer_state.json" +++ "b/trainer_state.json" @@ -1,14893 +1,6432 @@ { "best_metric": null, "best_model_checkpoint": null, - "epoch": 7.678410117434508, + "epoch": 7.0, "eval_steps": 500, - "global_step": 8500, + "global_step": 3661, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { - "epoch": 0.0036133694670280035, - "grad_norm": 0.294921875, - "learning_rate": 1.3483146067415729e-05, - "loss": 1.3019, + "epoch": 0.0076481835564053535, + "grad_norm": 0.291015625, + "learning_rate": 3.243243243243243e-05, + "loss": 1.3011, "step": 4 }, { - "epoch": 0.007226738934056007, - "grad_norm": 0.2890625, - "learning_rate": 2.6966292134831458e-05, - "loss": 1.3449, + "epoch": 0.015296367112810707, + "grad_norm": 0.26171875, + "learning_rate": 6.486486486486486e-05, + "loss": 1.3104, "step": 8 }, { - "epoch": 0.01084010840108401, - "grad_norm": 0.23046875, - "learning_rate": 4.0449438202247185e-05, - "loss": 1.2588, + "epoch": 0.022944550669216062, + "grad_norm": 0.2060546875, + "learning_rate": 9.72972972972973e-05, + "loss": 1.2309, "step": 12 }, { - "epoch": 0.014453477868112014, - "grad_norm": 0.1962890625, - "learning_rate": 5.3932584269662916e-05, - "loss": 1.2878, + "epoch": 0.030592734225621414, + "grad_norm": 0.2392578125, + "learning_rate": 0.00012972972972972972, + "loss": 1.2051, "step": 16 }, { - "epoch": 0.018066847335140017, - "grad_norm": 0.240234375, - "learning_rate": 6.741573033707865e-05, - "loss": 1.2136, + "epoch": 0.03824091778202677, + "grad_norm": 0.169921875, + "learning_rate": 0.00016216216216216215, + "loss": 1.1622, "step": 20 }, { - "epoch": 0.02168021680216802, - "grad_norm": 0.212890625, - "learning_rate": 8.089887640449437e-05, - "loss": 1.2275, + "epoch": 0.045889101338432124, + "grad_norm": 0.1748046875, + "learning_rate": 0.0001945945945945946, + "loss": 1.1749, "step": 24 }, { - "epoch": 0.025293586269196026, - "grad_norm": 0.2041015625, - "learning_rate": 9.438202247191011e-05, - "loss": 1.2667, + "epoch": 0.05353728489483748, + "grad_norm": 0.154296875, + "learning_rate": 0.00022702702702702703, + "loss": 1.149, "step": 28 }, { - "epoch": 0.028906955736224028, - "grad_norm": 0.1962890625, - "learning_rate": 0.00010786516853932583, - "loss": 1.1905, + "epoch": 0.06118546845124283, + "grad_norm": 0.1767578125, + "learning_rate": 0.00025945945945945944, + "loss": 1.1455, "step": 32 }, { - "epoch": 0.032520325203252036, - "grad_norm": 0.185546875, - "learning_rate": 0.00012134831460674156, - "loss": 1.2345, + "epoch": 0.06883365200764818, + "grad_norm": 0.1884765625, + "learning_rate": 0.0002918918918918919, + "loss": 1.1358, "step": 36 }, { - "epoch": 0.036133694670280034, - "grad_norm": 0.189453125, - "learning_rate": 0.0001348314606741573, - "loss": 1.2052, + "epoch": 0.07648183556405354, + "grad_norm": 0.181640625, + "learning_rate": 0.00029999949274434724, + "loss": 1.1201, "step": 40 }, { - "epoch": 0.03974706413730804, - "grad_norm": 0.1845703125, - "learning_rate": 0.00014831460674157303, - "loss": 1.146, + "epoch": 0.0841300191204589, + "grad_norm": 0.1708984375, + "learning_rate": 0.0002999972382816974, + "loss": 1.0549, "step": 44 }, { - "epoch": 0.04336043360433604, - "grad_norm": 0.2333984375, - "learning_rate": 0.00016179775280898874, - "loss": 1.1099, + "epoch": 0.09177820267686425, + "grad_norm": 0.19921875, + "learning_rate": 0.0002999931802773903, + "loss": 1.0946, "step": 48 }, { - "epoch": 0.04697380307136405, - "grad_norm": 0.19921875, - "learning_rate": 0.00017528089887640448, - "loss": 1.1519, + "epoch": 0.0994263862332696, + "grad_norm": 0.1826171875, + "learning_rate": 0.00029998731878021884, + "loss": 1.0841, "step": 52 }, { - "epoch": 0.05058717253839205, - "grad_norm": 0.19921875, - "learning_rate": 0.00018876404494382021, - "loss": 1.1701, + "epoch": 0.10707456978967496, + "grad_norm": 0.2001953125, + "learning_rate": 0.00029997965386066057, + "loss": 1.0904, "step": 56 }, { - "epoch": 0.05420054200542006, - "grad_norm": 0.19921875, - "learning_rate": 0.00020224719101123593, - "loss": 1.1546, + "epoch": 0.1147227533460803, + "grad_norm": 0.1826171875, + "learning_rate": 0.0002999701856108772, + "loss": 1.0673, "step": 60 }, { - "epoch": 0.057813911472448055, - "grad_norm": 0.2255859375, - "learning_rate": 0.00021573033707865166, - "loss": 1.1587, + "epoch": 0.12237093690248566, + "grad_norm": 0.1826171875, + "learning_rate": 0.00029995891414471334, + "loss": 1.0211, "step": 64 }, { - "epoch": 0.06142728093947606, - "grad_norm": 0.203125, - "learning_rate": 0.0002292134831460674, - "loss": 1.1414, + "epoch": 0.13001912045889102, + "grad_norm": 0.1865234375, + "learning_rate": 0.0002999458395976953, + "loss": 1.0497, "step": 68 }, { - "epoch": 0.06504065040650407, - "grad_norm": 0.21875, - "learning_rate": 0.0002426966292134831, - "loss": 1.1381, + "epoch": 0.13766730401529637, + "grad_norm": 0.208984375, + "learning_rate": 0.0002999309621270293, + "loss": 1.0711, "step": 72 }, { - "epoch": 0.06865401987353206, - "grad_norm": 0.2041015625, - "learning_rate": 0.0002561797752808988, - "loss": 1.121, + "epoch": 0.14531548757170173, + "grad_norm": 0.1923828125, + "learning_rate": 0.00029991428191159935, + "loss": 1.047, "step": 76 }, { - "epoch": 0.07226738934056007, - "grad_norm": 0.20703125, - "learning_rate": 0.0002696629213483146, - "loss": 1.053, + "epoch": 0.15296367112810708, + "grad_norm": 0.1962890625, + "learning_rate": 0.00029989579915196574, + "loss": 1.0442, "step": 80 }, { - "epoch": 0.07588075880758807, - "grad_norm": 0.201171875, - "learning_rate": 0.0002831460674157303, - "loss": 1.1037, + "epoch": 0.16061185468451242, + "grad_norm": 0.197265625, + "learning_rate": 0.000299875514070362, + "loss": 1.0197, "step": 84 }, { - "epoch": 0.07949412827461608, - "grad_norm": 0.205078125, - "learning_rate": 0.00029662921348314606, - "loss": 1.127, + "epoch": 0.1682600382409178, + "grad_norm": 0.2080078125, + "learning_rate": 0.00029985342691069255, + "loss": 1.0185, "step": 88 }, { - "epoch": 0.08310749774164408, - "grad_norm": 0.205078125, - "learning_rate": 0.0002999999133234806, - "loss": 1.0572, + "epoch": 0.17590822179732313, + "grad_norm": 0.2177734375, + "learning_rate": 0.0002998295379385297, + "loss": 1.0168, "step": 92 }, { - "epoch": 0.08672086720867209, - "grad_norm": 0.19140625, - "learning_rate": 0.00029999952809470757, - "loss": 1.1133, + "epoch": 0.1835564053537285, + "grad_norm": 0.2099609375, + "learning_rate": 0.00029980384744111047, + "loss": 1.0211, "step": 96 }, { - "epoch": 0.09033423667570009, - "grad_norm": 0.189453125, - "learning_rate": 0.00029999883468374714, - "loss": 1.1132, + "epoch": 0.19120458891013384, + "grad_norm": 0.208984375, + "learning_rate": 0.0002997763557273331, + "loss": 1.0178, "step": 100 }, { - "epoch": 0.0939476061427281, - "grad_norm": 0.20703125, - "learning_rate": 0.000299997833092024, - "loss": 1.0945, + "epoch": 0.1988527724665392, + "grad_norm": 0.2060546875, + "learning_rate": 0.0002997470631277533, + "loss": 0.9871, "step": 104 }, { - "epoch": 0.0975609756097561, - "grad_norm": 0.21875, - "learning_rate": 0.00029999652332159593, - "loss": 1.1265, + "epoch": 0.20650095602294455, + "grad_norm": 0.2314453125, + "learning_rate": 0.0002997159699945804, + "loss": 1.0197, "step": 108 }, { - "epoch": 0.1011743450767841, - "grad_norm": 0.2099609375, - "learning_rate": 0.00029999490537515396, - "loss": 1.0732, + "epoch": 0.21414913957934992, + "grad_norm": 0.240234375, + "learning_rate": 0.0002996830767016731, + "loss": 1.0, "step": 112 }, { - "epoch": 0.10478771454381211, - "grad_norm": 0.1953125, - "learning_rate": 0.00029999297925602226, - "loss": 1.1023, + "epoch": 0.22179732313575526, + "grad_norm": 0.236328125, + "learning_rate": 0.0002996483836445347, + "loss": 0.9885, "step": 116 }, { - "epoch": 0.10840108401084012, - "grad_norm": 0.20703125, - "learning_rate": 0.00029999074496815805, - "loss": 1.0428, + "epoch": 0.2294455066921606, + "grad_norm": 0.2099609375, + "learning_rate": 0.00029961189124030885, + "loss": 0.9664, "step": 120 }, { - "epoch": 0.1120144534778681, - "grad_norm": 0.203125, - "learning_rate": 0.00029998820251615196, - "loss": 1.0927, + "epoch": 0.23709369024856597, + "grad_norm": 0.236328125, + "learning_rate": 0.00029957359992777404, + "loss": 0.9831, "step": 124 }, { - "epoch": 0.11562782294489611, - "grad_norm": 0.2080078125, - "learning_rate": 0.00029998535190522746, - "loss": 1.0348, + "epoch": 0.2447418738049713, + "grad_norm": 0.25390625, + "learning_rate": 0.00029953351016733854, + "loss": 0.9718, "step": 128 }, { - "epoch": 0.11924119241192412, - "grad_norm": 0.2060546875, - "learning_rate": 0.00029998219314124135, - "loss": 1.0602, + "epoch": 0.25239005736137665, + "grad_norm": 0.205078125, + "learning_rate": 0.000299491622441035, + "loss": 0.9592, "step": 132 }, { - "epoch": 0.12285456187895212, - "grad_norm": 0.2099609375, - "learning_rate": 0.00029997872623068345, - "loss": 1.0567, + "epoch": 0.26003824091778205, + "grad_norm": 0.2109375, + "learning_rate": 0.00029944793725251436, + "loss": 0.9514, "step": 136 }, { - "epoch": 0.12646793134598014, - "grad_norm": 0.2099609375, - "learning_rate": 0.00029997495118067676, - "loss": 1.0755, + "epoch": 0.2676864244741874, + "grad_norm": 0.220703125, + "learning_rate": 0.00029940245512704005, + "loss": 0.9435, "step": 140 }, { - "epoch": 0.13008130081300814, - "grad_norm": 0.19921875, - "learning_rate": 0.00029997086799897726, - "loss": 1.0485, + "epoch": 0.27533460803059273, + "grad_norm": 0.2138671875, + "learning_rate": 0.0002993551766114815, + "loss": 0.9545, "step": 144 }, { - "epoch": 0.13369467028003612, - "grad_norm": 0.2060546875, - "learning_rate": 0.0002999664766939741, - "loss": 1.0222, + "epoch": 0.2829827915869981, + "grad_norm": 0.25, + "learning_rate": 0.00029930610227430767, + "loss": 0.9378, "step": 148 }, { - "epoch": 0.13730803974706413, - "grad_norm": 0.203125, - "learning_rate": 0.00029996177727468945, - "loss": 1.0111, + "epoch": 0.29063097514340347, + "grad_norm": 0.2314453125, + "learning_rate": 0.0002992552327055802, + "loss": 0.9085, "step": 152 }, { - "epoch": 0.14092140921409213, - "grad_norm": 0.21484375, - "learning_rate": 0.00029995676975077857, - "loss": 1.0359, + "epoch": 0.2982791586998088, + "grad_norm": 0.2353515625, + "learning_rate": 0.0002992025685169461, + "loss": 0.9482, "step": 156 }, { - "epoch": 0.14453477868112014, - "grad_norm": 0.2041015625, - "learning_rate": 0.0002999514541325296, - "loss": 1.005, + "epoch": 0.30592734225621415, + "grad_norm": 0.216796875, + "learning_rate": 0.00029914811034163096, + "loss": 0.949, "step": 160 }, { - "epoch": 0.14814814814814814, - "grad_norm": 0.22265625, - "learning_rate": 0.00029994583043086386, - "loss": 1.0534, + "epoch": 0.3135755258126195, + "grad_norm": 0.2421875, + "learning_rate": 0.00029909185883443063, + "loss": 0.9796, "step": 164 }, { - "epoch": 0.15176151761517614, - "grad_norm": 0.20703125, - "learning_rate": 0.00029993989865733555, - "loss": 1.0187, + "epoch": 0.32122370936902483, + "grad_norm": 0.2294921875, + "learning_rate": 0.0002990338146717039, + "loss": 0.9351, "step": 168 }, { - "epoch": 0.15537488708220415, - "grad_norm": 0.21484375, - "learning_rate": 0.0002999336588241317, - "loss": 1.0295, + "epoch": 0.32887189292543023, + "grad_norm": 0.220703125, + "learning_rate": 0.0002989739785513639, + "loss": 0.9166, "step": 172 }, { - "epoch": 0.15898825654923215, - "grad_norm": 0.2314453125, - "learning_rate": 0.00029992711094407247, - "loss": 1.043, + "epoch": 0.3365200764818356, + "grad_norm": 0.2158203125, + "learning_rate": 0.0002989123511928703, + "loss": 0.9204, "step": 176 }, { - "epoch": 0.16260162601626016, - "grad_norm": 0.2333984375, - "learning_rate": 0.0002999202550306109, - "loss": 1.035, + "epoch": 0.3441682600382409, + "grad_norm": 0.2255859375, + "learning_rate": 0.00029884893333722, + "loss": 0.9356, "step": 180 }, { - "epoch": 0.16621499548328816, - "grad_norm": 0.220703125, - "learning_rate": 0.0002999130910978327, - "loss": 0.9977, + "epoch": 0.35181644359464626, + "grad_norm": 0.2265625, + "learning_rate": 0.0002987837257469387, + "loss": 0.9342, "step": 184 }, { - "epoch": 0.16982836495031617, - "grad_norm": 0.2138671875, - "learning_rate": 0.0002999056191604567, - "loss": 0.9963, + "epoch": 0.35946462715105165, + "grad_norm": 0.2265625, + "learning_rate": 0.00029871672920607153, + "loss": 0.9026, "step": 188 }, { - "epoch": 0.17344173441734417, - "grad_norm": 0.2255859375, - "learning_rate": 0.0002998978392338342, - "loss": 1.0241, + "epoch": 0.367112810707457, + "grad_norm": 0.2314453125, + "learning_rate": 0.0002986479445201737, + "loss": 0.8983, "step": 192 }, { - "epoch": 0.17705510388437218, - "grad_norm": 0.2158203125, - "learning_rate": 0.0002998897513339497, - "loss": 1.0046, + "epoch": 0.37476099426386233, + "grad_norm": 0.259765625, + "learning_rate": 0.0002985773725163008, + "loss": 0.922, "step": 196 }, { - "epoch": 0.18066847335140018, - "grad_norm": 0.2421875, - "learning_rate": 0.0002998813554774202, - "loss": 1.0136, + "epoch": 0.3824091778202677, + "grad_norm": 0.251953125, + "learning_rate": 0.0002985050140429986, + "loss": 0.9099, "step": 200 }, { - "epoch": 0.1842818428184282, - "grad_norm": 0.2421875, - "learning_rate": 0.0002998726516814953, - "loss": 0.9941, + "epoch": 0.390057361376673, + "grad_norm": 0.25, + "learning_rate": 0.0002984308699702935, + "loss": 0.8825, "step": 204 }, { - "epoch": 0.1878952122854562, - "grad_norm": 0.22265625, - "learning_rate": 0.00029986363996405754, - "loss": 0.9964, + "epoch": 0.3977055449330784, + "grad_norm": 0.2373046875, + "learning_rate": 0.0002983549411896812, + "loss": 0.893, "step": 208 }, { - "epoch": 0.1915085817524842, - "grad_norm": 0.224609375, - "learning_rate": 0.00029985432034362197, - "loss": 0.9962, + "epoch": 0.40535372848948376, + "grad_norm": 0.2353515625, + "learning_rate": 0.0002982772286141167, + "loss": 0.9068, "step": 212 }, { - "epoch": 0.1951219512195122, - "grad_norm": 0.240234375, - "learning_rate": 0.0002998446928393362, - "loss": 1.0432, + "epoch": 0.4130019120458891, + "grad_norm": 0.2421875, + "learning_rate": 0.000298197733178003, + "loss": 0.8536, "step": 216 }, { - "epoch": 0.1987353206865402, - "grad_norm": 0.2109375, - "learning_rate": 0.0002998347574709805, - "loss": 0.9932, + "epoch": 0.42065009560229444, + "grad_norm": 0.2041015625, + "learning_rate": 0.00029811645583717987, + "loss": 0.8843, "step": 220 }, { - "epoch": 0.2023486901535682, - "grad_norm": 0.23046875, - "learning_rate": 0.0002998245142589677, - "loss": 0.9796, + "epoch": 0.42829827915869984, + "grad_norm": 0.263671875, + "learning_rate": 0.00029803339756891254, + "loss": 0.8627, "step": 224 }, { - "epoch": 0.20596205962059622, - "grad_norm": 0.24609375, - "learning_rate": 0.00029981396322434285, - "loss": 0.9699, + "epoch": 0.4359464627151052, + "grad_norm": 0.263671875, + "learning_rate": 0.00029794855937187963, + "loss": 0.8572, "step": 228 }, { - "epoch": 0.20957542908762422, - "grad_norm": 0.349609375, - "learning_rate": 0.0002998031043887838, - "loss": 0.9858, + "epoch": 0.4435946462715105, + "grad_norm": 0.2265625, + "learning_rate": 0.0002978619422661613, + "loss": 0.8255, "step": 232 }, { - "epoch": 0.21318879855465223, - "grad_norm": 0.2236328125, - "learning_rate": 0.00029979193777460053, - "loss": 0.968, + "epoch": 0.45124282982791586, + "grad_norm": 0.2197265625, + "learning_rate": 0.0002977735472932273, + "loss": 0.8274, "step": 236 }, { - "epoch": 0.21680216802168023, - "grad_norm": 0.2197265625, - "learning_rate": 0.0002997804634047355, - "loss": 0.9406, + "epoch": 0.4588910133843212, + "grad_norm": 0.2490234375, + "learning_rate": 0.00029768337551592394, + "loss": 0.8569, "step": 240 }, { - "epoch": 0.2204155374887082, - "grad_norm": 0.2392578125, - "learning_rate": 0.0002997686813027634, - "loss": 0.9795, + "epoch": 0.4665391969407266, + "grad_norm": 0.2255859375, + "learning_rate": 0.00029759142801846143, + "loss": 0.8791, "step": 244 }, { - "epoch": 0.2240289069557362, - "grad_norm": 0.2451171875, - "learning_rate": 0.0002997565914928912, - "loss": 0.9785, + "epoch": 0.47418738049713194, + "grad_norm": 0.248046875, + "learning_rate": 0.00029749770590640123, + "loss": 0.8215, "step": 248 }, { - "epoch": 0.22764227642276422, - "grad_norm": 0.220703125, - "learning_rate": 0.00029974419399995814, - "loss": 0.9495, + "epoch": 0.4818355640535373, + "grad_norm": 0.2470703125, + "learning_rate": 0.00029740221030664216, + "loss": 0.8546, "step": 252 }, { - "epoch": 0.23125564588979222, - "grad_norm": 0.2275390625, - "learning_rate": 0.0002997314888494354, - "loss": 0.9771, + "epoch": 0.4894837476099426, + "grad_norm": 0.2421875, + "learning_rate": 0.00029730494236740744, + "loss": 0.8518, "step": 256 }, { - "epoch": 0.23486901535682023, + "epoch": 0.497131931166348, "grad_norm": 0.251953125, - "learning_rate": 0.00029971847606742645, - "loss": 0.9933, + "learning_rate": 0.0002972059032582304, + "loss": 0.8115, "step": 260 }, { - "epoch": 0.23848238482384823, - "grad_norm": 0.267578125, - "learning_rate": 0.0002997051556806669, - "loss": 0.9808, + "epoch": 0.5047801147227533, + "grad_norm": 0.251953125, + "learning_rate": 0.0002971050941699407, + "loss": 0.8818, "step": 264 }, { - "epoch": 0.24209575429087624, - "grad_norm": 0.25, - "learning_rate": 0.000299691527716524, - "loss": 0.9665, + "epoch": 0.5124282982791587, + "grad_norm": 0.2314453125, + "learning_rate": 0.00029700251631464993, + "loss": 0.8834, "step": 268 }, { - "epoch": 0.24570912375790424, - "grad_norm": 0.244140625, - "learning_rate": 0.00029967759220299734, - "loss": 0.9489, + "epoch": 0.5200764818355641, + "grad_norm": 0.23828125, + "learning_rate": 0.000296898170925737, + "loss": 0.8502, "step": 272 }, { - "epoch": 0.24932249322493225, - "grad_norm": 0.263671875, - "learning_rate": 0.00029966334916871806, - "loss": 0.924, + "epoch": 0.5277246653919694, + "grad_norm": 0.2373046875, + "learning_rate": 0.0002967920592578335, + "loss": 0.854, "step": 276 }, { - "epoch": 0.2529358626919603, - "grad_norm": 0.2275390625, - "learning_rate": 0.00029964879864294926, - "loss": 0.9947, + "epoch": 0.5353728489483748, + "grad_norm": 0.240234375, + "learning_rate": 0.0002966841825868082, + "loss": 0.8164, "step": 280 }, { - "epoch": 0.2565492321589883, - "grad_norm": 0.2578125, - "learning_rate": 0.0002996339406555858, - "loss": 0.9652, + "epoch": 0.5430210325047801, + "grad_norm": 0.228515625, + "learning_rate": 0.00029657454220975216, + "loss": 0.9058, "step": 284 }, { - "epoch": 0.2601626016260163, - "grad_norm": 0.236328125, - "learning_rate": 0.0002996187752371543, - "loss": 0.9315, + "epoch": 0.5506692160611855, + "grad_norm": 0.2294921875, + "learning_rate": 0.00029646313944496297, + "loss": 0.8646, "step": 288 }, { - "epoch": 0.26377597109304424, - "grad_norm": 0.2353515625, - "learning_rate": 0.0002996033024188128, - "loss": 0.972, + "epoch": 0.5583173996175909, + "grad_norm": 0.2158203125, + "learning_rate": 0.00029634997563192866, + "loss": 0.8536, "step": 292 }, { - "epoch": 0.26738934056007224, - "grad_norm": 0.25390625, - "learning_rate": 0.00029958752223235114, - "loss": 0.9273, + "epoch": 0.5659655831739961, + "grad_norm": 0.2490234375, + "learning_rate": 0.0002962350521313122, + "loss": 0.8532, "step": 296 }, { - "epoch": 0.27100271002710025, - "grad_norm": 0.25, - "learning_rate": 0.00029957143471019053, - "loss": 0.9286, + "epoch": 0.5736137667304015, + "grad_norm": 0.2255859375, + "learning_rate": 0.0002961183703249342, + "loss": 0.8228, "step": 300 }, { - "epoch": 0.27461607949412825, - "grad_norm": 0.259765625, - "learning_rate": 0.0002995550398853837, - "loss": 0.949, + "epoch": 0.5812619502868069, + "grad_norm": 0.2392578125, + "learning_rate": 0.0002959999316157573, + "loss": 0.8088, "step": 304 }, { - "epoch": 0.27822944896115626, - "grad_norm": 0.232421875, - "learning_rate": 0.0002995383377916147, - "loss": 0.9447, + "epoch": 0.5889101338432122, + "grad_norm": 0.24609375, + "learning_rate": 0.00029587973742786875, + "loss": 0.8291, "step": 308 }, { - "epoch": 0.28184281842818426, - "grad_norm": 0.28125, - "learning_rate": 0.0002995213284631989, - "loss": 0.9322, + "epoch": 0.5965583173996176, + "grad_norm": 0.271484375, + "learning_rate": 0.0002957577892064632, + "loss": 0.8087, "step": 312 }, { - "epoch": 0.28545618789521227, - "grad_norm": 0.2392578125, - "learning_rate": 0.0002995040119350828, - "loss": 0.9584, + "epoch": 0.6042065009560229, + "grad_norm": 0.2373046875, + "learning_rate": 0.00029563408841782576, + "loss": 0.853, "step": 316 }, { - "epoch": 0.28906955736224027, - "grad_norm": 0.248046875, - "learning_rate": 0.0002994863882428442, - "loss": 0.9464, + "epoch": 0.6118546845124283, + "grad_norm": 0.251953125, + "learning_rate": 0.00029550863654931385, + "loss": 0.847, "step": 320 }, { - "epoch": 0.2926829268292683, - "grad_norm": 0.24609375, - "learning_rate": 0.000299468457422692, - "loss": 0.8938, + "epoch": 0.6195028680688337, + "grad_norm": 0.25390625, + "learning_rate": 0.0002953814351093398, + "loss": 0.8087, "step": 324 }, { - "epoch": 0.2962962962962963, - "grad_norm": 0.26953125, - "learning_rate": 0.00029945021951146596, - "loss": 0.9366, + "epoch": 0.627151051625239, + "grad_norm": 0.2421875, + "learning_rate": 0.0002952524856273524, + "loss": 0.8514, "step": 328 }, { - "epoch": 0.2999096657633243, - "grad_norm": 0.25, - "learning_rate": 0.00029943167454663693, - "loss": 0.9055, + "epoch": 0.6347992351816444, + "grad_norm": 0.212890625, + "learning_rate": 0.00029512178965381854, + "loss": 0.8501, "step": 332 }, { - "epoch": 0.3035230352303523, - "grad_norm": 0.2421875, - "learning_rate": 0.0002994128225663065, - "loss": 0.9295, + "epoch": 0.6424474187380497, + "grad_norm": 0.2353515625, + "learning_rate": 0.00029498934876020475, + "loss": 0.8029, "step": 336 }, { - "epoch": 0.3071364046973803, - "grad_norm": 0.228515625, - "learning_rate": 0.0002993936636092071, - "loss": 0.9431, + "epoch": 0.6500956022944551, + "grad_norm": 0.232421875, + "learning_rate": 0.00029485516453895826, + "loss": 0.8293, "step": 340 }, { - "epoch": 0.3107497741644083, - "grad_norm": 0.271484375, - "learning_rate": 0.00029937419771470185, - "loss": 0.9077, + "epoch": 0.6577437858508605, + "grad_norm": 0.2490234375, + "learning_rate": 0.0002947192386034874, + "loss": 0.8695, "step": 344 }, { - "epoch": 0.3143631436314363, - "grad_norm": 0.2392578125, - "learning_rate": 0.0002993544249227845, - "loss": 0.8974, + "epoch": 0.6653919694072657, + "grad_norm": 0.236328125, + "learning_rate": 0.00029458157258814316, + "loss": 0.8249, "step": 348 }, { - "epoch": 0.3179765130984643, - "grad_norm": 0.2734375, - "learning_rate": 0.00029933434527407924, - "loss": 0.9345, + "epoch": 0.6730401529636711, + "grad_norm": 0.234375, + "learning_rate": 0.00029444216814819834, + "loss": 0.8009, "step": 352 }, { - "epoch": 0.3215898825654923, - "grad_norm": 0.240234375, - "learning_rate": 0.000299313958809841, - "loss": 0.9295, + "epoch": 0.6806883365200764, + "grad_norm": 0.2490234375, + "learning_rate": 0.00029430102695982875, + "loss": 0.8642, "step": 356 }, { - "epoch": 0.3252032520325203, - "grad_norm": 0.2333984375, - "learning_rate": 0.0002992932655719547, - "loss": 0.9544, + "epoch": 0.6883365200764818, + "grad_norm": 0.228515625, + "learning_rate": 0.00029415815072009237, + "loss": 0.8562, "step": 360 }, { - "epoch": 0.3288166214995483, - "grad_norm": 0.251953125, - "learning_rate": 0.0002992722656029357, - "loss": 0.904, + "epoch": 0.6959847036328872, + "grad_norm": 0.240234375, + "learning_rate": 0.00029401354114690905, + "loss": 0.8274, "step": 364 }, { - "epoch": 0.3324299909665763, - "grad_norm": 0.236328125, - "learning_rate": 0.00029925095894592975, - "loss": 0.8913, + "epoch": 0.7036328871892925, + "grad_norm": 0.2099609375, + "learning_rate": 0.0002938671999790402, + "loss": 0.8214, "step": 368 }, { - "epoch": 0.33604336043360433, - "grad_norm": 0.234375, - "learning_rate": 0.00029922934564471243, - "loss": 0.9099, + "epoch": 0.7112810707456979, + "grad_norm": 0.24609375, + "learning_rate": 0.00029371912897606736, + "loss": 0.8537, "step": 372 }, { - "epoch": 0.33965672990063234, - "grad_norm": 0.2353515625, - "learning_rate": 0.0002992074257436894, - "loss": 0.9094, + "epoch": 0.7189292543021033, + "grad_norm": 0.2275390625, + "learning_rate": 0.00029356932991837163, + "loss": 0.8378, "step": 376 }, { - "epoch": 0.34327009936766034, - "grad_norm": 0.26953125, - "learning_rate": 0.0002991851992878964, - "loss": 0.9316, + "epoch": 0.7265774378585086, + "grad_norm": 0.251953125, + "learning_rate": 0.0002934178046071116, + "loss": 0.8064, "step": 380 }, { - "epoch": 0.34688346883468835, - "grad_norm": 0.263671875, - "learning_rate": 0.00029916266632299884, - "loss": 0.8649, + "epoch": 0.734225621414914, + "grad_norm": 0.259765625, + "learning_rate": 0.0002932645548642024, + "loss": 0.8427, "step": 384 }, { - "epoch": 0.35049683830171635, + "epoch": 0.7418738049713193, "grad_norm": 0.251953125, - "learning_rate": 0.0002991398268952919, - "loss": 0.933, + "learning_rate": 0.0002931095825322931, + "loss": 0.7602, "step": 388 }, { - "epoch": 0.35411020776874436, - "grad_norm": 0.248046875, - "learning_rate": 0.00029911668105170043, - "loss": 0.9069, + "epoch": 0.7495219885277247, + "grad_norm": 0.232421875, + "learning_rate": 0.00029295288947474513, + "loss": 0.881, "step": 392 }, { - "epoch": 0.35772357723577236, - "grad_norm": 0.2451171875, - "learning_rate": 0.00029909322883977884, - "loss": 0.8697, + "epoch": 0.7571701720841301, + "grad_norm": 0.2353515625, + "learning_rate": 0.0002927944775756095, + "loss": 0.865, "step": 396 }, { - "epoch": 0.36133694670280037, - "grad_norm": 0.2431640625, - "learning_rate": 0.00029906947030771097, - "loss": 0.875, + "epoch": 0.7648183556405354, + "grad_norm": 0.2353515625, + "learning_rate": 0.0002926343487396044, + "loss": 0.839, "step": 400 }, { - "epoch": 0.36495031616982837, - "grad_norm": 0.2373046875, - "learning_rate": 0.00029904540550431004, - "loss": 0.9392, + "epoch": 0.7724665391969407, + "grad_norm": 0.2412109375, + "learning_rate": 0.00029247250489209217, + "loss": 0.7939, "step": 404 }, { - "epoch": 0.3685636856368564, - "grad_norm": 0.220703125, - "learning_rate": 0.00029902103447901844, - "loss": 0.8518, + "epoch": 0.780114722753346, + "grad_norm": 0.2470703125, + "learning_rate": 0.00029230894797905595, + "loss": 0.7748, "step": 408 }, { - "epoch": 0.3721770551038844, - "grad_norm": 0.224609375, - "learning_rate": 0.00029899635728190783, - "loss": 0.9312, + "epoch": 0.7877629063097514, + "grad_norm": 0.255859375, + "learning_rate": 0.00029214367996707676, + "loss": 0.7829, "step": 412 }, { - "epoch": 0.3757904245709124, - "grad_norm": 0.27734375, - "learning_rate": 0.0002989713739636788, - "loss": 0.9395, + "epoch": 0.7954110898661568, + "grad_norm": 0.2392578125, + "learning_rate": 0.00029197670284330954, + "loss": 0.7867, "step": 416 }, { - "epoch": 0.3794037940379404, - "grad_norm": 0.251953125, - "learning_rate": 0.000298946084575661, - "loss": 0.9087, + "epoch": 0.8030592734225621, + "grad_norm": 0.2255859375, + "learning_rate": 0.00029180801861545906, + "loss": 0.7971, "step": 420 }, { - "epoch": 0.3830171635049684, - "grad_norm": 0.263671875, - "learning_rate": 0.00029892048916981283, - "loss": 0.8754, + "epoch": 0.8107074569789675, + "grad_norm": 0.25390625, + "learning_rate": 0.0002916376293117564, + "loss": 0.8241, "step": 424 }, { - "epoch": 0.3866305329719964, - "grad_norm": 0.236328125, - "learning_rate": 0.00029889458779872133, - "loss": 0.8802, + "epoch": 0.8183556405353728, + "grad_norm": 0.244140625, + "learning_rate": 0.00029146553698093387, + "loss": 0.8119, "step": 428 }, { - "epoch": 0.3902439024390244, - "grad_norm": 0.2451171875, - "learning_rate": 0.00029886838051560247, - "loss": 0.925, + "epoch": 0.8260038240917782, + "grad_norm": 0.271484375, + "learning_rate": 0.00029129174369220087, + "loss": 0.8048, "step": 432 }, { - "epoch": 0.3938572719060524, - "grad_norm": 0.2470703125, - "learning_rate": 0.0002988418673743004, - "loss": 0.9213, + "epoch": 0.8336520076481836, + "grad_norm": 0.232421875, + "learning_rate": 0.00029111625153521877, + "loss": 0.76, "step": 436 }, { - "epoch": 0.3974706413730804, - "grad_norm": 0.25390625, - "learning_rate": 0.00029881504842928783, - "loss": 0.8926, + "epoch": 0.8413001912045889, + "grad_norm": 0.271484375, + "learning_rate": 0.00029093906262007583, + "loss": 0.7833, "step": 440 }, { - "epoch": 0.4010840108401084, - "grad_norm": 0.248046875, - "learning_rate": 0.00029878792373566577, - "loss": 0.8691, + "epoch": 0.8489483747609943, + "grad_norm": 0.2158203125, + "learning_rate": 0.00029076017907726196, + "loss": 0.8027, "step": 444 }, { - "epoch": 0.4046973803071364, - "grad_norm": 0.283203125, - "learning_rate": 0.0002987604933491634, - "loss": 0.8952, + "epoch": 0.8565965583173997, + "grad_norm": 0.21875, + "learning_rate": 0.0002905796030576428, + "loss": 0.818, "step": 448 }, { - "epoch": 0.4083107497741644, - "grad_norm": 0.240234375, - "learning_rate": 0.0002987327573261379, - "loss": 0.8955, + "epoch": 0.864244741873805, + "grad_norm": 0.2373046875, + "learning_rate": 0.00029039733673243416, + "loss": 0.8358, "step": 452 }, { - "epoch": 0.41192411924119243, - "grad_norm": 0.234375, - "learning_rate": 0.0002987047157235745, - "loss": 0.8932, + "epoch": 0.8718929254302104, + "grad_norm": 0.2412109375, + "learning_rate": 0.0002902133822931759, + "loss": 0.7543, "step": 456 }, { - "epoch": 0.41553748870822044, - "grad_norm": 0.25390625, - "learning_rate": 0.0002986763685990861, - "loss": 0.8703, + "epoch": 0.8795411089866156, + "grad_norm": 0.2421875, + "learning_rate": 0.00029002774195170525, + "loss": 0.7765, "step": 460 }, { - "epoch": 0.41915085817524844, - "grad_norm": 0.2734375, - "learning_rate": 0.0002986477160109134, - "loss": 0.8803, + "epoch": 0.887189292543021, + "grad_norm": 0.2451171875, + "learning_rate": 0.0002898404179401306, + "loss": 0.8094, "step": 464 }, { - "epoch": 0.42276422764227645, - "grad_norm": 0.23828125, - "learning_rate": 0.0002986187580179248, - "loss": 0.9077, + "epoch": 0.8948374760994264, + "grad_norm": 0.2431640625, + "learning_rate": 0.0002896514125108045, + "loss": 0.7657, "step": 468 }, { - "epoch": 0.42637759710930445, - "grad_norm": 0.251953125, - "learning_rate": 0.00029858949467961605, - "loss": 0.8915, + "epoch": 0.9024856596558317, + "grad_norm": 0.2353515625, + "learning_rate": 0.0002894607279362966, + "loss": 0.7774, "step": 472 }, { - "epoch": 0.42999096657633246, - "grad_norm": 0.251953125, - "learning_rate": 0.0002985599260561102, - "loss": 0.826, + "epoch": 0.9101338432122371, + "grad_norm": 0.2255859375, + "learning_rate": 0.0002892683665093662, + "loss": 0.8148, "step": 476 }, { - "epoch": 0.43360433604336046, - "grad_norm": 0.255859375, - "learning_rate": 0.0002985300522081577, - "loss": 0.8754, + "epoch": 0.9177820267686424, + "grad_norm": 0.236328125, + "learning_rate": 0.0002890743305429348, + "loss": 0.7882, "step": 480 }, { - "epoch": 0.4372177055103884, - "grad_norm": 0.2490234375, - "learning_rate": 0.0002984998731971359, - "loss": 0.8972, + "epoch": 0.9254302103250478, + "grad_norm": 0.2294921875, + "learning_rate": 0.0002888786223700585, + "loss": 0.7656, "step": 484 }, { - "epoch": 0.4408310749774164, - "grad_norm": 0.2470703125, - "learning_rate": 0.00029846938908504927, - "loss": 0.8418, + "epoch": 0.9330783938814532, + "grad_norm": 0.224609375, + "learning_rate": 0.00028868124434389944, + "loss": 0.7802, "step": 488 }, { - "epoch": 0.4444444444444444, - "grad_norm": 0.2451171875, - "learning_rate": 0.0002984385999345291, - "loss": 0.8761, + "epoch": 0.9407265774378585, + "grad_norm": 0.240234375, + "learning_rate": 0.00028848219883769805, + "loss": 0.7773, "step": 492 }, { - "epoch": 0.4480578139114724, - "grad_norm": 0.232421875, - "learning_rate": 0.00029840750580883323, - "loss": 0.8642, + "epoch": 0.9483747609942639, + "grad_norm": 0.2314453125, + "learning_rate": 0.000288281488244744, + "loss": 0.7803, "step": 496 }, { - "epoch": 0.45167118337850043, - "grad_norm": 0.25390625, - "learning_rate": 0.0002983761067718464, - "loss": 0.8763, + "epoch": 0.9560229445506692, + "grad_norm": 0.255859375, + "learning_rate": 0.000288079114978348, + "loss": 0.8056, "step": 500 }, { - "epoch": 0.45528455284552843, - "grad_norm": 0.26953125, - "learning_rate": 0.0002983444028880796, - "loss": 0.8694, + "epoch": 0.9636711281070746, + "grad_norm": 0.240234375, + "learning_rate": 0.0002878750814718121, + "loss": 0.8309, "step": 504 }, { - "epoch": 0.45889792231255644, - "grad_norm": 0.2236328125, - "learning_rate": 0.0002983123942226701, - "loss": 0.8567, + "epoch": 0.97131931166348, + "grad_norm": 0.2294921875, + "learning_rate": 0.00028766939017840114, + "loss": 0.7737, "step": 508 }, { - "epoch": 0.46251129177958444, - "grad_norm": 0.294921875, - "learning_rate": 0.00029828008084138155, - "loss": 0.8337, + "epoch": 0.9789674952198852, + "grad_norm": 0.26171875, + "learning_rate": 0.00028746204357131273, + "loss": 0.8039, "step": 512 }, { - "epoch": 0.46612466124661245, - "grad_norm": 0.2333984375, - "learning_rate": 0.0002982474628106035, - "loss": 0.9067, + "epoch": 0.9866156787762906, + "grad_norm": 0.2255859375, + "learning_rate": 0.0002872530441436477, + "loss": 0.7341, "step": 516 }, { - "epoch": 0.46973803071364045, - "grad_norm": 0.251953125, - "learning_rate": 0.00029821454019735147, - "loss": 0.8444, + "epoch": 0.994263862332696, + "grad_norm": 0.228515625, + "learning_rate": 0.0002870423944083801, + "loss": 0.8122, "step": 520 }, { - "epoch": 0.47335140018066846, - "grad_norm": 0.25, - "learning_rate": 0.00029818131306926676, - "loss": 0.8703, + "epoch": 1.0019120458891013, + "grad_norm": 0.216796875, + "learning_rate": 0.0002868300968983271, + "loss": 0.7403, "step": 524 }, { - "epoch": 0.47696476964769646, - "grad_norm": 0.26171875, - "learning_rate": 0.0002981477814946163, - "loss": 0.8683, + "epoch": 1.0095602294455066, + "grad_norm": 0.24609375, + "learning_rate": 0.0002866161541661185, + "loss": 0.697, "step": 528 }, { - "epoch": 0.48057813911472447, - "grad_norm": 0.263671875, - "learning_rate": 0.00029811394554229254, - "loss": 0.8359, + "epoch": 1.0172084130019121, + "grad_norm": 0.23828125, + "learning_rate": 0.0002864005687841656, + "loss": 0.7442, "step": 532 }, { - "epoch": 0.48419150858175247, - "grad_norm": 0.259765625, - "learning_rate": 0.0002980798052818133, - "loss": 0.8693, + "epoch": 1.0248565965583174, + "grad_norm": 0.26953125, + "learning_rate": 0.0002861833433446312, + "loss": 0.6853, "step": 536 }, { - "epoch": 0.4878048780487805, - "grad_norm": 0.244140625, - "learning_rate": 0.0002980453607833216, - "loss": 0.8598, + "epoch": 1.0325047801147227, + "grad_norm": 0.23828125, + "learning_rate": 0.00028596448045939735, + "loss": 0.7398, "step": 540 }, { - "epoch": 0.4914182475158085, - "grad_norm": 0.267578125, - "learning_rate": 0.00029801061211758554, - "loss": 0.8279, + "epoch": 1.0401529636711282, + "grad_norm": 0.2490234375, + "learning_rate": 0.0002857439827600348, + "loss": 0.7912, "step": 544 }, { - "epoch": 0.4950316169828365, - "grad_norm": 0.279296875, - "learning_rate": 0.00029797555935599806, - "loss": 0.8298, + "epoch": 1.0478011472275335, + "grad_norm": 0.259765625, + "learning_rate": 0.0002855218528977709, + "loss": 0.7138, "step": 548 }, { - "epoch": 0.4986449864498645, - "grad_norm": 0.275390625, - "learning_rate": 0.00029794020257057713, - "loss": 0.8548, + "epoch": 1.0554493307839388, + "grad_norm": 0.2314453125, + "learning_rate": 0.00028529809354345794, + "loss": 0.726, "step": 552 }, { - "epoch": 0.5022583559168925, - "grad_norm": 0.2421875, - "learning_rate": 0.00029790454183396505, - "loss": 0.896, + "epoch": 1.063097514340344, + "grad_norm": 0.287109375, + "learning_rate": 0.0002850727073875409, + "loss": 0.7058, "step": 556 }, { - "epoch": 0.5058717253839206, - "grad_norm": 0.265625, - "learning_rate": 0.0002978685772194288, - "loss": 0.8971, + "epoch": 1.0707456978967496, + "grad_norm": 0.228515625, + "learning_rate": 0.00028484569714002517, + "loss": 0.7102, "step": 560 }, { - "epoch": 0.5094850948509485, - "grad_norm": 0.255859375, - "learning_rate": 0.0002978323088008596, - "loss": 0.8629, + "epoch": 1.0783938814531548, + "grad_norm": 0.2392578125, + "learning_rate": 0.0002846170655304438, + "loss": 0.6534, "step": 564 }, { - "epoch": 0.5130984643179766, - "grad_norm": 0.2578125, - "learning_rate": 0.0002977957366527729, - "loss": 0.8855, + "epoch": 1.0860420650095601, + "grad_norm": 0.23828125, + "learning_rate": 0.0002843868153078251, + "loss": 0.6918, "step": 568 }, { - "epoch": 0.5167118337850045, - "grad_norm": 0.2431640625, - "learning_rate": 0.00029775886085030816, - "loss": 0.8323, + "epoch": 1.0936902485659656, + "grad_norm": 0.251953125, + "learning_rate": 0.000284154949240659, + "loss": 0.6673, "step": 572 }, { - "epoch": 0.5203252032520326, - "grad_norm": 0.25, - "learning_rate": 0.0002977216814692288, - "loss": 0.8541, + "epoch": 1.101338432122371, + "grad_norm": 0.255859375, + "learning_rate": 0.0002839214701168644, + "loss": 0.6722, "step": 576 }, { - "epoch": 0.5239385727190605, - "grad_norm": 0.2490234375, - "learning_rate": 0.0002976841985859218, - "loss": 0.8808, + "epoch": 1.1089866156787762, + "grad_norm": 0.251953125, + "learning_rate": 0.00028368638074375516, + "loss": 0.7141, "step": 580 }, { - "epoch": 0.5275519421860885, - "grad_norm": 0.259765625, - "learning_rate": 0.00029764641227739776, - "loss": 0.835, + "epoch": 1.1166347992351817, + "grad_norm": 0.263671875, + "learning_rate": 0.0002834496839480063, + "loss": 0.665, "step": 584 }, { - "epoch": 0.5311653116531165, - "grad_norm": 0.28125, - "learning_rate": 0.0002976083226212908, - "loss": 0.8281, + "epoch": 1.124282982791587, + "grad_norm": 0.2421875, + "learning_rate": 0.00028321138257562066, + "loss": 0.6886, "step": 588 }, { - "epoch": 0.5347786811201445, - "grad_norm": 0.28125, - "learning_rate": 0.00029756992969585814, - "loss": 0.8838, + "epoch": 1.1319311663479923, + "grad_norm": 0.265625, + "learning_rate": 0.00028297147949189386, + "loss": 0.6997, "step": 592 }, { - "epoch": 0.5383920505871725, - "grad_norm": 0.2333984375, - "learning_rate": 0.0002975312335799801, - "loss": 0.8677, + "epoch": 1.1395793499043978, + "grad_norm": 0.2578125, + "learning_rate": 0.00028272997758138044, + "loss": 0.7051, "step": 596 }, { - "epoch": 0.5420054200542005, - "grad_norm": 0.228515625, - "learning_rate": 0.00029749223435315993, - "loss": 0.8747, + "epoch": 1.147227533460803, + "grad_norm": 0.248046875, + "learning_rate": 0.00028248687974785896, + "loss": 0.7188, "step": 600 }, { - "epoch": 0.5456187895212286, - "grad_norm": 0.2470703125, - "learning_rate": 0.0002974529320955238, - "loss": 0.8742, + "epoch": 1.1548757170172084, + "grad_norm": 0.275390625, + "learning_rate": 0.0002822421889142969, + "loss": 0.6757, "step": 604 }, { - "epoch": 0.5492321589882565, - "grad_norm": 0.259765625, - "learning_rate": 0.0002974133268878202, - "loss": 0.8547, + "epoch": 1.1625239005736137, + "grad_norm": 0.25, + "learning_rate": 0.00028199590802281595, + "loss": 0.7203, "step": 608 }, { - "epoch": 0.5528455284552846, - "grad_norm": 0.279296875, - "learning_rate": 0.0002973734188114203, - "loss": 0.8832, + "epoch": 1.1701720841300192, + "grad_norm": 0.255859375, + "learning_rate": 0.00028174804003465616, + "loss": 0.6943, "step": 612 }, { - "epoch": 0.5564588979223125, - "grad_norm": 0.248046875, - "learning_rate": 0.0002973332079483173, - "loss": 0.8457, + "epoch": 1.1778202676864244, + "grad_norm": 0.2412109375, + "learning_rate": 0.0002814985879301408, + "loss": 0.7037, "step": 616 }, { - "epoch": 0.5600722673893406, - "grad_norm": 0.265625, - "learning_rate": 0.00029729269438112665, - "loss": 0.8019, + "epoch": 1.1854684512428297, + "grad_norm": 0.26953125, + "learning_rate": 0.0002812475547086401, + "loss": 0.7215, "step": 620 }, { - "epoch": 0.5636856368563685, - "grad_norm": 0.2578125, - "learning_rate": 0.0002972518781930857, - "loss": 0.8371, + "epoch": 1.1931166347992352, + "grad_norm": 0.263671875, + "learning_rate": 0.00028099494338853554, + "loss": 0.6863, "step": 624 }, { - "epoch": 0.5672990063233966, - "grad_norm": 0.275390625, - "learning_rate": 0.00029721075946805345, - "loss": 0.8103, + "epoch": 1.2007648183556405, + "grad_norm": 0.240234375, + "learning_rate": 0.0002807407570071832, + "loss": 0.7432, "step": 628 }, { - "epoch": 0.5709123757904245, - "grad_norm": 0.259765625, - "learning_rate": 0.0002971693382905107, - "loss": 0.8604, + "epoch": 1.2084130019120458, + "grad_norm": 0.244140625, + "learning_rate": 0.00028048499862087757, + "loss": 0.7265, "step": 632 }, { - "epoch": 0.5745257452574526, - "grad_norm": 0.267578125, - "learning_rate": 0.0002971276147455594, - "loss": 0.8282, + "epoch": 1.2160611854684513, + "grad_norm": 0.259765625, + "learning_rate": 0.00028022767130481466, + "loss": 0.6848, "step": 636 }, { - "epoch": 0.5781391147244805, - "grad_norm": 0.2890625, - "learning_rate": 0.00029708558891892284, - "loss": 0.8315, + "epoch": 1.2237093690248566, + "grad_norm": 0.2470703125, + "learning_rate": 0.0002799687781530549, + "loss": 0.717, "step": 640 }, { - "epoch": 0.5817524841915086, - "grad_norm": 0.24609375, - "learning_rate": 0.0002970432608969454, - "loss": 0.856, + "epoch": 1.231357552581262, + "grad_norm": 0.259765625, + "learning_rate": 0.00027970832227848627, + "loss": 0.7011, "step": 644 }, { - "epoch": 0.5853658536585366, - "grad_norm": 0.25390625, - "learning_rate": 0.0002970006307665923, - "loss": 0.8755, + "epoch": 1.2390057361376674, + "grad_norm": 0.2578125, + "learning_rate": 0.0002794463068127866, + "loss": 0.7696, "step": 648 }, { - "epoch": 0.5889792231255646, - "grad_norm": 0.26953125, - "learning_rate": 0.00029695769861544944, - "loss": 0.8204, + "epoch": 1.2466539196940727, + "grad_norm": 0.2431640625, + "learning_rate": 0.00027918273490638574, + "loss": 0.6922, "step": 652 }, { - "epoch": 0.5925925925925926, - "grad_norm": 0.2578125, - "learning_rate": 0.0002969144645317233, - "loss": 0.8371, + "epoch": 1.254302103250478, + "grad_norm": 0.28515625, + "learning_rate": 0.0002789176097284283, + "loss": 0.6521, "step": 656 }, { - "epoch": 0.5962059620596206, - "grad_norm": 0.2451171875, - "learning_rate": 0.00029687092860424067, - "loss": 0.8409, + "epoch": 1.2619502868068833, + "grad_norm": 0.25390625, + "learning_rate": 0.0002786509344667349, + "loss": 0.6642, "step": 660 }, { - "epoch": 0.5998193315266486, - "grad_norm": 0.291015625, - "learning_rate": 0.0002968270909224484, - "loss": 0.8478, + "epoch": 1.2695984703632888, + "grad_norm": 0.25390625, + "learning_rate": 0.0002783827123277643, + "loss": 0.7773, "step": 664 }, { - "epoch": 0.6034327009936766, - "grad_norm": 0.2333984375, - "learning_rate": 0.0002967829515764134, - "loss": 0.8357, + "epoch": 1.277246653919694, + "grad_norm": 0.2431640625, + "learning_rate": 0.00027811294653657444, + "loss": 0.7314, "step": 668 }, { - "epoch": 0.6070460704607046, - "grad_norm": 0.267578125, - "learning_rate": 0.00029673851065682244, - "loss": 0.775, + "epoch": 1.2848948374760996, + "grad_norm": 0.2373046875, + "learning_rate": 0.000277841640336784, + "loss": 0.7461, "step": 672 }, { - "epoch": 0.6106594399277326, - "grad_norm": 0.265625, - "learning_rate": 0.00029669376825498177, - "loss": 0.8624, + "epoch": 1.2925430210325048, + "grad_norm": 0.236328125, + "learning_rate": 0.00027756879699053337, + "loss": 0.7426, "step": 676 }, { - "epoch": 0.6142728093947606, - "grad_norm": 0.255859375, - "learning_rate": 0.00029664872446281704, - "loss": 0.8696, + "epoch": 1.3001912045889101, + "grad_norm": 0.26171875, + "learning_rate": 0.0002772944197784451, + "loss": 0.6986, "step": 680 }, { - "epoch": 0.6178861788617886, - "grad_norm": 0.25, - "learning_rate": 0.0002966033793728732, - "loss": 0.8846, + "epoch": 1.3078393881453154, + "grad_norm": 0.2412109375, + "learning_rate": 0.0002770185119995849, + "loss": 0.7379, "step": 684 }, { - "epoch": 0.6214995483288166, - "grad_norm": 0.2451171875, - "learning_rate": 0.0002965577330783142, - "loss": 0.8533, + "epoch": 1.3154875717017207, + "grad_norm": 0.265625, + "learning_rate": 0.0002767410769714216, + "loss": 0.7146, "step": 688 }, { - "epoch": 0.6251129177958447, - "grad_norm": 0.2470703125, - "learning_rate": 0.00029651178567292294, - "loss": 0.8841, + "epoch": 1.3231357552581262, + "grad_norm": 0.2490234375, + "learning_rate": 0.0002764621180297875, + "loss": 0.7061, "step": 692 }, { - "epoch": 0.6287262872628726, - "grad_norm": 0.255859375, - "learning_rate": 0.0002964655372511007, - "loss": 0.8371, + "epoch": 1.3307839388145315, + "grad_norm": 0.271484375, + "learning_rate": 0.0002761816385288382, + "loss": 0.6547, "step": 696 }, { - "epoch": 0.6323396567299007, - "grad_norm": 0.259765625, - "learning_rate": 0.0002964189879078674, - "loss": 0.8183, + "epoch": 1.338432122370937, + "grad_norm": 0.271484375, + "learning_rate": 0.0002758996418410122, + "loss": 0.7018, "step": 700 }, { - "epoch": 0.6359530261969286, - "grad_norm": 0.2373046875, - "learning_rate": 0.0002963721377388613, - "loss": 0.7912, + "epoch": 1.3460803059273423, + "grad_norm": 0.2578125, + "learning_rate": 0.0002756161313569904, + "loss": 0.7062, "step": 704 }, { - "epoch": 0.6395663956639567, - "grad_norm": 0.263671875, - "learning_rate": 0.0002963249868403385, - "loss": 0.8244, + "epoch": 1.3537284894837476, + "grad_norm": 0.271484375, + "learning_rate": 0.00027533111048565537, + "loss": 0.778, "step": 708 }, { - "epoch": 0.6431797651309846, - "grad_norm": 0.255859375, - "learning_rate": 0.00029627753530917325, - "loss": 0.852, + "epoch": 1.3613766730401529, + "grad_norm": 0.271484375, + "learning_rate": 0.00027504458265405034, + "loss": 0.6916, "step": 712 }, { - "epoch": 0.6467931345980127, - "grad_norm": 0.259765625, - "learning_rate": 0.0002962297832428571, - "loss": 0.8214, + "epoch": 1.3690248565965584, + "grad_norm": 0.2333984375, + "learning_rate": 0.00027475655130733786, + "loss": 0.6862, "step": 716 }, { - "epoch": 0.6504065040650406, - "grad_norm": 0.271484375, - "learning_rate": 0.00029618173073949936, - "loss": 0.8618, + "epoch": 1.3766730401529637, + "grad_norm": 0.267578125, + "learning_rate": 0.00027446701990875864, + "loss": 0.7037, "step": 720 }, { - "epoch": 0.6540198735320687, - "grad_norm": 0.2275390625, - "learning_rate": 0.0002961333778978266, - "loss": 0.9167, + "epoch": 1.384321223709369, + "grad_norm": 0.2431640625, + "learning_rate": 0.00027417599193958964, + "loss": 0.6976, "step": 724 }, { - "epoch": 0.6576332429990966, - "grad_norm": 0.267578125, - "learning_rate": 0.0002960847248171822, - "loss": 0.8186, + "epoch": 1.3919694072657744, + "grad_norm": 0.24609375, + "learning_rate": 0.00027388347089910253, + "loss": 0.7209, "step": 728 }, { - "epoch": 0.6612466124661247, - "grad_norm": 0.234375, - "learning_rate": 0.00029603577159752664, - "loss": 0.8127, + "epoch": 1.3996175908221797, + "grad_norm": 0.251953125, + "learning_rate": 0.0002735894603045211, + "loss": 0.7009, "step": 732 }, { - "epoch": 0.6648599819331527, - "grad_norm": 0.27734375, - "learning_rate": 0.00029598651833943705, - "loss": 0.8128, + "epoch": 1.407265774378585, + "grad_norm": 0.2578125, + "learning_rate": 0.0002732939636909796, + "loss": 0.6583, "step": 736 }, { - "epoch": 0.6684733514001807, - "grad_norm": 0.25, - "learning_rate": 0.0002959369651441068, - "loss": 0.8373, + "epoch": 1.4149139579349903, + "grad_norm": 0.255859375, + "learning_rate": 0.00027299698461147966, + "loss": 0.6999, "step": 740 }, { - "epoch": 0.6720867208672087, - "grad_norm": 0.240234375, - "learning_rate": 0.0002958871121133456, - "loss": 0.7941, + "epoch": 1.4225621414913958, + "grad_norm": 0.267578125, + "learning_rate": 0.0002726985266368481, + "loss": 0.7269, "step": 744 }, { - "epoch": 0.6757000903342367, - "grad_norm": 0.283203125, - "learning_rate": 0.0002958369593495794, - "loss": 0.8336, + "epoch": 1.430210325047801, + "grad_norm": 0.2392578125, + "learning_rate": 0.0002723985933556936, + "loss": 0.6256, "step": 748 }, { - "epoch": 0.6793134598012647, - "grad_norm": 0.24609375, - "learning_rate": 0.00029578650695584956, - "loss": 0.8253, + "epoch": 1.4378585086042066, + "grad_norm": 0.271484375, + "learning_rate": 0.00027209718837436353, + "loss": 0.7129, "step": 752 }, { - "epoch": 0.6829268292682927, - "grad_norm": 0.2392578125, - "learning_rate": 0.0002957357550358134, - "loss": 0.7859, + "epoch": 1.445506692160612, + "grad_norm": 0.271484375, + "learning_rate": 0.000271794315316901, + "loss": 0.6623, "step": 756 }, { - "epoch": 0.6865401987353207, - "grad_norm": 0.27734375, - "learning_rate": 0.00029568470369374357, - "loss": 0.8203, + "epoch": 1.4531548757170172, + "grad_norm": 0.259765625, + "learning_rate": 0.00027148997782500085, + "loss": 0.6869, "step": 760 }, { - "epoch": 0.6901535682023487, - "grad_norm": 0.255859375, - "learning_rate": 0.0002956333530345276, - "loss": 0.8076, + "epoch": 1.4608030592734225, + "grad_norm": 0.2451171875, + "learning_rate": 0.0002711841795579661, + "loss": 0.7426, "step": 764 }, { - "epoch": 0.6937669376693767, - "grad_norm": 0.2353515625, - "learning_rate": 0.0002955817031636684, - "loss": 0.7564, + "epoch": 1.468451242829828, + "grad_norm": 0.2578125, + "learning_rate": 0.00027087692419266383, + "loss": 0.6731, "step": 768 }, { - "epoch": 0.6973803071364046, - "grad_norm": 0.287109375, - "learning_rate": 0.00029552975418728337, - "loss": 0.8361, + "epoch": 1.4760994263862333, + "grad_norm": 0.275390625, + "learning_rate": 0.00027056821542348114, + "loss": 0.7591, "step": 772 }, { - "epoch": 0.7009936766034327, - "grad_norm": 0.259765625, - "learning_rate": 0.0002954775062121045, - "loss": 0.8071, + "epoch": 1.4837476099426385, + "grad_norm": 0.25, + "learning_rate": 0.0002702580569622805, + "loss": 0.7129, "step": 776 }, { - "epoch": 0.7046070460704607, - "grad_norm": 0.259765625, - "learning_rate": 0.00029542495934547813, - "loss": 0.8453, + "epoch": 1.491395793499044, + "grad_norm": 0.2490234375, + "learning_rate": 0.0002699464525383552, + "loss": 0.7307, "step": 780 }, { - "epoch": 0.7082204155374887, - "grad_norm": 0.263671875, - "learning_rate": 0.0002953721136953646, - "loss": 0.8342, + "epoch": 1.4990439770554493, + "grad_norm": 0.2578125, + "learning_rate": 0.0002696334058983848, + "loss": 0.7317, "step": 784 }, { - "epoch": 0.7118337850045167, - "grad_norm": 0.27734375, - "learning_rate": 0.00029531896937033827, - "loss": 0.8209, + "epoch": 1.5066921606118546, + "grad_norm": 0.2578125, + "learning_rate": 0.0002693189208063894, + "loss": 0.6994, "step": 788 }, { - "epoch": 0.7154471544715447, - "grad_norm": 0.2451171875, - "learning_rate": 0.00029526552647958694, - "loss": 0.8183, + "epoch": 1.51434034416826, + "grad_norm": 0.255859375, + "learning_rate": 0.00026900300104368524, + "loss": 0.72, "step": 792 }, { - "epoch": 0.7190605239385727, + "epoch": 1.5219885277246654, "grad_norm": 0.259765625, - "learning_rate": 0.000295211785132912, - "loss": 0.8761, + "learning_rate": 0.0002686856504088385, + "loss": 0.7112, "step": 796 }, { - "epoch": 0.7226738934056007, - "grad_norm": 0.2890625, - "learning_rate": 0.00029515774544072796, - "loss": 0.85, + "epoch": 1.5296367112810707, + "grad_norm": 0.2490234375, + "learning_rate": 0.00026836687271762015, + "loss": 0.6912, "step": 800 }, { - "epoch": 0.7262872628726287, - "grad_norm": 0.251953125, - "learning_rate": 0.0002951034075140623, - "loss": 0.8058, + "epoch": 1.5372848948374762, + "grad_norm": 0.2421875, + "learning_rate": 0.0002680466718029596, + "loss": 0.6801, "step": 804 }, { - "epoch": 0.7299006323396567, - "grad_norm": 0.267578125, - "learning_rate": 0.00029504877146455523, - "loss": 0.8008, + "epoch": 1.5449330783938815, + "grad_norm": 0.265625, + "learning_rate": 0.00026772505151489897, + "loss": 0.7077, "step": 808 }, { - "epoch": 0.7335140018066847, - "grad_norm": 0.26171875, - "learning_rate": 0.00029499383740445943, - "loss": 0.8391, + "epoch": 1.5525812619502868, + "grad_norm": 0.283203125, + "learning_rate": 0.00026740201572054685, + "loss": 0.6926, "step": 812 }, { - "epoch": 0.7371273712737128, - "grad_norm": 0.2470703125, - "learning_rate": 0.00029493860544664005, - "loss": 0.816, + "epoch": 1.560229445506692, + "grad_norm": 0.27734375, + "learning_rate": 0.00026707756830403144, + "loss": 0.702, "step": 816 }, { - "epoch": 0.7407407407407407, - "grad_norm": 0.2734375, - "learning_rate": 0.0002948830757045741, - "loss": 0.8301, + "epoch": 1.5678776290630974, + "grad_norm": 0.271484375, + "learning_rate": 0.00026675171316645403, + "loss": 0.7178, "step": 820 }, { - "epoch": 0.7443541102077688, - "grad_norm": 0.25390625, - "learning_rate": 0.00029482724829235033, - "loss": 0.8092, + "epoch": 1.5755258126195029, + "grad_norm": 0.26171875, + "learning_rate": 0.00026642445422584224, + "loss": 0.6843, "step": 824 }, { - "epoch": 0.7479674796747967, - "grad_norm": 0.29296875, - "learning_rate": 0.0002947711233246694, - "loss": 0.8431, + "epoch": 1.5831739961759084, + "grad_norm": 0.25, + "learning_rate": 0.0002660957954171028, + "loss": 0.6722, "step": 828 }, { - "epoch": 0.7515808491418248, - "grad_norm": 0.232421875, - "learning_rate": 0.00029471470091684297, - "loss": 0.8175, + "epoch": 1.5908221797323137, + "grad_norm": 0.251953125, + "learning_rate": 0.00026576574069197406, + "loss": 0.6518, "step": 832 }, { - "epoch": 0.7551942186088527, - "grad_norm": 0.2578125, - "learning_rate": 0.000294657981184794, - "loss": 0.7798, + "epoch": 1.598470363288719, + "grad_norm": 0.263671875, + "learning_rate": 0.00026543429401897875, + "loss": 0.6998, "step": 836 }, { - "epoch": 0.7588075880758808, + "epoch": 1.6061185468451242, "grad_norm": 0.26171875, - "learning_rate": 0.00029460096424505626, - "loss": 0.8456, + "learning_rate": 0.0002651014593833762, + "loss": 0.6966, "step": 840 }, { - "epoch": 0.7624209575429087, - "grad_norm": 0.240234375, - "learning_rate": 0.0002945436502147741, - "loss": 0.8109, + "epoch": 1.6137667304015295, + "grad_norm": 0.25390625, + "learning_rate": 0.00026476724078711416, + "loss": 0.7054, "step": 844 }, { - "epoch": 0.7660343270099368, - "grad_norm": 0.26953125, - "learning_rate": 0.0002944860392117023, - "loss": 0.8604, + "epoch": 1.621414913957935, + "grad_norm": 0.279296875, + "learning_rate": 0.00026443164224878115, + "loss": 0.6655, "step": 848 }, { - "epoch": 0.7696476964769647, - "grad_norm": 0.251953125, - "learning_rate": 0.00029442813135420595, - "loss": 0.8226, + "epoch": 1.6290630975143403, + "grad_norm": 0.275390625, + "learning_rate": 0.0002640946678035576, + "loss": 0.7098, "step": 852 }, { - "epoch": 0.7732610659439928, - "grad_norm": 0.2412109375, - "learning_rate": 0.00029436992676125975, - "loss": 0.8219, + "epoch": 1.6367112810707458, + "grad_norm": 0.2451171875, + "learning_rate": 0.0002637563215031679, + "loss": 0.7099, "step": 856 }, { - "epoch": 0.7768744354110207, - "grad_norm": 0.236328125, - "learning_rate": 0.0002943114255524482, - "loss": 0.8126, + "epoch": 1.644359464627151, + "grad_norm": 0.2470703125, + "learning_rate": 0.00026341660741583127, + "loss": 0.6704, "step": 860 }, { - "epoch": 0.7804878048780488, - "grad_norm": 0.265625, - "learning_rate": 0.00029425262784796535, - "loss": 0.7987, + "epoch": 1.6520076481835564, + "grad_norm": 0.27734375, + "learning_rate": 0.00026307552962621293, + "loss": 0.686, "step": 864 }, { - "epoch": 0.7841011743450768, - "grad_norm": 0.248046875, - "learning_rate": 0.00029419353376861415, - "loss": 0.8254, + "epoch": 1.6596558317399617, + "grad_norm": 0.2470703125, + "learning_rate": 0.00026273309223537507, + "loss": 0.69, "step": 868 }, { - "epoch": 0.7877145438121048, - "grad_norm": 0.2412109375, - "learning_rate": 0.0002941341434358066, - "loss": 0.8161, + "epoch": 1.667304015296367, + "grad_norm": 0.275390625, + "learning_rate": 0.0002623892993607275, + "loss": 0.6957, "step": 872 }, { - "epoch": 0.7913279132791328, - "grad_norm": 0.2734375, - "learning_rate": 0.0002940744569715634, - "loss": 0.7733, + "epoch": 1.6749521988527725, + "grad_norm": 0.267578125, + "learning_rate": 0.00026204415513597813, + "loss": 0.7071, "step": 876 }, { - "epoch": 0.7949412827461608, - "grad_norm": 0.2734375, - "learning_rate": 0.0002940144744985137, - "loss": 0.8475, + "epoch": 1.682600382409178, + "grad_norm": 0.26953125, + "learning_rate": 0.0002616976637110832, + "loss": 0.6313, "step": 880 }, { - "epoch": 0.7985546522131888, - "grad_norm": 0.255859375, - "learning_rate": 0.0002939541961398947, - "loss": 0.8765, + "epoch": 1.6902485659655833, + "grad_norm": 0.267578125, + "learning_rate": 0.0002613498292521977, + "loss": 0.6809, "step": 884 }, { - "epoch": 0.8021680216802168, - "grad_norm": 0.265625, - "learning_rate": 0.0002938936220195515, - "loss": 0.847, + "epoch": 1.6978967495219885, + "grad_norm": 0.275390625, + "learning_rate": 0.00026100065594162475, + "loss": 0.6867, "step": 888 }, { - "epoch": 0.8057813911472448, - "grad_norm": 0.2734375, - "learning_rate": 0.0002938327522619371, - "loss": 0.8139, + "epoch": 1.7055449330783938, + "grad_norm": 0.26171875, + "learning_rate": 0.00026065014797776575, + "loss": 0.7065, "step": 892 }, { - "epoch": 0.8093947606142728, - "grad_norm": 0.2470703125, - "learning_rate": 0.0002937715869921115, - "loss": 0.8334, + "epoch": 1.7131931166347991, + "grad_norm": 0.296875, + "learning_rate": 0.0002602983095750698, + "loss": 0.6938, "step": 896 }, { - "epoch": 0.8130081300813008, - "grad_norm": 0.27734375, - "learning_rate": 0.00029371012633574225, - "loss": 0.8142, + "epoch": 1.7208413001912046, + "grad_norm": 0.25, + "learning_rate": 0.0002599451449639828, + "loss": 0.7138, "step": 900 }, { - "epoch": 0.8166214995483289, - "grad_norm": 0.2734375, - "learning_rate": 0.0002936483704191035, - "loss": 0.8294, + "epoch": 1.72848948374761, + "grad_norm": 0.2421875, + "learning_rate": 0.00025959065839089684, + "loss": 0.6976, "step": 904 }, { - "epoch": 0.8202348690153568, - "grad_norm": 0.2578125, - "learning_rate": 0.00029358631936907625, - "loss": 0.8177, + "epoch": 1.7361376673040154, + "grad_norm": 0.2431640625, + "learning_rate": 0.00025923485411809917, + "loss": 0.6792, "step": 908 }, { - "epoch": 0.8238482384823849, - "grad_norm": 0.267578125, - "learning_rate": 0.00029352397331314765, - "loss": 0.8487, + "epoch": 1.7437858508604207, + "grad_norm": 0.25390625, + "learning_rate": 0.00025887773642372064, + "loss": 0.7016, "step": 912 }, { - "epoch": 0.8274616079494128, - "grad_norm": 0.265625, - "learning_rate": 0.0002934613323794111, - "loss": 0.7998, + "epoch": 1.751434034416826, + "grad_norm": 0.2578125, + "learning_rate": 0.00025851930960168464, + "loss": 0.6845, "step": 916 }, { - "epoch": 0.8310749774164409, - "grad_norm": 0.2734375, - "learning_rate": 0.00029339839669656584, - "loss": 0.8358, + "epoch": 1.7590822179732313, + "grad_norm": 0.267578125, + "learning_rate": 0.0002581595779616552, + "loss": 0.6932, "step": 920 }, { - "epoch": 0.8346883468834688, + "epoch": 1.7667304015296366, "grad_norm": 0.263671875, - "learning_rate": 0.0002933351663939166, - "loss": 0.8398, + "learning_rate": 0.0002577985458289852, + "loss": 0.6911, "step": 924 }, { - "epoch": 0.8383017163504969, - "grad_norm": 0.244140625, - "learning_rate": 0.0002932716416013735, - "loss": 0.8188, + "epoch": 1.774378585086042, + "grad_norm": 0.259765625, + "learning_rate": 0.00025743621754466457, + "loss": 0.6611, "step": 928 }, { - "epoch": 0.8419150858175248, - "grad_norm": 0.2451171875, - "learning_rate": 0.00029320782244945167, - "loss": 0.841, + "epoch": 1.7820267686424476, + "grad_norm": 0.251953125, + "learning_rate": 0.0002570725974652679, + "loss": 0.7158, "step": 932 }, { - "epoch": 0.8455284552845529, - "grad_norm": 0.240234375, - "learning_rate": 0.00029314370906927107, - "loss": 0.8309, + "epoch": 1.7896749521988529, + "grad_norm": 0.2490234375, + "learning_rate": 0.0002567076899629021, + "loss": 0.7176, "step": 936 }, { - "epoch": 0.8491418247515808, - "grad_norm": 0.26171875, - "learning_rate": 0.000293079301592556, - "loss": 0.8081, + "epoch": 1.7973231357552581, + "grad_norm": 0.251953125, + "learning_rate": 0.0002563414994251538, + "loss": 0.7001, "step": 940 }, { - "epoch": 0.8527551942186089, - "grad_norm": 0.26953125, - "learning_rate": 0.0002930146001516353, - "loss": 0.8106, + "epoch": 1.8049713193116634, + "grad_norm": 0.267578125, + "learning_rate": 0.0002559740302550366, + "loss": 0.7583, "step": 944 }, { - "epoch": 0.8563685636856369, - "grad_norm": 0.255859375, - "learning_rate": 0.0002929496048794415, - "loss": 0.8113, + "epoch": 1.8126195028680687, + "grad_norm": 0.28125, + "learning_rate": 0.0002556052868709383, + "loss": 0.6374, "step": 948 }, { - "epoch": 0.8599819331526649, - "grad_norm": 0.2470703125, - "learning_rate": 0.00029288431590951094, - "loss": 0.8077, + "epoch": 1.8202676864244742, + "grad_norm": 0.279296875, + "learning_rate": 0.00025523527370656753, + "loss": 0.6771, "step": 952 }, { - "epoch": 0.8635953026196929, - "grad_norm": 0.26171875, - "learning_rate": 0.00029281873337598334, - "loss": 0.8226, + "epoch": 1.8279158699808795, + "grad_norm": 0.263671875, + "learning_rate": 0.0002548639952109006, + "loss": 0.6547, "step": 956 }, { - "epoch": 0.8672086720867209, - "grad_norm": 0.275390625, - "learning_rate": 0.0002927528574136016, - "loss": 0.7979, + "epoch": 1.835564053537285, + "grad_norm": 0.255859375, + "learning_rate": 0.0002544914558481279, + "loss": 0.7078, "step": 960 }, { - "epoch": 0.8708220415537489, - "grad_norm": 0.271484375, - "learning_rate": 0.0002926866881577115, - "loss": 0.7547, + "epoch": 1.8432122370936903, + "grad_norm": 0.23828125, + "learning_rate": 0.00025411766009760027, + "loss": 0.6913, "step": 964 }, { - "epoch": 0.8744354110207768, - "grad_norm": 0.25390625, - "learning_rate": 0.0002926202257442613, - "loss": 0.7724, + "epoch": 1.8508604206500956, + "grad_norm": 0.26171875, + "learning_rate": 0.00025374261245377525, + "loss": 0.725, "step": 968 }, { - "epoch": 0.8780487804878049, - "grad_norm": 0.244140625, - "learning_rate": 0.00029255347030980166, - "loss": 0.8242, + "epoch": 1.8585086042065009, + "grad_norm": 0.251953125, + "learning_rate": 0.0002533663174261628, + "loss": 0.6871, "step": 972 }, { - "epoch": 0.8816621499548328, - "grad_norm": 0.255859375, - "learning_rate": 0.0002924864219914853, - "loss": 0.7917, + "epoch": 1.8661567877629062, + "grad_norm": 0.283203125, + "learning_rate": 0.0002529887795392713, + "loss": 0.6748, "step": 976 }, { - "epoch": 0.8852755194218609, - "grad_norm": 0.25390625, - "learning_rate": 0.0002924190809270666, - "loss": 0.8284, + "epoch": 1.8738049713193117, + "grad_norm": 0.28515625, + "learning_rate": 0.00025261000333255305, + "loss": 0.6224, "step": 980 }, { - "epoch": 0.8888888888888888, - "grad_norm": 0.271484375, - "learning_rate": 0.00029235144725490146, - "loss": 0.7862, + "epoch": 1.8814531548757172, + "grad_norm": 0.255859375, + "learning_rate": 0.0002522299933603497, + "loss": 0.6914, "step": 984 }, { - "epoch": 0.8925022583559169, - "grad_norm": 0.2734375, - "learning_rate": 0.0002922835211139469, - "loss": 0.8303, + "epoch": 1.8891013384321225, + "grad_norm": 0.26953125, + "learning_rate": 0.0002518487541918374, + "loss": 0.684, "step": 988 }, { - "epoch": 0.8961156278229448, - "grad_norm": 0.27734375, - "learning_rate": 0.00029221530264376093, - "loss": 0.7848, + "epoch": 1.8967495219885278, + "grad_norm": 0.259765625, + "learning_rate": 0.000251466290410972, + "loss": 0.7131, "step": 992 }, { - "epoch": 0.8997289972899729, - "grad_norm": 0.259765625, - "learning_rate": 0.0002921467919845021, - "loss": 0.7719, + "epoch": 1.904397705544933, + "grad_norm": 0.267578125, + "learning_rate": 0.0002510826066164341, + "loss": 0.6956, "step": 996 }, { - "epoch": 0.9033423667570009, - "grad_norm": 0.2578125, - "learning_rate": 0.00029207798927692916, - "loss": 0.8385, + "epoch": 1.9120458891013383, + "grad_norm": 0.2470703125, + "learning_rate": 0.00025069770742157317, + "loss": 0.6683, "step": 1000 }, { - "epoch": 0.9069557362240289, - "grad_norm": 0.28125, - "learning_rate": 0.00029200889466240114, - "loss": 0.808, + "epoch": 1.9196940726577438, + "grad_norm": 0.251953125, + "learning_rate": 0.00025031159745435267, + "loss": 0.6941, "step": 1004 }, { - "epoch": 0.9105691056910569, - "grad_norm": 0.251953125, - "learning_rate": 0.0002919395082828767, - "loss": 0.7802, + "epoch": 1.9273422562141491, + "grad_norm": 0.248046875, + "learning_rate": 0.0002499242813572942, + "loss": 0.7458, "step": 1008 }, { - "epoch": 0.9141824751580849, - "grad_norm": 0.251953125, - "learning_rate": 0.00029186983028091384, - "loss": 0.8207, + "epoch": 1.9349904397705546, + "grad_norm": 0.2734375, + "learning_rate": 0.0002495357637874215, + "loss": 0.6724, "step": 1012 }, { - "epoch": 0.9177958446251129, - "grad_norm": 0.24609375, - "learning_rate": 0.00029179986079966985, - "loss": 0.8426, + "epoch": 1.94263862332696, + "grad_norm": 0.240234375, + "learning_rate": 0.0002491460494162048, + "loss": 0.6662, "step": 1016 }, { - "epoch": 0.9214092140921409, - "grad_norm": 0.275390625, - "learning_rate": 0.00029172959998290083, - "loss": 0.794, + "epoch": 1.9502868068833652, + "grad_norm": 0.279296875, + "learning_rate": 0.00024875514292950447, + "loss": 0.652, "step": 1020 }, { - "epoch": 0.9250225835591689, - "grad_norm": 0.275390625, - "learning_rate": 0.00029165904797496145, - "loss": 0.8051, + "epoch": 1.9579349904397705, + "grad_norm": 0.263671875, + "learning_rate": 0.00024836304902751445, + "loss": 0.7223, "step": 1024 }, { - "epoch": 0.928635953026197, - "grad_norm": 0.267578125, - "learning_rate": 0.0002915882049208047, - "loss": 0.834, + "epoch": 1.9655831739961758, + "grad_norm": 0.255859375, + "learning_rate": 0.0002479697724247062, + "loss": 0.7065, "step": 1028 }, { - "epoch": 0.9322493224932249, - "grad_norm": 0.25390625, - "learning_rate": 0.00029151707096598134, - "loss": 0.8063, + "epoch": 1.9732313575525813, + "grad_norm": 0.263671875, + "learning_rate": 0.0002475753178497716, + "loss": 0.7307, "step": 1032 }, { - "epoch": 0.935862691960253, - "grad_norm": 0.271484375, - "learning_rate": 0.0002914456462566401, - "loss": 0.8119, - "step": 1036 + "epoch": 1.9808795411089866, + "grad_norm": 0.279296875, + "learning_rate": 0.00024717969004556646, + "loss": 0.7086, + "step": 1036 }, { - "epoch": 0.9394760614272809, - "grad_norm": 0.26171875, - "learning_rate": 0.00029137393093952685, - "loss": 0.7962, + "epoch": 1.988527724665392, + "grad_norm": 0.259765625, + "learning_rate": 0.0002467828937690532, + "loss": 0.7051, "step": 1040 }, { - "epoch": 0.943089430894309, - "grad_norm": 0.27734375, - "learning_rate": 0.0002913019251619846, - "loss": 0.8037, + "epoch": 1.9961759082217974, + "grad_norm": 0.2734375, + "learning_rate": 0.0002463849337912437, + "loss": 0.7084, "step": 1044 }, { - "epoch": 0.9467028003613369, - "grad_norm": 0.28125, - "learning_rate": 0.0002912296290719532, - "loss": 0.803, + "epoch": 2.0038240917782026, + "grad_norm": 0.236328125, + "learning_rate": 0.00024598581489714206, + "loss": 0.5785, "step": 1048 }, { - "epoch": 0.950316169828365, - "grad_norm": 0.2578125, - "learning_rate": 0.0002911570428179689, - "loss": 0.8887, + "epoch": 2.011472275334608, + "grad_norm": 0.28125, + "learning_rate": 0.0002455855418856869, + "loss": 0.5793, "step": 1052 }, { - "epoch": 0.9539295392953929, - "grad_norm": 0.28125, - "learning_rate": 0.00029108416654916405, - "loss": 0.772, + "epoch": 2.019120458891013, + "grad_norm": 0.259765625, + "learning_rate": 0.000245184119569694, + "loss": 0.6212, "step": 1056 }, { - "epoch": 0.957542908762421, - "grad_norm": 0.259765625, - "learning_rate": 0.0002910110004152669, - "loss": 0.7948, + "epoch": 2.026768642447419, + "grad_norm": 0.26953125, + "learning_rate": 0.0002447815527757979, + "loss": 0.569, "step": 1060 }, { - "epoch": 0.9611562782294489, - "grad_norm": 0.251953125, - "learning_rate": 0.0002909375445666013, - "loss": 0.7825, + "epoch": 2.0344168260038242, + "grad_norm": 0.26953125, + "learning_rate": 0.0002443778463443944, + "loss": 0.5868, "step": 1064 }, { - "epoch": 0.964769647696477, - "grad_norm": 0.255859375, - "learning_rate": 0.00029086379915408626, - "loss": 0.8405, + "epoch": 2.0420650095602295, + "grad_norm": 0.236328125, + "learning_rate": 0.0002439730051295818, + "loss": 0.6195, "step": 1068 }, { - "epoch": 0.9683830171635049, - "grad_norm": 0.26171875, - "learning_rate": 0.00029078976432923576, - "loss": 0.7632, + "epoch": 2.049713193116635, + "grad_norm": 0.279296875, + "learning_rate": 0.0002435670339991031, + "loss": 0.5886, "step": 1072 }, { - "epoch": 0.971996386630533, - "grad_norm": 0.265625, - "learning_rate": 0.0002907154402441583, - "loss": 0.8101, + "epoch": 2.05736137667304, + "grad_norm": 0.28125, + "learning_rate": 0.00024315993783428718, + "loss": 0.6097, "step": 1076 }, { - "epoch": 0.975609756097561, - "grad_norm": 0.275390625, - "learning_rate": 0.00029064082705155685, - "loss": 0.7764, + "epoch": 2.0650095602294454, + "grad_norm": 0.3203125, + "learning_rate": 0.00024275172152999006, + "loss": 0.5727, "step": 1080 }, { - "epoch": 0.979223125564589, - "grad_norm": 0.2578125, - "learning_rate": 0.0002905659249047283, - "loss": 0.7553, + "epoch": 2.0726577437858507, + "grad_norm": 0.275390625, + "learning_rate": 0.00024234238999453614, + "loss": 0.6085, "step": 1084 }, { - "epoch": 0.982836495031617, - "grad_norm": 0.283203125, - "learning_rate": 0.00029049073395756304, - "loss": 0.7814, + "epoch": 2.0803059273422564, + "grad_norm": 0.29296875, + "learning_rate": 0.00024193194814965934, + "loss": 0.6144, "step": 1088 }, { - "epoch": 0.986449864498645, - "grad_norm": 0.265625, - "learning_rate": 0.00029041525436454503, - "loss": 0.7669, + "epoch": 2.0879541108986617, + "grad_norm": 0.279296875, + "learning_rate": 0.00024152040093044353, + "loss": 0.5868, "step": 1092 }, { - "epoch": 0.990063233965673, - "grad_norm": 0.24609375, - "learning_rate": 0.0002903394862807512, - "loss": 0.8187, + "epoch": 2.095602294455067, + "grad_norm": 0.2734375, + "learning_rate": 0.00024110775328526352, + "loss": 0.6278, "step": 1096 }, { - "epoch": 0.993676603432701, - "grad_norm": 0.3046875, - "learning_rate": 0.00029026342986185127, - "loss": 0.7845, + "epoch": 2.1032504780114722, + "grad_norm": 0.275390625, + "learning_rate": 0.00024069401017572543, + "loss": 0.5923, "step": 1100 }, { - "epoch": 0.997289972899729, - "grad_norm": 0.255859375, - "learning_rate": 0.00029018708526410715, - "loss": 0.7785, + "epoch": 2.1108986615678775, + "grad_norm": 0.265625, + "learning_rate": 0.00024027917657660713, + "loss": 0.5759, "step": 1104 }, { - "epoch": 1.000903342366757, - "grad_norm": 0.26171875, - "learning_rate": 0.00029011045264437305, - "loss": 0.7601, + "epoch": 2.118546845124283, + "grad_norm": 0.2734375, + "learning_rate": 0.00023986325747579824, + "loss": 0.6138, "step": 1108 }, { - "epoch": 1.004516711833785, - "grad_norm": 0.275390625, - "learning_rate": 0.0002900335321600949, - "loss": 0.7283, + "epoch": 2.126195028680688, + "grad_norm": 0.28515625, + "learning_rate": 0.0002394462578742403, + "loss": 0.5786, "step": 1112 }, { - "epoch": 1.008130081300813, - "grad_norm": 0.25, - "learning_rate": 0.0002899563239693099, - "loss": 0.7177, + "epoch": 2.133843212237094, + "grad_norm": 0.267578125, + "learning_rate": 0.0002390281827858668, + "loss": 0.64, "step": 1116 }, { - "epoch": 1.0117434507678411, - "grad_norm": 0.2451171875, - "learning_rate": 0.0002898788282306466, - "loss": 0.7408, + "epoch": 2.141491395793499, + "grad_norm": 0.279296875, + "learning_rate": 0.0002386090372375424, + "loss": 0.6307, "step": 1120 }, { - "epoch": 1.015356820234869, - "grad_norm": 0.259765625, - "learning_rate": 0.00028980104510332414, - "loss": 0.7268, + "epoch": 2.1491395793499044, + "grad_norm": 0.2578125, + "learning_rate": 0.00023818882626900294, + "loss": 0.5641, "step": 1124 }, { - "epoch": 1.018970189701897, - "grad_norm": 0.263671875, - "learning_rate": 0.0002897229747471521, - "loss": 0.7105, + "epoch": 2.1567877629063097, + "grad_norm": 0.28125, + "learning_rate": 0.00023776755493279473, + "loss": 0.623, "step": 1128 }, { - "epoch": 1.022583559168925, - "grad_norm": 0.271484375, - "learning_rate": 0.0002896446173225304, - "loss": 0.7627, + "epoch": 2.164435946462715, + "grad_norm": 0.275390625, + "learning_rate": 0.00023734522829421372, + "loss": 0.6022, "step": 1132 }, { - "epoch": 1.0261969286359531, - "grad_norm": 0.251953125, - "learning_rate": 0.0002895659729904487, - "loss": 0.7362, + "epoch": 2.1720841300191203, + "grad_norm": 0.2734375, + "learning_rate": 0.00023692185143124464, + "loss": 0.6121, "step": 1136 }, { - "epoch": 1.029810298102981, - "grad_norm": 0.255859375, - "learning_rate": 0.00028948704191248585, - "loss": 0.7749, + "epoch": 2.179732313575526, + "grad_norm": 0.267578125, + "learning_rate": 0.00023649742943449996, + "loss": 0.5878, "step": 1140 }, { - "epoch": 1.033423667570009, - "grad_norm": 0.283203125, - "learning_rate": 0.00028940782425081017, - "loss": 0.7324, + "epoch": 2.1873804971319313, + "grad_norm": 0.2578125, + "learning_rate": 0.00023607196740715858, + "loss": 0.6143, "step": 1144 }, { - "epoch": 1.037037037037037, - "grad_norm": 0.25, - "learning_rate": 0.00028932832016817864, - "loss": 0.7389, + "epoch": 2.1950286806883366, + "grad_norm": 0.28515625, + "learning_rate": 0.00023564547046490468, + "loss": 0.5655, "step": 1148 }, { - "epoch": 1.040650406504065, - "grad_norm": 0.2578125, - "learning_rate": 0.0002892485298279367, - "loss": 0.7309, + "epoch": 2.202676864244742, + "grad_norm": 0.26953125, + "learning_rate": 0.00023521794373586603, + "loss": 0.5685, "step": 1152 }, { - "epoch": 1.044263775971093, - "grad_norm": 0.248046875, - "learning_rate": 0.000289168453394018, - "loss": 0.7174, + "epoch": 2.210325047801147, + "grad_norm": 0.283203125, + "learning_rate": 0.00023478939236055228, + "loss": 0.5845, "step": 1156 }, { - "epoch": 1.047877145438121, - "grad_norm": 0.26953125, - "learning_rate": 0.0002890880910309438, - "loss": 0.7076, + "epoch": 2.2179732313575524, + "grad_norm": 0.2890625, + "learning_rate": 0.00023435982149179346, + "loss": 0.6108, "step": 1160 }, { - "epoch": 1.051490514905149, - "grad_norm": 0.26171875, - "learning_rate": 0.00028900744290382313, - "loss": 0.7059, + "epoch": 2.2256214149139577, + "grad_norm": 0.28125, + "learning_rate": 0.0002339292362946777, + "loss": 0.6221, "step": 1164 }, { - "epoch": 1.055103884372177, - "grad_norm": 0.25390625, - "learning_rate": 0.0002889265091783517, - "loss": 0.7081, + "epoch": 2.2332695984703634, + "grad_norm": 0.28515625, + "learning_rate": 0.0002334976419464892, + "loss": 0.5739, "step": 1168 }, { - "epoch": 1.0587172538392051, - "grad_norm": 0.287109375, - "learning_rate": 0.0002888452900208125, - "loss": 0.7433, + "epoch": 2.2409177820267687, + "grad_norm": 0.28515625, + "learning_rate": 0.00023306504363664613, + "loss": 0.5928, "step": 1172 }, { - "epoch": 1.062330623306233, - "grad_norm": 0.2734375, - "learning_rate": 0.00028876378559807464, - "loss": 0.6917, + "epoch": 2.248565965583174, + "grad_norm": 0.259765625, + "learning_rate": 0.00023263144656663801, + "loss": 0.5422, "step": 1176 }, { - "epoch": 1.065943992773261, - "grad_norm": 0.283203125, - "learning_rate": 0.00028868199607759323, - "loss": 0.6994, + "epoch": 2.2562141491395793, + "grad_norm": 0.271484375, + "learning_rate": 0.00023219685594996347, + "loss": 0.5815, "step": 1180 }, { - "epoch": 1.069557362240289, - "grad_norm": 0.25, - "learning_rate": 0.00028859992162740954, - "loss": 0.7196, + "epoch": 2.2638623326959846, + "grad_norm": 0.267578125, + "learning_rate": 0.00023176127701206713, + "loss": 0.5786, "step": 1184 }, { - "epoch": 1.0731707317073171, - "grad_norm": 0.2734375, - "learning_rate": 0.00028851756241614975, - "loss": 0.7514, + "epoch": 2.27151051625239, + "grad_norm": 0.263671875, + "learning_rate": 0.00023132471499027717, + "loss": 0.5634, "step": 1188 }, { - "epoch": 1.076784101174345, - "grad_norm": 0.26171875, - "learning_rate": 0.0002884349186130255, - "loss": 0.7266, + "epoch": 2.2791586998087956, + "grad_norm": 0.291015625, + "learning_rate": 0.0002308871751337422, + "loss": 0.5969, "step": 1192 }, { - "epoch": 1.080397470641373, - "grad_norm": 0.2578125, - "learning_rate": 0.00028835199038783295, - "loss": 0.6985, + "epoch": 2.286806883365201, + "grad_norm": 0.279296875, + "learning_rate": 0.00023044866270336822, + "loss": 0.587, "step": 1196 }, { - "epoch": 1.084010840108401, - "grad_norm": 0.259765625, - "learning_rate": 0.00028826877791095256, - "loss": 0.7516, + "epoch": 2.294455066921606, + "grad_norm": 0.310546875, + "learning_rate": 0.00023000918297175506, + "loss": 0.6312, "step": 1200 }, { - "epoch": 1.0876242095754292, - "grad_norm": 0.298828125, - "learning_rate": 0.000288185281353349, - "loss": 0.677, + "epoch": 2.3021032504780115, + "grad_norm": 0.28515625, + "learning_rate": 0.00022956874122313347, + "loss": 0.6268, "step": 1204 }, { - "epoch": 1.091237579042457, - "grad_norm": 0.26171875, - "learning_rate": 0.00028810150088657047, - "loss": 0.6898, + "epoch": 2.3097514340344167, + "grad_norm": 0.263671875, + "learning_rate": 0.00022912734275330117, + "loss": 0.586, "step": 1208 }, { - "epoch": 1.094850948509485, - "grad_norm": 0.271484375, - "learning_rate": 0.00028801743668274845, - "loss": 0.7318, + "epoch": 2.317399617590822, + "grad_norm": 0.28125, + "learning_rate": 0.00022868499286955943, + "loss": 0.571, "step": 1212 }, { - "epoch": 1.098464317976513, - "grad_norm": 0.25390625, - "learning_rate": 0.0002879330889145974, - "loss": 0.7356, + "epoch": 2.3250478011472273, + "grad_norm": 0.296875, + "learning_rate": 0.00022824169689064915, + "loss": 0.6062, "step": 1216 }, { - "epoch": 1.1020776874435412, - "grad_norm": 0.2578125, - "learning_rate": 0.0002878484577554144, - "loss": 0.7169, + "epoch": 2.332695984703633, + "grad_norm": 0.263671875, + "learning_rate": 0.00022779746014668683, + "loss": 0.5991, "step": 1220 }, { - "epoch": 1.1056910569105691, - "grad_norm": 0.28125, - "learning_rate": 0.0002877635433790789, - "loss": 0.7866, + "epoch": 2.3403441682600383, + "grad_norm": 0.27734375, + "learning_rate": 0.00022735228797910066, + "loss": 0.6193, "step": 1224 }, { - "epoch": 1.109304426377597, - "grad_norm": 0.27734375, - "learning_rate": 0.0002876783459600519, - "loss": 0.7008, + "epoch": 2.3479923518164436, + "grad_norm": 0.2890625, + "learning_rate": 0.0002269061857405662, + "loss": 0.5719, "step": 1228 }, { - "epoch": 1.112917795844625, - "grad_norm": 0.2734375, - "learning_rate": 0.00028759286567337633, - "loss": 0.6943, + "epoch": 2.355640535372849, + "grad_norm": 0.27734375, + "learning_rate": 0.00022645915879494202, + "loss": 0.6171, "step": 1232 }, { - "epoch": 1.1165311653116532, - "grad_norm": 0.28125, - "learning_rate": 0.00028750710269467595, - "loss": 0.6751, + "epoch": 2.363288718929254, + "grad_norm": 0.267578125, + "learning_rate": 0.00022601121251720514, + "loss": 0.6213, "step": 1236 }, { - "epoch": 1.1201445347786811, - "grad_norm": 0.255859375, - "learning_rate": 0.0002874210572001555, - "loss": 0.7603, + "epoch": 2.3709369024856595, + "grad_norm": 0.271484375, + "learning_rate": 0.0002255623522933866, + "loss": 0.5933, "step": 1240 }, { - "epoch": 1.123757904245709, - "grad_norm": 0.275390625, - "learning_rate": 0.00028733472936660014, - "loss": 0.7299, + "epoch": 2.378585086042065, + "grad_norm": 0.2890625, + "learning_rate": 0.00022511258352050649, + "loss": 0.5598, "step": 1244 }, { - "epoch": 1.127371273712737, - "grad_norm": 0.271484375, - "learning_rate": 0.0002872481193713751, - "loss": 0.7056, + "epoch": 2.3862332695984705, + "grad_norm": 0.2890625, + "learning_rate": 0.00022466191160650916, + "loss": 0.6435, "step": 1248 }, { - "epoch": 1.1309846431797652, - "grad_norm": 0.296875, - "learning_rate": 0.00028716122739242533, - "loss": 0.7833, + "epoch": 2.3938814531548758, + "grad_norm": 0.28515625, + "learning_rate": 0.00022421034197019822, + "loss": 0.5553, "step": 1252 }, { - "epoch": 1.1345980126467932, - "grad_norm": 0.263671875, - "learning_rate": 0.00028707405360827506, - "loss": 0.7327, + "epoch": 2.401529636711281, + "grad_norm": 0.25390625, + "learning_rate": 0.00022375788004117128, + "loss": 0.6193, "step": 1256 }, { - "epoch": 1.1382113821138211, - "grad_norm": 0.2578125, - "learning_rate": 0.00028698659819802766, - "loss": 0.6559, + "epoch": 2.4091778202676863, + "grad_norm": 0.29296875, + "learning_rate": 0.00022330453125975474, + "loss": 0.6117, "step": 1260 }, { - "epoch": 1.141824751580849, - "grad_norm": 0.279296875, - "learning_rate": 0.000286898861341365, - "loss": 0.7023, + "epoch": 2.4168260038240916, + "grad_norm": 0.318359375, + "learning_rate": 0.0002228503010769384, + "loss": 0.5761, "step": 1264 }, { - "epoch": 1.1454381210478772, - "grad_norm": 0.2734375, - "learning_rate": 0.0002868108432185472, - "loss": 0.7236, + "epoch": 2.424474187380497, + "grad_norm": 0.2890625, + "learning_rate": 0.0002223951949543098, + "loss": 0.6343, "step": 1268 }, { - "epoch": 1.1490514905149052, - "grad_norm": 0.2890625, - "learning_rate": 0.0002867225440104123, - "loss": 0.7181, + "epoch": 2.4321223709369026, + "grad_norm": 0.28515625, + "learning_rate": 0.00022193921836398875, + "loss": 0.5763, "step": 1272 }, { - "epoch": 1.1526648599819331, - "grad_norm": 0.259765625, - "learning_rate": 0.0002866339638983758, - "loss": 0.7769, + "epoch": 2.439770554493308, + "grad_norm": 0.2890625, + "learning_rate": 0.00022148237678856138, + "loss": 0.5807, "step": 1276 }, { - "epoch": 1.156278229448961, - "grad_norm": 0.2392578125, - "learning_rate": 0.00028654510306443034, - "loss": 0.7196, + "epoch": 2.447418738049713, + "grad_norm": 0.279296875, + "learning_rate": 0.0002210246757210142, + "loss": 0.62, "step": 1280 }, { - "epoch": 1.1598915989159893, - "grad_norm": 0.279296875, - "learning_rate": 0.0002864559616911454, - "loss": 0.7283, + "epoch": 2.4550669216061185, + "grad_norm": 0.28515625, + "learning_rate": 0.00022056612066466817, + "loss": 0.6255, "step": 1284 }, { - "epoch": 1.1635049683830172, - "grad_norm": 0.2578125, - "learning_rate": 0.00028636653996166677, - "loss": 0.7252, + "epoch": 2.462715105162524, + "grad_norm": 0.3203125, + "learning_rate": 0.00022010671713311238, + "loss": 0.6786, "step": 1288 }, { - "epoch": 1.1671183378500452, - "grad_norm": 0.271484375, - "learning_rate": 0.0002862768380597162, - "loss": 0.7131, + "epoch": 2.470363288718929, + "grad_norm": 0.2890625, + "learning_rate": 0.000219646470650138, + "loss": 0.6173, "step": 1292 }, { - "epoch": 1.170731707317073, + "epoch": 2.478011472275335, "grad_norm": 0.2734375, - "learning_rate": 0.0002861868561695912, - "loss": 0.7274, + "learning_rate": 0.00021918538674967156, + "loss": 0.6113, "step": 1296 }, { - "epoch": 1.174345076784101, - "grad_norm": 0.271484375, - "learning_rate": 0.0002860965944761644, - "loss": 0.7314, + "epoch": 2.48565965583174, + "grad_norm": 0.265625, + "learning_rate": 0.0002187234709757087, + "loss": 0.5949, "step": 1300 }, { - "epoch": 1.1779584462511292, - "grad_norm": 0.291015625, - "learning_rate": 0.00028600605316488336, - "loss": 0.7095, + "epoch": 2.4933078393881454, + "grad_norm": 0.296875, + "learning_rate": 0.00021826072888224716, + "loss": 0.6248, "step": 1304 }, { - "epoch": 1.1815718157181572, - "grad_norm": 0.271484375, - "learning_rate": 0.0002859152324217701, - "loss": 0.6816, + "epoch": 2.5009560229445507, + "grad_norm": 0.283203125, + "learning_rate": 0.00021779716603322034, + "loss": 0.5849, "step": 1308 }, { - "epoch": 1.1851851851851851, - "grad_norm": 0.2578125, - "learning_rate": 0.00028582413243342095, - "loss": 0.7008, + "epoch": 2.508604206500956, + "grad_norm": 0.26953125, + "learning_rate": 0.0002173327880024303, + "loss": 0.5947, "step": 1312 }, { - "epoch": 1.1887985546522133, - "grad_norm": 0.28515625, - "learning_rate": 0.0002857327533870055, - "loss": 0.7118, + "epoch": 2.5162523900573612, + "grad_norm": 0.294921875, + "learning_rate": 0.00021686760037348065, + "loss": 0.5689, "step": 1316 }, { - "epoch": 1.1924119241192412, - "grad_norm": 0.267578125, - "learning_rate": 0.0002856410954702672, - "loss": 0.7472, + "epoch": 2.5239005736137665, + "grad_norm": 0.2890625, + "learning_rate": 0.00021640160873970954, + "loss": 0.6302, "step": 1320 }, { - "epoch": 1.1960252935862692, - "grad_norm": 0.248046875, - "learning_rate": 0.0002855491588715222, - "loss": 0.7236, + "epoch": 2.5315487571701722, + "grad_norm": 0.275390625, + "learning_rate": 0.00021593481870412217, + "loss": 0.6117, "step": 1324 }, { - "epoch": 1.1996386630532971, - "grad_norm": 0.310546875, - "learning_rate": 0.0002854569437796591, - "loss": 0.7762, + "epoch": 2.5391969407265775, + "grad_norm": 0.298828125, + "learning_rate": 0.0002154672358793238, + "loss": 0.6024, "step": 1328 }, { - "epoch": 1.203252032520325, - "grad_norm": 0.271484375, - "learning_rate": 0.0002853644503841389, - "loss": 0.7694, + "epoch": 2.546845124282983, + "grad_norm": 0.298828125, + "learning_rate": 0.00021499886588745195, + "loss": 0.5417, "step": 1332 }, { - "epoch": 1.2068654019873533, + "epoch": 2.554493307839388, "grad_norm": 0.29296875, - "learning_rate": 0.00028527167887499444, - "loss": 0.7559, + "learning_rate": 0.00021452971436010886, + "loss": 0.5975, "step": 1336 }, { - "epoch": 1.2104787714543812, - "grad_norm": 0.275390625, - "learning_rate": 0.00028517862944282964, - "loss": 0.7358, + "epoch": 2.5621414913957934, + "grad_norm": 0.287109375, + "learning_rate": 0.00021405978693829397, + "loss": 0.5997, "step": 1340 }, { - "epoch": 1.2140921409214092, - "grad_norm": 0.271484375, - "learning_rate": 0.0002850853022788197, - "loss": 0.7076, + "epoch": 2.569789674952199, + "grad_norm": 0.296875, + "learning_rate": 0.00021358908927233576, + "loss": 0.6047, "step": 1344 }, { - "epoch": 1.2177055103884373, - "grad_norm": 0.267578125, - "learning_rate": 0.00028499169757471035, - "loss": 0.7354, + "epoch": 2.5774378585086044, + "grad_norm": 0.283203125, + "learning_rate": 0.00021311762702182414, + "loss": 0.6135, "step": 1348 }, { - "epoch": 1.2213188798554653, + "epoch": 2.5850860420650097, "grad_norm": 0.2890625, - "learning_rate": 0.00028489781552281756, - "loss": 0.7191, + "learning_rate": 0.00021264540585554215, + "loss": 0.6251, "step": 1352 }, { - "epoch": 1.2249322493224932, - "grad_norm": 0.279296875, - "learning_rate": 0.00028480365631602706, - "loss": 0.7447, + "epoch": 2.592734225621415, + "grad_norm": 0.275390625, + "learning_rate": 0.00021217243145139802, + "loss": 0.6308, "step": 1356 }, { - "epoch": 1.2285456187895212, - "grad_norm": 0.25390625, - "learning_rate": 0.0002847092201477941, - "loss": 0.7253, + "epoch": 2.6003824091778203, + "grad_norm": 0.2890625, + "learning_rate": 0.0002116987094963567, + "loss": 0.5828, "step": 1360 }, { - "epoch": 1.2321589882565491, - "grad_norm": 0.291015625, - "learning_rate": 0.00028461450721214293, - "loss": 0.7286, + "epoch": 2.6080305927342256, + "grad_norm": 0.30078125, + "learning_rate": 0.00021122424568637157, + "loss": 0.6057, "step": 1364 }, { - "epoch": 1.2357723577235773, - "grad_norm": 0.271484375, - "learning_rate": 0.0002845195177036664, - "loss": 0.7333, + "epoch": 2.615678776290631, + "grad_norm": 0.291015625, + "learning_rate": 0.00021074904572631606, + "loss": 0.6435, "step": 1368 }, { - "epoch": 1.2393857271906052, - "grad_norm": 0.3046875, - "learning_rate": 0.0002844242518175256, - "loss": 0.7299, + "epoch": 2.623326959847036, + "grad_norm": 0.294921875, + "learning_rate": 0.00021027311532991475, + "loss": 0.6201, "step": 1372 }, { - "epoch": 1.2429990966576332, - "grad_norm": 0.279296875, - "learning_rate": 0.0002843287097494496, - "loss": 0.755, + "epoch": 2.6309751434034414, + "grad_norm": 0.27734375, + "learning_rate": 0.00020979646021967503, + "loss": 0.6192, "step": 1376 }, { - "epoch": 1.2466124661246614, - "grad_norm": 0.2734375, - "learning_rate": 0.00028423289169573465, - "loss": 0.6887, + "epoch": 2.638623326959847, + "grad_norm": 0.287109375, + "learning_rate": 0.00020931908612681805, + "loss": 0.6072, "step": 1380 }, { - "epoch": 1.2502258355916893, - "grad_norm": 0.263671875, - "learning_rate": 0.00028413679785324413, - "loss": 0.7668, + "epoch": 2.6462715105162524, + "grad_norm": 0.287109375, + "learning_rate": 0.00020884099879120993, + "loss": 0.5332, "step": 1384 }, { - "epoch": 1.2538392050587173, - "grad_norm": 0.263671875, - "learning_rate": 0.00028404042841940813, - "loss": 0.7157, + "epoch": 2.6539196940726577, + "grad_norm": 0.291015625, + "learning_rate": 0.00020836220396129265, + "loss": 0.5923, "step": 1388 }, { - "epoch": 1.2574525745257452, - "grad_norm": 0.28125, - "learning_rate": 0.00028394378359222294, - "loss": 0.7247, + "epoch": 2.661567877629063, + "grad_norm": 0.2734375, + "learning_rate": 0.00020788270739401505, + "loss": 0.6293, "step": 1392 }, { - "epoch": 1.2610659439927732, - "grad_norm": 0.26171875, - "learning_rate": 0.0002838468635702505, - "loss": 0.6949, + "epoch": 2.6692160611854687, + "grad_norm": 0.275390625, + "learning_rate": 0.00020740251485476345, + "loss": 0.5851, "step": 1396 }, { - "epoch": 1.2646793134598013, - "grad_norm": 0.287109375, - "learning_rate": 0.0002837496685526183, - "loss": 0.7799, + "epoch": 2.676864244741874, + "grad_norm": 0.30078125, + "learning_rate": 0.00020692163211729253, + "loss": 0.6088, "step": 1400 }, { - "epoch": 1.2682926829268293, - "grad_norm": 0.2890625, - "learning_rate": 0.00028365219873901885, - "loss": 0.6787, + "epoch": 2.6845124282982793, + "grad_norm": 0.306640625, + "learning_rate": 0.0002064400649636557, + "loss": 0.6033, "step": 1404 }, { - "epoch": 1.2719060523938572, - "grad_norm": 0.283203125, - "learning_rate": 0.00028355445432970915, - "loss": 0.7171, + "epoch": 2.6921606118546846, + "grad_norm": 0.3046875, + "learning_rate": 0.0002059578191841357, + "loss": 0.5948, "step": 1408 }, { - "epoch": 1.2755194218608854, - "grad_norm": 0.271484375, - "learning_rate": 0.00028345643552551037, - "loss": 0.722, + "epoch": 2.69980879541109, + "grad_norm": 0.29296875, + "learning_rate": 0.00020547490057717499, + "loss": 0.6287, "step": 1412 }, { - "epoch": 1.2791327913279134, - "grad_norm": 0.26171875, - "learning_rate": 0.0002833581425278075, - "loss": 0.7501, + "epoch": 2.707456978967495, + "grad_norm": 0.296875, + "learning_rate": 0.00020499131494930602, + "loss": 0.5736, "step": 1416 }, { - "epoch": 1.2827461607949413, - "grad_norm": 0.267578125, - "learning_rate": 0.00028325957553854885, - "loss": 0.7811, + "epoch": 2.7151051625239004, + "grad_norm": 0.287109375, + "learning_rate": 0.0002045070681150813, + "loss": 0.6496, "step": 1420 }, { - "epoch": 1.2863595302619693, - "grad_norm": 0.26953125, - "learning_rate": 0.00028316073476024567, - "loss": 0.7448, + "epoch": 2.7227533460803057, + "grad_norm": 0.279296875, + "learning_rate": 0.00020402216589700362, + "loss": 0.5993, "step": 1424 }, { - "epoch": 1.2899728997289972, + "epoch": 2.730401529636711, "grad_norm": 0.296875, - "learning_rate": 0.00028306162039597166, - "loss": 0.7354, + "learning_rate": 0.00020353661412545598, + "loss": 0.596, "step": 1428 }, { - "epoch": 1.2935862691960254, - "grad_norm": 0.24609375, - "learning_rate": 0.00028296223264936277, - "loss": 0.6888, + "epoch": 2.7380497131931167, + "grad_norm": 0.28515625, + "learning_rate": 0.00020305041863863152, + "loss": 0.639, "step": 1432 }, { - "epoch": 1.2971996386630533, - "grad_norm": 0.279296875, - "learning_rate": 0.00028286257172461637, - "loss": 0.7354, + "epoch": 2.745697896749522, + "grad_norm": 0.27734375, + "learning_rate": 0.00020256358528246334, + "loss": 0.5703, "step": 1436 }, { - "epoch": 1.3008130081300813, - "grad_norm": 0.267578125, - "learning_rate": 0.0002827626378264914, - "loss": 0.7097, + "epoch": 2.7533460803059273, + "grad_norm": 0.279296875, + "learning_rate": 0.00020207611991055407, + "loss": 0.5838, "step": 1440 }, { - "epoch": 1.3044263775971092, - "grad_norm": 0.27734375, - "learning_rate": 0.00028266243116030753, - "loss": 0.7167, + "epoch": 2.7609942638623326, + "grad_norm": 0.3046875, + "learning_rate": 0.0002015880283841057, + "loss": 0.5845, "step": 1444 }, { - "epoch": 1.3080397470641372, - "grad_norm": 0.2734375, - "learning_rate": 0.00028256195193194464, - "loss": 0.7424, + "epoch": 2.768642447418738, + "grad_norm": 0.287109375, + "learning_rate": 0.00020109931657184894, + "loss": 0.6169, "step": 1448 }, { - "epoch": 1.3116531165311653, - "grad_norm": 0.27734375, - "learning_rate": 0.0002824612003478428, - "loss": 0.7203, + "epoch": 2.7762906309751436, + "grad_norm": 0.296875, + "learning_rate": 0.0002006099903499727, + "loss": 0.6026, "step": 1452 }, { - "epoch": 1.3152664859981933, - "grad_norm": 0.275390625, - "learning_rate": 0.00028236017661500176, - "loss": 0.7034, + "epoch": 2.783938814531549, + "grad_norm": 0.291015625, + "learning_rate": 0.00020012005560205356, + "loss": 0.6278, "step": 1456 }, { - "epoch": 1.3188798554652212, - "grad_norm": 0.302734375, - "learning_rate": 0.0002822588809409801, - "loss": 0.7605, + "epoch": 2.791586998087954, + "grad_norm": 0.28515625, + "learning_rate": 0.0001996295182189847, + "loss": 0.6273, "step": 1460 }, { - "epoch": 1.3224932249322494, - "grad_norm": 0.2734375, - "learning_rate": 0.0002821573135338954, - "loss": 0.7303, + "epoch": 2.7992351816443595, + "grad_norm": 0.291015625, + "learning_rate": 0.00019913838409890548, + "loss": 0.6084, "step": 1464 }, { - "epoch": 1.3261065943992774, - "grad_norm": 0.291015625, - "learning_rate": 0.00028205547460242336, - "loss": 0.7412, + "epoch": 2.8068833652007648, + "grad_norm": 0.30859375, + "learning_rate": 0.00019864665914713024, + "loss": 0.6295, "step": 1468 }, { - "epoch": 1.3297199638663053, - "grad_norm": 0.28515625, - "learning_rate": 0.00028195336435579764, - "loss": 0.7397, + "epoch": 2.81453154875717, + "grad_norm": 0.29296875, + "learning_rate": 0.0001981543492760774, + "loss": 0.5889, "step": 1472 }, { - "epoch": 1.3333333333333333, - "grad_norm": 0.271484375, - "learning_rate": 0.0002818509830038093, - "loss": 0.7335, + "epoch": 2.8221797323135753, + "grad_norm": 0.287109375, + "learning_rate": 0.00019766146040519836, + "loss": 0.6064, "step": 1476 }, { - "epoch": 1.3369467028003612, - "grad_norm": 0.2734375, - "learning_rate": 0.0002817483307568064, - "loss": 0.7167, + "epoch": 2.8298279158699806, + "grad_norm": 0.28515625, + "learning_rate": 0.00019716799846090634, + "loss": 0.6269, "step": 1480 }, { - "epoch": 1.3405600722673894, - "grad_norm": 0.26953125, - "learning_rate": 0.00028164540782569343, - "loss": 0.7075, + "epoch": 2.8374760994263863, + "grad_norm": 0.287109375, + "learning_rate": 0.00019667396937650506, + "loss": 0.5742, "step": 1484 }, { - "epoch": 1.3441734417344173, - "grad_norm": 0.283203125, - "learning_rate": 0.00028154221442193135, - "loss": 0.729, + "epoch": 2.8451242829827916, + "grad_norm": 0.33203125, + "learning_rate": 0.0001961793790921174, + "loss": 0.5701, "step": 1488 }, { - "epoch": 1.3477868112014453, - "grad_norm": 0.287109375, - "learning_rate": 0.00028143875075753655, - "loss": 0.695, + "epoch": 2.852772466539197, + "grad_norm": 0.2734375, + "learning_rate": 0.00019568423355461402, + "loss": 0.5973, "step": 1492 }, { - "epoch": 1.3514001806684734, - "grad_norm": 0.27734375, - "learning_rate": 0.0002813350170450807, - "loss": 0.748, + "epoch": 2.860420650095602, + "grad_norm": 0.296875, + "learning_rate": 0.00019518853871754204, + "loss": 0.609, "step": 1496 }, { - "epoch": 1.3550135501355014, - "grad_norm": 0.2890625, - "learning_rate": 0.00028123101349769033, - "loss": 0.741, + "epoch": 2.8680688336520075, + "grad_norm": 0.30078125, + "learning_rate": 0.00019469230054105295, + "loss": 0.5944, "step": 1500 }, { - "epoch": 1.3586269196025293, - "grad_norm": 0.267578125, - "learning_rate": 0.0002811267403290465, - "loss": 0.7294, + "epoch": 2.875717017208413, + "grad_norm": 0.302734375, + "learning_rate": 0.0001941955249918315, + "loss": 0.5914, "step": 1504 }, { - "epoch": 1.3622402890695573, - "grad_norm": 0.267578125, - "learning_rate": 0.00028102219775338406, - "loss": 0.743, + "epoch": 2.8833652007648185, + "grad_norm": 0.283203125, + "learning_rate": 0.00019369821804302365, + "loss": 0.6191, "step": 1508 }, { - "epoch": 1.3658536585365852, - "grad_norm": 0.306640625, - "learning_rate": 0.00028091738598549145, - "loss": 0.7018, + "epoch": 2.891013384321224, + "grad_norm": 0.296875, + "learning_rate": 0.00019320038567416484, + "loss": 0.6409, "step": 1512 }, { - "epoch": 1.3694670280036134, - "grad_norm": 0.27734375, - "learning_rate": 0.00028081230524071, - "loss": 0.7036, + "epoch": 2.898661567877629, + "grad_norm": 0.294921875, + "learning_rate": 0.00019270203387110798, + "loss": 0.5779, "step": 1516 }, { - "epoch": 1.3730803974706414, - "grad_norm": 0.263671875, - "learning_rate": 0.00028070695573493394, - "loss": 0.7261, + "epoch": 2.9063097514340344, + "grad_norm": 0.29296875, + "learning_rate": 0.00019220316862595167, + "loss": 0.5956, "step": 1520 }, { - "epoch": 1.3766937669376693, + "epoch": 2.9139579349904396, "grad_norm": 0.3046875, - "learning_rate": 0.00028060133768460956, - "loss": 0.7342, + "learning_rate": 0.00019170379593696802, + "loss": 0.5916, "step": 1524 }, { - "epoch": 1.3803071364046975, - "grad_norm": 0.29296875, - "learning_rate": 0.0002804954513067349, - "loss": 0.7539, + "epoch": 2.921606118546845, + "grad_norm": 0.2890625, + "learning_rate": 0.00019120392180853058, + "loss": 0.6069, "step": 1528 }, { - "epoch": 1.3839205058717254, - "grad_norm": 0.26953125, - "learning_rate": 0.0002803892968188592, - "loss": 0.7647, + "epoch": 2.92925430210325, + "grad_norm": 0.279296875, + "learning_rate": 0.0001907035522510421, + "loss": 0.6029, "step": 1532 }, { - "epoch": 1.3875338753387534, - "grad_norm": 0.267578125, - "learning_rate": 0.0002802828744390826, - "loss": 0.704, + "epoch": 2.936902485659656, + "grad_norm": 0.310546875, + "learning_rate": 0.00019020269328086226, + "loss": 0.5706, "step": 1536 }, { - "epoch": 1.3911472448057813, - "grad_norm": 0.26171875, - "learning_rate": 0.00028017618438605583, - "loss": 0.7364, + "epoch": 2.9445506692160612, + "grad_norm": 0.29296875, + "learning_rate": 0.0001897013509202354, + "loss": 0.6024, "step": 1540 }, { - "epoch": 1.3947606142728093, - "grad_norm": 0.2890625, - "learning_rate": 0.0002800692268789793, - "loss": 0.6884, + "epoch": 2.9521988527724665, + "grad_norm": 0.287109375, + "learning_rate": 0.00018919953119721808, + "loss": 0.6326, "step": 1544 }, { - "epoch": 1.3983739837398375, - "grad_norm": 0.26171875, - "learning_rate": 0.000279962002137603, - "loss": 0.7081, + "epoch": 2.959847036328872, + "grad_norm": 0.296875, + "learning_rate": 0.0001886972401456065, + "loss": 0.5744, "step": 1548 }, { - "epoch": 1.4019873532068654, - "grad_norm": 0.283203125, - "learning_rate": 0.000279854510382226, - "loss": 0.7041, + "epoch": 2.967495219885277, + "grad_norm": 0.28515625, + "learning_rate": 0.00018819448380486413, + "loss": 0.5679, "step": 1552 }, { - "epoch": 1.4056007226738934, - "grad_norm": 0.265625, - "learning_rate": 0.0002797467518336961, - "loss": 0.6753, + "epoch": 2.975143403441683, + "grad_norm": 0.27734375, + "learning_rate": 0.00018769126822004898, + "loss": 0.5992, "step": 1556 }, { - "epoch": 1.4092140921409215, - "grad_norm": 0.310546875, - "learning_rate": 0.00027963872671340887, - "loss": 0.7222, + "epoch": 2.982791586998088, + "grad_norm": 0.30078125, + "learning_rate": 0.00018718759944174086, + "loss": 0.5981, "step": 1560 }, { - "epoch": 1.4128274616079495, - "grad_norm": 0.279296875, - "learning_rate": 0.000279530435243308, - "loss": 0.7318, + "epoch": 2.9904397705544934, + "grad_norm": 0.29296875, + "learning_rate": 0.0001866834835259688, + "loss": 0.6188, "step": 1564 }, { - "epoch": 1.4164408310749774, - "grad_norm": 0.25, - "learning_rate": 0.00027942187764588405, - "loss": 0.7012, + "epoch": 2.9980879541108987, + "grad_norm": 0.33984375, + "learning_rate": 0.0001861789265341381, + "loss": 0.617, "step": 1568 }, { - "epoch": 1.4200542005420054, - "grad_norm": 0.255859375, - "learning_rate": 0.0002793130541441746, - "loss": 0.7813, + "epoch": 3.005736137667304, + "grad_norm": 0.322265625, + "learning_rate": 0.00018567393453295742, + "loss": 0.4644, "step": 1572 }, { - "epoch": 1.4236675700090333, - "grad_norm": 0.255859375, - "learning_rate": 0.0002792039649617634, - "loss": 0.696, + "epoch": 3.0133843212237093, + "grad_norm": 0.298828125, + "learning_rate": 0.00018516851359436602, + "loss": 0.4965, "step": 1576 }, { - "epoch": 1.4272809394760615, - "grad_norm": 0.310546875, - "learning_rate": 0.00027909461032278013, - "loss": 0.7043, + "epoch": 3.0210325047801145, + "grad_norm": 0.29296875, + "learning_rate": 0.00018466266979546057, + "loss": 0.501, "step": 1580 }, { - "epoch": 1.4308943089430894, - "grad_norm": 0.2578125, - "learning_rate": 0.0002789849904518999, - "loss": 0.7521, + "epoch": 3.0286806883365203, + "grad_norm": 0.28125, + "learning_rate": 0.0001841564092184221, + "loss": 0.4787, "step": 1584 }, { - "epoch": 1.4345076784101174, - "grad_norm": 0.29296875, - "learning_rate": 0.00027887510557434255, - "loss": 0.7258, + "epoch": 3.0363288718929256, + "grad_norm": 0.298828125, + "learning_rate": 0.00018364973795044294, + "loss": 0.5116, "step": 1588 }, { - "epoch": 1.4381210478771456, - "grad_norm": 0.271484375, - "learning_rate": 0.0002787649559158727, - "loss": 0.7201, + "epoch": 3.043977055449331, + "grad_norm": 0.28125, + "learning_rate": 0.00018314266208365357, + "loss": 0.5309, "step": 1592 }, { - "epoch": 1.4417344173441735, - "grad_norm": 0.279296875, - "learning_rate": 0.0002786545417027987, - "loss": 0.7215, + "epoch": 3.051625239005736, + "grad_norm": 0.30078125, + "learning_rate": 0.00018263518771504924, + "loss": 0.4979, "step": 1596 }, { - "epoch": 1.4453477868112015, - "grad_norm": 0.251953125, - "learning_rate": 0.0002785438631619726, - "loss": 0.6907, + "epoch": 3.0592734225621414, + "grad_norm": 0.296875, + "learning_rate": 0.00018212732094641666, + "loss": 0.4647, "step": 1600 }, { - "epoch": 1.4489611562782294, - "grad_norm": 0.2734375, - "learning_rate": 0.00027843292052078943, - "loss": 0.7227, + "epoch": 3.0669216061185467, + "grad_norm": 0.294921875, + "learning_rate": 0.00018161906788426076, + "loss": 0.5367, "step": 1604 }, { - "epoch": 1.4525745257452574, - "grad_norm": 0.296875, - "learning_rate": 0.000278321714007187, - "loss": 0.6769, + "epoch": 3.0745697896749524, + "grad_norm": 0.30078125, + "learning_rate": 0.00018111043463973122, + "loss": 0.5095, "step": 1608 }, { - "epoch": 1.4561878952122855, - "grad_norm": 0.275390625, - "learning_rate": 0.000278210243849645, - "loss": 0.7191, + "epoch": 3.0822179732313577, + "grad_norm": 0.27734375, + "learning_rate": 0.00018060142732854894, + "loss": 0.4615, "step": 1612 }, { - "epoch": 1.4598012646793135, - "grad_norm": 0.283203125, - "learning_rate": 0.000278098510277185, - "loss": 0.6908, + "epoch": 3.089866156787763, + "grad_norm": 0.267578125, + "learning_rate": 0.00018009205207093252, + "loss": 0.5105, "step": 1616 }, { - "epoch": 1.4634146341463414, - "grad_norm": 0.29296875, - "learning_rate": 0.0002779865135193697, - "loss": 0.7349, + "epoch": 3.0975143403441683, + "grad_norm": 0.287109375, + "learning_rate": 0.00017958231499152463, + "loss": 0.5326, "step": 1620 }, { - "epoch": 1.4670280036133696, - "grad_norm": 0.25, - "learning_rate": 0.00027787425380630253, - "loss": 0.6604, + "epoch": 3.1051625239005736, + "grad_norm": 0.3125, + "learning_rate": 0.0001790722222193186, + "loss": 0.5383, "step": 1624 }, { - "epoch": 1.4706413730803976, - "grad_norm": 0.28125, - "learning_rate": 0.00027776173136862713, - "loss": 0.6614, + "epoch": 3.112810707456979, + "grad_norm": 0.296875, + "learning_rate": 0.00017856177988758438, + "loss": 0.5192, "step": 1628 }, { - "epoch": 1.4742547425474255, - "grad_norm": 0.287109375, - "learning_rate": 0.00027764894643752706, - "loss": 0.7078, + "epoch": 3.120458891013384, + "grad_norm": 0.3046875, + "learning_rate": 0.00017805099413379508, + "loss": 0.5029, "step": 1632 }, { - "epoch": 1.4778681120144534, - "grad_norm": 0.291015625, - "learning_rate": 0.000277535899244725, - "loss": 0.7213, + "epoch": 3.12810707456979, + "grad_norm": 0.318359375, + "learning_rate": 0.00017753987109955297, + "loss": 0.4896, "step": 1636 }, { - "epoch": 1.4814814814814814, - "grad_norm": 0.296875, - "learning_rate": 0.00027742259002248263, - "loss": 0.7703, + "epoch": 3.135755258126195, + "grad_norm": 0.310546875, + "learning_rate": 0.00017702841693051577, + "loss": 0.5254, "step": 1640 }, { - "epoch": 1.4850948509485096, - "grad_norm": 0.275390625, - "learning_rate": 0.00027730901900359985, - "loss": 0.7415, + "epoch": 3.1434034416826004, + "grad_norm": 0.330078125, + "learning_rate": 0.0001765166377763227, + "loss": 0.4829, "step": 1644 }, { - "epoch": 1.4887082204155375, - "grad_norm": 0.267578125, - "learning_rate": 0.0002771951864214145, - "loss": 0.7169, + "epoch": 3.1510516252390057, + "grad_norm": 0.3046875, + "learning_rate": 0.00017600453979052055, + "loss": 0.5461, "step": 1648 }, { - "epoch": 1.4923215898825655, - "grad_norm": 0.271484375, - "learning_rate": 0.00027708109250980183, - "loss": 0.6944, + "epoch": 3.158699808795411, + "grad_norm": 0.298828125, + "learning_rate": 0.0001754921291304897, + "loss": 0.5415, "step": 1652 }, { - "epoch": 1.4959349593495934, - "grad_norm": 0.271484375, - "learning_rate": 0.000276966737503174, - "loss": 0.6798, + "epoch": 3.1663479923518163, + "grad_norm": 0.31640625, + "learning_rate": 0.00017497941195737004, + "loss": 0.5501, "step": 1656 }, { - "epoch": 1.4995483288166214, - "grad_norm": 0.291015625, - "learning_rate": 0.00027685212163647955, - "loss": 0.7072, + "epoch": 3.173996175908222, + "grad_norm": 0.314453125, + "learning_rate": 0.00017446639443598696, + "loss": 0.4964, "step": 1660 }, { - "epoch": 1.5031616982836495, - "grad_norm": 0.251953125, - "learning_rate": 0.00027673724514520306, - "loss": 0.7185, + "epoch": 3.1816443594646273, + "grad_norm": 0.3046875, + "learning_rate": 0.00017395308273477714, + "loss": 0.4938, "step": 1664 }, { - "epoch": 1.5067750677506775, - "grad_norm": 0.271484375, - "learning_rate": 0.0002766221082653645, - "loss": 0.707, + "epoch": 3.1892925430210326, + "grad_norm": 0.30859375, + "learning_rate": 0.00017343948302571446, + "loss": 0.5409, "step": 1668 }, { - "epoch": 1.5103884372177054, - "grad_norm": 0.2890625, - "learning_rate": 0.00027650671123351884, - "loss": 0.7042, + "epoch": 3.196940726577438, + "grad_norm": 0.302734375, + "learning_rate": 0.00017292560148423578, + "loss": 0.4844, "step": 1672 }, { - "epoch": 1.5140018066847336, - "grad_norm": 0.275390625, - "learning_rate": 0.00027639105428675556, - "loss": 0.7221, + "epoch": 3.204588910133843, + "grad_norm": 0.3046875, + "learning_rate": 0.00017241144428916655, + "loss": 0.539, "step": 1676 }, { - "epoch": 1.5176151761517616, - "grad_norm": 0.291015625, - "learning_rate": 0.0002762751376626982, - "loss": 0.7181, + "epoch": 3.2122370936902485, + "grad_norm": 0.3046875, + "learning_rate": 0.00017189701762264687, + "loss": 0.4974, "step": 1680 }, { - "epoch": 1.5212285456187895, - "grad_norm": 0.271484375, - "learning_rate": 0.00027615896159950367, - "loss": 0.7511, + "epoch": 3.2198852772466537, + "grad_norm": 0.337890625, + "learning_rate": 0.0001713823276700567, + "loss": 0.5443, "step": 1684 }, { - "epoch": 1.5248419150858177, - "grad_norm": 0.2734375, - "learning_rate": 0.0002760425263358621, - "loss": 0.6876, + "epoch": 3.2275334608030595, + "grad_norm": 0.29296875, + "learning_rate": 0.00017086738061994176, + "loss": 0.4936, "step": 1688 }, { - "epoch": 1.5284552845528454, - "grad_norm": 0.291015625, - "learning_rate": 0.000275925832110996, - "loss": 0.7231, + "epoch": 3.2351816443594648, + "grad_norm": 0.298828125, + "learning_rate": 0.00017035218266393918, + "loss": 0.5027, "step": 1692 }, { - "epoch": 1.5320686540198736, - "grad_norm": 0.251953125, - "learning_rate": 0.00027580887916466007, - "loss": 0.7025, + "epoch": 3.24282982791587, + "grad_norm": 0.326171875, + "learning_rate": 0.00016983673999670273, + "loss": 0.5352, "step": 1696 }, { - "epoch": 1.5356820234869015, - "grad_norm": 0.25390625, - "learning_rate": 0.00027569166773714043, - "loss": 0.6513, + "epoch": 3.2504780114722753, + "grad_norm": 0.310546875, + "learning_rate": 0.0001693210588158287, + "loss": 0.5147, "step": 1700 }, { - "epoch": 1.5392953929539295, - "grad_norm": 0.26953125, - "learning_rate": 0.00027557419806925436, - "loss": 0.7532, + "epoch": 3.2581261950286806, + "grad_norm": 0.296875, + "learning_rate": 0.00016880514532178123, + "loss": 0.5013, "step": 1704 }, { - "epoch": 1.5429087624209576, - "grad_norm": 0.279296875, - "learning_rate": 0.0002754564704023497, - "loss": 0.7641, + "epoch": 3.265774378585086, + "grad_norm": 0.3125, + "learning_rate": 0.00016828900571781767, + "loss": 0.5408, "step": 1708 }, { - "epoch": 1.5465221318879856, - "grad_norm": 0.291015625, - "learning_rate": 0.00027533848497830434, - "loss": 0.7356, + "epoch": 3.2734225621414916, + "grad_norm": 0.3046875, + "learning_rate": 0.00016777264620991414, + "loss": 0.4758, "step": 1712 }, { - "epoch": 1.5501355013550135, - "grad_norm": 0.271484375, - "learning_rate": 0.00027522024203952575, - "loss": 0.7366, + "epoch": 3.281070745697897, + "grad_norm": 0.3046875, + "learning_rate": 0.00016725607300669087, + "loss": 0.5154, "step": 1716 }, { - "epoch": 1.5537488708220417, - "grad_norm": 0.26953125, - "learning_rate": 0.00027510174182895046, - "loss": 0.6814, + "epoch": 3.288718929254302, + "grad_norm": 0.302734375, + "learning_rate": 0.0001667392923193375, + "loss": 0.4882, "step": 1720 }, { - "epoch": 1.5573622402890694, - "grad_norm": 0.291015625, - "learning_rate": 0.00027498298459004356, - "loss": 0.6878, + "epoch": 3.2963671128107075, + "grad_norm": 0.322265625, + "learning_rate": 0.00016622231036153836, + "loss": 0.5423, "step": 1724 }, { - "epoch": 1.5609756097560976, - "grad_norm": 0.267578125, - "learning_rate": 0.00027486397056679835, - "loss": 0.7133, + "epoch": 3.3040152963671128, + "grad_norm": 0.328125, + "learning_rate": 0.0001657051333493978, + "loss": 0.509, "step": 1728 }, { - "epoch": 1.5645889792231256, - "grad_norm": 0.26953125, - "learning_rate": 0.00027474470000373553, - "loss": 0.7425, + "epoch": 3.311663479923518, + "grad_norm": 0.349609375, + "learning_rate": 0.00016518776750136578, + "loss": 0.5447, "step": 1732 }, { - "epoch": 1.5682023486901535, - "grad_norm": 0.259765625, - "learning_rate": 0.000274625173145903, - "loss": 0.724, + "epoch": 3.3193116634799233, + "grad_norm": 0.33203125, + "learning_rate": 0.00016467021903816237, + "loss": 0.5048, "step": 1736 }, { - "epoch": 1.5718157181571817, - "grad_norm": 0.27734375, - "learning_rate": 0.00027450539023887515, - "loss": 0.7249, + "epoch": 3.3269598470363286, + "grad_norm": 0.287109375, + "learning_rate": 0.00016415249418270364, + "loss": 0.5183, "step": 1740 }, { - "epoch": 1.5754290876242094, - "grad_norm": 0.318359375, - "learning_rate": 0.00027438535152875254, - "loss": 0.7067, + "epoch": 3.3346080305927344, + "grad_norm": 0.3125, + "learning_rate": 0.00016363459916002643, + "loss": 0.4915, "step": 1744 }, { - "epoch": 1.5790424570912376, - "grad_norm": 0.271484375, - "learning_rate": 0.0002742650572621612, - "loss": 0.7092, + "epoch": 3.3422562141491396, + "grad_norm": 0.349609375, + "learning_rate": 0.00016311654019721377, + "loss": 0.5016, "step": 1748 }, { - "epoch": 1.5826558265582655, - "grad_norm": 0.26953125, - "learning_rate": 0.0002741445076862522, - "loss": 0.6639, + "epoch": 3.349904397705545, + "grad_norm": 0.30859375, + "learning_rate": 0.00016259832352331978, + "loss": 0.5276, "step": 1752 }, { - "epoch": 1.5862691960252935, - "grad_norm": 0.33984375, - "learning_rate": 0.00027402370304870126, - "loss": 0.6936, + "epoch": 3.35755258126195, + "grad_norm": 0.34375, + "learning_rate": 0.0001620799553692949, + "loss": 0.5436, "step": 1756 }, { - "epoch": 1.5898825654923217, - "grad_norm": 0.28515625, - "learning_rate": 0.00027390264359770797, - "loss": 0.721, + "epoch": 3.3652007648183555, + "grad_norm": 0.29296875, + "learning_rate": 0.00016156144196791103, + "loss": 0.5152, "step": 1760 }, { - "epoch": 1.5934959349593496, - "grad_norm": 0.283203125, - "learning_rate": 0.00027378132958199577, - "loss": 0.7253, + "epoch": 3.3728489483747612, + "grad_norm": 0.28515625, + "learning_rate": 0.0001610427895536863, + "loss": 0.4845, "step": 1764 }, { - "epoch": 1.5971093044263776, - "grad_norm": 0.29296875, - "learning_rate": 0.00027365976125081064, - "loss": 0.7269, + "epoch": 3.3804971319311665, + "grad_norm": 0.33203125, + "learning_rate": 0.00016052400436281046, + "loss": 0.51, "step": 1768 }, { - "epoch": 1.6007226738934057, - "grad_norm": 0.267578125, - "learning_rate": 0.00027353793885392155, - "loss": 0.7555, + "epoch": 3.388145315487572, + "grad_norm": 0.306640625, + "learning_rate": 0.00016000509263306976, + "loss": 0.5163, "step": 1772 }, { - "epoch": 1.6043360433604335, - "grad_norm": 0.271484375, - "learning_rate": 0.00027341586264161905, - "loss": 0.7226, + "epoch": 3.395793499043977, + "grad_norm": 0.34765625, + "learning_rate": 0.0001594860606037719, + "loss": 0.559, "step": 1776 }, { - "epoch": 1.6079494128274616, - "grad_norm": 0.2890625, - "learning_rate": 0.00027329353286471554, - "loss": 0.7104, + "epoch": 3.4034416826003824, + "grad_norm": 0.3359375, + "learning_rate": 0.0001589669145156709, + "loss": 0.523, "step": 1780 }, { - "epoch": 1.6115627822944896, - "grad_norm": 0.26953125, - "learning_rate": 0.0002731709497745441, - "loss": 0.7279, + "epoch": 3.4110898661567877, + "grad_norm": 0.3359375, + "learning_rate": 0.00015844766061089241, + "loss": 0.4994, "step": 1784 }, { - "epoch": 1.6151761517615175, - "grad_norm": 0.294921875, - "learning_rate": 0.0002730481136229583, - "loss": 0.7066, + "epoch": 3.418738049713193, + "grad_norm": 0.32421875, + "learning_rate": 0.00015792830513285838, + "loss": 0.5259, "step": 1788 }, { - "epoch": 1.6187895212285457, - "grad_norm": 0.337890625, - "learning_rate": 0.00027292502466233184, - "loss": 0.7169, + "epoch": 3.4263862332695982, + "grad_norm": 0.314453125, + "learning_rate": 0.000157408854326212, + "loss": 0.4734, "step": 1792 }, { - "epoch": 1.6224028906955736, - "grad_norm": 0.279296875, - "learning_rate": 0.0002728016831455575, - "loss": 0.6813, + "epoch": 3.434034416826004, + "grad_norm": 0.330078125, + "learning_rate": 0.00015688931443674276, + "loss": 0.5163, "step": 1796 }, { - "epoch": 1.6260162601626016, - "grad_norm": 0.275390625, - "learning_rate": 0.0002726780893260473, - "loss": 0.6798, + "epoch": 3.4416826003824093, + "grad_norm": 0.337890625, + "learning_rate": 0.0001563696917113112, + "loss": 0.514, "step": 1800 }, { - "epoch": 1.6296296296296298, - "grad_norm": 0.275390625, - "learning_rate": 0.0002725542434577314, - "loss": 0.6663, + "epoch": 3.4493307839388145, + "grad_norm": 0.322265625, + "learning_rate": 0.00015584999239777393, + "loss": 0.5691, "step": 1804 }, { - "epoch": 1.6332429990966575, - "grad_norm": 0.294921875, - "learning_rate": 0.0002724301457950578, - "loss": 0.6847, + "epoch": 3.45697896749522, + "grad_norm": 0.3515625, + "learning_rate": 0.0001553302227449084, + "loss": 0.5365, "step": 1808 }, { - "epoch": 1.6368563685636857, - "grad_norm": 0.287109375, - "learning_rate": 0.0002723057965929921, - "loss": 0.7253, + "epoch": 3.464627151051625, + "grad_norm": 0.318359375, + "learning_rate": 0.0001548103890023378, + "loss": 0.5111, "step": 1812 }, { - "epoch": 1.6404697380307136, - "grad_norm": 0.27734375, - "learning_rate": 0.0002721811961070163, - "loss": 0.7399, + "epoch": 3.472275334608031, + "grad_norm": 0.318359375, + "learning_rate": 0.00015429049742045591, + "loss": 0.5272, "step": 1816 }, { - "epoch": 1.6440831074977416, - "grad_norm": 0.302734375, - "learning_rate": 0.00027205634459312884, - "loss": 0.7394, + "epoch": 3.479923518164436, + "grad_norm": 0.30078125, + "learning_rate": 0.000153770554250352, + "loss": 0.4746, "step": 1820 }, { - "epoch": 1.6476964769647697, - "grad_norm": 0.28515625, - "learning_rate": 0.00027193124230784414, - "loss": 0.7702, + "epoch": 3.4875717017208414, + "grad_norm": 0.328125, + "learning_rate": 0.00015325056574373564, + "loss": 0.5091, "step": 1824 }, { - "epoch": 1.6513098464317977, - "grad_norm": 0.265625, - "learning_rate": 0.00027180588950819157, - "loss": 0.6959, + "epoch": 3.4952198852772467, + "grad_norm": 0.322265625, + "learning_rate": 0.00015273053815286153, + "loss": 0.5043, "step": 1828 }, { - "epoch": 1.6549232158988256, + "epoch": 3.502868068833652, "grad_norm": 0.310546875, - "learning_rate": 0.00027168028645171516, - "loss": 0.7626, + "learning_rate": 0.00015221047773045424, + "loss": 0.5157, "step": 1832 }, { - "epoch": 1.6585365853658538, - "grad_norm": 0.275390625, - "learning_rate": 0.00027155443339647335, - "loss": 0.7254, + "epoch": 3.5105162523900573, + "grad_norm": 0.310546875, + "learning_rate": 0.00015169039072963312, + "loss": 0.525, "step": 1836 }, { - "epoch": 1.6621499548328815, - "grad_norm": 0.263671875, - "learning_rate": 0.00027142833060103807, - "loss": 0.728, + "epoch": 3.5181644359464626, + "grad_norm": 0.333984375, + "learning_rate": 0.00015117028340383713, + "loss": 0.536, "step": 1840 }, { - "epoch": 1.6657633242999097, - "grad_norm": 0.28125, - "learning_rate": 0.0002713019783244944, - "loss": 0.717, + "epoch": 3.525812619502868, + "grad_norm": 0.326171875, + "learning_rate": 0.00015065016200674963, + "loss": 0.556, "step": 1844 }, { - "epoch": 1.6693766937669376, - "grad_norm": 0.25390625, - "learning_rate": 0.0002711753768264398, - "loss": 0.729, + "epoch": 3.5334608030592736, + "grad_norm": 0.345703125, + "learning_rate": 0.00015013003279222312, + "loss": 0.5199, "step": 1848 }, { - "epoch": 1.6729900632339656, - "grad_norm": 0.26171875, - "learning_rate": 0.0002710485263669841, - "loss": 0.7737, + "epoch": 3.541108986615679, + "grad_norm": 0.30859375, + "learning_rate": 0.0001496099020142041, + "loss": 0.5381, "step": 1852 }, { - "epoch": 1.6766034327009938, - "grad_norm": 0.265625, - "learning_rate": 0.0002709214272067484, - "loss": 0.7194, + "epoch": 3.548757170172084, + "grad_norm": 0.33203125, + "learning_rate": 0.00014908977592665787, + "loss": 0.5092, "step": 1856 }, { - "epoch": 1.6802168021680217, - "grad_norm": 0.28515625, - "learning_rate": 0.00027079407960686487, - "loss": 0.7001, + "epoch": 3.5564053537284894, + "grad_norm": 0.328125, + "learning_rate": 0.00014856966078349339, + "loss": 0.5101, "step": 1860 }, { - "epoch": 1.6838301716350497, - "grad_norm": 0.275390625, - "learning_rate": 0.00027066648382897604, - "loss": 0.7283, + "epoch": 3.5640535372848947, + "grad_norm": 0.353515625, + "learning_rate": 0.00014804956283848793, + "loss": 0.5093, "step": 1864 }, { - "epoch": 1.6874435411020778, - "grad_norm": 0.283203125, - "learning_rate": 0.0002705386401352344, - "loss": 0.6987, + "epoch": 3.5717017208413004, + "grad_norm": 0.306640625, + "learning_rate": 0.00014752948834521206, + "loss": 0.499, "step": 1868 }, { - "epoch": 1.6910569105691056, - "grad_norm": 0.283203125, - "learning_rate": 0.00027041054878830176, - "loss": 0.7261, + "epoch": 3.5793499043977057, + "grad_norm": 0.33984375, + "learning_rate": 0.00014700944355695432, + "loss": 0.4342, "step": 1872 }, { - "epoch": 1.6946702800361337, - "grad_norm": 0.287109375, - "learning_rate": 0.0002702822100513487, - "loss": 0.7475, + "epoch": 3.586998087954111, + "grad_norm": 0.333984375, + "learning_rate": 0.00014648943472664612, + "loss": 0.541, "step": 1876 }, { - "epoch": 1.6982836495031617, - "grad_norm": 0.28515625, - "learning_rate": 0.00027015362418805424, - "loss": 0.7149, + "epoch": 3.5946462715105163, + "grad_norm": 0.349609375, + "learning_rate": 0.00014596946810678646, + "loss": 0.5089, "step": 1880 }, { - "epoch": 1.7018970189701896, - "grad_norm": 0.28125, - "learning_rate": 0.00027002479146260504, - "loss": 0.7151, + "epoch": 3.6022944550669216, + "grad_norm": 0.30859375, + "learning_rate": 0.00014544954994936689, + "loss": 0.4995, "step": 1884 }, { - "epoch": 1.7055103884372178, - "grad_norm": 0.275390625, - "learning_rate": 0.0002698957121396948, - "loss": 0.778, + "epoch": 3.609942638623327, + "grad_norm": 0.318359375, + "learning_rate": 0.0001449296865057962, + "loss": 0.5299, "step": 1888 }, { - "epoch": 1.7091237579042458, - "grad_norm": 0.30859375, - "learning_rate": 0.00026976638648452413, - "loss": 0.7065, + "epoch": 3.617590822179732, + "grad_norm": 0.337890625, + "learning_rate": 0.00014440988402682526, + "loss": 0.5933, "step": 1892 }, { - "epoch": 1.7127371273712737, - "grad_norm": 0.28515625, - "learning_rate": 0.00026963681476279956, - "loss": 0.6981, + "epoch": 3.6252390057361374, + "grad_norm": 0.3359375, + "learning_rate": 0.00014389014876247205, + "loss": 0.5045, "step": 1896 }, { - "epoch": 1.7163504968383019, - "grad_norm": 0.30078125, - "learning_rate": 0.00026950699724073333, - "loss": 0.6818, + "epoch": 3.632887189292543, + "grad_norm": 0.3125, + "learning_rate": 0.00014337048696194625, + "loss": 0.4814, "step": 1900 }, { - "epoch": 1.7199638663053296, - "grad_norm": 0.263671875, - "learning_rate": 0.00026937693418504246, - "loss": 0.7459, + "epoch": 3.6405353728489485, + "grad_norm": 0.32421875, + "learning_rate": 0.00014285090487357427, + "loss": 0.5416, "step": 1904 }, { - "epoch": 1.7235772357723578, - "grad_norm": 0.2890625, - "learning_rate": 0.0002692466258629486, - "loss": 0.7163, + "epoch": 3.6481835564053537, + "grad_norm": 0.32421875, + "learning_rate": 0.0001423314087447241, + "loss": 0.5236, "step": 1908 }, { - "epoch": 1.7271906052393857, - "grad_norm": 0.2578125, - "learning_rate": 0.0002691160725421774, - "loss": 0.7107, + "epoch": 3.655831739961759, + "grad_norm": 0.34765625, + "learning_rate": 0.00014181200482173015, + "loss": 0.5281, "step": 1912 }, { - "epoch": 1.7308039747064137, - "grad_norm": 0.2890625, - "learning_rate": 0.00026898527449095765, - "loss": 0.7582, + "epoch": 3.6634799235181643, + "grad_norm": 0.3046875, + "learning_rate": 0.00014129269934981802, + "loss": 0.5446, "step": 1916 }, { - "epoch": 1.7344173441734418, - "grad_norm": 0.283203125, - "learning_rate": 0.0002688542319780211, - "loss": 0.7266, + "epoch": 3.67112810707457, + "grad_norm": 0.359375, + "learning_rate": 0.00014077349857302983, + "loss": 0.4949, "step": 1920 }, { - "epoch": 1.7380307136404698, - "grad_norm": 0.2734375, - "learning_rate": 0.0002687229452726017, - "loss": 0.7625, + "epoch": 3.6787762906309753, + "grad_norm": 0.31640625, + "learning_rate": 0.00014025440873414863, + "loss": 0.4875, "step": 1924 }, { - "epoch": 1.7416440831074977, - "grad_norm": 0.263671875, - "learning_rate": 0.00026859141464443515, - "loss": 0.7086, + "epoch": 3.6864244741873806, + "grad_norm": 0.3203125, + "learning_rate": 0.0001397354360746237, + "loss": 0.528, "step": 1928 }, { - "epoch": 1.7452574525745257, - "grad_norm": 0.296875, - "learning_rate": 0.00026845964036375825, - "loss": 0.7283, + "epoch": 3.694072657743786, + "grad_norm": 0.298828125, + "learning_rate": 0.0001392165868344953, + "loss": 0.4827, "step": 1932 }, { - "epoch": 1.7488708220415536, - "grad_norm": 0.279296875, - "learning_rate": 0.00026832762270130844, - "loss": 0.7321, + "epoch": 3.701720841300191, + "grad_norm": 0.337890625, + "learning_rate": 0.0001386978672523198, + "loss": 0.5073, "step": 1936 }, { - "epoch": 1.7524841915085818, - "grad_norm": 0.279296875, - "learning_rate": 0.0002681953619283232, - "loss": 0.6968, + "epoch": 3.7093690248565965, + "grad_norm": 0.3203125, + "learning_rate": 0.0001381792835650945, + "loss": 0.5036, "step": 1940 }, { - "epoch": 1.7560975609756098, - "grad_norm": 0.2890625, - "learning_rate": 0.00026806285831653943, - "loss": 0.7761, + "epoch": 3.7170172084130018, + "grad_norm": 0.30859375, + "learning_rate": 0.00013766084200818272, + "loss": 0.5396, "step": 1944 }, { - "epoch": 1.7597109304426377, - "grad_norm": 0.27734375, - "learning_rate": 0.00026793011213819304, - "loss": 0.7347, + "epoch": 3.724665391969407, + "grad_norm": 0.345703125, + "learning_rate": 0.0001371425488152389, + "loss": 0.4815, "step": 1948 }, { - "epoch": 1.7633242999096659, - "grad_norm": 0.26953125, - "learning_rate": 0.0002677971236660183, - "loss": 0.7694, + "epoch": 3.7323135755258128, + "grad_norm": 0.3203125, + "learning_rate": 0.0001366244102181335, + "loss": 0.5306, "step": 1952 }, { - "epoch": 1.7669376693766936, - "grad_norm": 0.26171875, - "learning_rate": 0.0002676638931732472, - "loss": 0.7586, + "epoch": 3.739961759082218, + "grad_norm": 0.34375, + "learning_rate": 0.00013610643244687826, + "loss": 0.5419, "step": 1956 }, { - "epoch": 1.7705510388437218, - "grad_norm": 0.26953125, - "learning_rate": 0.0002675304209336091, - "loss": 0.7129, + "epoch": 3.7476099426386233, + "grad_norm": 0.3203125, + "learning_rate": 0.00013558862172955105, + "loss": 0.5204, "step": 1960 }, { - "epoch": 1.7741644083107497, - "grad_norm": 0.2890625, - "learning_rate": 0.00026739670722132993, - "loss": 0.7161, + "epoch": 3.7552581261950286, + "grad_norm": 0.3203125, + "learning_rate": 0.00013507098429222115, + "loss": 0.4982, "step": 1964 }, { - "epoch": 1.7777777777777777, - "grad_norm": 0.291015625, - "learning_rate": 0.00026726275231113173, - "loss": 0.7078, + "epoch": 3.762906309751434, + "grad_norm": 0.3203125, + "learning_rate": 0.00013455352635887438, + "loss": 0.4667, "step": 1968 }, { - "epoch": 1.7813911472448059, - "grad_norm": 0.26171875, - "learning_rate": 0.0002671285564782323, - "loss": 0.7319, + "epoch": 3.7705544933078396, + "grad_norm": 0.34765625, + "learning_rate": 0.00013403625415133824, + "loss": 0.4302, "step": 1972 }, { - "epoch": 1.7850045167118338, - "grad_norm": 0.27734375, - "learning_rate": 0.0002669941199983441, - "loss": 0.7176, + "epoch": 3.778202676864245, + "grad_norm": 0.310546875, + "learning_rate": 0.00013351917388920704, + "loss": 0.4545, "step": 1976 }, { - "epoch": 1.7886178861788617, - "grad_norm": 0.287109375, - "learning_rate": 0.0002668594431476743, - "loss": 0.757, + "epoch": 3.78585086042065, + "grad_norm": 0.337890625, + "learning_rate": 0.00013300229178976722, + "loss": 0.4953, "step": 1980 }, { - "epoch": 1.79223125564589, - "grad_norm": 0.296875, - "learning_rate": 0.0002667245262029238, - "loss": 0.7165, + "epoch": 3.7934990439770555, + "grad_norm": 0.3359375, + "learning_rate": 0.0001324856140679225, + "loss": 0.4966, "step": 1984 }, { - "epoch": 1.7958446251129176, - "grad_norm": 0.26953125, - "learning_rate": 0.0002665893694412868, - "loss": 0.7306, + "epoch": 3.801147227533461, + "grad_norm": 0.3203125, + "learning_rate": 0.0001319691469361193, + "loss": 0.5236, "step": 1988 }, { - "epoch": 1.7994579945799458, - "grad_norm": 0.283203125, - "learning_rate": 0.0002664539731404502, - "loss": 0.6908, + "epoch": 3.808795411089866, + "grad_norm": 0.328125, + "learning_rate": 0.00013145289660427173, + "loss": 0.5244, "step": 1992 }, { - "epoch": 1.8030713640469738, - "grad_norm": 0.296875, - "learning_rate": 0.00026631833757859304, - "loss": 0.6951, + "epoch": 3.8164435946462714, + "grad_norm": 0.3359375, + "learning_rate": 0.00013093686927968738, + "loss": 0.4982, "step": 1996 }, { - "epoch": 1.8066847335140017, - "grad_norm": 0.263671875, - "learning_rate": 0.000266182463034386, - "loss": 0.6831, + "epoch": 3.8240917782026767, + "grad_norm": 0.345703125, + "learning_rate": 0.00013042107116699228, + "loss": 0.4899, "step": 2000 }, { - "epoch": 1.8102981029810299, - "grad_norm": 0.291015625, - "learning_rate": 0.00026604634978699075, - "loss": 0.7225, + "epoch": 3.8317399617590824, + "grad_norm": 0.3125, + "learning_rate": 0.00012990550846805654, + "loss": 0.5296, "step": 2004 }, { - "epoch": 1.8139114724480578, - "grad_norm": 0.27734375, - "learning_rate": 0.0002659099981160592, - "loss": 0.7541, + "epoch": 3.8393881453154877, + "grad_norm": 0.322265625, + "learning_rate": 0.0001293901873819196, + "loss": 0.5331, "step": 2008 }, { - "epoch": 1.8175248419150858, - "grad_norm": 0.287109375, - "learning_rate": 0.0002657734083017335, - "loss": 0.6766, + "epoch": 3.847036328871893, + "grad_norm": 0.333984375, + "learning_rate": 0.00012887511410471589, + "loss": 0.501, "step": 2012 }, { - "epoch": 1.821138211382114, - "grad_norm": 0.2734375, - "learning_rate": 0.00026563658062464464, - "loss": 0.7134, + "epoch": 3.8546845124282982, + "grad_norm": 0.328125, + "learning_rate": 0.00012836029482960018, + "loss": 0.5254, "step": 2016 }, { - "epoch": 1.8247515808491417, - "grad_norm": 0.29296875, - "learning_rate": 0.00026549951536591264, - "loss": 0.7577, + "epoch": 3.8623326959847035, + "grad_norm": 0.337890625, + "learning_rate": 0.00012784573574667316, + "loss": 0.5009, "step": 2020 }, { - "epoch": 1.8283649503161699, - "grad_norm": 0.26953125, - "learning_rate": 0.0002653622128071455, - "loss": 0.6919, + "epoch": 3.8699808795411093, + "grad_norm": 0.353515625, + "learning_rate": 0.00012733144304290697, + "loss": 0.5107, "step": 2024 }, { - "epoch": 1.8319783197831978, - "grad_norm": 0.265625, - "learning_rate": 0.00026522467323043884, - "loss": 0.7221, + "epoch": 3.8776290630975145, + "grad_norm": 0.326171875, + "learning_rate": 0.0001268174229020709, + "loss": 0.5025, "step": 2028 }, { - "epoch": 1.8355916892502258, - "grad_norm": 0.27734375, - "learning_rate": 0.00026508689691837517, - "loss": 0.6636, + "epoch": 3.88527724665392, + "grad_norm": 0.318359375, + "learning_rate": 0.0001263036815046571, + "loss": 0.5239, "step": 2032 }, { - "epoch": 1.839205058717254, - "grad_norm": 0.322265625, - "learning_rate": 0.00026494888415402336, - "loss": 0.7046, + "epoch": 3.892925430210325, + "grad_norm": 0.333984375, + "learning_rate": 0.00012579022502780596, + "loss": 0.5112, "step": 2036 }, { - "epoch": 1.8428184281842819, - "grad_norm": 0.26953125, - "learning_rate": 0.0002648106352209382, - "loss": 0.7064, + "epoch": 3.9005736137667304, + "grad_norm": 0.322265625, + "learning_rate": 0.00012527705964523209, + "loss": 0.5182, "step": 2040 }, { - "epoch": 1.8464317976513098, - "grad_norm": 0.251953125, - "learning_rate": 0.00026467215040315963, - "loss": 0.6979, + "epoch": 3.9082217973231357, + "grad_norm": 0.322265625, + "learning_rate": 0.00012476419152715007, + "loss": 0.5505, "step": 2044 }, { - "epoch": 1.850045167118338, - "grad_norm": 0.296875, - "learning_rate": 0.0002645334299852122, - "loss": 0.7182, + "epoch": 3.915869980879541, + "grad_norm": 0.3125, + "learning_rate": 0.00012425162684020024, + "loss": 0.4957, "step": 2048 }, { - "epoch": 1.8536585365853657, - "grad_norm": 0.291015625, - "learning_rate": 0.0002643944742521046, - "loss": 0.7247, + "epoch": 3.9235181644359463, + "grad_norm": 0.322265625, + "learning_rate": 0.0001237393717473745, + "loss": 0.5132, "step": 2052 }, { - "epoch": 1.857271906052394, - "grad_norm": 0.28515625, - "learning_rate": 0.000264255283489329, - "loss": 0.6865, + "epoch": 3.9311663479923515, + "grad_norm": 0.3125, + "learning_rate": 0.0001232274324079422, + "loss": 0.516, "step": 2056 }, { - "epoch": 1.8608852755194218, - "grad_norm": 0.27734375, - "learning_rate": 0.0002641158579828602, - "loss": 0.6904, + "epoch": 3.9388145315487573, + "grad_norm": 0.33203125, + "learning_rate": 0.00012271581497737619, + "loss": 0.5156, "step": 2060 }, { - "epoch": 1.8644986449864498, - "grad_norm": 0.27734375, - "learning_rate": 0.0002639761980191557, - "loss": 0.6798, + "epoch": 3.9464627151051626, + "grad_norm": 0.314453125, + "learning_rate": 0.00012220452560727875, + "loss": 0.4621, "step": 2064 }, { - "epoch": 1.868112014453478, - "grad_norm": 0.271484375, - "learning_rate": 0.00026383630388515433, - "loss": 0.6785, + "epoch": 3.954110898661568, + "grad_norm": 0.33203125, + "learning_rate": 0.00012169357044530758, + "loss": 0.5206, "step": 2068 }, { - "epoch": 1.871725383920506, - "grad_norm": 0.267578125, - "learning_rate": 0.00026369617586827627, - "loss": 0.7089, + "epoch": 3.961759082217973, + "grad_norm": 0.322265625, + "learning_rate": 0.0001211829556351019, + "loss": 0.4511, "step": 2072 }, { - "epoch": 1.8753387533875339, - "grad_norm": 0.275390625, - "learning_rate": 0.0002635558142564221, - "loss": 0.729, + "epoch": 3.969407265774379, + "grad_norm": 0.328125, + "learning_rate": 0.00012067268731620861, + "loss": 0.5047, "step": 2076 }, { - "epoch": 1.878952122854562, - "grad_norm": 0.279296875, - "learning_rate": 0.0002634152193379725, - "loss": 0.7456, + "epoch": 3.977055449330784, + "grad_norm": 0.3359375, + "learning_rate": 0.00012016277162400848, + "loss": 0.5295, "step": 2080 }, { - "epoch": 1.8825654923215898, - "grad_norm": 0.279296875, - "learning_rate": 0.00026327439140178726, - "loss": 0.6812, + "epoch": 3.9847036328871894, + "grad_norm": 0.333984375, + "learning_rate": 0.00011965321468964237, + "loss": 0.5204, "step": 2084 }, { - "epoch": 1.886178861788618, - "grad_norm": 0.275390625, - "learning_rate": 0.00026313333073720507, - "loss": 0.6964, + "epoch": 3.9923518164435947, + "grad_norm": 0.333984375, + "learning_rate": 0.00011914402263993745, + "loss": 0.5064, "step": 2088 }, { - "epoch": 1.8897922312556459, - "grad_norm": 0.271484375, - "learning_rate": 0.0002629920376340427, - "loss": 0.7094, + "epoch": 4.0, + "grad_norm": 0.87109375, + "learning_rate": 0.00011863520159733357, + "loss": 0.53, "step": 2092 }, { - "epoch": 1.8934056007226738, - "grad_norm": 0.2578125, - "learning_rate": 0.00026285051238259465, - "loss": 0.7041, + "epoch": 4.007648183556405, + "grad_norm": 0.3046875, + "learning_rate": 0.00011812675767980972, + "loss": 0.4532, "step": 2096 }, { - "epoch": 1.897018970189702, - "grad_norm": 0.30078125, - "learning_rate": 0.0002627087552736321, - "loss": 0.6752, + "epoch": 4.015296367112811, + "grad_norm": 0.3046875, + "learning_rate": 0.00011761869700081036, + "loss": 0.4748, "step": 2100 }, { - "epoch": 1.90063233965673, - "grad_norm": 0.2470703125, - "learning_rate": 0.00026256676659840275, - "loss": 0.72, + "epoch": 4.022944550669216, + "grad_norm": 0.314453125, + "learning_rate": 0.00011711102566917194, + "loss": 0.4188, "step": 2104 }, { - "epoch": 1.904245709123758, - "grad_norm": 0.279296875, - "learning_rate": 0.00026242454664863015, - "loss": 0.7093, + "epoch": 4.030592734225621, + "grad_norm": 0.32421875, + "learning_rate": 0.00011660374978904947, + "loss": 0.4466, "step": 2108 }, { - "epoch": 1.907859078590786, - "grad_norm": 0.279296875, - "learning_rate": 0.0002622820957165128, - "loss": 0.6896, + "epoch": 4.038240917782026, + "grad_norm": 0.32421875, + "learning_rate": 0.00011609687545984315, + "loss": 0.4112, "step": 2112 }, { - "epoch": 1.9114724480578138, - "grad_norm": 0.291015625, - "learning_rate": 0.000262139414094724, - "loss": 0.7046, + "epoch": 4.045889101338432, + "grad_norm": 0.345703125, + "learning_rate": 0.00011559040877612497, + "loss": 0.4566, "step": 2116 }, { - "epoch": 1.915085817524842, - "grad_norm": 0.306640625, - "learning_rate": 0.00026199650207641085, - "loss": 0.729, + "epoch": 4.053537284894838, + "grad_norm": 0.33203125, + "learning_rate": 0.00011508435582756545, + "loss": 0.4413, "step": 2120 }, { - "epoch": 1.91869918699187, - "grad_norm": 0.283203125, - "learning_rate": 0.00026185335995519384, - "loss": 0.7202, + "epoch": 4.061185468451243, + "grad_norm": 0.34765625, + "learning_rate": 0.00011457872269886043, + "loss": 0.4435, "step": 2124 }, { - "epoch": 1.9223125564588979, - "grad_norm": 0.279296875, - "learning_rate": 0.00026170998802516624, - "loss": 0.6912, + "epoch": 4.0688336520076485, + "grad_norm": 0.341796875, + "learning_rate": 0.00011407351546965796, + "loss": 0.4568, "step": 2128 }, { - "epoch": 1.925925925925926, - "grad_norm": 0.2890625, - "learning_rate": 0.0002615663865808935, - "loss": 0.7145, + "epoch": 4.076481835564054, + "grad_norm": 0.328125, + "learning_rate": 0.00011356874021448506, + "loss": 0.4247, "step": 2132 }, { - "epoch": 1.9295392953929538, - "grad_norm": 0.31640625, - "learning_rate": 0.00026142255591741267, - "loss": 0.722, + "epoch": 4.084130019120459, + "grad_norm": 0.302734375, + "learning_rate": 0.00011306440300267482, + "loss": 0.3762, "step": 2136 }, { - "epoch": 1.933152664859982, - "grad_norm": 0.279296875, - "learning_rate": 0.00026127849633023157, - "loss": 0.6763, + "epoch": 4.091778202676864, + "grad_norm": 0.337890625, + "learning_rate": 0.00011256050989829337, + "loss": 0.4713, "step": 2140 }, { - "epoch": 1.9367660343270099, - "grad_norm": 0.296875, - "learning_rate": 0.0002611342081153284, - "loss": 0.7073, + "epoch": 4.09942638623327, + "grad_norm": 0.3203125, + "learning_rate": 0.00011205706696006698, + "loss": 0.4178, "step": 2144 }, { - "epoch": 1.9403794037940378, - "grad_norm": 0.287109375, - "learning_rate": 0.0002609896915691513, - "loss": 0.6981, + "epoch": 4.107074569789675, + "grad_norm": 0.353515625, + "learning_rate": 0.00011155408024130921, + "loss": 0.4266, "step": 2148 }, { - "epoch": 1.943992773261066, - "grad_norm": 0.283203125, - "learning_rate": 0.00026084494698861723, - "loss": 0.7239, + "epoch": 4.11472275334608, + "grad_norm": 0.31640625, + "learning_rate": 0.00011105155578984795, + "loss": 0.4242, "step": 2152 }, { - "epoch": 1.947606142728094, - "grad_norm": 0.283203125, - "learning_rate": 0.000260699974671112, - "loss": 0.7198, + "epoch": 4.1223709369024855, + "grad_norm": 0.310546875, + "learning_rate": 0.00011054949964795307, + "loss": 0.464, "step": 2156 }, { - "epoch": 1.951219512195122, - "grad_norm": 0.2890625, - "learning_rate": 0.0002605547749144889, - "loss": 0.6963, + "epoch": 4.130019120458891, + "grad_norm": 0.31640625, + "learning_rate": 0.00011004791785226347, + "loss": 0.4216, "step": 2160 }, { - "epoch": 1.95483288166215, - "grad_norm": 0.2734375, - "learning_rate": 0.00026040934801706874, - "loss": 0.7097, + "epoch": 4.137667304015296, + "grad_norm": 0.34375, + "learning_rate": 0.00010954681643371462, + "loss": 0.4222, "step": 2164 }, { - "epoch": 1.9584462511291778, - "grad_norm": 0.29296875, - "learning_rate": 0.00026026369427763923, - "loss": 0.7057, + "epoch": 4.145315487571701, + "grad_norm": 0.3203125, + "learning_rate": 0.00010904620141746601, + "loss": 0.4321, "step": 2168 }, { - "epoch": 1.962059620596206, - "grad_norm": 0.27734375, - "learning_rate": 0.00026011781399545365, - "loss": 0.6887, + "epoch": 4.1529636711281075, + "grad_norm": 0.330078125, + "learning_rate": 0.0001085460788228287, + "loss": 0.441, "step": 2172 }, { - "epoch": 1.965672990063234, - "grad_norm": 0.291015625, - "learning_rate": 0.00025997170747023106, - "loss": 0.6859, + "epoch": 4.160611854684513, + "grad_norm": 0.330078125, + "learning_rate": 0.00010804645466319292, + "loss": 0.4468, "step": 2176 }, { - "epoch": 1.9692863595302619, - "grad_norm": 0.287109375, - "learning_rate": 0.00025982537500215524, - "loss": 0.6541, + "epoch": 4.168260038240918, + "grad_norm": 0.33203125, + "learning_rate": 0.0001075473349459559, + "loss": 0.3948, "step": 2180 }, { - "epoch": 1.97289972899729, - "grad_norm": 0.275390625, - "learning_rate": 0.00025967881689187424, - "loss": 0.6997, + "epoch": 4.175908221797323, + "grad_norm": 0.33203125, + "learning_rate": 0.00010704872567244948, + "loss": 0.4233, "step": 2184 }, { - "epoch": 1.976513098464318, - "grad_norm": 0.28125, - "learning_rate": 0.00025953203344049965, - "loss": 0.6942, + "epoch": 4.183556405353729, + "grad_norm": 0.34765625, + "learning_rate": 0.00010655063283786795, + "loss": 0.4227, "step": 2188 }, { - "epoch": 1.980126467931346, - "grad_norm": 0.287109375, - "learning_rate": 0.00025938502494960607, - "loss": 0.77, + "epoch": 4.191204588910134, + "grad_norm": 0.337890625, + "learning_rate": 0.00010605306243119617, + "loss": 0.4242, "step": 2192 }, { - "epoch": 1.9837398373983741, - "grad_norm": 0.255859375, - "learning_rate": 0.00025923779172123035, - "loss": 0.6931, + "epoch": 4.198852772466539, + "grad_norm": 0.337890625, + "learning_rate": 0.00010555602043513724, + "loss": 0.4428, "step": 2196 }, { - "epoch": 1.9873532068654018, - "grad_norm": 0.29296875, - "learning_rate": 0.00025909033405787133, - "loss": 0.6612, + "epoch": 4.2065009560229445, + "grad_norm": 0.34375, + "learning_rate": 0.00010505951282604088, + "loss": 0.4132, "step": 2200 }, { - "epoch": 1.99096657633243, - "grad_norm": 0.29296875, - "learning_rate": 0.0002589426522624886, - "loss": 0.691, + "epoch": 4.21414913957935, + "grad_norm": 0.33984375, + "learning_rate": 0.00010456354557383139, + "loss": 0.4046, "step": 2204 }, { - "epoch": 1.994579945799458, - "grad_norm": 0.25390625, - "learning_rate": 0.0002587947466385026, - "loss": 0.7166, + "epoch": 4.221797323135755, + "grad_norm": 0.345703125, + "learning_rate": 0.00010406812464193584, + "loss": 0.4527, "step": 2208 }, { - "epoch": 1.998193315266486, - "grad_norm": 0.27734375, - "learning_rate": 0.0002586466174897934, - "loss": 0.6862, + "epoch": 4.22944550669216, + "grad_norm": 0.33203125, + "learning_rate": 0.00010357325598721255, + "loss": 0.4107, "step": 2212 }, { - "epoch": 2.001806684733514, - "grad_norm": 0.27734375, - "learning_rate": 0.0002584982651207005, - "loss": 0.6855, + "epoch": 4.237093690248566, + "grad_norm": 0.31640625, + "learning_rate": 0.00010307894555987927, + "loss": 0.4375, "step": 2216 }, { - "epoch": 2.005420054200542, - "grad_norm": 0.28515625, - "learning_rate": 0.00025834968983602175, - "loss": 0.606, + "epoch": 4.244741873804971, + "grad_norm": 0.328125, + "learning_rate": 0.00010258519930344179, + "loss": 0.4328, "step": 2220 }, { - "epoch": 2.00903342366757, - "grad_norm": 0.2890625, - "learning_rate": 0.00025820089194101317, - "loss": 0.6066, + "epoch": 4.252390057361376, + "grad_norm": 0.34375, + "learning_rate": 0.0001020920231546223, + "loss": 0.4388, "step": 2224 }, { - "epoch": 2.012646793134598, - "grad_norm": 0.275390625, - "learning_rate": 0.0002580518717413882, - "loss": 0.647, + "epoch": 4.260038240917782, + "grad_norm": 0.33984375, + "learning_rate": 0.00010159942304328819, + "loss": 0.4443, "step": 2228 }, { - "epoch": 2.016260162601626, - "grad_norm": 0.26953125, - "learning_rate": 0.00025790262954331696, - "loss": 0.6474, + "epoch": 4.267686424474188, + "grad_norm": 0.361328125, + "learning_rate": 0.00010110740489238066, + "loss": 0.4446, "step": 2232 }, { - "epoch": 2.019873532068654, - "grad_norm": 0.287109375, - "learning_rate": 0.00025775316565342565, - "loss": 0.64, + "epoch": 4.275334608030593, + "grad_norm": 0.341796875, + "learning_rate": 0.00010061597461784346, + "loss": 0.4269, "step": 2236 }, { - "epoch": 2.0234869015356822, - "grad_norm": 0.294921875, - "learning_rate": 0.0002576034803787959, - "loss": 0.651, + "epoch": 4.282982791586998, + "grad_norm": 0.33984375, + "learning_rate": 0.00010012513812855191, + "loss": 0.4326, "step": 2240 }, { - "epoch": 2.02710027100271, - "grad_norm": 0.279296875, - "learning_rate": 0.00025745357402696424, - "loss": 0.6223, + "epoch": 4.2906309751434035, + "grad_norm": 0.35546875, + "learning_rate": 9.963490132624169e-05, + "loss": 0.4342, "step": 2244 }, { - "epoch": 2.030713640469738, - "grad_norm": 0.267578125, - "learning_rate": 0.0002573034469059215, - "loss": 0.6823, + "epoch": 4.298279158699809, + "grad_norm": 0.3515625, + "learning_rate": 9.914527010543795e-05, + "loss": 0.4157, "step": 2248 }, { - "epoch": 2.034327009936766, - "grad_norm": 0.275390625, - "learning_rate": 0.00025715309932411196, - "loss": 0.6411, + "epoch": 4.305927342256214, + "grad_norm": 0.3359375, + "learning_rate": 9.865625035338447e-05, + "loss": 0.4237, "step": 2252 }, { - "epoch": 2.037940379403794, - "grad_norm": 0.30078125, - "learning_rate": 0.00025700253159043296, - "loss": 0.6004, + "epoch": 4.313575525812619, + "grad_norm": 0.314453125, + "learning_rate": 9.816784794997275e-05, + "loss": 0.4033, "step": 2256 }, { - "epoch": 2.041553748870822, - "grad_norm": 0.29296875, - "learning_rate": 0.000256851744014234, - "loss": 0.6005, + "epoch": 4.321223709369025, + "grad_norm": 0.359375, + "learning_rate": 9.76800687676715e-05, + "loss": 0.4436, "step": 2260 }, { - "epoch": 2.04516711833785, - "grad_norm": 0.271484375, - "learning_rate": 0.0002567007369053164, - "loss": 0.6616, + "epoch": 4.32887189292543, + "grad_norm": 0.3359375, + "learning_rate": 9.719291867145583e-05, + "loss": 0.4384, "step": 2264 }, { - "epoch": 2.048780487804878, - "grad_norm": 0.30859375, - "learning_rate": 0.00025654951057393254, - "loss": 0.5758, + "epoch": 4.336520076481835, + "grad_norm": 0.341796875, + "learning_rate": 9.670640351873688e-05, + "loss": 0.4512, "step": 2268 }, { - "epoch": 2.0523938572719063, - "grad_norm": 0.287109375, - "learning_rate": 0.0002563980653307851, - "loss": 0.6309, + "epoch": 4.3441682600382405, + "grad_norm": 0.34375, + "learning_rate": 9.62205291592913e-05, + "loss": 0.4259, "step": 2272 }, { - "epoch": 2.056007226738934, - "grad_norm": 0.3125, - "learning_rate": 0.00025624640148702647, - "loss": 0.6173, + "epoch": 4.351816443594647, + "grad_norm": 0.337890625, + "learning_rate": 9.573530143519098e-05, + "loss": 0.4178, "step": 2276 }, { - "epoch": 2.059620596205962, - "grad_norm": 0.291015625, - "learning_rate": 0.00025609451935425844, - "loss": 0.6838, + "epoch": 4.359464627151052, + "grad_norm": 0.3671875, + "learning_rate": 9.525072618073277e-05, + "loss": 0.3608, "step": 2280 }, { - "epoch": 2.06323396567299, - "grad_norm": 0.30078125, - "learning_rate": 0.00025594241924453096, - "loss": 0.6917, + "epoch": 4.367112810707457, + "grad_norm": 0.32421875, + "learning_rate": 9.476680922236831e-05, + "loss": 0.4489, "step": 2284 }, { - "epoch": 2.066847335140018, - "grad_norm": 0.28515625, - "learning_rate": 0.0002557901014703421, - "loss": 0.6417, + "epoch": 4.374760994263863, + "grad_norm": 0.330078125, + "learning_rate": 9.428355637863402e-05, + "loss": 0.4346, "step": 2288 }, { - "epoch": 2.0704607046070462, - "grad_norm": 0.29296875, - "learning_rate": 0.0002556375663446369, - "loss": 0.6295, + "epoch": 4.382409177820268, + "grad_norm": 0.33203125, + "learning_rate": 9.380097346008112e-05, + "loss": 0.4542, "step": 2292 }, { - "epoch": 2.074074074074074, - "grad_norm": 0.287109375, - "learning_rate": 0.00025548481418080713, - "loss": 0.6559, + "epoch": 4.390057361376673, + "grad_norm": 0.341796875, + "learning_rate": 9.331906626920576e-05, + "loss": 0.4395, "step": 2296 }, { - "epoch": 2.077687443541102, - "grad_norm": 0.27734375, - "learning_rate": 0.0002553318452926904, - "loss": 0.6531, + "epoch": 4.397705544933078, + "grad_norm": 0.341796875, + "learning_rate": 9.283784060037921e-05, + "loss": 0.4858, "step": 2300 }, { - "epoch": 2.08130081300813, - "grad_norm": 0.3046875, - "learning_rate": 0.00025517865999456963, - "loss": 0.6311, + "epoch": 4.405353728489484, + "grad_norm": 0.30859375, + "learning_rate": 9.235730223977837e-05, + "loss": 0.4148, "step": 2304 }, { - "epoch": 2.084914182475158, - "grad_norm": 0.28515625, - "learning_rate": 0.0002550252586011723, - "loss": 0.6364, + "epoch": 4.413001912045889, + "grad_norm": 0.36328125, + "learning_rate": 9.187745696531584e-05, + "loss": 0.4579, "step": 2308 }, { - "epoch": 2.088527551942186, - "grad_norm": 0.28515625, - "learning_rate": 0.00025487164142767, - "loss": 0.6485, + "epoch": 4.420650095602294, + "grad_norm": 0.357421875, + "learning_rate": 9.139831054657081e-05, + "loss": 0.457, "step": 2312 }, { - "epoch": 2.092140921409214, - "grad_norm": 0.298828125, - "learning_rate": 0.00025471780878967744, - "loss": 0.6544, + "epoch": 4.4282982791587, + "grad_norm": 0.341796875, + "learning_rate": 9.091986874471956e-05, + "loss": 0.4257, "step": 2316 }, { - "epoch": 2.095754290876242, - "grad_norm": 0.29296875, - "learning_rate": 0.00025456376100325224, - "loss": 0.6185, + "epoch": 4.435946462715105, + "grad_norm": 0.3515625, + "learning_rate": 9.044213731246614e-05, + "loss": 0.4287, "step": 2320 }, { - "epoch": 2.0993676603432703, - "grad_norm": 0.28125, - "learning_rate": 0.00025440949838489394, - "loss": 0.6092, + "epoch": 4.44359464627151, + "grad_norm": 0.3359375, + "learning_rate": 8.99651219939732e-05, + "loss": 0.4482, "step": 2324 }, { - "epoch": 2.102981029810298, - "grad_norm": 0.328125, - "learning_rate": 0.0002542550212515435, - "loss": 0.5907, + "epoch": 4.451242829827915, + "grad_norm": 0.345703125, + "learning_rate": 8.948882852479305e-05, + "loss": 0.4772, "step": 2328 }, { - "epoch": 2.106594399277326, - "grad_norm": 0.30078125, - "learning_rate": 0.0002541003299205825, - "loss": 0.6177, + "epoch": 4.458891013384322, + "grad_norm": 0.33203125, + "learning_rate": 8.901326263179851e-05, + "loss": 0.4067, "step": 2332 }, { - "epoch": 2.110207768744354, - "grad_norm": 0.3125, - "learning_rate": 0.0002539454247098328, - "loss": 0.6671, + "epoch": 4.466539196940727, + "grad_norm": 0.373046875, + "learning_rate": 8.85384300331142e-05, + "loss": 0.4399, "step": 2336 }, { - "epoch": 2.113821138211382, - "grad_norm": 0.283203125, - "learning_rate": 0.00025379030593755545, - "loss": 0.5961, + "epoch": 4.474187380497132, + "grad_norm": 0.3671875, + "learning_rate": 8.80643364380477e-05, + "loss": 0.4726, "step": 2340 }, { - "epoch": 2.1174345076784102, - "grad_norm": 0.279296875, - "learning_rate": 0.0002536349739224505, - "loss": 0.6423, + "epoch": 4.4818355640535374, + "grad_norm": 0.349609375, + "learning_rate": 8.759098754702099e-05, + "loss": 0.4514, "step": 2344 }, { - "epoch": 2.121047877145438, - "grad_norm": 0.314453125, - "learning_rate": 0.00025347942898365603, - "loss": 0.5715, + "epoch": 4.489483747609943, + "grad_norm": 0.36328125, + "learning_rate": 8.711838905150179e-05, + "loss": 0.4502, "step": 2348 }, { - "epoch": 2.124661246612466, - "grad_norm": 0.275390625, - "learning_rate": 0.0002533236714407475, - "loss": 0.623, + "epoch": 4.497131931166348, + "grad_norm": 0.328125, + "learning_rate": 8.664654663393516e-05, + "loss": 0.4366, "step": 2352 }, { - "epoch": 2.1282746160794943, - "grad_norm": 0.30859375, - "learning_rate": 0.0002531677016137374, - "loss": 0.6196, + "epoch": 4.504780114722753, + "grad_norm": 0.341796875, + "learning_rate": 8.617546596767534e-05, + "loss": 0.437, "step": 2356 }, { - "epoch": 2.131887985546522, - "grad_norm": 0.28515625, - "learning_rate": 0.000253011519823074, - "loss": 0.6454, + "epoch": 4.512428298279159, + "grad_norm": 0.337890625, + "learning_rate": 8.570515271691723e-05, + "loss": 0.4313, "step": 2360 }, { - "epoch": 2.13550135501355, - "grad_norm": 0.3046875, - "learning_rate": 0.00025285512638964145, - "loss": 0.6169, + "epoch": 4.520076481835564, + "grad_norm": 0.34765625, + "learning_rate": 8.523561253662864e-05, + "loss": 0.447, "step": 2364 }, { - "epoch": 2.139114724480578, - "grad_norm": 0.30078125, - "learning_rate": 0.0002526985216347585, - "loss": 0.6842, + "epoch": 4.527724665391969, + "grad_norm": 0.357421875, + "learning_rate": 8.476685107248197e-05, + "loss": 0.4488, "step": 2368 }, { - "epoch": 2.142728093947606, - "grad_norm": 0.306640625, - "learning_rate": 0.0002525417058801781, - "loss": 0.6178, + "epoch": 4.5353728489483744, + "grad_norm": 0.33203125, + "learning_rate": 8.429887396078655e-05, + "loss": 0.469, "step": 2372 }, { - "epoch": 2.1463414634146343, - "grad_norm": 0.310546875, - "learning_rate": 0.0002523846794480869, - "loss": 0.6206, + "epoch": 4.54302103250478, + "grad_norm": 0.3515625, + "learning_rate": 8.38316868284207e-05, + "loss": 0.4886, "step": 2376 }, { - "epoch": 2.149954832881662, - "grad_norm": 0.294921875, - "learning_rate": 0.0002522274426611041, - "loss": 0.6268, + "epoch": 4.550669216061186, + "grad_norm": 0.341796875, + "learning_rate": 8.336529529276421e-05, + "loss": 0.433, "step": 2380 }, { - "epoch": 2.15356820234869, - "grad_norm": 0.296875, - "learning_rate": 0.00025206999584228125, - "loss": 0.6667, + "epoch": 4.558317399617591, + "grad_norm": 0.33984375, + "learning_rate": 8.289970496163085e-05, + "loss": 0.4029, "step": 2384 }, { - "epoch": 2.1571815718157183, - "grad_norm": 0.302734375, - "learning_rate": 0.00025191233931510143, - "loss": 0.646, + "epoch": 4.5659655831739965, + "grad_norm": 0.322265625, + "learning_rate": 8.243492143320058e-05, + "loss": 0.4198, "step": 2388 }, { - "epoch": 2.160794941282746, - "grad_norm": 0.291015625, - "learning_rate": 0.00025175447340347856, - "loss": 0.6293, + "epoch": 4.573613766730402, + "grad_norm": 0.341796875, + "learning_rate": 8.197095029595276e-05, + "loss": 0.4377, "step": 2392 }, { - "epoch": 2.1644083107497742, - "grad_norm": 0.3046875, - "learning_rate": 0.0002515963984317567, - "loss": 0.6158, + "epoch": 4.581261950286807, + "grad_norm": 0.359375, + "learning_rate": 8.150779712859854e-05, + "loss": 0.4263, "step": 2396 }, { - "epoch": 2.168021680216802, - "grad_norm": 0.314453125, - "learning_rate": 0.00025143811472470943, - "loss": 0.6386, + "epoch": 4.588910133843212, + "grad_norm": 0.349609375, + "learning_rate": 8.104546750001402e-05, + "loss": 0.4784, "step": 2400 }, { - "epoch": 2.17163504968383, - "grad_norm": 0.3203125, - "learning_rate": 0.00025127962260753934, - "loss": 0.6441, + "epoch": 4.596558317399618, + "grad_norm": 0.337890625, + "learning_rate": 8.05839669691732e-05, + "loss": 0.4549, "step": 2404 }, { - "epoch": 2.1752484191508583, - "grad_norm": 0.30078125, - "learning_rate": 0.0002511209224058771, - "loss": 0.6115, + "epoch": 4.604206500956023, + "grad_norm": 0.33203125, + "learning_rate": 8.01233010850811e-05, + "loss": 0.4519, "step": 2408 }, { - "epoch": 2.178861788617886, - "grad_norm": 0.298828125, - "learning_rate": 0.0002509620144457808, - "loss": 0.6351, + "epoch": 4.611854684512428, + "grad_norm": 0.37109375, + "learning_rate": 7.966347538670712e-05, + "loss": 0.4242, "step": 2412 }, { - "epoch": 2.182475158084914, - "grad_norm": 0.294921875, - "learning_rate": 0.0002508028990537356, - "loss": 0.6807, + "epoch": 4.6195028680688335, + "grad_norm": 0.333984375, + "learning_rate": 7.92044954029184e-05, + "loss": 0.4647, "step": 2416 }, { - "epoch": 2.1860885275519424, - "grad_norm": 0.279296875, - "learning_rate": 0.0002506435765566527, - "loss": 0.6026, + "epoch": 4.627151051625239, + "grad_norm": 0.341796875, + "learning_rate": 7.874636665241335e-05, + "loss": 0.46, "step": 2420 }, { - "epoch": 2.18970189701897, - "grad_norm": 0.302734375, - "learning_rate": 0.0002504840472818687, - "loss": 0.6566, + "epoch": 4.634799235181644, + "grad_norm": 0.34375, + "learning_rate": 7.828909464365531e-05, + "loss": 0.4169, "step": 2424 }, { - "epoch": 2.1933152664859983, - "grad_norm": 0.298828125, - "learning_rate": 0.0002503243115571454, - "loss": 0.6153, + "epoch": 4.642447418738049, + "grad_norm": 0.349609375, + "learning_rate": 7.783268487480626e-05, + "loss": 0.4251, "step": 2428 }, { - "epoch": 2.196928635953026, - "grad_norm": 0.32421875, - "learning_rate": 0.00025016436971066837, - "loss": 0.632, + "epoch": 4.650095602294455, + "grad_norm": 0.337890625, + "learning_rate": 7.73771428336608e-05, + "loss": 0.4232, "step": 2432 }, { - "epoch": 2.200542005420054, - "grad_norm": 0.296875, - "learning_rate": 0.00025000422207104684, - "loss": 0.6327, + "epoch": 4.657743785850861, + "grad_norm": 0.3671875, + "learning_rate": 7.692247399758008e-05, + "loss": 0.4836, "step": 2436 }, { - "epoch": 2.2041553748870824, - "grad_norm": 0.296875, - "learning_rate": 0.0002498438689673129, - "loss": 0.6611, + "epoch": 4.665391969407266, + "grad_norm": 0.341796875, + "learning_rate": 7.6468683833426e-05, + "loss": 0.3917, "step": 2440 }, { - "epoch": 2.20776874435411, - "grad_norm": 0.294921875, - "learning_rate": 0.0002496833107289207, - "loss": 0.6004, + "epoch": 4.673040152963671, + "grad_norm": 0.345703125, + "learning_rate": 7.601577779749545e-05, + "loss": 0.4153, "step": 2444 }, { - "epoch": 2.2113821138211383, - "grad_norm": 0.318359375, - "learning_rate": 0.00024952254768574584, - "loss": 0.581, + "epoch": 4.680688336520077, + "grad_norm": 0.33984375, + "learning_rate": 7.55637613354547e-05, + "loss": 0.4103, "step": 2448 }, { - "epoch": 2.2149954832881664, - "grad_norm": 0.30078125, - "learning_rate": 0.0002493615801680848, - "loss": 0.6382, + "epoch": 4.688336520076482, + "grad_norm": 0.34765625, + "learning_rate": 7.511263988227397e-05, + "loss": 0.4858, "step": 2452 }, { - "epoch": 2.218608852755194, - "grad_norm": 0.30859375, - "learning_rate": 0.0002492004085066541, - "loss": 0.6183, + "epoch": 4.695984703632887, + "grad_norm": 0.369140625, + "learning_rate": 7.466241886216198e-05, + "loss": 0.468, "step": 2456 }, { - "epoch": 2.2222222222222223, - "grad_norm": 0.31640625, - "learning_rate": 0.0002490390330325896, - "loss": 0.6295, + "epoch": 4.7036328871892925, + "grad_norm": 0.34765625, + "learning_rate": 7.421310368850085e-05, + "loss": 0.4817, "step": 2460 }, { - "epoch": 2.22583559168925, - "grad_norm": 0.298828125, - "learning_rate": 0.00024887745407744605, - "loss": 0.6535, + "epoch": 4.711281070745698, + "grad_norm": 0.35546875, + "learning_rate": 7.376469976378094e-05, + "loss": 0.3923, "step": 2464 }, { - "epoch": 2.229448961156278, - "grad_norm": 0.283203125, - "learning_rate": 0.00024871567197319616, - "loss": 0.6478, + "epoch": 4.718929254302103, + "grad_norm": 0.337890625, + "learning_rate": 7.33172124795359e-05, + "loss": 0.3954, "step": 2468 }, { - "epoch": 2.2330623306233064, - "grad_norm": 0.291015625, - "learning_rate": 0.0002485536870522301, - "loss": 0.6159, + "epoch": 4.726577437858508, + "grad_norm": 0.345703125, + "learning_rate": 7.287064721627782e-05, + "loss": 0.4606, "step": 2472 }, { - "epoch": 2.236675700090334, - "grad_norm": 0.30078125, - "learning_rate": 0.0002483914996473547, - "loss": 0.6679, + "epoch": 4.734225621414914, + "grad_norm": 0.357421875, + "learning_rate": 7.242500934343262e-05, + "loss": 0.4253, "step": 2476 }, { - "epoch": 2.2402890695573623, - "grad_norm": 0.306640625, - "learning_rate": 0.00024822911009179276, - "loss": 0.6093, + "epoch": 4.741873804971319, + "grad_norm": 0.3671875, + "learning_rate": 7.19803042192754e-05, + "loss": 0.4431, "step": 2480 }, { - "epoch": 2.2439024390243905, - "grad_norm": 0.28515625, - "learning_rate": 0.0002480665187191825, - "loss": 0.6462, + "epoch": 4.749521988527725, + "grad_norm": 0.322265625, + "learning_rate": 7.153653719086604e-05, + "loss": 0.4097, "step": 2484 }, { - "epoch": 2.247515808491418, - "grad_norm": 0.294921875, - "learning_rate": 0.00024790372586357666, - "loss": 0.6456, + "epoch": 4.75717017208413, + "grad_norm": 0.322265625, + "learning_rate": 7.109371359398493e-05, + "loss": 0.4208, "step": 2488 }, { - "epoch": 2.2511291779584464, - "grad_norm": 0.3046875, - "learning_rate": 0.0002477407318594421, - "loss": 0.6216, + "epoch": 4.764818355640536, + "grad_norm": 0.345703125, + "learning_rate": 7.06518387530688e-05, + "loss": 0.4396, "step": 2492 }, { - "epoch": 2.254742547425474, - "grad_norm": 0.30078125, - "learning_rate": 0.0002475775370416589, - "loss": 0.6621, + "epoch": 4.772466539196941, + "grad_norm": 0.328125, + "learning_rate": 7.021091798114667e-05, + "loss": 0.4212, "step": 2496 }, { - "epoch": 2.2583559168925023, - "grad_norm": 0.302734375, - "learning_rate": 0.0002474141417455195, - "loss": 0.6094, + "epoch": 4.780114722753346, + "grad_norm": 0.33203125, + "learning_rate": 6.977095657977603e-05, + "loss": 0.4742, "step": 2500 }, { - "epoch": 2.2619692863595304, - "grad_norm": 0.337890625, - "learning_rate": 0.00024725054630672866, - "loss": 0.6318, + "epoch": 4.7877629063097515, + "grad_norm": 0.341796875, + "learning_rate": 6.933195983897905e-05, + "loss": 0.4198, "step": 2504 }, { - "epoch": 2.265582655826558, - "grad_norm": 0.291015625, - "learning_rate": 0.0002470867510614019, - "loss": 0.6301, + "epoch": 4.795411089866157, + "grad_norm": 0.361328125, + "learning_rate": 6.889393303717898e-05, + "loss": 0.4401, "step": 2508 }, { - "epoch": 2.2691960252935863, - "grad_norm": 0.3125, - "learning_rate": 0.00024692275634606564, - "loss": 0.6492, + "epoch": 4.803059273422562, + "grad_norm": 0.34765625, + "learning_rate": 6.845688144113663e-05, + "loss": 0.4273, "step": 2512 }, { - "epoch": 2.272809394760614, - "grad_norm": 0.306640625, - "learning_rate": 0.0002467585624976558, - "loss": 0.6677, + "epoch": 4.810707456978967, + "grad_norm": 0.322265625, + "learning_rate": 6.802081030588722e-05, + "loss": 0.4589, "step": 2516 }, { - "epoch": 2.2764227642276422, - "grad_norm": 0.31640625, - "learning_rate": 0.00024659416985351763, - "loss": 0.6629, + "epoch": 4.818355640535373, + "grad_norm": 0.34765625, + "learning_rate": 6.758572487467698e-05, + "loss": 0.4835, "step": 2520 }, { - "epoch": 2.2800361336946704, - "grad_norm": 0.30078125, - "learning_rate": 0.0002464295787514047, - "loss": 0.6153, + "epoch": 4.826003824091778, + "grad_norm": 0.34375, + "learning_rate": 6.715163037890021e-05, + "loss": 0.4493, "step": 2524 }, { - "epoch": 2.283649503161698, - "grad_norm": 0.2890625, - "learning_rate": 0.00024626478952947847, - "loss": 0.5951, + "epoch": 4.833652007648183, + "grad_norm": 0.3515625, + "learning_rate": 6.671853203803641e-05, + "loss": 0.4032, "step": 2528 }, { - "epoch": 2.2872628726287263, - "grad_norm": 0.294921875, - "learning_rate": 0.00024609980252630735, - "loss": 0.6298, + "epoch": 4.8413001912045885, + "grad_norm": 0.34765625, + "learning_rate": 6.628643505958742e-05, + "loss": 0.4498, "step": 2532 }, { - "epoch": 2.2908762420957545, - "grad_norm": 0.3046875, - "learning_rate": 0.000245934618080866, - "loss": 0.6122, + "epoch": 4.848948374760994, + "grad_norm": 0.34375, + "learning_rate": 6.585534463901493e-05, + "loss": 0.4255, "step": 2536 }, { - "epoch": 2.294489611562782, - "grad_norm": 0.29296875, - "learning_rate": 0.0002457692365325349, - "loss": 0.6431, + "epoch": 4.8565965583174, + "grad_norm": 0.388671875, + "learning_rate": 6.542526595967795e-05, + "loss": 0.4285, "step": 2540 }, { - "epoch": 2.2981029810298104, - "grad_norm": 0.291015625, - "learning_rate": 0.00024560365822109953, - "loss": 0.6434, + "epoch": 4.864244741873805, + "grad_norm": 0.337890625, + "learning_rate": 6.499620419277036e-05, + "loss": 0.4507, "step": 2544 }, { - "epoch": 2.3017163504968385, - "grad_norm": 0.330078125, - "learning_rate": 0.0002454378834867494, - "loss": 0.6324, + "epoch": 4.871892925430211, + "grad_norm": 0.333984375, + "learning_rate": 6.456816449725892e-05, + "loss": 0.4196, "step": 2548 }, { - "epoch": 2.3053297199638663, - "grad_norm": 0.33203125, - "learning_rate": 0.0002452719126700778, - "loss": 0.6332, + "epoch": 4.879541108986616, + "grad_norm": 0.345703125, + "learning_rate": 6.414115201982134e-05, + "loss": 0.4061, "step": 2552 }, { - "epoch": 2.3089430894308944, - "grad_norm": 0.326171875, - "learning_rate": 0.0002451057461120808, - "loss": 0.676, + "epoch": 4.887189292543021, + "grad_norm": 0.33984375, + "learning_rate": 6.371517189478403e-05, + "loss": 0.4199, "step": 2556 }, { - "epoch": 2.312556458897922, - "grad_norm": 0.30078125, - "learning_rate": 0.00024493938415415657, - "loss": 0.6499, + "epoch": 4.894837476099426, + "grad_norm": 0.3515625, + "learning_rate": 6.329022924406061e-05, + "loss": 0.452, "step": 2560 }, { - "epoch": 2.3161698283649503, - "grad_norm": 0.30078125, - "learning_rate": 0.00024477282713810484, - "loss": 0.5823, + "epoch": 4.902485659655832, + "grad_norm": 0.345703125, + "learning_rate": 6.286632917709031e-05, + "loss": 0.4516, "step": 2564 }, { - "epoch": 2.3197831978319785, - "grad_norm": 0.326171875, - "learning_rate": 0.00024460607540612603, - "loss": 0.6632, + "epoch": 4.910133843212237, + "grad_norm": 0.33984375, + "learning_rate": 6.244347679077651e-05, + "loss": 0.4622, "step": 2568 }, { - "epoch": 2.3233965672990062, - "grad_norm": 0.294921875, - "learning_rate": 0.0002444391293008207, - "loss": 0.5896, + "epoch": 4.917782026768642, + "grad_norm": 0.333984375, + "learning_rate": 6.202167716942543e-05, + "loss": 0.4387, "step": 2572 }, { - "epoch": 2.3270099367660344, - "grad_norm": 0.298828125, - "learning_rate": 0.0002442719891651886, - "loss": 0.6416, + "epoch": 4.925430210325048, + "grad_norm": 0.35546875, + "learning_rate": 6.160093538468505e-05, + "loss": 0.4064, "step": 2576 }, { - "epoch": 2.330623306233062, - "grad_norm": 0.291015625, - "learning_rate": 0.00024410465534262834, - "loss": 0.6521, + "epoch": 4.933078393881453, + "grad_norm": 0.3671875, + "learning_rate": 6.118125649548405e-05, + "loss": 0.4728, "step": 2580 }, { - "epoch": 2.3342366757000903, - "grad_norm": 0.306640625, - "learning_rate": 0.0002439371281769363, - "loss": 0.5804, + "epoch": 4.940726577437858, + "grad_norm": 0.35546875, + "learning_rate": 6.076264554797112e-05, + "loss": 0.4555, "step": 2584 }, { - "epoch": 2.3378500451671185, - "grad_norm": 0.298828125, - "learning_rate": 0.00024376940801230614, - "loss": 0.6091, + "epoch": 4.948374760994264, + "grad_norm": 0.337890625, + "learning_rate": 6.0345107575454105e-05, + "loss": 0.427, "step": 2588 }, { - "epoch": 2.341463414634146, - "grad_norm": 0.2890625, - "learning_rate": 0.00024360149519332808, - "loss": 0.6804, + "epoch": 4.95602294455067, + "grad_norm": 0.341796875, + "learning_rate": 5.992864759833963e-05, + "loss": 0.3841, "step": 2592 }, { - "epoch": 2.3450767841011744, - "grad_norm": 0.2890625, - "learning_rate": 0.00024343339006498813, - "loss": 0.6016, + "epoch": 4.963671128107075, + "grad_norm": 0.330078125, + "learning_rate": 5.9513270624072655e-05, + "loss": 0.4289, "step": 2596 }, { - "epoch": 2.348690153568202, - "grad_norm": 0.30078125, - "learning_rate": 0.00024326509297266746, - "loss": 0.6019, + "epoch": 4.97131931166348, + "grad_norm": 0.353515625, + "learning_rate": 5.9098981647076345e-05, + "loss": 0.4538, "step": 2600 }, { - "epoch": 2.3523035230352303, - "grad_norm": 0.298828125, - "learning_rate": 0.00024309660426214154, - "loss": 0.6194, + "epoch": 4.9789674952198855, + "grad_norm": 0.33203125, + "learning_rate": 5.8685785648691894e-05, + "loss": 0.3954, "step": 2604 }, { - "epoch": 2.3559168925022584, - "grad_norm": 0.2890625, - "learning_rate": 0.00024292792427957962, - "loss": 0.6322, + "epoch": 4.986615678776291, + "grad_norm": 0.34765625, + "learning_rate": 5.8273687597118765e-05, + "loss": 0.4614, "step": 2608 }, { - "epoch": 2.359530261969286, - "grad_norm": 0.302734375, - "learning_rate": 0.00024275905337154386, - "loss": 0.6148, + "epoch": 4.994263862332696, + "grad_norm": 0.3359375, + "learning_rate": 5.786269244735488e-05, + "loss": 0.4429, "step": 2612 }, { - "epoch": 2.3631436314363143, - "grad_norm": 0.296875, - "learning_rate": 0.00024258999188498878, - "loss": 0.6429, + "epoch": 5.001912045889101, + "grad_norm": 0.30859375, + "learning_rate": 5.7452805141137034e-05, + "loss": 0.3411, "step": 2616 }, { - "epoch": 2.3667570009033425, - "grad_norm": 0.30859375, - "learning_rate": 0.00024242074016726035, - "loss": 0.6343, + "epoch": 5.009560229445507, + "grad_norm": 0.318359375, + "learning_rate": 5.704403060688158e-05, + "loss": 0.3973, "step": 2620 }, { - "epoch": 2.3703703703703702, + "epoch": 5.017208413001912, "grad_norm": 0.314453125, - "learning_rate": 0.00024225129856609545, - "loss": 0.6325, + "learning_rate": 5.663637375962489e-05, + "loss": 0.4319, "step": 2624 }, { - "epoch": 2.3739837398373984, - "grad_norm": 0.298828125, - "learning_rate": 0.00024208166742962107, - "loss": 0.6194, + "epoch": 5.024856596558317, + "grad_norm": 0.3203125, + "learning_rate": 5.6229839500964635e-05, + "loss": 0.3768, "step": 2628 }, { - "epoch": 2.3775971093044266, - "grad_norm": 0.298828125, - "learning_rate": 0.00024191184710635358, - "loss": 0.6261, + "epoch": 5.0325047801147225, + "grad_norm": 0.314453125, + "learning_rate": 5.582443271900063e-05, + "loss": 0.3683, "step": 2632 }, { - "epoch": 2.3812104787714543, - "grad_norm": 0.306640625, - "learning_rate": 0.0002417418379451981, - "loss": 0.5721, + "epoch": 5.040152963671128, + "grad_norm": 0.35546875, + "learning_rate": 5.542015828827609e-05, + "loss": 0.4145, "step": 2636 }, { - "epoch": 2.3848238482384825, - "grad_norm": 0.322265625, - "learning_rate": 0.00024157164029544765, - "loss": 0.6571, + "epoch": 5.047801147227533, + "grad_norm": 0.35546875, + "learning_rate": 5.5017021069719014e-05, + "loss": 0.4116, "step": 2640 }, { - "epoch": 2.38843721770551, - "grad_norm": 0.302734375, - "learning_rate": 0.0002414012545067826, - "loss": 0.6082, + "epoch": 5.055449330783939, + "grad_norm": 0.3359375, + "learning_rate": 5.4615025910583756e-05, + "loss": 0.4069, "step": 2644 }, { - "epoch": 2.3920505871725384, - "grad_norm": 0.314453125, - "learning_rate": 0.0002412306809292698, - "loss": 0.5889, + "epoch": 5.0630975143403445, + "grad_norm": 0.35546875, + "learning_rate": 5.421417764439276e-05, + "loss": 0.375, "step": 2648 }, { - "epoch": 2.3956639566395665, - "grad_norm": 0.3125, - "learning_rate": 0.00024105991991336197, - "loss": 0.6266, + "epoch": 5.07074569789675, + "grad_norm": 0.333984375, + "learning_rate": 5.3814481090878374e-05, + "loss": 0.3576, "step": 2652 }, { - "epoch": 2.3992773261065943, - "grad_norm": 0.310546875, - "learning_rate": 0.00024088897180989683, - "loss": 0.6239, + "epoch": 5.078393881453155, + "grad_norm": 0.337890625, + "learning_rate": 5.3415941055924974e-05, + "loss": 0.4027, "step": 2656 }, { - "epoch": 2.4028906955736224, - "grad_norm": 0.283203125, - "learning_rate": 0.00024071783697009666, - "loss": 0.6489, + "epoch": 5.08604206500956, + "grad_norm": 0.328125, + "learning_rate": 5.301856233151123e-05, + "loss": 0.3981, "step": 2660 }, { - "epoch": 2.40650406504065, - "grad_norm": 0.306640625, - "learning_rate": 0.00024054651574556717, - "loss": 0.654, + "epoch": 5.093690248565966, + "grad_norm": 0.318359375, + "learning_rate": 5.262234969565226e-05, + "loss": 0.3928, "step": 2664 }, { - "epoch": 2.4101174345076783, - "grad_norm": 0.310546875, - "learning_rate": 0.00024037500848829725, - "loss": 0.5871, + "epoch": 5.101338432122371, + "grad_norm": 0.33203125, + "learning_rate": 5.222730791234246e-05, + "loss": 0.3813, "step": 2668 }, { - "epoch": 2.4137308039747065, - "grad_norm": 0.314453125, - "learning_rate": 0.0002402033155506578, - "loss": 0.6203, + "epoch": 5.108986615678776, + "grad_norm": 0.3671875, + "learning_rate": 5.183344173149798e-05, + "loss": 0.4151, "step": 2672 }, { - "epoch": 2.4173441734417342, - "grad_norm": 0.310546875, - "learning_rate": 0.00024003143728540136, - "loss": 0.6587, + "epoch": 5.1166347992351815, + "grad_norm": 0.35546875, + "learning_rate": 5.14407558888998e-05, + "loss": 0.4051, "step": 2676 }, { - "epoch": 2.4209575429087624, - "grad_norm": 0.291015625, - "learning_rate": 0.00023985937404566115, - "loss": 0.6234, + "epoch": 5.124282982791587, + "grad_norm": 0.330078125, + "learning_rate": 5.104925510613668e-05, + "loss": 0.3973, "step": 2680 }, { - "epoch": 2.4245709123757906, - "grad_norm": 0.310546875, - "learning_rate": 0.00023968712618495044, - "loss": 0.6166, + "epoch": 5.131931166347992, + "grad_norm": 0.3203125, + "learning_rate": 5.0658944090548436e-05, + "loss": 0.3958, "step": 2684 }, { - "epoch": 2.4281842818428183, - "grad_norm": 0.291015625, - "learning_rate": 0.0002395146940571618, - "loss": 0.6067, + "epoch": 5.139579349904397, + "grad_norm": 0.34765625, + "learning_rate": 5.0269827535169306e-05, + "loss": 0.3731, "step": 2688 }, { - "epoch": 2.4317976513098465, - "grad_norm": 0.298828125, - "learning_rate": 0.00023934207801656651, - "loss": 0.6243, + "epoch": 5.147227533460803, + "grad_norm": 0.33984375, + "learning_rate": 4.988191011867153e-05, + "loss": 0.4028, "step": 2692 }, { - "epoch": 2.4354110207768747, - "grad_norm": 0.296875, - "learning_rate": 0.00023916927841781356, - "loss": 0.6855, + "epoch": 5.154875717017209, + "grad_norm": 0.36328125, + "learning_rate": 4.9495196505309196e-05, + "loss": 0.4201, "step": 2696 }, { - "epoch": 2.4390243902439024, - "grad_norm": 0.283203125, - "learning_rate": 0.00023899629561592903, - "loss": 0.6323, + "epoch": 5.162523900573614, + "grad_norm": 0.345703125, + "learning_rate": 4.9109691344861886e-05, + "loss": 0.3815, "step": 2700 }, { - "epoch": 2.4426377597109306, - "grad_norm": 0.30078125, - "learning_rate": 0.00023882312996631566, - "loss": 0.6511, + "epoch": 5.170172084130019, + "grad_norm": 0.326171875, + "learning_rate": 4.8725399272579075e-05, + "loss": 0.4043, "step": 2704 }, { - "epoch": 2.4462511291779583, - "grad_norm": 0.296875, - "learning_rate": 0.00023864978182475154, - "loss": 0.6638, + "epoch": 5.177820267686425, + "grad_norm": 0.330078125, + "learning_rate": 4.8342324909124256e-05, + "loss": 0.3828, "step": 2708 }, { - "epoch": 2.4498644986449865, - "grad_norm": 0.291015625, - "learning_rate": 0.00023847625154738999, - "loss": 0.6051, + "epoch": 5.18546845124283, + "grad_norm": 0.34765625, + "learning_rate": 4.7960472860519365e-05, + "loss": 0.4199, "step": 2712 }, { - "epoch": 2.4534778681120146, - "grad_norm": 0.28515625, - "learning_rate": 0.00023830253949075827, - "loss": 0.6598, + "epoch": 5.193116634799235, + "grad_norm": 0.330078125, + "learning_rate": 4.757984771808947e-05, + "loss": 0.4071, "step": 2716 }, { - "epoch": 2.4570912375790424, - "grad_norm": 0.29296875, - "learning_rate": 0.00023812864601175735, - "loss": 0.618, + "epoch": 5.2007648183556405, + "grad_norm": 0.34765625, + "learning_rate": 4.72004540584075e-05, + "loss": 0.4288, "step": 2720 }, { - "epoch": 2.4607046070460705, - "grad_norm": 0.326171875, - "learning_rate": 0.00023795457146766078, - "loss": 0.6816, + "epoch": 5.208413001912046, + "grad_norm": 0.353515625, + "learning_rate": 4.682229644323922e-05, + "loss": 0.3935, "step": 2724 }, { - "epoch": 2.4643179765130983, - "grad_norm": 0.33203125, - "learning_rate": 0.0002377803162161142, - "loss": 0.6624, + "epoch": 5.216061185468451, + "grad_norm": 0.37109375, + "learning_rate": 4.6445379419488436e-05, + "loss": 0.3801, "step": 2728 }, { - "epoch": 2.4679313459801264, - "grad_norm": 0.291015625, - "learning_rate": 0.0002376058806151345, - "loss": 0.6298, + "epoch": 5.223709369024856, + "grad_norm": 0.33203125, + "learning_rate": 4.606970751914229e-05, + "loss": 0.4181, "step": 2732 }, { - "epoch": 2.4715447154471546, - "grad_norm": 0.291015625, - "learning_rate": 0.00023743126502310914, - "loss": 0.5964, + "epoch": 5.231357552581262, + "grad_norm": 0.341796875, + "learning_rate": 4.569528525921672e-05, + "loss": 0.3816, "step": 2736 }, { - "epoch": 2.4751580849141823, - "grad_norm": 0.3125, - "learning_rate": 0.00023725646979879528, - "loss": 0.6039, + "epoch": 5.239005736137667, + "grad_norm": 0.349609375, + "learning_rate": 4.532211714170229e-05, + "loss": 0.4331, "step": 2740 }, { - "epoch": 2.4787714543812105, - "grad_norm": 0.31640625, - "learning_rate": 0.00023708149530131923, - "loss": 0.6232, + "epoch": 5.246653919694072, + "grad_norm": 0.365234375, + "learning_rate": 4.495020765350988e-05, + "loss": 0.4331, "step": 2744 }, { - "epoch": 2.4823848238482387, - "grad_norm": 0.318359375, - "learning_rate": 0.00023690634189017567, - "loss": 0.6234, + "epoch": 5.254302103250478, + "grad_norm": 0.359375, + "learning_rate": 4.4579561266416855e-05, + "loss": 0.3889, "step": 2748 }, { - "epoch": 2.4859981933152664, - "grad_norm": 0.3203125, - "learning_rate": 0.00023673100992522674, - "loss": 0.5961, + "epoch": 5.261950286806884, + "grad_norm": 0.328125, + "learning_rate": 4.421018243701327e-05, + "loss": 0.3804, "step": 2752 }, { - "epoch": 2.4896115627822946, - "grad_norm": 0.322265625, - "learning_rate": 0.00023655549976670152, - "loss": 0.6586, + "epoch": 5.269598470363289, + "grad_norm": 0.55859375, + "learning_rate": 4.384207560664825e-05, + "loss": 0.4386, "step": 2756 }, { - "epoch": 2.4932249322493227, - "grad_norm": 0.30859375, - "learning_rate": 0.0002363798117751952, - "loss": 0.6162, + "epoch": 5.277246653919694, + "grad_norm": 0.34375, + "learning_rate": 4.347524520137667e-05, + "loss": 0.3547, "step": 2760 }, { - "epoch": 2.4968383017163505, - "grad_norm": 0.3046875, - "learning_rate": 0.00023620394631166828, - "loss": 0.6455, + "epoch": 5.2848948374761, + "grad_norm": 0.337890625, + "learning_rate": 4.310969563190578e-05, + "loss": 0.3719, "step": 2764 }, { - "epoch": 2.5004516711833786, - "grad_norm": 0.29296875, - "learning_rate": 0.00023602790373744594, - "loss": 0.6159, + "epoch": 5.292543021032505, + "grad_norm": 0.328125, + "learning_rate": 4.274543129354245e-05, + "loss": 0.3611, "step": 2768 }, { - "epoch": 2.5040650406504064, - "grad_norm": 0.29296875, - "learning_rate": 0.0002358516844142172, - "loss": 0.6267, + "epoch": 5.30019120458891, + "grad_norm": 0.3125, + "learning_rate": 4.2382456566139985e-05, + "loss": 0.3173, "step": 2772 }, { - "epoch": 2.5076784101174345, - "grad_norm": 0.326171875, - "learning_rate": 0.00023567528870403425, - "loss": 0.6341, + "epoch": 5.307839388145315, + "grad_norm": 0.345703125, + "learning_rate": 4.202077581404574e-05, + "loss": 0.3848, "step": 2776 }, { - "epoch": 2.5112917795844627, - "grad_norm": 0.310546875, - "learning_rate": 0.00023549871696931167, - "loss": 0.6187, + "epoch": 5.315487571701721, + "grad_norm": 0.330078125, + "learning_rate": 4.166039338604838e-05, + "loss": 0.3921, "step": 2780 }, { - "epoch": 2.5149051490514904, - "grad_norm": 0.322265625, - "learning_rate": 0.00023532196957282568, - "loss": 0.6156, + "epoch": 5.323135755258126, + "grad_norm": 0.318359375, + "learning_rate": 4.130131361532586e-05, + "loss": 0.4049, "step": 2784 }, { - "epoch": 2.5185185185185186, - "grad_norm": 0.447265625, - "learning_rate": 0.00023514504687771346, - "loss": 0.5952, + "epoch": 5.330783938814531, + "grad_norm": 0.35546875, + "learning_rate": 4.094354081939317e-05, + "loss": 0.3866, "step": 2788 }, { - "epoch": 2.5221318879855463, - "grad_norm": 0.3046875, - "learning_rate": 0.00023496794924747228, - "loss": 0.6097, + "epoch": 5.338432122370937, + "grad_norm": 0.365234375, + "learning_rate": 4.058707930005048e-05, + "loss": 0.3727, "step": 2792 }, { - "epoch": 2.5257452574525745, - "grad_norm": 0.3359375, - "learning_rate": 0.0002347906770459588, - "loss": 0.6406, + "epoch": 5.346080305927342, + "grad_norm": 0.341796875, + "learning_rate": 4.023193334333132e-05, + "loss": 0.3957, "step": 2796 }, { - "epoch": 2.5293586269196027, - "grad_norm": 0.310546875, - "learning_rate": 0.0002346132306373885, - "loss": 0.6756, + "epoch": 5.353728489483748, + "grad_norm": 0.341796875, + "learning_rate": 3.9878107219451206e-05, + "loss": 0.344, "step": 2800 }, { - "epoch": 2.5329719963866304, - "grad_norm": 0.337890625, - "learning_rate": 0.0002344356103863346, - "loss": 0.6112, + "epoch": 5.361376673040153, + "grad_norm": 0.375, + "learning_rate": 3.9525605182756134e-05, + "loss": 0.4048, "step": 2804 }, { - "epoch": 2.5365853658536586, - "grad_norm": 0.328125, - "learning_rate": 0.00023425781665772758, - "loss": 0.6363, + "epoch": 5.369024856596559, + "grad_norm": 0.349609375, + "learning_rate": 3.917443147167152e-05, + "loss": 0.4343, "step": 2808 }, { - "epoch": 2.5401987353206863, - "grad_norm": 0.31640625, - "learning_rate": 0.00023407984981685436, - "loss": 0.6156, + "epoch": 5.376673040152964, + "grad_norm": 0.345703125, + "learning_rate": 3.882459030865124e-05, + "loss": 0.3908, "step": 2812 }, { - "epoch": 2.5438121047877145, - "grad_norm": 0.328125, - "learning_rate": 0.00023390171022935746, - "loss": 0.6203, + "epoch": 5.384321223709369, + "grad_norm": 0.34765625, + "learning_rate": 3.8476085900126776e-05, + "loss": 0.3491, "step": 2816 }, { - "epoch": 2.5474254742547426, - "grad_norm": 0.326171875, - "learning_rate": 0.00023372339826123432, - "loss": 0.5995, + "epoch": 5.3919694072657744, + "grad_norm": 0.3515625, + "learning_rate": 3.8128922436456766e-05, + "loss": 0.4092, "step": 2820 }, { - "epoch": 2.551038843721771, - "grad_norm": 0.30859375, - "learning_rate": 0.00023354491427883664, - "loss": 0.6362, + "epoch": 5.39961759082218, + "grad_norm": 0.375, + "learning_rate": 3.7783104091876524e-05, + "loss": 0.3766, "step": 2824 }, { - "epoch": 2.5546522131887985, - "grad_norm": 0.310546875, - "learning_rate": 0.00023336625864886943, - "loss": 0.6234, + "epoch": 5.407265774378585, + "grad_norm": 0.34375, + "learning_rate": 3.743863502444783e-05, + "loss": 0.3835, "step": 2828 }, { - "epoch": 2.5582655826558267, - "grad_norm": 0.296875, - "learning_rate": 0.0002331874317383904, - "loss": 0.6176, + "epoch": 5.41491395793499, + "grad_norm": 0.33203125, + "learning_rate": 3.709551937600909e-05, + "loss": 0.3733, "step": 2832 }, { - "epoch": 2.5618789521228544, - "grad_norm": 0.318359375, - "learning_rate": 0.00023300843391480916, - "loss": 0.5973, + "epoch": 5.422562141491396, + "grad_norm": 0.34375, + "learning_rate": 3.675376127212532e-05, + "loss": 0.4127, "step": 2836 }, { - "epoch": 2.5654923215898826, - "grad_norm": 0.296875, - "learning_rate": 0.0002328292655458865, - "loss": 0.6319, + "epoch": 5.430210325047801, + "grad_norm": 0.361328125, + "learning_rate": 3.64133648220387e-05, + "loss": 0.4091, "step": 2840 }, { - "epoch": 2.569105691056911, - "grad_norm": 0.283203125, - "learning_rate": 0.00023264992699973357, - "loss": 0.6267, + "epoch": 5.437858508604206, + "grad_norm": 0.365234375, + "learning_rate": 3.607433411861912e-05, + "loss": 0.4612, "step": 2844 }, { - "epoch": 2.5727190605239385, - "grad_norm": 0.30859375, - "learning_rate": 0.0002324704186448111, - "loss": 0.6169, + "epoch": 5.4455066921606115, + "grad_norm": 0.330078125, + "learning_rate": 3.5736673238314914e-05, + "loss": 0.3668, "step": 2848 }, { - "epoch": 2.5763324299909667, - "grad_norm": 0.3125, - "learning_rate": 0.00023229074084992884, - "loss": 0.6681, + "epoch": 5.453154875717018, + "grad_norm": 0.3359375, + "learning_rate": 3.5400386241103946e-05, + "loss": 0.3645, "step": 2852 }, { - "epoch": 2.5799457994579944, - "grad_norm": 0.322265625, - "learning_rate": 0.00023211089398424457, - "loss": 0.6289, + "epoch": 5.460803059273423, + "grad_norm": 0.33984375, + "learning_rate": 3.506547717044472e-05, + "loss": 0.348, "step": 2856 }, { - "epoch": 2.5835591689250226, - "grad_norm": 0.322265625, - "learning_rate": 0.00023193087841726347, - "loss": 0.66, + "epoch": 5.468451242829828, + "grad_norm": 0.33984375, + "learning_rate": 3.473195005322776e-05, + "loss": 0.3812, "step": 2860 }, { - "epoch": 2.5871725383920507, - "grad_norm": 0.296875, - "learning_rate": 0.00023175069451883727, - "loss": 0.6165, + "epoch": 5.4760994263862335, + "grad_norm": 0.333984375, + "learning_rate": 3.439980889972723e-05, + "loss": 0.3705, "step": 2864 }, { - "epoch": 2.5907859078590785, - "grad_norm": 0.287109375, - "learning_rate": 0.00023157034265916364, - "loss": 0.6014, + "epoch": 5.483747609942639, + "grad_norm": 0.35546875, + "learning_rate": 3.406905770355274e-05, + "loss": 0.3687, "step": 2868 }, { - "epoch": 2.5943992773261066, - "grad_norm": 0.333984375, - "learning_rate": 0.0002313898232087852, - "loss": 0.5906, + "epoch": 5.491395793499044, + "grad_norm": 0.32421875, + "learning_rate": 3.373970044160121e-05, + "loss": 0.4348, "step": 2872 }, { - "epoch": 2.5980126467931344, - "grad_norm": 0.306640625, - "learning_rate": 0.0002312091365385891, - "loss": 0.6539, + "epoch": 5.499043977055449, + "grad_norm": 0.36328125, + "learning_rate": 3.341174107400916e-05, + "loss": 0.4039, "step": 2876 }, { - "epoch": 2.6016260162601625, - "grad_norm": 0.330078125, - "learning_rate": 0.00023102828301980582, - "loss": 0.6245, + "epoch": 5.506692160611855, + "grad_norm": 0.3359375, + "learning_rate": 3.30851835441051e-05, + "loss": 0.4072, "step": 2880 }, { - "epoch": 2.6052393857271907, - "grad_norm": 0.33984375, - "learning_rate": 0.00023084726302400883, - "loss": 0.6089, + "epoch": 5.51434034416826, + "grad_norm": 0.353515625, + "learning_rate": 3.276003177836203e-05, + "loss": 0.3796, "step": 2884 }, { - "epoch": 2.6088527551942184, - "grad_norm": 0.32421875, - "learning_rate": 0.00023066607692311347, - "loss": 0.6076, + "epoch": 5.521988527724665, + "grad_norm": 0.345703125, + "learning_rate": 3.2436289686350285e-05, + "loss": 0.3647, "step": 2888 }, { - "epoch": 2.6124661246612466, - "grad_norm": 0.314453125, - "learning_rate": 0.00023048472508937652, - "loss": 0.6506, + "epoch": 5.5296367112810705, + "grad_norm": 0.341796875, + "learning_rate": 3.211396116069055e-05, + "loss": 0.3895, "step": 2892 }, { - "epoch": 2.6160794941282743, - "grad_norm": 0.302734375, - "learning_rate": 0.0002303032078953951, - "loss": 0.6154, + "epoch": 5.537284894837476, + "grad_norm": 0.34375, + "learning_rate": 3.179305007700697e-05, + "loss": 0.3689, "step": 2896 }, { - "epoch": 2.6196928635953025, - "grad_norm": 0.302734375, - "learning_rate": 0.00023012152571410627, - "loss": 0.6112, + "epoch": 5.544933078393882, + "grad_norm": 0.349609375, + "learning_rate": 3.147356029388067e-05, + "loss": 0.3782, "step": 2900 }, { - "epoch": 2.6233062330623307, - "grad_norm": 0.3203125, - "learning_rate": 0.00022993967891878585, - "loss": 0.6254, + "epoch": 5.552581261950287, + "grad_norm": 0.349609375, + "learning_rate": 3.115549565280325e-05, + "loss": 0.3985, "step": 2904 }, { - "epoch": 2.626919602529359, - "grad_norm": 0.3359375, - "learning_rate": 0.000229757667883048, - "loss": 0.6436, + "epoch": 5.5602294455066925, + "grad_norm": 0.341796875, + "learning_rate": 3.083885997813066e-05, + "loss": 0.4289, "step": 2908 }, { - "epoch": 2.6305329719963866, - "grad_norm": 0.30859375, - "learning_rate": 0.00022957549298084433, - "loss": 0.6586, + "epoch": 5.567877629063098, + "grad_norm": 0.33984375, + "learning_rate": 3.052365707703718e-05, + "loss": 0.436, "step": 2912 }, { - "epoch": 2.6341463414634148, - "grad_norm": 0.353515625, - "learning_rate": 0.000229393154586463, - "loss": 0.6565, + "epoch": 5.575525812619503, + "grad_norm": 0.341796875, + "learning_rate": 3.0209890739469693e-05, + "loss": 0.387, "step": 2916 }, { - "epoch": 2.6377597109304425, - "grad_norm": 0.314453125, - "learning_rate": 0.00022921065307452825, - "loss": 0.6168, + "epoch": 5.583173996175908, + "grad_norm": 0.326171875, + "learning_rate": 2.989756473810203e-05, + "loss": 0.4034, "step": 2920 }, { - "epoch": 2.6413730803974707, - "grad_norm": 0.306640625, - "learning_rate": 0.0002290279888199993, - "loss": 0.6327, + "epoch": 5.590822179732314, + "grad_norm": 0.3671875, + "learning_rate": 2.9586682828289738e-05, + "loss": 0.4206, "step": 2924 }, { - "epoch": 2.644986449864499, - "grad_norm": 0.3046875, - "learning_rate": 0.0002288451621981698, - "loss": 0.5869, + "epoch": 5.598470363288719, + "grad_norm": 0.36328125, + "learning_rate": 2.9277248748024763e-05, + "loss": 0.4529, "step": 2928 }, { - "epoch": 2.6485998193315266, - "grad_norm": 0.314453125, - "learning_rate": 0.00022866217358466704, - "loss": 0.6394, + "epoch": 5.606118546845124, + "grad_norm": 0.330078125, + "learning_rate": 2.8969266217890648e-05, + "loss": 0.3527, "step": 2932 }, { - "epoch": 2.6522131887985547, - "grad_norm": 0.3046875, - "learning_rate": 0.00022847902335545097, - "loss": 0.6287, + "epoch": 5.6137667304015295, + "grad_norm": 0.365234375, + "learning_rate": 2.866273894101776e-05, + "loss": 0.389, "step": 2936 }, { - "epoch": 2.6558265582655824, - "grad_norm": 0.3203125, - "learning_rate": 0.00022829571188681382, - "loss": 0.6535, + "epoch": 5.621414913957935, + "grad_norm": 0.361328125, + "learning_rate": 2.835767060303865e-05, + "loss": 0.4017, "step": 2940 }, { - "epoch": 2.6594399277326106, - "grad_norm": 0.32421875, - "learning_rate": 0.00022811223955537886, - "loss": 0.676, + "epoch": 5.62906309751434, + "grad_norm": 0.353515625, + "learning_rate": 2.8054064872043917e-05, + "loss": 0.3973, "step": 2944 }, { - "epoch": 2.663053297199639, - "grad_norm": 0.314453125, - "learning_rate": 0.00022792860673810005, - "loss": 0.6782, + "epoch": 5.636711281070745, + "grad_norm": 0.375, + "learning_rate": 2.7751925398537993e-05, + "loss": 0.4281, "step": 2948 }, { - "epoch": 2.6666666666666665, - "grad_norm": 0.369140625, - "learning_rate": 0.00022774481381226095, - "loss": 0.6651, + "epoch": 5.644359464627151, + "grad_norm": 0.330078125, + "learning_rate": 2.745125581539523e-05, + "loss": 0.4071, "step": 2952 }, { - "epoch": 2.6702800361336947, - "grad_norm": 0.322265625, - "learning_rate": 0.00022756086115547416, - "loss": 0.624, + "epoch": 5.652007648183556, + "grad_norm": 0.373046875, + "learning_rate": 2.7152059737816395e-05, + "loss": 0.3866, "step": 2956 }, { - "epoch": 2.6738934056007224, - "grad_norm": 0.296875, - "learning_rate": 0.00022737674914568039, - "loss": 0.6237, + "epoch": 5.659655831739962, + "grad_norm": 0.32421875, + "learning_rate": 2.6854340763284954e-05, + "loss": 0.4029, "step": 2960 }, { - "epoch": 2.6775067750677506, - "grad_norm": 0.322265625, - "learning_rate": 0.00022719247816114783, - "loss": 0.6197, + "epoch": 5.667304015296367, + "grad_norm": 0.369140625, + "learning_rate": 2.6558102471523975e-05, + "loss": 0.4207, "step": 2964 }, { - "epoch": 2.6811201445347788, - "grad_norm": 0.30859375, - "learning_rate": 0.0002270080485804711, - "loss": 0.6159, + "epoch": 5.674952198852773, + "grad_norm": 0.365234375, + "learning_rate": 2.6263348424453012e-05, + "loss": 0.3769, "step": 2968 }, { - "epoch": 2.684733514001807, - "grad_norm": 0.333984375, - "learning_rate": 0.000226823460782571, - "loss": 0.6186, + "epoch": 5.682600382409178, + "grad_norm": 0.337890625, + "learning_rate": 2.597008216614534e-05, + "loss": 0.3527, "step": 2972 }, { - "epoch": 2.6883468834688347, - "grad_norm": 0.310546875, - "learning_rate": 0.00022663871514669304, - "loss": 0.6349, + "epoch": 5.690248565965583, + "grad_norm": 0.345703125, + "learning_rate": 2.5678307222785315e-05, + "loss": 0.4091, "step": 2976 }, { - "epoch": 2.691960252935863, - "grad_norm": 0.298828125, - "learning_rate": 0.00022645381205240722, - "loss": 0.6122, + "epoch": 5.6978967495219885, + "grad_norm": 0.328125, + "learning_rate": 2.5388027102625945e-05, + "loss": 0.4175, "step": 2980 }, { - "epoch": 2.6955736224028906, - "grad_norm": 0.337890625, - "learning_rate": 0.00022626875187960703, - "loss": 0.6423, + "epoch": 5.705544933078394, + "grad_norm": 0.349609375, + "learning_rate": 2.5099245295946764e-05, + "loss": 0.3557, "step": 2984 }, { - "epoch": 2.6991869918699187, - "grad_norm": 0.32421875, - "learning_rate": 0.00022608353500850863, - "loss": 0.6259, + "epoch": 5.713193116634799, + "grad_norm": 0.328125, + "learning_rate": 2.4811965275011825e-05, + "loss": 0.4021, "step": 2988 }, { - "epoch": 2.702800361336947, - "grad_norm": 0.3125, - "learning_rate": 0.00022589816181965022, - "loss": 0.6267, + "epoch": 5.720841300191204, + "grad_norm": 0.33203125, + "learning_rate": 2.4526190494027953e-05, + "loss": 0.3868, "step": 2992 }, { - "epoch": 2.7064137308039746, - "grad_norm": 0.29296875, - "learning_rate": 0.000225712632693891, - "loss": 0.6657, + "epoch": 5.72848948374761, + "grad_norm": 0.369140625, + "learning_rate": 2.4241924389103227e-05, + "loss": 0.4399, "step": 2996 }, { - "epoch": 2.710027100271003, - "grad_norm": 0.34375, - "learning_rate": 0.00022552694801241066, - "loss": 0.6439, + "epoch": 5.736137667304015, + "grad_norm": 0.3359375, + "learning_rate": 2.395917037820566e-05, + "loss": 0.3907, "step": 3000 }, { - "epoch": 2.7136404697380305, - "grad_norm": 0.3125, - "learning_rate": 0.00022534110815670855, - "loss": 0.6932, + "epoch": 5.743785850860421, + "grad_norm": 0.349609375, + "learning_rate": 2.3677931861122084e-05, + "loss": 0.4195, "step": 3004 }, { - "epoch": 2.7172538392050587, - "grad_norm": 0.30078125, - "learning_rate": 0.0002251551135086027, - "loss": 0.6355, + "epoch": 5.751434034416826, + "grad_norm": 0.3359375, + "learning_rate": 2.339821221941731e-05, + "loss": 0.3867, "step": 3008 }, { - "epoch": 2.720867208672087, + "epoch": 5.759082217973232, "grad_norm": 0.3359375, - "learning_rate": 0.00022496896445022924, - "loss": 0.6464, + "learning_rate": 2.312001481639348e-05, + "loss": 0.3583, "step": 3012 }, { - "epoch": 2.7244805781391146, - "grad_norm": 0.30078125, - "learning_rate": 0.00022478266136404153, - "loss": 0.6115, + "epoch": 5.766730401529637, + "grad_norm": 0.3359375, + "learning_rate": 2.2843342997049445e-05, + "loss": 0.3527, "step": 3016 }, { - "epoch": 2.7280939476061428, - "grad_norm": 0.32421875, - "learning_rate": 0.0002245962046328094, - "loss": 0.633, + "epoch": 5.774378585086042, + "grad_norm": 0.30859375, + "learning_rate": 2.2568200088040867e-05, + "loss": 0.3393, "step": 3020 }, { - "epoch": 2.7317073170731705, - "grad_norm": 0.32421875, - "learning_rate": 0.00022440959463961832, - "loss": 0.6263, + "epoch": 5.782026768642448, + "grad_norm": 0.359375, + "learning_rate": 2.2294589397639978e-05, + "loss": 0.4225, "step": 3024 }, { - "epoch": 2.7353206865401987, - "grad_norm": 0.3203125, - "learning_rate": 0.00022422283176786871, - "loss": 0.6686, + "epoch": 5.789674952198853, + "grad_norm": 0.357421875, + "learning_rate": 2.2022514215695842e-05, + "loss": 0.4191, "step": 3028 }, { - "epoch": 2.738934056007227, - "grad_norm": 0.298828125, - "learning_rate": 0.00022403591640127505, - "loss": 0.6192, + "epoch": 5.797323135755258, + "grad_norm": 0.341796875, + "learning_rate": 2.175197781359485e-05, + "loss": 0.3792, "step": 3032 }, { - "epoch": 2.742547425474255, - "grad_norm": 0.298828125, - "learning_rate": 0.00022384884892386509, - "loss": 0.6031, + "epoch": 5.804971319311663, + "grad_norm": 0.337890625, + "learning_rate": 2.1482983444221402e-05, + "loss": 0.3942, "step": 3036 }, { - "epoch": 2.7461607949412827, - "grad_norm": 0.330078125, - "learning_rate": 0.00022366162971997915, - "loss": 0.645, + "epoch": 5.812619502868069, + "grad_norm": 0.326171875, + "learning_rate": 2.1215534341918707e-05, + "loss": 0.3753, "step": 3040 }, { - "epoch": 2.749774164408311, - "grad_norm": 0.306640625, - "learning_rate": 0.00022347425917426927, - "loss": 0.6432, + "epoch": 5.820267686424474, + "grad_norm": 0.33984375, + "learning_rate": 2.0949633722449915e-05, + "loss": 0.4234, "step": 3044 }, { - "epoch": 2.7533875338753386, - "grad_norm": 0.306640625, - "learning_rate": 0.00022328673767169841, - "loss": 0.626, + "epoch": 5.827915869980879, + "grad_norm": 0.357421875, + "learning_rate": 2.0685284782959566e-05, + "loss": 0.3925, "step": 3048 }, { - "epoch": 2.757000903342367, - "grad_norm": 0.318359375, - "learning_rate": 0.00022309906559753965, - "loss": 0.6708, + "epoch": 5.835564053537285, + "grad_norm": 0.349609375, + "learning_rate": 2.0422490701934996e-05, + "loss": 0.412, "step": 3052 }, { - "epoch": 2.760614272809395, - "grad_norm": 0.287109375, - "learning_rate": 0.0002229112433373756, - "loss": 0.6306, + "epoch": 5.84321223709369, + "grad_norm": 0.328125, + "learning_rate": 2.0161254639168183e-05, + "loss": 0.3981, "step": 3056 }, { - "epoch": 2.7642276422764227, - "grad_norm": 0.310546875, - "learning_rate": 0.00022272327127709717, - "loss": 0.6488, + "epoch": 5.850860420650095, + "grad_norm": 0.3515625, + "learning_rate": 1.9901579735717743e-05, + "loss": 0.4204, "step": 3060 }, { - "epoch": 2.767841011743451, - "grad_norm": 0.29296875, - "learning_rate": 0.00022253514980290326, - "loss": 0.6636, + "epoch": 5.858508604206501, + "grad_norm": 0.333984375, + "learning_rate": 1.964346911387127e-05, + "loss": 0.3878, "step": 3064 }, { - "epoch": 2.7714543812104786, - "grad_norm": 0.3046875, - "learning_rate": 0.0002223468793012997, - "loss": 0.6511, + "epoch": 5.866156787762907, + "grad_norm": 0.373046875, + "learning_rate": 1.9386925877107585e-05, + "loss": 0.4047, "step": 3068 }, { - "epoch": 2.7750677506775068, - "grad_norm": 0.30859375, - "learning_rate": 0.00022215846015909835, - "loss": 0.6433, + "epoch": 5.873804971319312, + "grad_norm": 0.3671875, + "learning_rate": 1.913195311005959e-05, + "loss": 0.3899, "step": 3072 }, { - "epoch": 2.778681120144535, - "grad_norm": 0.322265625, - "learning_rate": 0.0002219698927634167, - "loss": 0.6107, + "epoch": 5.881453154875717, + "grad_norm": 0.353515625, + "learning_rate": 1.8878553878477105e-05, + "loss": 0.4179, "step": 3076 }, { - "epoch": 2.7822944896115627, - "grad_norm": 0.3125, - "learning_rate": 0.00022178117750167663, - "loss": 0.6756, + "epoch": 5.8891013384321225, + "grad_norm": 0.341796875, + "learning_rate": 1.8626731229190016e-05, + "loss": 0.403, "step": 3080 }, { - "epoch": 2.785907859078591, - "grad_norm": 0.3125, - "learning_rate": 0.00022159231476160402, - "loss": 0.6706, + "epoch": 5.896749521988528, + "grad_norm": 0.31640625, + "learning_rate": 1.8376488190071666e-05, + "loss": 0.3925, "step": 3084 }, { - "epoch": 2.7895212285456186, - "grad_norm": 0.31640625, - "learning_rate": 0.00022140330493122752, - "loss": 0.6185, + "epoch": 5.904397705544933, + "grad_norm": 0.3359375, + "learning_rate": 1.8127827770002423e-05, + "loss": 0.3647, "step": 3088 }, { - "epoch": 2.7931345980126467, - "grad_norm": 0.333984375, - "learning_rate": 0.00022121414839887813, - "loss": 0.6646, + "epoch": 5.912045889101338, + "grad_norm": 0.353515625, + "learning_rate": 1.7880752958833543e-05, + "loss": 0.4301, "step": 3092 }, { - "epoch": 2.796747967479675, - "grad_norm": 0.287109375, - "learning_rate": 0.00022102484555318834, - "loss": 0.6591, + "epoch": 5.919694072657744, + "grad_norm": 0.35546875, + "learning_rate": 1.7635266727351092e-05, + "loss": 0.3988, "step": 3096 }, { - "epoch": 2.8003613369467026, - "grad_norm": 0.306640625, - "learning_rate": 0.00022083539678309098, - "loss": 0.6143, + "epoch": 5.927342256214149, + "grad_norm": 0.34375, + "learning_rate": 1.73913720272404e-05, + "loss": 0.4028, "step": 3100 }, { - "epoch": 2.803974706413731, - "grad_norm": 0.32421875, - "learning_rate": 0.00022064580247781903, - "loss": 0.6596, + "epoch": 5.934990439770554, + "grad_norm": 0.328125, + "learning_rate": 1.714907179105049e-05, + "loss": 0.3756, "step": 3104 }, { - "epoch": 2.8075880758807585, - "grad_norm": 0.318359375, - "learning_rate": 0.00022045606302690412, - "loss": 0.6317, + "epoch": 5.9426386233269595, + "grad_norm": 0.341796875, + "learning_rate": 1.6908368932158777e-05, + "loss": 0.4023, "step": 3108 }, { - "epoch": 2.8112014453477867, - "grad_norm": 0.322265625, - "learning_rate": 0.00022026617882017638, - "loss": 0.6856, + "epoch": 5.950286806883366, + "grad_norm": 0.345703125, + "learning_rate": 1.6669266344736104e-05, + "loss": 0.3784, "step": 3112 }, { - "epoch": 2.814814814814815, - "grad_norm": 0.341796875, - "learning_rate": 0.0002200761502477632, - "loss": 0.6679, + "epoch": 5.957934990439771, + "grad_norm": 0.34375, + "learning_rate": 1.6431766903711914e-05, + "loss": 0.3622, "step": 3116 }, { - "epoch": 2.818428184281843, - "grad_norm": 0.31640625, - "learning_rate": 0.0002198859777000886, - "loss": 0.6058, + "epoch": 5.965583173996176, + "grad_norm": 0.359375, + "learning_rate": 1.6195873464739702e-05, + "loss": 0.416, "step": 3120 }, { - "epoch": 2.822041553748871, - "grad_norm": 0.3125, - "learning_rate": 0.00021969566156787248, - "loss": 0.6112, + "epoch": 5.9732313575525815, + "grad_norm": 0.3359375, + "learning_rate": 1.5961588864162627e-05, + "loss": 0.4191, "step": 3124 }, { - "epoch": 2.825654923215899, - "grad_norm": 0.306640625, - "learning_rate": 0.00021950520224212967, - "loss": 0.6011, + "epoch": 5.980879541108987, + "grad_norm": 0.328125, + "learning_rate": 1.5728915918979477e-05, + "loss": 0.3683, "step": 3128 }, { - "epoch": 2.8292682926829267, - "grad_norm": 0.3046875, - "learning_rate": 0.00021931460011416913, - "loss": 0.6622, + "epoch": 5.988527724665392, + "grad_norm": 0.337890625, + "learning_rate": 1.5497857426810756e-05, + "loss": 0.365, "step": 3132 }, { - "epoch": 2.832881662149955, - "grad_norm": 0.306640625, - "learning_rate": 0.0002191238555755934, - "loss": 0.6222, + "epoch": 5.996175908221797, + "grad_norm": 0.3671875, + "learning_rate": 1.5268416165865055e-05, + "loss": 0.401, "step": 3136 }, { - "epoch": 2.836495031616983, - "grad_norm": 0.2890625, - "learning_rate": 0.00021893296901829743, - "loss": 0.608, + "epoch": 6.003824091778203, + "grad_norm": 0.33203125, + "learning_rate": 1.5040594894905628e-05, + "loss": 0.3805, "step": 3140 }, { - "epoch": 2.8401084010840107, - "grad_norm": 0.32421875, - "learning_rate": 0.0002187419408344681, - "loss": 0.6183, + "epoch": 6.011472275334608, + "grad_norm": 0.34765625, + "learning_rate": 1.481439635321729e-05, + "loss": 0.3635, "step": 3144 }, { - "epoch": 2.843721770551039, - "grad_norm": 0.310546875, - "learning_rate": 0.00021855077141658317, - "loss": 0.6442, + "epoch": 6.019120458891013, + "grad_norm": 0.33984375, + "learning_rate": 1.458982326057338e-05, + "loss": 0.37, "step": 3148 }, { - "epoch": 2.8473351400180666, - "grad_norm": 0.328125, - "learning_rate": 0.00021835946115741057, - "loss": 0.6435, + "epoch": 6.0267686424474185, + "grad_norm": 0.34765625, + "learning_rate": 1.436687831720314e-05, + "loss": 0.3981, "step": 3152 }, { - "epoch": 2.850948509485095, - "grad_norm": 0.46484375, - "learning_rate": 0.00021816801045000767, - "loss": 0.643, + "epoch": 6.034416826003824, + "grad_norm": 0.345703125, + "learning_rate": 1.4145564203759219e-05, + "loss": 0.4484, "step": 3156 }, { - "epoch": 2.854561878952123, - "grad_norm": 0.330078125, - "learning_rate": 0.00021797641968772038, - "loss": 0.6622, + "epoch": 6.042065009560229, + "grad_norm": 0.3671875, + "learning_rate": 1.3925883581285401e-05, + "loss": 0.3988, "step": 3160 }, { - "epoch": 2.8581752484191507, - "grad_norm": 0.30078125, - "learning_rate": 0.00021778468926418227, - "loss": 0.6305, + "epoch": 6.049713193116634, + "grad_norm": 0.357421875, + "learning_rate": 1.3707839091184702e-05, + "loss": 0.3422, "step": 3164 }, { - "epoch": 2.861788617886179, - "grad_norm": 0.326171875, - "learning_rate": 0.000217592819573314, - "loss": 0.6354, + "epoch": 6.0573613766730405, + "grad_norm": 0.35546875, + "learning_rate": 1.349143335518752e-05, + "loss": 0.3994, "step": 3168 }, { - "epoch": 2.8654019873532066, - "grad_norm": 0.31640625, - "learning_rate": 0.00021740081100932225, - "loss": 0.6947, + "epoch": 6.065009560229446, + "grad_norm": 0.341796875, + "learning_rate": 1.3276668975320165e-05, + "loss": 0.4038, "step": 3172 }, { - "epoch": 2.869015356820235, - "grad_norm": 0.29296875, - "learning_rate": 0.00021720866396669904, - "loss": 0.6697, + "epoch": 6.072657743785851, + "grad_norm": 0.328125, + "learning_rate": 1.3063548533873536e-05, + "loss": 0.4065, "step": 3176 }, { - "epoch": 2.872628726287263, - "grad_norm": 0.291015625, - "learning_rate": 0.00021701637884022094, - "loss": 0.5763, + "epoch": 6.080305927342256, + "grad_norm": 0.33984375, + "learning_rate": 1.2852074593372142e-05, + "loss": 0.3902, "step": 3180 }, { - "epoch": 2.876242095754291, - "grad_norm": 0.3203125, - "learning_rate": 0.00021682395602494817, - "loss": 0.6741, + "epoch": 6.087954110898662, + "grad_norm": 0.328125, + "learning_rate": 1.2642249696543178e-05, + "loss": 0.388, "step": 3184 }, { - "epoch": 2.879855465221319, - "grad_norm": 0.31640625, - "learning_rate": 0.0002166313959162239, - "loss": 0.6487, + "epoch": 6.095602294455067, + "grad_norm": 0.34375, + "learning_rate": 1.243407636628605e-05, + "loss": 0.3979, "step": 3188 }, { - "epoch": 2.883468834688347, - "grad_norm": 0.302734375, - "learning_rate": 0.00021643869890967336, - "loss": 0.6081, + "epoch": 6.103250478011472, + "grad_norm": 0.3359375, + "learning_rate": 1.2227557105642e-05, + "loss": 0.404, "step": 3192 }, { - "epoch": 2.8870822041553748, - "grad_norm": 0.330078125, - "learning_rate": 0.00021624586540120296, - "loss": 0.6229, + "epoch": 6.1108986615678775, + "grad_norm": 0.34765625, + "learning_rate": 1.2022694397763993e-05, + "loss": 0.4059, "step": 3196 }, { - "epoch": 2.890695573622403, - "grad_norm": 0.296875, - "learning_rate": 0.0002160528957869997, - "loss": 0.6686, + "epoch": 6.118546845124283, + "grad_norm": 0.359375, + "learning_rate": 1.1819490705886914e-05, + "loss": 0.3766, "step": 3200 }, { - "epoch": 2.894308943089431, - "grad_norm": 0.2890625, - "learning_rate": 0.00021585979046353008, - "loss": 0.638, + "epoch": 6.126195028680688, + "grad_norm": 0.337890625, + "learning_rate": 1.16179484732979e-05, + "loss": 0.3629, "step": 3204 }, { - "epoch": 2.897922312556459, - "grad_norm": 0.29296875, - "learning_rate": 0.00021566654982753955, - "loss": 0.5879, + "epoch": 6.133843212237093, + "grad_norm": 0.375, + "learning_rate": 1.1418070123306989e-05, + "loss": 0.3733, "step": 3208 }, { - "epoch": 2.901535682023487, - "grad_norm": 0.296875, - "learning_rate": 0.00021547317427605148, - "loss": 0.5981, + "epoch": 6.141491395793499, + "grad_norm": 0.337890625, + "learning_rate": 1.1219858059217951e-05, + "loss": 0.4169, "step": 3212 }, { - "epoch": 2.9051490514905147, - "grad_norm": 0.3203125, - "learning_rate": 0.00021527966420636644, - "loss": 0.6293, + "epoch": 6.149139579349905, + "grad_norm": 0.33984375, + "learning_rate": 1.1023314664299455e-05, + "loss": 0.4131, "step": 3216 }, { - "epoch": 2.908762420957543, - "grad_norm": 0.330078125, - "learning_rate": 0.00021508602001606145, - "loss": 0.6284, + "epoch": 6.15678776290631, + "grad_norm": 0.318359375, + "learning_rate": 1.0828442301756312e-05, + "loss": 0.3505, "step": 3220 }, { - "epoch": 2.912375790424571, - "grad_norm": 0.2890625, - "learning_rate": 0.00021489224210298894, - "loss": 0.5764, + "epoch": 6.164435946462715, + "grad_norm": 0.337890625, + "learning_rate": 1.0635243314701163e-05, + "loss": 0.3978, "step": 3224 }, { - "epoch": 2.915989159891599, - "grad_norm": 0.33203125, - "learning_rate": 0.00021469833086527624, - "loss": 0.6103, + "epoch": 6.172084130019121, + "grad_norm": 0.33984375, + "learning_rate": 1.0443720026126273e-05, + "loss": 0.3982, "step": 3228 }, { - "epoch": 2.919602529358627, - "grad_norm": 0.3359375, - "learning_rate": 0.00021450428670132448, - "loss": 0.6215, + "epoch": 6.179732313575526, + "grad_norm": 0.357421875, + "learning_rate": 1.025387473887554e-05, + "loss": 0.3935, "step": 3232 }, { - "epoch": 2.9232158988256547, - "grad_norm": 0.330078125, - "learning_rate": 0.000214310110009808, - "loss": 0.6189, + "epoch": 6.187380497131931, + "grad_norm": 0.337890625, + "learning_rate": 1.0065709735616917e-05, + "loss": 0.3709, "step": 3236 }, { - "epoch": 2.926829268292683, - "grad_norm": 0.31640625, - "learning_rate": 0.00021411580118967335, - "loss": 0.6523, + "epoch": 6.195028680688337, + "grad_norm": 0.3515625, + "learning_rate": 9.87922727881484e-06, + "loss": 0.3922, "step": 3240 }, { - "epoch": 2.930442637759711, - "grad_norm": 0.3046875, - "learning_rate": 0.0002139213606401386, - "loss": 0.651, + "epoch": 6.202676864244742, + "grad_norm": 0.328125, + "learning_rate": 9.694429610703153e-06, + "loss": 0.3719, "step": 3244 }, { - "epoch": 2.934056007226739, - "grad_norm": 0.3046875, - "learning_rate": 0.00021372678876069236, - "loss": 0.6704, + "epoch": 6.210325047801147, + "grad_norm": 0.333984375, + "learning_rate": 9.511318953258013e-06, + "loss": 0.2995, "step": 3248 }, { - "epoch": 2.937669376693767, - "grad_norm": 0.33203125, - "learning_rate": 0.00021353208595109317, - "loss": 0.6155, + "epoch": 6.217973231357552, + "grad_norm": 0.34765625, + "learning_rate": 9.329897508171296e-06, + "loss": 0.3932, "step": 3252 }, { - "epoch": 2.941282746160795, - "grad_norm": 0.283203125, - "learning_rate": 0.00021333725261136856, - "loss": 0.6136, + "epoch": 6.225621414913958, + "grad_norm": 0.34765625, + "learning_rate": 9.150167456824065e-06, + "loss": 0.4016, "step": 3256 }, { - "epoch": 2.944896115627823, - "grad_norm": 0.341796875, - "learning_rate": 0.00021314228914181418, - "loss": 0.6262, + "epoch": 6.233269598470363, + "grad_norm": 0.33984375, + "learning_rate": 8.972130960260326e-06, + "loss": 0.3622, "step": 3260 }, { - "epoch": 2.948509485094851, - "grad_norm": 0.3359375, - "learning_rate": 0.00021294719594299315, - "loss": 0.6668, + "epoch": 6.240917782026768, + "grad_norm": 0.337890625, + "learning_rate": 8.795790159161098e-06, + "loss": 0.4053, "step": 3264 }, { - "epoch": 2.952122854561879, - "grad_norm": 0.302734375, - "learning_rate": 0.000212751973415735, - "loss": 0.6747, + "epoch": 6.248565965583174, + "grad_norm": 0.3203125, + "learning_rate": 8.621147173818587e-06, + "loss": 0.3517, "step": 3268 }, { - "epoch": 2.955736224028907, - "grad_norm": 0.296875, - "learning_rate": 0.00021255662196113502, - "loss": 0.6239, + "epoch": 6.25621414913958, + "grad_norm": 0.35546875, + "learning_rate": 8.448204104110818e-06, + "loss": 0.3448, "step": 3272 }, { - "epoch": 2.959349593495935, - "grad_norm": 0.3203125, - "learning_rate": 0.00021236114198055342, - "loss": 0.6522, + "epoch": 6.263862332695985, + "grad_norm": 0.349609375, + "learning_rate": 8.276963029476275e-06, + "loss": 0.3758, "step": 3276 }, { - "epoch": 2.962962962962963, - "grad_norm": 0.3359375, - "learning_rate": 0.00021216553387561445, - "loss": 0.6569, + "epoch": 6.27151051625239, + "grad_norm": 0.37109375, + "learning_rate": 8.107426008888934e-06, + "loss": 0.3911, "step": 3280 }, { - "epoch": 2.966576332429991, - "grad_norm": 0.3203125, - "learning_rate": 0.00021196979804820562, - "loss": 0.6878, + "epoch": 6.279158699808796, + "grad_norm": 0.322265625, + "learning_rate": 7.93959508083351e-06, + "loss": 0.3491, "step": 3284 }, { - "epoch": 2.970189701897019, - "grad_norm": 0.298828125, - "learning_rate": 0.00021177393490047678, - "loss": 0.6494, + "epoch": 6.286806883365201, + "grad_norm": 0.34375, + "learning_rate": 7.773472263280977e-06, + "loss": 0.3954, "step": 3288 }, { - "epoch": 2.973803071364047, - "grad_norm": 0.28515625, - "learning_rate": 0.00021157794483483947, - "loss": 0.6326, + "epoch": 6.294455066921606, + "grad_norm": 0.34375, + "learning_rate": 7.609059553664254e-06, + "loss": 0.4018, "step": 3292 }, { - "epoch": 2.977416440831075, - "grad_norm": 0.31640625, - "learning_rate": 0.00021138182825396594, - "loss": 0.6859, + "epoch": 6.3021032504780115, + "grad_norm": 0.330078125, + "learning_rate": 7.446358928854207e-06, + "loss": 0.3823, "step": 3296 }, { - "epoch": 2.9810298102981028, - "grad_norm": 0.298828125, - "learning_rate": 0.00021118558556078827, - "loss": 0.6028, + "epoch": 6.309751434034417, + "grad_norm": 0.345703125, + "learning_rate": 7.2853723451358705e-06, + "loss": 0.4097, "step": 3300 }, { - "epoch": 2.984643179765131, - "grad_norm": 0.296875, - "learning_rate": 0.00021098921715849786, - "loss": 0.6063, + "epoch": 6.317399617590822, + "grad_norm": 0.345703125, + "learning_rate": 7.126101738184964e-06, + "loss": 0.3676, "step": 3304 }, { - "epoch": 2.988256549232159, - "grad_norm": 0.306640625, - "learning_rate": 0.00021079272345054422, - "loss": 0.6191, + "epoch": 6.325047801147227, + "grad_norm": 0.337890625, + "learning_rate": 6.9685490230445615e-06, + "loss": 0.3901, "step": 3308 }, { - "epoch": 2.991869918699187, - "grad_norm": 0.298828125, - "learning_rate": 0.00021059610484063437, - "loss": 0.5924, + "epoch": 6.332695984703633, + "grad_norm": 0.32421875, + "learning_rate": 6.812716094102128e-06, + "loss": 0.3652, "step": 3312 }, { - "epoch": 2.995483288166215, - "grad_norm": 0.318359375, - "learning_rate": 0.00021039936173273196, - "loss": 0.6085, + "epoch": 6.340344168260038, + "grad_norm": 0.35546875, + "learning_rate": 6.658604825066683e-06, + "loss": 0.3631, "step": 3316 }, { - "epoch": 2.9990966576332427, - "grad_norm": 0.30859375, - "learning_rate": 0.00021020249453105637, - "loss": 0.5925, + "epoch": 6.347992351816444, + "grad_norm": 0.341796875, + "learning_rate": 6.50621706894629e-06, + "loss": 0.4053, "step": 3320 }, { - "epoch": 3.002710027100271, - "grad_norm": 0.267578125, - "learning_rate": 0.00021000550364008198, - "loss": 0.5536, + "epoch": 6.355640535372849, + "grad_norm": 0.33203125, + "learning_rate": 6.355554658025791e-06, + "loss": 0.3451, "step": 3324 }, { - "epoch": 3.006323396567299, - "grad_norm": 0.28125, - "learning_rate": 0.00020980838946453727, - "loss": 0.5074, + "epoch": 6.363288718929255, + "grad_norm": 0.34375, + "learning_rate": 6.206619403844804e-06, + "loss": 0.4109, "step": 3328 }, { - "epoch": 3.009936766034327, - "grad_norm": 0.302734375, - "learning_rate": 0.00020961115240940404, - "loss": 0.5476, + "epoch": 6.37093690248566, + "grad_norm": 0.318359375, + "learning_rate": 6.059413097175808e-06, + "loss": 0.3897, "step": 3332 }, { - "epoch": 3.013550135501355, - "grad_norm": 0.349609375, - "learning_rate": 0.00020941379287991657, - "loss": 0.5756, + "epoch": 6.378585086042065, + "grad_norm": 0.341796875, + "learning_rate": 5.913937508002797e-06, + "loss": 0.3573, "step": 3336 }, { - "epoch": 3.017163504968383, - "grad_norm": 0.30078125, - "learning_rate": 0.00020921631128156075, - "loss": 0.5602, + "epoch": 6.3862332695984705, + "grad_norm": 0.341796875, + "learning_rate": 5.770194385499877e-06, + "loss": 0.3928, "step": 3340 }, { - "epoch": 3.020776874435411, - "grad_norm": 0.318359375, - "learning_rate": 0.00020901870802007323, - "loss": 0.5553, + "epoch": 6.393881453154876, + "grad_norm": 0.330078125, + "learning_rate": 5.628185458010248e-06, + "loss": 0.3662, "step": 3344 }, { - "epoch": 3.024390243902439, - "grad_norm": 0.3671875, - "learning_rate": 0.00020882098350144066, - "loss": 0.5324, + "epoch": 6.401529636711281, + "grad_norm": 0.349609375, + "learning_rate": 5.487912433025493e-06, + "loss": 0.3974, "step": 3348 }, { - "epoch": 3.028003613369467, - "grad_norm": 0.30859375, - "learning_rate": 0.00020862313813189886, - "loss": 0.5725, + "epoch": 6.409177820267686, + "grad_norm": 0.32421875, + "learning_rate": 5.349376997164923e-06, + "loss": 0.3477, "step": 3352 }, { - "epoch": 3.031616982836495, - "grad_norm": 0.3359375, - "learning_rate": 0.0002084251723179318, - "loss": 0.5595, + "epoch": 6.416826003824092, + "grad_norm": 0.330078125, + "learning_rate": 5.212580816155426e-06, + "loss": 0.3992, "step": 3356 }, { - "epoch": 3.035230352303523, - "grad_norm": 0.318359375, - "learning_rate": 0.00020822708646627117, - "loss": 0.5841, + "epoch": 6.424474187380497, + "grad_norm": 0.345703125, + "learning_rate": 5.077525534811339e-06, + "loss": 0.3881, "step": 3360 }, { - "epoch": 3.038843721770551, - "grad_norm": 0.306640625, - "learning_rate": 0.00020802888098389493, - "loss": 0.5697, + "epoch": 6.432122370936902, + "grad_norm": 0.322265625, + "learning_rate": 4.9442127770147385e-06, + "loss": 0.3767, "step": 3364 }, { - "epoch": 3.042457091237579, - "grad_norm": 0.318359375, - "learning_rate": 0.0002078305562780272, - "loss": 0.5432, + "epoch": 6.4397705544933075, + "grad_norm": 0.34765625, + "learning_rate": 4.812644145695915e-06, + "loss": 0.3939, "step": 3368 }, { - "epoch": 3.046070460704607, - "grad_norm": 0.3203125, - "learning_rate": 0.00020763211275613678, - "loss": 0.5347, + "epoch": 6.447418738049713, + "grad_norm": 0.3359375, + "learning_rate": 4.682821222813998e-06, + "loss": 0.3322, "step": 3372 }, { - "epoch": 3.049683830171635, - "grad_norm": 0.318359375, - "learning_rate": 0.00020743355082593662, - "loss": 0.5611, + "epoch": 6.455066921606119, + "grad_norm": 0.34375, + "learning_rate": 4.554745569338092e-06, + "loss": 0.3414, "step": 3376 }, { - "epoch": 3.053297199638663, - "grad_norm": 0.30859375, - "learning_rate": 0.00020723487089538317, - "loss": 0.5526, + "epoch": 6.462715105162524, + "grad_norm": 0.35546875, + "learning_rate": 4.428418725228372e-06, + "loss": 0.3982, "step": 3380 }, { - "epoch": 3.0569105691056913, - "grad_norm": 0.306640625, - "learning_rate": 0.000207036073372675, - "loss": 0.5453, + "epoch": 6.4703632887189295, + "grad_norm": 0.33203125, + "learning_rate": 4.303842209417652e-06, + "loss": 0.3522, "step": 3384 }, { - "epoch": 3.060523938572719, - "grad_norm": 0.3359375, - "learning_rate": 0.00020683715866625253, - "loss": 0.5596, + "epoch": 6.478011472275335, + "grad_norm": 0.330078125, + "learning_rate": 4.181017519793079e-06, + "loss": 0.4167, "step": 3388 }, { - "epoch": 3.064137308039747, - "grad_norm": 0.3046875, - "learning_rate": 0.00020663812718479675, - "loss": 0.5554, + "epoch": 6.48565965583174, + "grad_norm": 0.349609375, + "learning_rate": 4.059946133178132e-06, + "loss": 0.3875, "step": 3392 }, { - "epoch": 3.067750677506775, - "grad_norm": 0.314453125, - "learning_rate": 0.00020643897933722875, - "loss": 0.5474, + "epoch": 6.493307839388145, + "grad_norm": 0.33203125, + "learning_rate": 3.94062950531489e-06, + "loss": 0.3662, "step": 3396 }, { - "epoch": 3.071364046973803, - "grad_norm": 0.298828125, - "learning_rate": 0.0002062397155327086, - "loss": 0.5386, + "epoch": 6.500956022944551, + "grad_norm": 0.34375, + "learning_rate": 3.823069070846474e-06, + "loss": 0.3625, "step": 3400 }, { - "epoch": 3.0749774164408312, - "grad_norm": 0.326171875, - "learning_rate": 0.00020604033618063454, - "loss": 0.538, + "epoch": 6.508604206500956, + "grad_norm": 0.318359375, + "learning_rate": 3.707266243299861e-06, + "loss": 0.3445, "step": 3404 }, { - "epoch": 3.078590785907859, - "grad_norm": 0.3203125, - "learning_rate": 0.0002058408416906424, - "loss": 0.5471, + "epoch": 6.516252390057361, + "grad_norm": 0.353515625, + "learning_rate": 3.5932224150688526e-06, + "loss": 0.3705, "step": 3408 }, { - "epoch": 3.082204155374887, - "grad_norm": 0.33203125, - "learning_rate": 0.00020564123247260432, - "loss": 0.573, + "epoch": 6.5239005736137665, + "grad_norm": 0.35546875, + "learning_rate": 3.4809389573973e-06, + "loss": 0.396, "step": 3412 }, { - "epoch": 3.0858175248419153, - "grad_norm": 0.296875, - "learning_rate": 0.00020544150893662846, - "loss": 0.5385, + "epoch": 6.531548757170172, + "grad_norm": 0.359375, + "learning_rate": 3.3704172203627035e-06, + "loss": 0.3689, "step": 3416 }, { - "epoch": 3.089430894308943, - "grad_norm": 0.330078125, - "learning_rate": 0.00020524167149305758, - "loss": 0.5666, + "epoch": 6.539196940726577, + "grad_norm": 0.33203125, + "learning_rate": 3.2616585328599065e-06, + "loss": 0.4253, "step": 3420 }, { - "epoch": 3.093044263775971, - "grad_norm": 0.306640625, - "learning_rate": 0.0002050417205524686, - "loss": 0.5123, + "epoch": 6.546845124282983, + "grad_norm": 0.3359375, + "learning_rate": 3.154664202585128e-06, + "loss": 0.387, "step": 3424 }, { - "epoch": 3.096657633242999, - "grad_norm": 0.318359375, - "learning_rate": 0.00020484165652567157, - "loss": 0.5652, + "epoch": 6.5544933078393885, + "grad_norm": 0.353515625, + "learning_rate": 3.049435516020271e-06, + "loss": 0.3585, "step": 3428 }, { - "epoch": 3.100271002710027, - "grad_norm": 0.3359375, - "learning_rate": 0.00020464147982370896, - "loss": 0.5819, + "epoch": 6.562141491395794, + "grad_norm": 0.34765625, + "learning_rate": 2.94597373841744e-06, + "loss": 0.3912, "step": 3432 }, { - "epoch": 3.1038843721770553, - "grad_norm": 0.314453125, - "learning_rate": 0.00020444119085785467, - "loss": 0.5436, + "epoch": 6.569789674952199, + "grad_norm": 0.337890625, + "learning_rate": 2.844280113783698e-06, + "loss": 0.3863, "step": 3436 }, { - "epoch": 3.107497741644083, - "grad_norm": 0.337890625, - "learning_rate": 0.00020424079003961328, - "loss": 0.5348, + "epoch": 6.577437858508604, + "grad_norm": 0.353515625, + "learning_rate": 2.7443558648661656e-06, + "loss": 0.4214, "step": 3440 }, { - "epoch": 3.111111111111111, - "grad_norm": 0.322265625, - "learning_rate": 0.0002040402777807192, - "loss": 0.5701, + "epoch": 6.58508604206501, + "grad_norm": 0.33203125, + "learning_rate": 2.646202193137248e-06, + "loss": 0.429, "step": 3444 }, { - "epoch": 3.114724480578139, - "grad_norm": 0.306640625, - "learning_rate": 0.00020383965449313567, - "loss": 0.5627, + "epoch": 6.592734225621415, + "grad_norm": 0.330078125, + "learning_rate": 2.549820278780246e-06, + "loss": 0.3523, "step": 3448 }, { - "epoch": 3.118337850045167, - "grad_norm": 0.34375, - "learning_rate": 0.00020363892058905422, - "loss": 0.5389, + "epoch": 6.60038240917782, + "grad_norm": 0.33984375, + "learning_rate": 2.455211280675168e-06, + "loss": 0.4001, "step": 3452 }, { - "epoch": 3.1219512195121952, - "grad_norm": 0.3359375, - "learning_rate": 0.00020343807648089357, - "loss": 0.5448, + "epoch": 6.6080305927342256, + "grad_norm": 0.376953125, + "learning_rate": 2.3623763363847246e-06, + "loss": 0.3819, "step": 3456 }, { - "epoch": 3.125564588979223, - "grad_norm": 0.33984375, - "learning_rate": 0.00020323712258129884, - "loss": 0.5465, + "epoch": 6.615678776290631, + "grad_norm": 0.361328125, + "learning_rate": 2.271316562140757e-06, + "loss": 0.3632, "step": 3460 }, { - "epoch": 3.129177958446251, + "epoch": 6.623326959847036, "grad_norm": 0.357421875, - "learning_rate": 0.00020303605930314077, - "loss": 0.5307, + "learning_rate": 2.182033052830695e-06, + "loss": 0.3842, "step": 3464 }, { - "epoch": 3.1327913279132793, - "grad_norm": 0.30078125, - "learning_rate": 0.00020283488705951475, - "loss": 0.5384, + "epoch": 6.630975143403441, + "grad_norm": 0.34375, + "learning_rate": 2.094526881984521e-06, + "loss": 0.3844, "step": 3468 }, { - "epoch": 3.136404697380307, + "epoch": 6.638623326959847, "grad_norm": 0.328125, - "learning_rate": 0.0002026336062637402, - "loss": 0.5594, + "learning_rate": 2.0087991017617598e-06, + "loss": 0.3697, "step": 3472 }, { - "epoch": 3.140018066847335, - "grad_norm": 0.326171875, - "learning_rate": 0.00020243221732935932, - "loss": 0.529, + "epoch": 6.646271510516252, + "grad_norm": 0.3515625, + "learning_rate": 1.924850742938894e-06, + "loss": 0.3886, "step": 3476 }, { - "epoch": 3.1436314363143634, - "grad_norm": 0.337890625, - "learning_rate": 0.00020223072067013673, - "loss": 0.5833, + "epoch": 6.653919694072657, + "grad_norm": 0.349609375, + "learning_rate": 1.8426828148969008e-06, + "loss": 0.363, "step": 3480 }, { - "epoch": 3.147244805781391, - "grad_norm": 0.322265625, - "learning_rate": 0.00020202911670005823, - "loss": 0.5319, + "epoch": 6.661567877629063, + "grad_norm": 0.33984375, + "learning_rate": 1.7622963056091843e-06, + "loss": 0.3671, "step": 3484 }, { - "epoch": 3.1508581752484193, - "grad_norm": 0.314453125, - "learning_rate": 0.00020182740583333012, - "loss": 0.5311, + "epoch": 6.669216061185469, + "grad_norm": 0.341796875, + "learning_rate": 1.6836921816296644e-06, + "loss": 0.4043, "step": 3488 }, { - "epoch": 3.154471544715447, - "grad_norm": 0.3125, - "learning_rate": 0.00020162558848437842, - "loss": 0.5411, + "epoch": 6.676864244741874, + "grad_norm": 0.33203125, + "learning_rate": 1.6068713880811546e-06, + "loss": 0.398, "step": 3492 }, { - "epoch": 3.158084914182475, - "grad_norm": 0.3125, - "learning_rate": 0.0002014236650678478, - "loss": 0.5302, + "epoch": 6.684512428298279, + "grad_norm": 0.330078125, + "learning_rate": 1.531834848643987e-06, + "loss": 0.3847, "step": 3496 }, { - "epoch": 3.1616982836495033, - "grad_norm": 0.306640625, - "learning_rate": 0.00020122163599860086, - "loss": 0.5373, + "epoch": 6.692160611854685, + "grad_norm": 0.34765625, + "learning_rate": 1.4585834655449547e-06, + "loss": 0.4041, "step": 3500 }, { - "epoch": 3.165311653116531, - "grad_norm": 0.34765625, - "learning_rate": 0.00020101950169171737, - "loss": 0.5507, + "epoch": 6.69980879541109, + "grad_norm": 0.333984375, + "learning_rate": 1.3871181195464042e-06, + "loss": 0.3924, "step": 3504 }, { - "epoch": 3.1689250225835592, - "grad_norm": 0.33984375, - "learning_rate": 0.0002008172625624932, - "loss": 0.5737, + "epoch": 6.707456978967495, + "grad_norm": 0.32421875, + "learning_rate": 1.3174396699356937e-06, + "loss": 0.3557, "step": 3508 }, { - "epoch": 3.172538392050587, - "grad_norm": 0.333984375, - "learning_rate": 0.00020061491902643972, - "loss": 0.5877, + "epoch": 6.7151051625239, + "grad_norm": 0.33984375, + "learning_rate": 1.2495489545148008e-06, + "loss": 0.3767, "step": 3512 }, { - "epoch": 3.176151761517615, - "grad_norm": 0.328125, - "learning_rate": 0.00020041247149928267, - "loss": 0.5461, + "epoch": 6.722753346080306, + "grad_norm": 0.31640625, + "learning_rate": 1.1834467895903476e-06, + "loss": 0.4122, "step": 3516 }, { - "epoch": 3.1797651309846433, - "grad_norm": 0.353515625, - "learning_rate": 0.00020020992039696152, - "loss": 0.5582, + "epoch": 6.730401529636711, + "grad_norm": 0.3515625, + "learning_rate": 1.1191339699636426e-06, + "loss": 0.3873, "step": 3520 }, { - "epoch": 3.183378500451671, - "grad_norm": 0.337890625, - "learning_rate": 0.00020000726613562852, - "loss": 0.5615, + "epoch": 6.738049713193116, + "grad_norm": 0.3515625, + "learning_rate": 1.0566112689213035e-06, + "loss": 0.4039, "step": 3524 }, { - "epoch": 3.186991869918699, - "grad_norm": 0.322265625, - "learning_rate": 0.00019980450913164785, - "loss": 0.5438, + "epoch": 6.7456978967495225, + "grad_norm": 0.341796875, + "learning_rate": 9.958794382257663e-07, + "loss": 0.4338, "step": 3528 }, { - "epoch": 3.1906052393857274, - "grad_norm": 0.333984375, - "learning_rate": 0.00019960164980159484, - "loss": 0.5615, + "epoch": 6.753346080305928, + "grad_norm": 0.337890625, + "learning_rate": 9.369392081063908e-07, + "loss": 0.3474, "step": 3532 }, { - "epoch": 3.194218608852755, - "grad_norm": 0.337890625, - "learning_rate": 0.000199398688562255, - "loss": 0.5521, + "epoch": 6.760994263862333, + "grad_norm": 0.330078125, + "learning_rate": 8.797912872506529e-07, + "loss": 0.3553, "step": 3536 }, { - "epoch": 3.1978319783197833, - "grad_norm": 0.322265625, - "learning_rate": 0.0001991956258306232, - "loss": 0.5773, + "epoch": 6.768642447418738, + "grad_norm": 0.35546875, + "learning_rate": 8.244363627955664e-07, + "loss": 0.3955, "step": 3540 }, { - "epoch": 3.201445347786811, - "grad_norm": 0.310546875, - "learning_rate": 0.00019899246202390294, - "loss": 0.5693, + "epoch": 6.776290630975144, + "grad_norm": 0.357421875, + "learning_rate": 7.708751003194569e-07, + "loss": 0.3925, "step": 3544 }, { - "epoch": 3.205058717253839, - "grad_norm": 0.31640625, - "learning_rate": 0.0001987891975595052, - "loss": 0.5533, + "epoch": 6.783938814531549, + "grad_norm": 0.333984375, + "learning_rate": 7.191081438339685e-07, + "loss": 0.4088, "step": 3548 }, { - "epoch": 3.2086720867208673, - "grad_norm": 0.33203125, - "learning_rate": 0.000198585832855048, - "loss": 0.5302, + "epoch": 6.791586998087954, + "grad_norm": 0.322265625, + "learning_rate": 6.691361157763198e-07, + "loss": 0.3916, "step": 3552 }, { - "epoch": 3.212285456187895, - "grad_norm": 0.34765625, - "learning_rate": 0.0001983823683283551, - "loss": 0.5504, + "epoch": 6.7992351816443595, + "grad_norm": 0.349609375, + "learning_rate": 6.209596170018094e-07, + "loss": 0.4524, "step": 3556 }, { - "epoch": 3.2158988256549232, - "grad_norm": 0.31640625, - "learning_rate": 0.00019817880439745546, - "loss": 0.5243, + "epoch": 6.806883365200765, + "grad_norm": 0.357421875, + "learning_rate": 5.745792267765559e-07, + "loss": 0.398, "step": 3560 }, { - "epoch": 3.2195121951219514, - "grad_norm": 0.3203125, - "learning_rate": 0.00019797514148058224, - "loss": 0.5859, + "epoch": 6.81453154875717, + "grad_norm": 0.32421875, + "learning_rate": 5.29995502770636e-07, + "loss": 0.4172, "step": 3564 }, { - "epoch": 3.223125564588979, - "grad_norm": 0.34375, - "learning_rate": 0.000197771379996172, - "loss": 0.588, + "epoch": 6.822179732313575, + "grad_norm": 0.33203125, + "learning_rate": 4.87208981051257e-07, + "loss": 0.3622, "step": 3568 }, { - "epoch": 3.2267389340560073, - "grad_norm": 0.345703125, - "learning_rate": 0.00019756752036286375, - "loss": 0.5883, + "epoch": 6.829827915869981, + "grad_norm": 0.365234375, + "learning_rate": 4.462201760763784e-07, + "loss": 0.4181, "step": 3572 }, { - "epoch": 3.230352303523035, - "grad_norm": 0.314453125, - "learning_rate": 0.0001973635629994982, - "loss": 0.514, + "epoch": 6.837476099426386, + "grad_norm": 0.35546875, + "learning_rate": 4.0702958068853373e-07, + "loss": 0.3955, "step": 3576 }, { - "epoch": 3.233965672990063, - "grad_norm": 0.337890625, - "learning_rate": 0.0001971595083251169, - "loss": 0.5427, + "epoch": 6.845124282982791, + "grad_norm": 0.349609375, + "learning_rate": 3.696376661088685e-07, + "loss": 0.3865, "step": 3580 }, { - "epoch": 3.2375790424570914, - "grad_norm": 0.33203125, - "learning_rate": 0.00019695535675896115, - "loss": 0.513, + "epoch": 6.8527724665391965, + "grad_norm": 0.333984375, + "learning_rate": 3.340448819315111e-07, + "loss": 0.3909, "step": 3584 }, { - "epoch": 3.241192411924119, - "grad_norm": 0.333984375, - "learning_rate": 0.0001967511087204716, - "loss": 0.5919, + "epoch": 6.860420650095603, + "grad_norm": 0.341796875, + "learning_rate": 3.002516561181112e-07, + "loss": 0.3462, "step": 3588 }, { - "epoch": 3.2448057813911473, - "grad_norm": 0.3203125, - "learning_rate": 0.0001965467646292868, - "loss": 0.5145, + "epoch": 6.868068833652008, + "grad_norm": 0.373046875, + "learning_rate": 2.6825839499277634e-07, + "loss": 0.3655, "step": 3592 }, { - "epoch": 3.2484191508581755, - "grad_norm": 0.33203125, - "learning_rate": 0.00019634232490524297, - "loss": 0.5296, + "epoch": 6.875717017208413, + "grad_norm": 0.32421875, + "learning_rate": 2.3806548323710984e-07, + "loss": 0.4164, "step": 3596 }, { - "epoch": 3.252032520325203, - "grad_norm": 0.328125, - "learning_rate": 0.00019613778996837252, - "loss": 0.5701, + "epoch": 6.8833652007648185, + "grad_norm": 0.365234375, + "learning_rate": 2.096732838856141e-07, + "loss": 0.3865, "step": 3600 }, { - "epoch": 3.2556458897922314, - "grad_norm": 0.333984375, - "learning_rate": 0.00019593316023890358, - "loss": 0.5738, + "epoch": 6.891013384321224, + "grad_norm": 0.33984375, + "learning_rate": 1.8308213832134434e-07, + "loss": 0.4012, "step": 3604 }, { - "epoch": 3.259259259259259, - "grad_norm": 0.341796875, - "learning_rate": 0.00019572843613725917, - "loss": 0.5653, + "epoch": 6.898661567877629, + "grad_norm": 0.333984375, + "learning_rate": 1.5829236627177833e-07, + "loss": 0.3761, "step": 3608 }, { - "epoch": 3.2628726287262872, - "grad_norm": 0.31640625, - "learning_rate": 0.00019552361808405596, - "loss": 0.5736, + "epoch": 6.906309751434034, + "grad_norm": 0.341796875, + "learning_rate": 1.3530426580496968e-07, + "loss": 0.3674, "step": 3612 }, { - "epoch": 3.2664859981933154, - "grad_norm": 0.328125, - "learning_rate": 0.0001953187065001039, - "loss": 0.584, + "epoch": 6.91395793499044, + "grad_norm": 0.3359375, + "learning_rate": 1.1411811332600051e-07, + "loss": 0.3708, "step": 3616 }, { - "epoch": 3.270099367660343, - "grad_norm": 0.34375, - "learning_rate": 0.00019511370180640486, - "loss": 0.5677, + "epoch": 6.921606118546845, + "grad_norm": 0.333984375, + "learning_rate": 9.473416357361762e-08, + "loss": 0.4026, "step": 3620 }, { - "epoch": 3.2737127371273713, - "grad_norm": 0.326171875, - "learning_rate": 0.00019490860442415226, - "loss": 0.5222, + "epoch": 6.92925430210325, + "grad_norm": 0.35546875, + "learning_rate": 7.715264961718481e-08, + "loss": 0.4073, "step": 3624 }, { - "epoch": 3.2773261065943995, - "grad_norm": 0.353515625, - "learning_rate": 0.0001947034147747297, - "loss": 0.5507, + "epoch": 6.9369024856596555, + "grad_norm": 0.3359375, + "learning_rate": 6.137378285386851e-08, + "loss": 0.3944, "step": 3628 }, { - "epoch": 3.280939476061427, - "grad_norm": 0.322265625, - "learning_rate": 0.00019449813327971057, - "loss": 0.5827, + "epoch": 6.944550669216062, + "grad_norm": 0.337890625, + "learning_rate": 4.739775300612314e-08, + "loss": 0.352, "step": 3632 }, { - "epoch": 3.2845528455284554, - "grad_norm": 0.3203125, - "learning_rate": 0.0001942927603608568, - "loss": 0.5695, + "epoch": 6.952198852772467, + "grad_norm": 0.333984375, + "learning_rate": 3.522472811939292e-08, + "loss": 0.3563, "step": 3636 }, { - "epoch": 3.288166214995483, - "grad_norm": 0.3359375, - "learning_rate": 0.00019408729644011826, - "loss": 0.5417, + "epoch": 6.959847036328872, + "grad_norm": 0.34375, + "learning_rate": 2.4854854560096837e-08, + "loss": 0.3673, "step": 3640 }, { - "epoch": 3.2917795844625113, - "grad_norm": 0.3203125, - "learning_rate": 0.00019388174193963174, - "loss": 0.5685, + "epoch": 6.9674952198852775, + "grad_norm": 0.353515625, + "learning_rate": 1.6288257013830075e-08, + "loss": 0.4149, "step": 3644 }, { - "epoch": 3.2953929539295395, - "grad_norm": 0.34765625, - "learning_rate": 0.00019367609728172016, - "loss": 0.6039, + "epoch": 6.975143403441683, + "grad_norm": 0.3515625, + "learning_rate": 9.52503848396513e-09, + "loss": 0.4061, "step": 3648 }, { - "epoch": 3.299006323396567, - "grad_norm": 0.33203125, - "learning_rate": 0.0001934703628888916, - "loss": 0.5437, + "epoch": 6.982791586998088, + "grad_norm": 0.32421875, + "learning_rate": 4.565280290269591e-09, + "loss": 0.3616, "step": 3652 }, { - "epoch": 3.3026196928635954, - "grad_norm": 0.33984375, - "learning_rate": 0.0001932645391838386, - "loss": 0.521, + "epoch": 6.990439770554493, + "grad_norm": 0.3671875, + "learning_rate": 1.4090420680734627e-09, + "loss": 0.3985, "step": 3656 }, { - "epoch": 3.306233062330623, + "epoch": 6.998087954110899, "grad_norm": 0.3125, - "learning_rate": 0.0001930586265894371, - "loss": 0.563, + "learning_rate": 5.6361767436507997e-11, + "loss": 0.3654, "step": 3660 }, { - "epoch": 3.3098464317976513, - "grad_norm": 0.337890625, - "learning_rate": 0.0001928526255287458, - "loss": 0.5217, - "step": 3664 - }, - { - "epoch": 3.3134598012646794, - "grad_norm": 0.34375, - "learning_rate": 0.000192646536425005, - "loss": 0.5684, - "step": 3668 - }, - { - "epoch": 3.317073170731707, - "grad_norm": 0.345703125, - "learning_rate": 0.00019244035970163604, - "loss": 0.5654, - "step": 3672 - }, - { - "epoch": 3.3206865401987353, - "grad_norm": 0.328125, - "learning_rate": 0.00019223409578224022, - "loss": 0.5945, - "step": 3676 - }, - { - "epoch": 3.3242999096657635, - "grad_norm": 0.341796875, - "learning_rate": 0.00019202774509059789, - "loss": 0.5358, - "step": 3680 - }, - { - "epoch": 3.3279132791327912, - "grad_norm": 0.37890625, - "learning_rate": 0.00019182130805066785, - "loss": 0.5368, - "step": 3684 - }, - { - "epoch": 3.3315266485998194, - "grad_norm": 0.3359375, - "learning_rate": 0.00019161478508658615, - "loss": 0.5286, - "step": 3688 - }, - { - "epoch": 3.3351400180668476, - "grad_norm": 0.34375, - "learning_rate": 0.00019140817662266556, - "loss": 0.5547, - "step": 3692 - }, - { - "epoch": 3.3387533875338753, - "grad_norm": 0.322265625, - "learning_rate": 0.0001912014830833943, - "loss": 0.5808, - "step": 3696 - }, - { - "epoch": 3.3423667570009035, - "grad_norm": 0.328125, - "learning_rate": 0.00019099470489343558, - "loss": 0.5341, - "step": 3700 - }, - { - "epoch": 3.345980126467931, - "grad_norm": 0.341796875, - "learning_rate": 0.00019078784247762633, - "loss": 0.6313, - "step": 3704 - }, - { - "epoch": 3.3495934959349594, - "grad_norm": 0.36328125, - "learning_rate": 0.00019058089626097668, - "loss": 0.6229, - "step": 3708 - }, - { - "epoch": 3.3532068654019875, - "grad_norm": 0.337890625, - "learning_rate": 0.0001903738666686689, - "loss": 0.5368, - "step": 3712 - }, - { - "epoch": 3.3568202348690153, - "grad_norm": 0.345703125, - "learning_rate": 0.00019016675412605652, - "loss": 0.5718, - "step": 3716 - }, - { - "epoch": 3.3604336043360434, - "grad_norm": 0.302734375, - "learning_rate": 0.0001899595590586636, - "loss": 0.5737, - "step": 3720 - }, - { - "epoch": 3.364046973803071, - "grad_norm": 0.359375, - "learning_rate": 0.00018975228189218354, - "loss": 0.5556, - "step": 3724 - }, - { - "epoch": 3.3676603432700993, - "grad_norm": 0.35546875, - "learning_rate": 0.00018954492305247865, - "loss": 0.5544, - "step": 3728 - }, - { - "epoch": 3.3712737127371275, - "grad_norm": 0.345703125, - "learning_rate": 0.0001893374829655789, - "loss": 0.5282, - "step": 3732 - }, - { - "epoch": 3.3748870822041552, - "grad_norm": 0.34765625, - "learning_rate": 0.00018912996205768125, - "loss": 0.5734, - "step": 3736 - }, - { - "epoch": 3.3785004516711834, - "grad_norm": 0.357421875, - "learning_rate": 0.00018892236075514869, - "loss": 0.6076, - "step": 3740 - }, - { - "epoch": 3.3821138211382116, - "grad_norm": 0.345703125, - "learning_rate": 0.00018871467948450936, - "loss": 0.5469, - "step": 3744 - }, - { - "epoch": 3.3857271906052393, - "grad_norm": 0.36328125, - "learning_rate": 0.00018850691867245572, - "loss": 0.5731, - "step": 3748 - }, - { - "epoch": 3.3893405600722675, - "grad_norm": 0.337890625, - "learning_rate": 0.00018829907874584376, - "loss": 0.5417, - "step": 3752 - }, - { - "epoch": 3.392953929539295, - "grad_norm": 0.345703125, - "learning_rate": 0.00018809116013169177, - "loss": 0.5449, - "step": 3756 - }, - { - "epoch": 3.3965672990063234, - "grad_norm": 0.345703125, - "learning_rate": 0.00018788316325717993, - "loss": 0.537, - "step": 3760 - }, - { - "epoch": 3.4001806684733515, - "grad_norm": 0.333984375, - "learning_rate": 0.00018767508854964908, - "loss": 0.5881, - "step": 3764 - }, - { - "epoch": 3.4037940379403793, - "grad_norm": 0.32421875, - "learning_rate": 0.00018746693643660007, - "loss": 0.5387, - "step": 3768 - }, - { - "epoch": 3.4074074074074074, - "grad_norm": 0.33984375, - "learning_rate": 0.0001872587073456927, - "loss": 0.5376, - "step": 3772 - }, - { - "epoch": 3.4110207768744356, - "grad_norm": 0.330078125, - "learning_rate": 0.00018705040170474495, - "loss": 0.56, - "step": 3776 - }, - { - "epoch": 3.4146341463414633, - "grad_norm": 0.31640625, - "learning_rate": 0.00018684201994173216, - "loss": 0.6074, - "step": 3780 - }, - { - "epoch": 3.4182475158084915, - "grad_norm": 0.337890625, - "learning_rate": 0.00018663356248478592, - "loss": 0.5337, - "step": 3784 - }, - { - "epoch": 3.4218608852755192, - "grad_norm": 0.31640625, - "learning_rate": 0.0001864250297621935, - "loss": 0.5546, - "step": 3788 - }, - { - "epoch": 3.4254742547425474, - "grad_norm": 0.31640625, - "learning_rate": 0.00018621642220239653, - "loss": 0.5278, - "step": 3792 - }, - { - "epoch": 3.4290876242095756, - "grad_norm": 0.349609375, - "learning_rate": 0.0001860077402339908, - "loss": 0.5097, - "step": 3796 - }, - { - "epoch": 3.4327009936766033, - "grad_norm": 0.341796875, - "learning_rate": 0.00018579898428572466, - "loss": 0.5837, - "step": 3800 - }, - { - "epoch": 3.4363143631436315, - "grad_norm": 0.328125, - "learning_rate": 0.00018559015478649858, - "loss": 0.5655, - "step": 3804 - }, - { - "epoch": 3.439927732610659, - "grad_norm": 0.326171875, - "learning_rate": 0.00018538125216536412, - "loss": 0.5507, - "step": 3808 - }, - { - "epoch": 3.4435411020776874, - "grad_norm": 0.359375, - "learning_rate": 0.00018517227685152304, - "loss": 0.5943, - "step": 3812 - }, - { - "epoch": 3.4471544715447155, - "grad_norm": 0.322265625, - "learning_rate": 0.00018496322927432653, - "loss": 0.5815, - "step": 3816 - }, - { - "epoch": 3.4507678410117433, - "grad_norm": 0.326171875, - "learning_rate": 0.00018475410986327416, - "loss": 0.5206, - "step": 3820 - }, - { - "epoch": 3.4543812104787714, - "grad_norm": 0.34765625, - "learning_rate": 0.00018454491904801315, - "loss": 0.5722, - "step": 3824 - }, - { - "epoch": 3.4579945799457996, - "grad_norm": 0.322265625, - "learning_rate": 0.00018433565725833753, - "loss": 0.5577, - "step": 3828 - }, - { - "epoch": 3.4616079494128273, - "grad_norm": 0.314453125, - "learning_rate": 0.00018412632492418682, - "loss": 0.5678, - "step": 3832 - }, - { - "epoch": 3.4652213188798555, - "grad_norm": 0.32421875, - "learning_rate": 0.00018391692247564582, - "loss": 0.5566, - "step": 3836 - }, - { - "epoch": 3.4688346883468837, - "grad_norm": 0.345703125, - "learning_rate": 0.00018370745034294316, - "loss": 0.5937, - "step": 3840 - }, - { - "epoch": 3.4724480578139114, - "grad_norm": 0.35546875, - "learning_rate": 0.00018349790895645082, - "loss": 0.5795, - "step": 3844 - }, - { - "epoch": 3.4760614272809396, - "grad_norm": 0.35546875, - "learning_rate": 0.0001832882987466829, - "loss": 0.5501, - "step": 3848 - }, - { - "epoch": 3.4796747967479673, - "grad_norm": 0.37109375, - "learning_rate": 0.000183078620144295, - "loss": 0.5379, - "step": 3852 - }, - { - "epoch": 3.4832881662149955, - "grad_norm": 0.333984375, - "learning_rate": 0.00018286887358008323, - "loss": 0.5591, - "step": 3856 - }, - { - "epoch": 3.4869015356820237, - "grad_norm": 0.353515625, - "learning_rate": 0.0001826590594849833, - "loss": 0.5886, - "step": 3860 - }, - { - "epoch": 3.4905149051490514, - "grad_norm": 0.32421875, - "learning_rate": 0.00018244917829006964, - "loss": 0.5821, - "step": 3864 - }, - { - "epoch": 3.4941282746160796, - "grad_norm": 0.330078125, - "learning_rate": 0.0001822392304265546, - "loss": 0.5541, - "step": 3868 - }, - { - "epoch": 3.4977416440831073, - "grad_norm": 0.32421875, - "learning_rate": 0.0001820292163257875, - "loss": 0.5662, - "step": 3872 - }, - { - "epoch": 3.5013550135501355, - "grad_norm": 0.33984375, - "learning_rate": 0.00018181913641925377, - "loss": 0.5446, - "step": 3876 - }, - { - "epoch": 3.5049683830171636, - "grad_norm": 0.3359375, - "learning_rate": 0.00018160899113857392, - "loss": 0.5613, - "step": 3880 - }, - { - "epoch": 3.5085817524841914, - "grad_norm": 0.33984375, - "learning_rate": 0.00018139878091550293, - "loss": 0.5821, - "step": 3884 - }, - { - "epoch": 3.5121951219512195, - "grad_norm": 0.353515625, - "learning_rate": 0.00018118850618192904, - "loss": 0.5521, - "step": 3888 - }, - { - "epoch": 3.5158084914182473, - "grad_norm": 0.35546875, - "learning_rate": 0.0001809781673698732, - "loss": 0.5405, - "step": 3892 - }, - { - "epoch": 3.5194218608852754, - "grad_norm": 0.328125, - "learning_rate": 0.00018076776491148794, - "loss": 0.5545, - "step": 3896 - }, - { - "epoch": 3.5230352303523036, - "grad_norm": 0.33984375, - "learning_rate": 0.0001805572992390565, - "loss": 0.5163, - "step": 3900 - }, - { - "epoch": 3.5266485998193318, - "grad_norm": 0.328125, - "learning_rate": 0.00018034677078499216, - "loss": 0.5589, - "step": 3904 - }, - { - "epoch": 3.5302619692863595, - "grad_norm": 0.333984375, - "learning_rate": 0.0001801361799818369, - "loss": 0.6153, - "step": 3908 - }, - { - "epoch": 3.5338753387533877, - "grad_norm": 0.3203125, - "learning_rate": 0.0001799255272622611, - "loss": 0.5782, - "step": 3912 - }, - { - "epoch": 3.5374887082204154, - "grad_norm": 0.34375, - "learning_rate": 0.00017971481305906222, - "loss": 0.5632, - "step": 3916 - }, - { - "epoch": 3.5411020776874436, - "grad_norm": 0.328125, - "learning_rate": 0.00017950403780516398, - "loss": 0.557, - "step": 3920 - }, - { - "epoch": 3.5447154471544717, - "grad_norm": 0.333984375, - "learning_rate": 0.00017929320193361574, - "loss": 0.5593, - "step": 3924 - }, - { - "epoch": 3.5483288166214995, - "grad_norm": 0.349609375, - "learning_rate": 0.00017908230587759108, - "loss": 0.5818, - "step": 3928 - }, - { - "epoch": 3.5519421860885276, - "grad_norm": 0.33984375, - "learning_rate": 0.00017887135007038762, - "loss": 0.5731, - "step": 3932 - }, - { - "epoch": 3.5555555555555554, - "grad_norm": 0.353515625, - "learning_rate": 0.00017866033494542535, - "loss": 0.5172, - "step": 3936 - }, - { - "epoch": 3.5591689250225835, - "grad_norm": 0.333984375, - "learning_rate": 0.00017844926093624644, - "loss": 0.5789, - "step": 3940 - }, - { - "epoch": 3.5627822944896117, - "grad_norm": 0.33984375, - "learning_rate": 0.00017823812847651382, - "loss": 0.5574, - "step": 3944 - }, - { - "epoch": 3.5663956639566394, - "grad_norm": 0.341796875, - "learning_rate": 0.0001780269380000107, - "loss": 0.5353, - "step": 3948 - }, - { - "epoch": 3.5700090334236676, - "grad_norm": 0.345703125, - "learning_rate": 0.00017781568994063937, - "loss": 0.5513, - "step": 3952 - }, - { - "epoch": 3.5736224028906953, - "grad_norm": 0.322265625, - "learning_rate": 0.0001776043847324205, - "loss": 0.557, - "step": 3956 - }, - { - "epoch": 3.5772357723577235, - "grad_norm": 0.357421875, - "learning_rate": 0.000177393022809492, - "loss": 0.5666, - "step": 3960 - }, - { - "epoch": 3.5808491418247517, - "grad_norm": 0.35546875, - "learning_rate": 0.00017718160460610858, - "loss": 0.579, - "step": 3964 - }, - { - "epoch": 3.58446251129178, - "grad_norm": 0.318359375, - "learning_rate": 0.00017697013055664035, - "loss": 0.57, - "step": 3968 - }, - { - "epoch": 3.5880758807588076, - "grad_norm": 0.35546875, - "learning_rate": 0.00017675860109557225, - "loss": 0.5668, - "step": 3972 - }, - { - "epoch": 3.5916892502258357, - "grad_norm": 0.33984375, - "learning_rate": 0.00017654701665750306, - "loss": 0.5952, - "step": 3976 - }, - { - "epoch": 3.5953026196928635, - "grad_norm": 0.365234375, - "learning_rate": 0.00017633537767714462, - "loss": 0.5357, - "step": 3980 - }, - { - "epoch": 3.5989159891598916, - "grad_norm": 0.361328125, - "learning_rate": 0.00017612368458932052, - "loss": 0.5801, - "step": 3984 - }, - { - "epoch": 3.60252935862692, - "grad_norm": 0.31640625, - "learning_rate": 0.0001759119378289659, - "loss": 0.5527, - "step": 3988 - }, - { - "epoch": 3.6061427280939475, - "grad_norm": 0.34765625, - "learning_rate": 0.00017570013783112592, - "loss": 0.5276, - "step": 3992 - }, - { - "epoch": 3.6097560975609757, - "grad_norm": 0.333984375, - "learning_rate": 0.00017548828503095518, - "loss": 0.5595, - "step": 3996 - }, - { - "epoch": 3.6133694670280034, - "grad_norm": 0.3359375, - "learning_rate": 0.00017527637986371685, - "loss": 0.5526, - "step": 4000 - }, - { - "epoch": 3.6169828364950316, - "grad_norm": 0.35546875, - "learning_rate": 0.00017506442276478152, - "loss": 0.5703, - "step": 4004 - }, - { - "epoch": 3.6205962059620598, - "grad_norm": 0.341796875, - "learning_rate": 0.00017485241416962668, - "loss": 0.5836, - "step": 4008 - }, - { - "epoch": 3.6242095754290875, - "grad_norm": 0.3203125, - "learning_rate": 0.0001746403545138354, - "loss": 0.5687, - "step": 4012 - }, - { - "epoch": 3.6278229448961157, - "grad_norm": 0.35546875, - "learning_rate": 0.00017442824423309592, - "loss": 0.575, - "step": 4016 - }, - { - "epoch": 3.6314363143631434, - "grad_norm": 0.341796875, - "learning_rate": 0.0001742160837632002, - "loss": 0.5646, - "step": 4020 - }, - { - "epoch": 3.6350496838301716, - "grad_norm": 0.330078125, - "learning_rate": 0.00017400387354004357, - "loss": 0.5465, - "step": 4024 - }, - { - "epoch": 3.6386630532971997, - "grad_norm": 0.314453125, - "learning_rate": 0.0001737916139996234, - "loss": 0.5504, - "step": 4028 - }, - { - "epoch": 3.642276422764228, - "grad_norm": 0.349609375, - "learning_rate": 0.00017357930557803853, - "loss": 0.5749, - "step": 4032 - }, - { - "epoch": 3.6458897922312556, - "grad_norm": 0.349609375, - "learning_rate": 0.00017336694871148808, - "loss": 0.5385, - "step": 4036 - }, - { - "epoch": 3.649503161698284, - "grad_norm": 0.33984375, - "learning_rate": 0.00017315454383627082, - "loss": 0.5664, - "step": 4040 - }, - { - "epoch": 3.6531165311653115, - "grad_norm": 0.349609375, - "learning_rate": 0.0001729420913887841, - "loss": 0.5726, - "step": 4044 - }, - { - "epoch": 3.6567299006323397, - "grad_norm": 0.33203125, - "learning_rate": 0.0001727295918055231, - "loss": 0.5914, - "step": 4048 - }, - { - "epoch": 3.660343270099368, - "grad_norm": 0.353515625, - "learning_rate": 0.00017251704552307965, - "loss": 0.5505, - "step": 4052 - }, - { - "epoch": 3.6639566395663956, - "grad_norm": 0.357421875, - "learning_rate": 0.00017230445297814175, - "loss": 0.5353, - "step": 4056 - }, - { - "epoch": 3.667570009033424, - "grad_norm": 0.34375, - "learning_rate": 0.00017209181460749228, - "loss": 0.5662, - "step": 4060 - }, - { - "epoch": 3.6711833785004515, - "grad_norm": 0.349609375, - "learning_rate": 0.00017187913084800837, - "loss": 0.5594, - "step": 4064 - }, - { - "epoch": 3.6747967479674797, - "grad_norm": 0.3125, - "learning_rate": 0.0001716664021366603, - "loss": 0.532, - "step": 4068 - }, - { - "epoch": 3.678410117434508, - "grad_norm": 0.37109375, - "learning_rate": 0.00017145362891051084, - "loss": 0.5822, - "step": 4072 - }, - { - "epoch": 3.6820234869015356, - "grad_norm": 0.326171875, - "learning_rate": 0.0001712408116067142, - "loss": 0.5387, - "step": 4076 - }, - { - "epoch": 3.6856368563685638, - "grad_norm": 0.314453125, - "learning_rate": 0.00017102795066251502, - "loss": 0.5367, - "step": 4080 - }, - { - "epoch": 3.6892502258355915, - "grad_norm": 0.353515625, - "learning_rate": 0.00017081504651524773, - "loss": 0.5445, - "step": 4084 - }, - { - "epoch": 3.6928635953026197, - "grad_norm": 0.357421875, - "learning_rate": 0.00017060209960233544, - "loss": 0.5815, - "step": 4088 - }, - { - "epoch": 3.696476964769648, - "grad_norm": 0.33984375, - "learning_rate": 0.0001703891103612892, - "loss": 0.5598, - "step": 4092 - }, - { - "epoch": 3.7000903342366755, - "grad_norm": 0.345703125, - "learning_rate": 0.00017017607922970693, - "loss": 0.5526, - "step": 4096 - }, - { - "epoch": 3.7037037037037037, - "grad_norm": 0.369140625, - "learning_rate": 0.00016996300664527274, - "loss": 0.5901, - "step": 4100 - }, - { - "epoch": 3.7073170731707314, - "grad_norm": 0.333984375, - "learning_rate": 0.00016974989304575583, - "loss": 0.5884, - "step": 4104 - }, - { - "epoch": 3.7109304426377596, - "grad_norm": 0.392578125, - "learning_rate": 0.00016953673886900967, - "loss": 0.6022, - "step": 4108 - }, - { - "epoch": 3.714543812104788, - "grad_norm": 0.3359375, - "learning_rate": 0.00016932354455297113, - "loss": 0.5683, - "step": 4112 - }, - { - "epoch": 3.718157181571816, - "grad_norm": 0.34765625, - "learning_rate": 0.00016911031053565943, - "loss": 0.5815, - "step": 4116 - }, - { - "epoch": 3.7217705510388437, - "grad_norm": 0.341796875, - "learning_rate": 0.00016889703725517553, - "loss": 0.5582, - "step": 4120 - }, - { - "epoch": 3.725383920505872, - "grad_norm": 0.33203125, - "learning_rate": 0.00016868372514970104, - "loss": 0.5383, - "step": 4124 - }, - { - "epoch": 3.7289972899728996, - "grad_norm": 0.333984375, - "learning_rate": 0.00016847037465749713, - "loss": 0.5236, - "step": 4128 - }, - { - "epoch": 3.7326106594399278, - "grad_norm": 0.34375, - "learning_rate": 0.00016825698621690419, - "loss": 0.5677, - "step": 4132 - }, - { - "epoch": 3.736224028906956, - "grad_norm": 0.345703125, - "learning_rate": 0.00016804356026634017, - "loss": 0.5743, - "step": 4136 - }, - { - "epoch": 3.7398373983739837, - "grad_norm": 0.33984375, - "learning_rate": 0.00016783009724430042, - "loss": 0.5377, - "step": 4140 - }, - { - "epoch": 3.743450767841012, - "grad_norm": 0.33203125, - "learning_rate": 0.00016761659758935624, - "loss": 0.5916, - "step": 4144 - }, - { - "epoch": 3.7470641373080396, - "grad_norm": 0.33984375, - "learning_rate": 0.00016740306174015435, - "loss": 0.5635, - "step": 4148 - }, - { - "epoch": 3.7506775067750677, - "grad_norm": 0.365234375, - "learning_rate": 0.00016718949013541578, - "loss": 0.5483, - "step": 4152 - }, - { - "epoch": 3.754290876242096, - "grad_norm": 0.359375, - "learning_rate": 0.00016697588321393494, - "loss": 0.5791, - "step": 4156 - }, - { - "epoch": 3.7579042457091236, - "grad_norm": 0.341796875, - "learning_rate": 0.00016676224141457893, - "loss": 0.6157, - "step": 4160 - }, - { - "epoch": 3.761517615176152, - "grad_norm": 0.353515625, - "learning_rate": 0.0001665485651762864, - "loss": 0.5631, - "step": 4164 - }, - { - "epoch": 3.7651309846431795, - "grad_norm": 0.337890625, - "learning_rate": 0.00016633485493806676, - "loss": 0.5893, - "step": 4168 - }, - { - "epoch": 3.7687443541102077, - "grad_norm": 0.333984375, - "learning_rate": 0.00016612111113899943, - "loss": 0.5555, - "step": 4172 - }, - { - "epoch": 3.772357723577236, - "grad_norm": 0.36328125, - "learning_rate": 0.00016590733421823254, - "loss": 0.5636, - "step": 4176 - }, - { - "epoch": 3.775971093044264, - "grad_norm": 0.330078125, - "learning_rate": 0.0001656935246149825, - "loss": 0.56, - "step": 4180 - }, - { - "epoch": 3.7795844625112918, - "grad_norm": 0.345703125, - "learning_rate": 0.00016547968276853274, - "loss": 0.5431, - "step": 4184 - }, - { - "epoch": 3.78319783197832, - "grad_norm": 0.34765625, - "learning_rate": 0.000165265809118233, - "loss": 0.5402, - "step": 4188 - }, - { - "epoch": 3.7868112014453477, - "grad_norm": 0.345703125, - "learning_rate": 0.00016505190410349817, - "loss": 0.5214, - "step": 4192 - }, - { - "epoch": 3.790424570912376, - "grad_norm": 0.31640625, - "learning_rate": 0.0001648379681638079, - "loss": 0.5794, - "step": 4196 - }, - { - "epoch": 3.794037940379404, - "grad_norm": 0.337890625, - "learning_rate": 0.00016462400173870514, - "loss": 0.579, - "step": 4200 - }, - { - "epoch": 3.7976513098464317, - "grad_norm": 0.3359375, - "learning_rate": 0.00016441000526779556, - "loss": 0.5955, - "step": 4204 - }, - { - "epoch": 3.80126467931346, - "grad_norm": 0.318359375, - "learning_rate": 0.00016419597919074656, - "loss": 0.5505, - "step": 4208 - }, - { - "epoch": 3.8048780487804876, - "grad_norm": 0.35546875, - "learning_rate": 0.0001639819239472864, - "loss": 0.5759, - "step": 4212 - }, - { - "epoch": 3.808491418247516, - "grad_norm": 0.34765625, - "learning_rate": 0.0001637678399772031, - "loss": 0.5646, - "step": 4216 - }, - { - "epoch": 3.812104787714544, - "grad_norm": 0.365234375, - "learning_rate": 0.00016355372772034388, - "loss": 0.5929, - "step": 4220 - }, - { - "epoch": 3.8157181571815717, - "grad_norm": 0.375, - "learning_rate": 0.00016333958761661398, - "loss": 0.5522, - "step": 4224 - }, - { - "epoch": 3.8193315266486, - "grad_norm": 0.3359375, - "learning_rate": 0.00016312542010597593, - "loss": 0.5343, - "step": 4228 - }, - { - "epoch": 3.8229448961156276, - "grad_norm": 0.3515625, - "learning_rate": 0.00016291122562844855, - "loss": 0.5524, - "step": 4232 - }, - { - "epoch": 3.8265582655826558, - "grad_norm": 0.365234375, - "learning_rate": 0.00016269700462410603, - "loss": 0.539, - "step": 4236 - }, - { - "epoch": 3.830171635049684, - "grad_norm": 0.34765625, - "learning_rate": 0.00016248275753307702, - "loss": 0.5513, - "step": 4240 - }, - { - "epoch": 3.8337850045167117, - "grad_norm": 0.34765625, - "learning_rate": 0.00016226848479554382, - "loss": 0.5407, - "step": 4244 - }, - { - "epoch": 3.83739837398374, - "grad_norm": 0.337890625, - "learning_rate": 0.0001620541868517415, - "loss": 0.5228, - "step": 4248 - }, - { - "epoch": 3.8410117434507676, - "grad_norm": 0.33984375, - "learning_rate": 0.00016183986414195677, - "loss": 0.5958, - "step": 4252 - }, - { - "epoch": 3.8446251129177957, - "grad_norm": 0.337890625, - "learning_rate": 0.00016162551710652734, - "loss": 0.5598, - "step": 4256 - }, - { - "epoch": 3.848238482384824, - "grad_norm": 0.330078125, - "learning_rate": 0.0001614111461858408, - "loss": 0.5138, - "step": 4260 - }, - { - "epoch": 3.851851851851852, - "grad_norm": 0.341796875, - "learning_rate": 0.00016119675182033391, - "loss": 0.5419, - "step": 4264 - }, - { - "epoch": 3.85546522131888, - "grad_norm": 0.373046875, - "learning_rate": 0.0001609823344504915, - "loss": 0.569, - "step": 4268 - }, - { - "epoch": 3.859078590785908, - "grad_norm": 0.32421875, - "learning_rate": 0.00016076789451684574, - "loss": 0.549, - "step": 4272 - }, - { - "epoch": 3.8626919602529357, - "grad_norm": 0.365234375, - "learning_rate": 0.00016055343245997518, - "loss": 0.5565, - "step": 4276 - }, - { - "epoch": 3.866305329719964, - "grad_norm": 0.34765625, - "learning_rate": 0.00016033894872050368, - "loss": 0.5619, - "step": 4280 - }, - { - "epoch": 3.869918699186992, - "grad_norm": 0.361328125, - "learning_rate": 0.00016012444373909987, - "loss": 0.5603, - "step": 4284 - }, - { - "epoch": 3.8735320686540198, - "grad_norm": 0.337890625, - "learning_rate": 0.00015990991795647582, - "loss": 0.5487, - "step": 4288 - }, - { - "epoch": 3.877145438121048, - "grad_norm": 0.369140625, - "learning_rate": 0.0001596953718133864, - "loss": 0.5858, - "step": 4292 - }, - { - "epoch": 3.8807588075880757, - "grad_norm": 0.3515625, - "learning_rate": 0.0001594808057506284, - "loss": 0.5938, - "step": 4296 - }, - { - "epoch": 3.884372177055104, - "grad_norm": 0.361328125, - "learning_rate": 0.0001592662202090394, - "loss": 0.546, - "step": 4300 - }, - { - "epoch": 3.887985546522132, - "grad_norm": 0.3359375, - "learning_rate": 0.00015905161562949712, - "loss": 0.5365, - "step": 4304 - }, - { - "epoch": 3.8915989159891597, - "grad_norm": 0.3671875, - "learning_rate": 0.00015883699245291822, - "loss": 0.6163, - "step": 4308 - }, - { - "epoch": 3.895212285456188, - "grad_norm": 0.337890625, - "learning_rate": 0.0001586223511202579, - "loss": 0.5412, - "step": 4312 - }, - { - "epoch": 3.8988256549232156, - "grad_norm": 0.349609375, - "learning_rate": 0.00015840769207250825, - "loss": 0.5812, - "step": 4316 - }, - { - "epoch": 3.902439024390244, - "grad_norm": 0.349609375, - "learning_rate": 0.00015819301575069802, - "loss": 0.5649, - "step": 4320 - }, - { - "epoch": 3.906052393857272, - "grad_norm": 0.369140625, - "learning_rate": 0.0001579783225958914, - "loss": 0.5503, - "step": 4324 - }, - { - "epoch": 3.9096657633243, - "grad_norm": 0.35546875, - "learning_rate": 0.00015776361304918715, - "loss": 0.5589, - "step": 4328 - }, - { - "epoch": 3.913279132791328, - "grad_norm": 0.349609375, - "learning_rate": 0.00015754888755171773, - "loss": 0.6127, - "step": 4332 - }, - { - "epoch": 3.916892502258356, - "grad_norm": 0.369140625, - "learning_rate": 0.00015733414654464835, - "loss": 0.6025, - "step": 4336 - }, - { - "epoch": 3.920505871725384, - "grad_norm": 0.322265625, - "learning_rate": 0.00015711939046917606, - "loss": 0.5519, - "step": 4340 - }, - { - "epoch": 3.924119241192412, - "grad_norm": 0.34765625, - "learning_rate": 0.00015690461976652884, - "loss": 0.613, - "step": 4344 - }, - { - "epoch": 3.92773261065944, - "grad_norm": 0.341796875, - "learning_rate": 0.00015668983487796486, - "loss": 0.5976, - "step": 4348 - }, - { - "epoch": 3.931345980126468, - "grad_norm": 0.326171875, - "learning_rate": 0.00015647503624477137, - "loss": 0.563, - "step": 4352 - }, - { - "epoch": 3.934959349593496, - "grad_norm": 0.333984375, - "learning_rate": 0.00015626022430826375, - "loss": 0.5501, - "step": 4356 - }, - { - "epoch": 3.9385727190605238, - "grad_norm": 0.33984375, - "learning_rate": 0.0001560453995097849, - "loss": 0.5709, - "step": 4360 - }, - { - "epoch": 3.942186088527552, - "grad_norm": 0.35546875, - "learning_rate": 0.00015583056229070405, - "loss": 0.5044, - "step": 4364 - }, - { - "epoch": 3.94579945799458, - "grad_norm": 0.3515625, - "learning_rate": 0.00015561571309241585, - "loss": 0.5778, - "step": 4368 - }, - { - "epoch": 3.949412827461608, - "grad_norm": 0.328125, - "learning_rate": 0.00015540085235633973, - "loss": 0.5547, - "step": 4372 - }, - { - "epoch": 3.953026196928636, - "grad_norm": 0.35546875, - "learning_rate": 0.00015518598052391874, - "loss": 0.5934, - "step": 4376 - }, - { - "epoch": 3.9566395663956637, - "grad_norm": 0.365234375, - "learning_rate": 0.0001549710980366188, - "loss": 0.5204, - "step": 4380 - }, - { - "epoch": 3.960252935862692, - "grad_norm": 0.380859375, - "learning_rate": 0.00015475620533592758, - "loss": 0.5592, - "step": 4384 - }, - { - "epoch": 3.96386630532972, - "grad_norm": 0.333984375, - "learning_rate": 0.0001545413028633539, - "loss": 0.5684, - "step": 4388 - }, - { - "epoch": 3.9674796747967482, - "grad_norm": 0.34375, - "learning_rate": 0.00015432639106042646, - "loss": 0.5568, - "step": 4392 - }, - { - "epoch": 3.971093044263776, - "grad_norm": 0.337890625, - "learning_rate": 0.00015411147036869334, - "loss": 0.5771, - "step": 4396 - }, - { - "epoch": 3.974706413730804, - "grad_norm": 0.361328125, - "learning_rate": 0.0001538965412297208, - "loss": 0.5472, - "step": 4400 - }, - { - "epoch": 3.978319783197832, - "grad_norm": 0.3515625, - "learning_rate": 0.00015368160408509238, - "loss": 0.5926, - "step": 4404 - }, - { - "epoch": 3.98193315266486, - "grad_norm": 0.345703125, - "learning_rate": 0.00015346665937640818, - "loss": 0.5737, - "step": 4408 - }, - { - "epoch": 3.985546522131888, - "grad_norm": 0.333984375, - "learning_rate": 0.00015325170754528376, - "loss": 0.4884, - "step": 4412 - }, - { - "epoch": 3.989159891598916, - "grad_norm": 0.380859375, - "learning_rate": 0.0001530367490333494, - "loss": 0.5861, - "step": 4416 - }, - { - "epoch": 3.992773261065944, - "grad_norm": 0.3515625, - "learning_rate": 0.000152821784282249, - "loss": 0.5417, - "step": 4420 - }, - { - "epoch": 3.996386630532972, - "grad_norm": 0.337890625, - "learning_rate": 0.00015260681373363932, - "loss": 0.5862, - "step": 4424 - }, - { - "epoch": 4.0, - "grad_norm": 0.50390625, - "learning_rate": 0.0001523918378291891, - "loss": 0.5813, - "step": 4428 - }, - { - "epoch": 4.003613369467028, - "grad_norm": 0.32421875, - "learning_rate": 0.00015217685701057792, - "loss": 0.5454, - "step": 4432 - }, - { - "epoch": 4.007226738934056, - "grad_norm": 0.310546875, - "learning_rate": 0.00015196187171949573, - "loss": 0.5102, - "step": 4436 - }, - { - "epoch": 4.010840108401084, - "grad_norm": 0.35546875, - "learning_rate": 0.00015174688239764136, - "loss": 0.4544, - "step": 4440 - }, - { - "epoch": 4.014453477868112, - "grad_norm": 0.34765625, - "learning_rate": 0.0001515318894867221, - "loss": 0.4868, - "step": 4444 - }, - { - "epoch": 4.01806684733514, - "grad_norm": 0.375, - "learning_rate": 0.0001513168934284526, - "loss": 0.486, - "step": 4448 - }, - { - "epoch": 4.021680216802168, - "grad_norm": 0.353515625, - "learning_rate": 0.00015110189466455395, - "loss": 0.5037, - "step": 4452 - }, - { - "epoch": 4.025293586269196, - "grad_norm": 0.341796875, - "learning_rate": 0.0001508868936367528, - "loss": 0.4714, - "step": 4456 - }, - { - "epoch": 4.028906955736224, - "grad_norm": 0.36328125, - "learning_rate": 0.0001506718907867804, - "loss": 0.4837, - "step": 4460 - }, - { - "epoch": 4.032520325203252, - "grad_norm": 0.330078125, - "learning_rate": 0.0001504568865563719, - "loss": 0.5177, - "step": 4464 - }, - { - "epoch": 4.03613369467028, - "grad_norm": 0.369140625, - "learning_rate": 0.00015024188138726508, - "loss": 0.4873, - "step": 4468 - }, - { - "epoch": 4.039747064137308, - "grad_norm": 0.328125, - "learning_rate": 0.00015002687572119978, - "loss": 0.4903, - "step": 4472 - }, - { - "epoch": 4.043360433604336, - "grad_norm": 0.359375, - "learning_rate": 0.00014981186999991694, - "loss": 0.5045, - "step": 4476 - }, - { - "epoch": 4.0469738030713645, - "grad_norm": 0.349609375, - "learning_rate": 0.00014959686466515729, - "loss": 0.508, - "step": 4480 - }, - { - "epoch": 4.050587172538392, - "grad_norm": 0.34375, - "learning_rate": 0.0001493818601586612, - "loss": 0.4853, - "step": 4484 - }, - { - "epoch": 4.05420054200542, - "grad_norm": 0.3359375, - "learning_rate": 0.00014916685692216696, - "loss": 0.4733, - "step": 4488 - }, - { - "epoch": 4.057813911472448, - "grad_norm": 0.357421875, - "learning_rate": 0.00014895185539741054, - "loss": 0.4786, - "step": 4492 - }, - { - "epoch": 4.061427280939476, - "grad_norm": 0.333984375, - "learning_rate": 0.00014873685602612415, - "loss": 0.4783, - "step": 4496 - }, - { - "epoch": 4.065040650406504, - "grad_norm": 0.369140625, - "learning_rate": 0.00014852185925003577, - "loss": 0.4978, - "step": 4500 - }, - { - "epoch": 4.068654019873532, - "grad_norm": 0.341796875, - "learning_rate": 0.000148306865510868, - "loss": 0.5434, - "step": 4504 - }, - { - "epoch": 4.07226738934056, - "grad_norm": 0.34375, - "learning_rate": 0.000148091875250337, - "loss": 0.4549, - "step": 4508 - }, - { - "epoch": 4.075880758807588, - "grad_norm": 0.36328125, - "learning_rate": 0.00014787688891015218, - "loss": 0.4809, - "step": 4512 - }, - { - "epoch": 4.079494128274616, - "grad_norm": 0.328125, - "learning_rate": 0.00014766190693201443, - "loss": 0.4999, - "step": 4516 - }, - { - "epoch": 4.083107497741644, - "grad_norm": 0.345703125, - "learning_rate": 0.0001474469297576161, - "loss": 0.48, - "step": 4520 - }, - { - "epoch": 4.086720867208672, - "grad_norm": 0.361328125, - "learning_rate": 0.00014723195782863937, - "loss": 0.491, - "step": 4524 - }, - { - "epoch": 4.0903342366757, - "grad_norm": 0.3671875, - "learning_rate": 0.00014701699158675588, - "loss": 0.4985, - "step": 4528 - }, - { - "epoch": 4.093947606142728, - "grad_norm": 0.341796875, - "learning_rate": 0.0001468020314736254, - "loss": 0.466, - "step": 4532 - }, - { - "epoch": 4.097560975609756, - "grad_norm": 0.3359375, - "learning_rate": 0.00014658707793089505, - "loss": 0.5192, - "step": 4536 - }, - { - "epoch": 4.101174345076784, - "grad_norm": 0.35546875, - "learning_rate": 0.00014637213140019877, - "loss": 0.4989, - "step": 4540 - }, - { - "epoch": 4.1047877145438125, - "grad_norm": 0.37109375, - "learning_rate": 0.00014615719232315575, - "loss": 0.4902, - "step": 4544 - }, - { - "epoch": 4.10840108401084, - "grad_norm": 0.3515625, - "learning_rate": 0.00014594226114137014, - "loss": 0.5096, - "step": 4548 - }, - { - "epoch": 4.112014453477868, - "grad_norm": 0.349609375, - "learning_rate": 0.00014572733829642962, - "loss": 0.4781, - "step": 4552 - }, - { - "epoch": 4.115627822944896, - "grad_norm": 0.39453125, - "learning_rate": 0.00014551242422990496, - "loss": 0.4662, - "step": 4556 - }, - { - "epoch": 4.119241192411924, - "grad_norm": 0.34765625, - "learning_rate": 0.00014529751938334876, - "loss": 0.4736, - "step": 4560 - }, - { - "epoch": 4.1228545618789525, - "grad_norm": 0.328125, - "learning_rate": 0.00014508262419829463, - "loss": 0.4831, - "step": 4564 - }, - { - "epoch": 4.12646793134598, - "grad_norm": 0.3671875, - "learning_rate": 0.00014486773911625655, - "loss": 0.4733, - "step": 4568 - }, - { - "epoch": 4.130081300813008, - "grad_norm": 0.3359375, - "learning_rate": 0.0001446528645787275, - "loss": 0.5184, - "step": 4572 - }, - { - "epoch": 4.133694670280036, - "grad_norm": 0.392578125, - "learning_rate": 0.000144438001027179, - "loss": 0.5103, - "step": 4576 - }, - { - "epoch": 4.137308039747064, - "grad_norm": 0.361328125, - "learning_rate": 0.0001442231489030599, - "loss": 0.5411, - "step": 4580 - }, - { - "epoch": 4.1409214092140925, - "grad_norm": 0.359375, - "learning_rate": 0.00014400830864779539, - "loss": 0.4805, - "step": 4584 - }, - { - "epoch": 4.14453477868112, - "grad_norm": 0.365234375, - "learning_rate": 0.00014379348070278668, - "loss": 0.5105, - "step": 4588 - }, - { - "epoch": 4.148148148148148, - "grad_norm": 0.337890625, - "learning_rate": 0.0001435786655094093, - "loss": 0.4928, - "step": 4592 - }, - { - "epoch": 4.151761517615176, - "grad_norm": 0.3515625, - "learning_rate": 0.00014336386350901289, - "loss": 0.4993, - "step": 4596 - }, - { - "epoch": 4.155374887082204, - "grad_norm": 0.328125, - "learning_rate": 0.0001431490751429197, - "loss": 0.4553, - "step": 4600 - }, - { - "epoch": 4.158988256549232, - "grad_norm": 0.37890625, - "learning_rate": 0.00014293430085242424, - "loss": 0.4709, - "step": 4604 - }, - { - "epoch": 4.16260162601626, - "grad_norm": 0.37109375, - "learning_rate": 0.00014271954107879196, - "loss": 0.5244, - "step": 4608 - }, - { - "epoch": 4.166214995483288, - "grad_norm": 0.369140625, - "learning_rate": 0.00014250479626325832, - "loss": 0.4795, - "step": 4612 - }, - { - "epoch": 4.169828364950316, - "grad_norm": 0.37109375, - "learning_rate": 0.0001422900668470284, - "loss": 0.5038, - "step": 4616 - }, - { - "epoch": 4.173441734417344, - "grad_norm": 0.357421875, - "learning_rate": 0.0001420753532712754, - "loss": 0.4542, - "step": 4620 - }, - { - "epoch": 4.177055103884372, - "grad_norm": 0.3515625, - "learning_rate": 0.00014186065597714004, - "loss": 0.5415, - "step": 4624 - }, - { - "epoch": 4.180668473351401, - "grad_norm": 0.345703125, - "learning_rate": 0.00014164597540572958, - "loss": 0.4913, - "step": 4628 - }, - { - "epoch": 4.184281842818428, - "grad_norm": 0.375, - "learning_rate": 0.00014143131199811695, - "loss": 0.507, - "step": 4632 - }, - { - "epoch": 4.187895212285456, - "grad_norm": 0.359375, - "learning_rate": 0.00014121666619533986, - "loss": 0.4978, - "step": 4636 - }, - { - "epoch": 4.191508581752484, - "grad_norm": 0.36328125, - "learning_rate": 0.00014100203843839956, - "loss": 0.4906, - "step": 4640 - }, - { - "epoch": 4.195121951219512, - "grad_norm": 0.333984375, - "learning_rate": 0.00014078742916826068, - "loss": 0.4673, - "step": 4644 - }, - { - "epoch": 4.1987353206865405, - "grad_norm": 0.361328125, - "learning_rate": 0.00014057283882584951, - "loss": 0.5296, - "step": 4648 - }, - { - "epoch": 4.202348690153568, - "grad_norm": 0.35546875, - "learning_rate": 0.00014035826785205365, - "loss": 0.4449, - "step": 4652 - }, - { - "epoch": 4.205962059620596, - "grad_norm": 0.359375, - "learning_rate": 0.00014014371668772085, - "loss": 0.4812, - "step": 4656 - }, - { - "epoch": 4.209575429087624, - "grad_norm": 0.33984375, - "learning_rate": 0.000139929185773658, - "loss": 0.506, - "step": 4660 - }, - { - "epoch": 4.213188798554652, - "grad_norm": 0.35546875, - "learning_rate": 0.00013971467555063064, - "loss": 0.4995, - "step": 4664 - }, - { - "epoch": 4.2168021680216805, - "grad_norm": 0.3671875, - "learning_rate": 0.0001395001864593617, - "loss": 0.4825, - "step": 4668 - }, - { - "epoch": 4.220415537488708, - "grad_norm": 0.361328125, - "learning_rate": 0.00013928571894053065, - "loss": 0.5069, - "step": 4672 - }, - { - "epoch": 4.224028906955736, - "grad_norm": 0.33984375, - "learning_rate": 0.00013907127343477265, - "loss": 0.5236, - "step": 4676 - }, - { - "epoch": 4.227642276422764, - "grad_norm": 0.357421875, - "learning_rate": 0.00013885685038267774, - "loss": 0.4994, - "step": 4680 - }, - { - "epoch": 4.231255645889792, - "grad_norm": 0.353515625, - "learning_rate": 0.00013864245022478968, - "loss": 0.4814, - "step": 4684 - }, - { - "epoch": 4.2348690153568205, - "grad_norm": 0.34375, - "learning_rate": 0.00013842807340160514, - "loss": 0.4836, - "step": 4688 - }, - { - "epoch": 4.238482384823849, - "grad_norm": 0.35546875, - "learning_rate": 0.00013821372035357316, - "loss": 0.4885, - "step": 4692 - }, - { - "epoch": 4.242095754290876, - "grad_norm": 0.349609375, - "learning_rate": 0.00013799939152109363, - "loss": 0.4962, - "step": 4696 - }, - { - "epoch": 4.245709123757904, - "grad_norm": 0.384765625, - "learning_rate": 0.00013778508734451682, - "loss": 0.4996, - "step": 4700 - }, - { - "epoch": 4.249322493224932, - "grad_norm": 0.3828125, - "learning_rate": 0.00013757080826414227, - "loss": 0.4894, - "step": 4704 - }, - { - "epoch": 4.2529358626919604, - "grad_norm": 0.349609375, - "learning_rate": 0.00013735655472021807, - "loss": 0.4748, - "step": 4708 - }, - { - "epoch": 4.256549232158989, - "grad_norm": 0.36328125, - "learning_rate": 0.00013714232715293978, - "loss": 0.4613, - "step": 4712 - }, - { - "epoch": 4.260162601626016, - "grad_norm": 0.35546875, - "learning_rate": 0.00013692812600244944, - "loss": 0.532, - "step": 4716 - }, - { - "epoch": 4.263775971093044, - "grad_norm": 0.365234375, - "learning_rate": 0.00013671395170883514, - "loss": 0.54, - "step": 4720 - }, - { - "epoch": 4.267389340560072, - "grad_norm": 0.376953125, - "learning_rate": 0.00013649980471212953, - "loss": 0.4708, - "step": 4724 - }, - { - "epoch": 4.2710027100271, - "grad_norm": 0.35546875, - "learning_rate": 0.00013628568545230927, - "loss": 0.4768, - "step": 4728 - }, - { - "epoch": 4.274616079494129, - "grad_norm": 0.3671875, - "learning_rate": 0.0001360715943692941, - "loss": 0.5263, - "step": 4732 - }, - { - "epoch": 4.278229448961156, - "grad_norm": 0.3828125, - "learning_rate": 0.0001358575319029456, - "loss": 0.4815, - "step": 4736 - }, - { - "epoch": 4.281842818428184, - "grad_norm": 0.373046875, - "learning_rate": 0.00013564349849306694, - "loss": 0.5291, - "step": 4740 - }, - { - "epoch": 4.285456187895212, - "grad_norm": 0.353515625, - "learning_rate": 0.00013542949457940126, - "loss": 0.4996, - "step": 4744 - }, - { - "epoch": 4.28906955736224, - "grad_norm": 0.33984375, - "learning_rate": 0.00013521552060163138, - "loss": 0.5285, - "step": 4748 - }, - { - "epoch": 4.2926829268292686, - "grad_norm": 0.36328125, - "learning_rate": 0.00013500157699937836, - "loss": 0.4899, - "step": 4752 - }, - { - "epoch": 4.296296296296296, - "grad_norm": 0.353515625, - "learning_rate": 0.000134787664212201, - "loss": 0.5122, - "step": 4756 - }, - { - "epoch": 4.299909665763324, - "grad_norm": 0.353515625, - "learning_rate": 0.00013457378267959483, - "loss": 0.5029, - "step": 4760 - }, - { - "epoch": 4.303523035230352, - "grad_norm": 0.37890625, - "learning_rate": 0.00013435993284099091, - "loss": 0.4732, - "step": 4764 - }, - { - "epoch": 4.30713640469738, - "grad_norm": 0.375, - "learning_rate": 0.00013414611513575553, - "loss": 0.5427, - "step": 4768 - }, - { - "epoch": 4.3107497741644085, - "grad_norm": 0.3671875, - "learning_rate": 0.0001339323300031887, - "loss": 0.504, - "step": 4772 - }, - { - "epoch": 4.314363143631437, - "grad_norm": 0.353515625, - "learning_rate": 0.00013371857788252366, - "loss": 0.4896, - "step": 4776 - }, - { - "epoch": 4.317976513098464, - "grad_norm": 0.38671875, - "learning_rate": 0.0001335048592129257, - "loss": 0.4816, - "step": 4780 - }, - { - "epoch": 4.321589882565492, - "grad_norm": 0.369140625, - "learning_rate": 0.00013329117443349153, - "loss": 0.4952, - "step": 4784 - }, - { - "epoch": 4.32520325203252, - "grad_norm": 0.337890625, - "learning_rate": 0.00013307752398324807, - "loss": 0.452, - "step": 4788 - }, - { - "epoch": 4.3288166214995485, - "grad_norm": 0.373046875, - "learning_rate": 0.0001328639083011518, - "loss": 0.5189, - "step": 4792 - }, - { - "epoch": 4.332429990966577, - "grad_norm": 0.35546875, - "learning_rate": 0.00013265032782608786, - "loss": 0.4767, - "step": 4796 - }, - { - "epoch": 4.336043360433604, - "grad_norm": 0.357421875, - "learning_rate": 0.0001324367829968688, - "loss": 0.4891, - "step": 4800 - }, - { - "epoch": 4.339656729900632, - "grad_norm": 0.361328125, - "learning_rate": 0.0001322232742522342, - "loss": 0.4857, - "step": 4804 - }, - { - "epoch": 4.34327009936766, - "grad_norm": 0.33984375, - "learning_rate": 0.00013200980203084936, - "loss": 0.5197, - "step": 4808 - }, - { - "epoch": 4.3468834688346885, - "grad_norm": 0.3671875, - "learning_rate": 0.00013179636677130446, - "loss": 0.4855, - "step": 4812 - }, - { - "epoch": 4.350496838301717, - "grad_norm": 0.357421875, - "learning_rate": 0.000131582968912114, - "loss": 0.5269, - "step": 4816 - }, - { - "epoch": 4.354110207768745, - "grad_norm": 0.357421875, - "learning_rate": 0.00013136960889171536, - "loss": 0.483, - "step": 4820 - }, - { - "epoch": 4.357723577235772, - "grad_norm": 0.3515625, - "learning_rate": 0.00013115628714846842, - "loss": 0.4785, - "step": 4824 - }, - { - "epoch": 4.3613369467028, - "grad_norm": 0.373046875, - "learning_rate": 0.00013094300412065424, - "loss": 0.4587, - "step": 4828 - }, - { - "epoch": 4.364950316169828, - "grad_norm": 0.37890625, - "learning_rate": 0.0001307297602464744, - "loss": 0.5134, - "step": 4832 - }, - { - "epoch": 4.368563685636857, - "grad_norm": 0.37109375, - "learning_rate": 0.00013051655596405006, - "loss": 0.5088, - "step": 4836 - }, - { - "epoch": 4.372177055103885, - "grad_norm": 0.375, - "learning_rate": 0.00013030339171142094, - "loss": 0.5073, - "step": 4840 - }, - { - "epoch": 4.375790424570912, - "grad_norm": 0.353515625, - "learning_rate": 0.0001300902679265447, - "loss": 0.5079, - "step": 4844 - }, - { - "epoch": 4.37940379403794, - "grad_norm": 0.390625, - "learning_rate": 0.00012987718504729563, - "loss": 0.5293, - "step": 4848 - }, - { - "epoch": 4.383017163504968, - "grad_norm": 0.380859375, - "learning_rate": 0.0001296641435114642, - "loss": 0.4757, - "step": 4852 - }, - { - "epoch": 4.386630532971997, - "grad_norm": 0.380859375, - "learning_rate": 0.00012945114375675573, - "loss": 0.4939, - "step": 4856 - }, - { - "epoch": 4.390243902439025, - "grad_norm": 0.365234375, - "learning_rate": 0.0001292381862207899, - "loss": 0.493, - "step": 4860 - }, - { - "epoch": 4.393857271906052, - "grad_norm": 0.365234375, - "learning_rate": 0.0001290252713410995, - "loss": 0.5238, - "step": 4864 - }, - { - "epoch": 4.39747064137308, - "grad_norm": 0.37109375, - "learning_rate": 0.0001288123995551297, - "loss": 0.5016, - "step": 4868 - }, - { - "epoch": 4.401084010840108, - "grad_norm": 0.373046875, - "learning_rate": 0.0001285995713002372, - "loss": 0.4808, - "step": 4872 - }, - { - "epoch": 4.4046973803071365, - "grad_norm": 0.376953125, - "learning_rate": 0.00012838678701368916, - "loss": 0.4892, - "step": 4876 - }, - { - "epoch": 4.408310749774165, - "grad_norm": 0.359375, - "learning_rate": 0.0001281740471326626, - "loss": 0.4923, - "step": 4880 - }, - { - "epoch": 4.411924119241192, - "grad_norm": 0.373046875, - "learning_rate": 0.0001279613520942431, - "loss": 0.4856, - "step": 4884 - }, - { - "epoch": 4.41553748870822, - "grad_norm": 0.365234375, - "learning_rate": 0.00012774870233542407, - "loss": 0.5078, - "step": 4888 - }, - { - "epoch": 4.419150858175248, - "grad_norm": 0.359375, - "learning_rate": 0.00012753609829310616, - "loss": 0.4861, - "step": 4892 - }, - { - "epoch": 4.4227642276422765, - "grad_norm": 0.345703125, - "learning_rate": 0.0001273235404040959, - "loss": 0.5044, - "step": 4896 - }, - { - "epoch": 4.426377597109305, - "grad_norm": 0.359375, - "learning_rate": 0.000127111029105105, - "loss": 0.5363, - "step": 4900 - }, - { - "epoch": 4.429990966576333, - "grad_norm": 0.3515625, - "learning_rate": 0.0001268985648327495, - "loss": 0.4659, - "step": 4904 - }, - { - "epoch": 4.43360433604336, - "grad_norm": 0.400390625, - "learning_rate": 0.0001266861480235488, - "loss": 0.5122, - "step": 4908 - }, - { - "epoch": 4.437217705510388, - "grad_norm": 0.359375, - "learning_rate": 0.0001264737791139248, - "loss": 0.5029, - "step": 4912 - }, - { - "epoch": 4.4408310749774165, - "grad_norm": 0.369140625, - "learning_rate": 0.00012626145854020092, - "loss": 0.4561, - "step": 4916 - }, - { - "epoch": 4.444444444444445, - "grad_norm": 0.40234375, - "learning_rate": 0.0001260491867386014, - "loss": 0.504, - "step": 4920 - }, - { - "epoch": 4.448057813911473, - "grad_norm": 0.375, - "learning_rate": 0.00012583696414525012, - "loss": 0.5015, - "step": 4924 - }, - { - "epoch": 4.4516711833785, - "grad_norm": 0.3515625, - "learning_rate": 0.00012562479119617, - "loss": 0.4928, - "step": 4928 - }, - { - "epoch": 4.455284552845528, - "grad_norm": 0.365234375, - "learning_rate": 0.00012541266832728187, - "loss": 0.4672, - "step": 4932 - }, - { - "epoch": 4.458897922312556, - "grad_norm": 0.3671875, - "learning_rate": 0.00012520059597440373, - "loss": 0.4683, - "step": 4936 - }, - { - "epoch": 4.462511291779585, - "grad_norm": 0.3828125, - "learning_rate": 0.00012498857457324974, - "loss": 0.4727, - "step": 4940 - }, - { - "epoch": 4.466124661246613, - "grad_norm": 0.36328125, - "learning_rate": 0.0001247766045594294, - "loss": 0.5256, - "step": 4944 - }, - { - "epoch": 4.46973803071364, - "grad_norm": 0.375, - "learning_rate": 0.00012456468636844664, - "loss": 0.485, - "step": 4948 - }, - { - "epoch": 4.473351400180668, - "grad_norm": 0.390625, - "learning_rate": 0.00012435282043569883, - "loss": 0.483, - "step": 4952 - }, - { - "epoch": 4.476964769647696, - "grad_norm": 0.35546875, - "learning_rate": 0.0001241410071964762, - "loss": 0.4951, - "step": 4956 - }, - { - "epoch": 4.480578139114725, - "grad_norm": 0.361328125, - "learning_rate": 0.0001239292470859605, - "loss": 0.4883, - "step": 4960 - }, - { - "epoch": 4.484191508581753, - "grad_norm": 0.37109375, - "learning_rate": 0.00012371754053922427, - "loss": 0.4958, - "step": 4964 - }, - { - "epoch": 4.487804878048781, - "grad_norm": 0.35546875, - "learning_rate": 0.00012350588799123032, - "loss": 0.488, - "step": 4968 - }, - { - "epoch": 4.491418247515808, - "grad_norm": 0.369140625, - "learning_rate": 0.0001232942898768302, - "loss": 0.5583, - "step": 4972 - }, - { - "epoch": 4.495031616982836, - "grad_norm": 0.35546875, - "learning_rate": 0.00012308274663076378, - "loss": 0.4998, - "step": 4976 - }, - { - "epoch": 4.4986449864498645, - "grad_norm": 0.35546875, - "learning_rate": 0.00012287125868765814, - "loss": 0.4579, - "step": 4980 - }, - { - "epoch": 4.502258355916893, - "grad_norm": 0.365234375, - "learning_rate": 0.0001226598264820268, - "loss": 0.5235, - "step": 4984 - }, - { - "epoch": 4.505871725383921, - "grad_norm": 0.365234375, - "learning_rate": 0.0001224484504482687, - "loss": 0.5149, - "step": 4988 - }, - { - "epoch": 4.509485094850948, - "grad_norm": 0.35546875, - "learning_rate": 0.00012223713102066738, - "loss": 0.5022, - "step": 4992 - }, - { - "epoch": 4.513098464317976, - "grad_norm": 0.365234375, - "learning_rate": 0.00012202586863339012, - "loss": 0.5239, - "step": 4996 - }, - { - "epoch": 4.5167118337850045, - "grad_norm": 0.357421875, - "learning_rate": 0.00012181466372048696, - "loss": 0.4847, - "step": 5000 - }, - { - "epoch": 4.520325203252033, - "grad_norm": 0.35546875, - "learning_rate": 0.00012160351671588992, - "loss": 0.4884, - "step": 5004 - }, - { - "epoch": 4.523938572719061, - "grad_norm": 0.373046875, - "learning_rate": 0.00012139242805341194, - "loss": 0.5007, - "step": 5008 - }, - { - "epoch": 4.527551942186088, - "grad_norm": 0.3828125, - "learning_rate": 0.00012118139816674621, - "loss": 0.502, - "step": 5012 - }, - { - "epoch": 4.531165311653116, - "grad_norm": 0.37109375, - "learning_rate": 0.00012097042748946512, - "loss": 0.488, - "step": 5016 - }, - { - "epoch": 4.5347786811201445, - "grad_norm": 0.408203125, - "learning_rate": 0.00012075951645501932, - "loss": 0.5187, - "step": 5020 - }, - { - "epoch": 4.538392050587173, - "grad_norm": 0.359375, - "learning_rate": 0.00012054866549673708, - "loss": 0.5105, - "step": 5024 - }, - { - "epoch": 4.542005420054201, - "grad_norm": 0.369140625, - "learning_rate": 0.00012033787504782311, - "loss": 0.5026, - "step": 5028 - }, - { - "epoch": 4.545618789521228, - "grad_norm": 0.369140625, - "learning_rate": 0.00012012714554135788, - "loss": 0.5304, - "step": 5032 - }, - { - "epoch": 4.549232158988256, - "grad_norm": 0.41015625, - "learning_rate": 0.0001199164774102966, - "loss": 0.5012, - "step": 5036 - }, - { - "epoch": 4.5528455284552845, - "grad_norm": 0.3515625, - "learning_rate": 0.00011970587108746841, - "loss": 0.5098, - "step": 5040 - }, - { - "epoch": 4.556458897922313, - "grad_norm": 0.3671875, - "learning_rate": 0.00011949532700557547, - "loss": 0.4814, - "step": 5044 - }, - { - "epoch": 4.560072267389341, - "grad_norm": 0.373046875, - "learning_rate": 0.000119284845597192, - "loss": 0.5322, - "step": 5048 - }, - { - "epoch": 4.563685636856368, - "grad_norm": 0.3828125, - "learning_rate": 0.00011907442729476354, - "loss": 0.4925, - "step": 5052 - }, - { - "epoch": 4.567299006323396, - "grad_norm": 0.349609375, - "learning_rate": 0.00011886407253060592, - "loss": 0.4805, - "step": 5056 - }, - { - "epoch": 4.570912375790424, - "grad_norm": 0.37890625, - "learning_rate": 0.00011865378173690446, - "loss": 0.4974, - "step": 5060 - }, - { - "epoch": 4.574525745257453, - "grad_norm": 0.36328125, - "learning_rate": 0.00011844355534571302, - "loss": 0.5354, - "step": 5064 - }, - { - "epoch": 4.578139114724481, - "grad_norm": 0.375, - "learning_rate": 0.00011823339378895312, - "loss": 0.4368, - "step": 5068 - }, - { - "epoch": 4.581752484191509, - "grad_norm": 0.3515625, - "learning_rate": 0.00011802329749841316, - "loss": 0.4979, - "step": 5072 - }, - { - "epoch": 4.585365853658536, - "grad_norm": 0.41796875, - "learning_rate": 0.00011781326690574733, - "loss": 0.5221, - "step": 5076 - }, - { - "epoch": 4.588979223125564, - "grad_norm": 0.373046875, - "learning_rate": 0.000117603302442475, - "loss": 0.4517, - "step": 5080 - }, - { - "epoch": 4.592592592592593, - "grad_norm": 0.365234375, - "learning_rate": 0.00011739340453997946, - "loss": 0.5323, - "step": 5084 - }, - { - "epoch": 4.596205962059621, - "grad_norm": 0.35546875, - "learning_rate": 0.00011718357362950747, - "loss": 0.4959, - "step": 5088 - }, - { - "epoch": 4.599819331526649, - "grad_norm": 0.373046875, - "learning_rate": 0.000116973810142168, - "loss": 0.4846, - "step": 5092 - }, - { - "epoch": 4.603432700993677, - "grad_norm": 0.40234375, - "learning_rate": 0.00011676411450893152, - "loss": 0.5178, - "step": 5096 - }, - { - "epoch": 4.607046070460704, - "grad_norm": 0.357421875, - "learning_rate": 0.00011655448716062916, - "loss": 0.4228, - "step": 5100 - }, - { - "epoch": 4.6106594399277325, - "grad_norm": 0.357421875, - "learning_rate": 0.00011634492852795165, - "loss": 0.4944, - "step": 5104 - }, - { - "epoch": 4.614272809394761, - "grad_norm": 0.376953125, - "learning_rate": 0.00011613543904144865, - "loss": 0.4711, - "step": 5108 - }, - { - "epoch": 4.617886178861789, - "grad_norm": 0.373046875, - "learning_rate": 0.00011592601913152766, - "loss": 0.4716, - "step": 5112 - }, - { - "epoch": 4.621499548328817, - "grad_norm": 0.369140625, - "learning_rate": 0.00011571666922845334, - "loss": 0.4811, - "step": 5116 - }, - { - "epoch": 4.625112917795844, - "grad_norm": 0.390625, - "learning_rate": 0.00011550738976234637, - "loss": 0.5702, - "step": 5120 - }, - { - "epoch": 4.6287262872628725, - "grad_norm": 0.369140625, - "learning_rate": 0.00011529818116318281, - "loss": 0.5025, - "step": 5124 - }, - { - "epoch": 4.632339656729901, - "grad_norm": 0.345703125, - "learning_rate": 0.00011508904386079317, - "loss": 0.4702, - "step": 5128 - }, - { - "epoch": 4.635953026196929, - "grad_norm": 0.369140625, - "learning_rate": 0.00011487997828486129, - "loss": 0.5187, - "step": 5132 - }, - { - "epoch": 4.639566395663957, - "grad_norm": 0.384765625, - "learning_rate": 0.00011467098486492389, - "loss": 0.5601, - "step": 5136 - }, - { - "epoch": 4.643179765130984, - "grad_norm": 0.353515625, - "learning_rate": 0.00011446206403036927, - "loss": 0.5089, - "step": 5140 - }, - { - "epoch": 4.6467931345980125, - "grad_norm": 0.373046875, - "learning_rate": 0.00011425321621043659, - "loss": 0.4851, - "step": 5144 - }, - { - "epoch": 4.650406504065041, - "grad_norm": 0.35546875, - "learning_rate": 0.00011404444183421515, - "loss": 0.4992, - "step": 5148 - }, - { - "epoch": 4.654019873532069, - "grad_norm": 0.3828125, - "learning_rate": 0.00011383574133064315, - "loss": 0.4739, - "step": 5152 - }, - { - "epoch": 4.657633242999097, - "grad_norm": 0.376953125, - "learning_rate": 0.00011362711512850723, - "loss": 0.4987, - "step": 5156 - }, - { - "epoch": 4.661246612466124, - "grad_norm": 0.361328125, - "learning_rate": 0.0001134185636564412, - "loss": 0.5039, - "step": 5160 - }, - { - "epoch": 4.664859981933152, - "grad_norm": 0.369140625, - "learning_rate": 0.00011321008734292544, - "loss": 0.4694, - "step": 5164 - }, - { - "epoch": 4.668473351400181, - "grad_norm": 0.4140625, - "learning_rate": 0.00011300168661628587, - "loss": 0.4751, - "step": 5168 - }, - { - "epoch": 4.672086720867209, - "grad_norm": 0.396484375, - "learning_rate": 0.00011279336190469305, - "loss": 0.5174, - "step": 5172 - }, - { - "epoch": 4.675700090334237, - "grad_norm": 0.353515625, - "learning_rate": 0.00011258511363616151, - "loss": 0.4763, - "step": 5176 - }, - { - "epoch": 4.679313459801264, - "grad_norm": 0.373046875, - "learning_rate": 0.0001123769422385486, - "loss": 0.4803, - "step": 5180 - }, - { - "epoch": 4.682926829268292, - "grad_norm": 0.39453125, - "learning_rate": 0.0001121688481395538, - "loss": 0.5152, - "step": 5184 - }, - { - "epoch": 4.686540198735321, - "grad_norm": 0.361328125, - "learning_rate": 0.0001119608317667177, - "loss": 0.4553, - "step": 5188 - }, - { - "epoch": 4.690153568202349, - "grad_norm": 0.3515625, - "learning_rate": 0.00011175289354742134, - "loss": 0.4471, - "step": 5192 - }, - { - "epoch": 4.693766937669377, - "grad_norm": 0.3671875, - "learning_rate": 0.00011154503390888503, - "loss": 0.4766, - "step": 5196 - }, - { - "epoch": 4.697380307136404, - "grad_norm": 0.345703125, - "learning_rate": 0.00011133725327816771, - "loss": 0.4628, - "step": 5200 - }, - { - "epoch": 4.700993676603432, - "grad_norm": 0.3671875, - "learning_rate": 0.00011112955208216601, - "loss": 0.5262, - "step": 5204 - }, - { - "epoch": 4.7046070460704605, - "grad_norm": 0.3671875, - "learning_rate": 0.00011092193074761329, - "loss": 0.5055, - "step": 5208 - }, - { - "epoch": 4.708220415537489, - "grad_norm": 0.35546875, - "learning_rate": 0.00011071438970107893, - "loss": 0.4822, - "step": 5212 - }, - { - "epoch": 4.711833785004517, - "grad_norm": 0.3359375, - "learning_rate": 0.00011050692936896723, - "loss": 0.469, - "step": 5216 - }, - { - "epoch": 4.715447154471545, - "grad_norm": 0.359375, - "learning_rate": 0.00011029955017751677, - "loss": 0.4591, - "step": 5220 - }, - { - "epoch": 4.719060523938572, - "grad_norm": 0.357421875, - "learning_rate": 0.00011009225255279938, - "loss": 0.5104, - "step": 5224 - }, - { - "epoch": 4.7226738934056005, - "grad_norm": 0.396484375, - "learning_rate": 0.00010988503692071926, - "loss": 0.4972, - "step": 5228 - }, - { - "epoch": 4.726287262872629, - "grad_norm": 0.400390625, - "learning_rate": 0.00010967790370701223, - "loss": 0.4893, - "step": 5232 - }, - { - "epoch": 4.729900632339657, - "grad_norm": 0.3671875, - "learning_rate": 0.00010947085333724475, - "loss": 0.4707, - "step": 5236 - }, - { - "epoch": 4.733514001806685, - "grad_norm": 0.380859375, - "learning_rate": 0.00010926388623681304, - "loss": 0.4775, - "step": 5240 - }, - { - "epoch": 4.737127371273713, - "grad_norm": 0.34765625, - "learning_rate": 0.00010905700283094226, - "loss": 0.4952, - "step": 5244 - }, - { - "epoch": 4.7407407407407405, - "grad_norm": 0.375, - "learning_rate": 0.00010885020354468563, - "loss": 0.5275, - "step": 5248 - }, - { - "epoch": 4.744354110207769, - "grad_norm": 0.35546875, - "learning_rate": 0.00010864348880292354, - "loss": 0.5259, - "step": 5252 - }, - { - "epoch": 4.747967479674797, - "grad_norm": 0.3671875, - "learning_rate": 0.00010843685903036262, - "loss": 0.4457, - "step": 5256 - }, - { - "epoch": 4.751580849141825, - "grad_norm": 0.330078125, - "learning_rate": 0.00010823031465153505, - "loss": 0.4538, - "step": 5260 - }, - { - "epoch": 4.755194218608853, - "grad_norm": 0.39453125, - "learning_rate": 0.00010802385609079739, - "loss": 0.5377, - "step": 5264 - }, - { - "epoch": 4.7588075880758804, - "grad_norm": 0.357421875, - "learning_rate": 0.0001078174837723301, - "loss": 0.5104, - "step": 5268 - }, - { - "epoch": 4.762420957542909, - "grad_norm": 0.3828125, - "learning_rate": 0.00010761119812013625, - "loss": 0.5585, - "step": 5272 - }, - { - "epoch": 4.766034327009937, - "grad_norm": 0.375, - "learning_rate": 0.00010740499955804091, - "loss": 0.4972, - "step": 5276 - }, - { - "epoch": 4.769647696476965, - "grad_norm": 0.388671875, - "learning_rate": 0.00010719888850969031, - "loss": 0.5246, - "step": 5280 - }, - { - "epoch": 4.773261065943993, - "grad_norm": 0.376953125, - "learning_rate": 0.00010699286539855076, - "loss": 0.514, - "step": 5284 - }, - { - "epoch": 4.77687443541102, - "grad_norm": 0.365234375, - "learning_rate": 0.00010678693064790798, - "loss": 0.4796, - "step": 5288 - }, - { - "epoch": 4.780487804878049, - "grad_norm": 0.375, - "learning_rate": 0.00010658108468086611, - "loss": 0.4739, - "step": 5292 - }, - { - "epoch": 4.784101174345077, - "grad_norm": 0.357421875, - "learning_rate": 0.0001063753279203468, - "loss": 0.5086, - "step": 5296 - }, - { - "epoch": 4.787714543812105, - "grad_norm": 0.35546875, - "learning_rate": 0.00010616966078908861, - "loss": 0.5169, - "step": 5300 - }, - { - "epoch": 4.791327913279133, - "grad_norm": 0.361328125, - "learning_rate": 0.00010596408370964579, - "loss": 0.5098, - "step": 5304 - }, - { - "epoch": 4.79494128274616, - "grad_norm": 0.361328125, - "learning_rate": 0.00010575859710438765, - "loss": 0.4885, - "step": 5308 - }, - { - "epoch": 4.7985546522131886, - "grad_norm": 0.3671875, - "learning_rate": 0.00010555320139549755, - "loss": 0.5206, - "step": 5312 - }, - { - "epoch": 4.802168021680217, - "grad_norm": 0.390625, - "learning_rate": 0.00010534789700497226, - "loss": 0.4649, - "step": 5316 - }, - { - "epoch": 4.805781391147245, - "grad_norm": 0.36328125, - "learning_rate": 0.00010514268435462069, - "loss": 0.4921, - "step": 5320 - }, - { - "epoch": 4.809394760614273, - "grad_norm": 0.361328125, - "learning_rate": 0.00010493756386606341, - "loss": 0.5084, - "step": 5324 - }, - { - "epoch": 4.8130081300813, - "grad_norm": 0.3671875, - "learning_rate": 0.00010473253596073169, - "loss": 0.4661, - "step": 5328 - }, - { - "epoch": 4.8166214995483285, - "grad_norm": 0.359375, - "learning_rate": 0.00010452760105986643, - "loss": 0.4978, - "step": 5332 - }, - { - "epoch": 4.820234869015357, - "grad_norm": 0.392578125, - "learning_rate": 0.0001043227595845176, - "loss": 0.5176, - "step": 5336 - }, - { - "epoch": 4.823848238482385, - "grad_norm": 0.36328125, - "learning_rate": 0.00010411801195554308, - "loss": 0.514, - "step": 5340 - }, - { - "epoch": 4.827461607949413, - "grad_norm": 0.388671875, - "learning_rate": 0.00010391335859360806, - "loss": 0.5451, - "step": 5344 - }, - { - "epoch": 4.831074977416441, - "grad_norm": 0.375, - "learning_rate": 0.00010370879991918398, - "loss": 0.5037, - "step": 5348 - }, - { - "epoch": 4.8346883468834685, - "grad_norm": 0.3671875, - "learning_rate": 0.00010350433635254775, - "loss": 0.4899, - "step": 5352 - }, - { - "epoch": 4.838301716350497, - "grad_norm": 0.3671875, - "learning_rate": 0.00010329996831378091, - "loss": 0.5386, - "step": 5356 - }, - { - "epoch": 4.841915085817525, - "grad_norm": 0.376953125, - "learning_rate": 0.00010309569622276866, - "loss": 0.5056, - "step": 5360 - }, - { - "epoch": 4.845528455284553, - "grad_norm": 0.40625, - "learning_rate": 0.00010289152049919922, - "loss": 0.4911, - "step": 5364 - }, - { - "epoch": 4.849141824751581, - "grad_norm": 0.376953125, - "learning_rate": 0.00010268744156256264, - "loss": 0.4855, - "step": 5368 - }, - { - "epoch": 4.852755194218609, - "grad_norm": 0.3515625, - "learning_rate": 0.00010248345983215015, - "loss": 0.513, - "step": 5372 - }, - { - "epoch": 4.856368563685637, - "grad_norm": 0.380859375, - "learning_rate": 0.00010227957572705342, - "loss": 0.4928, - "step": 5376 - }, - { - "epoch": 4.859981933152665, - "grad_norm": 0.375, - "learning_rate": 0.00010207578966616334, - "loss": 0.4715, - "step": 5380 - }, - { - "epoch": 4.863595302619693, - "grad_norm": 0.390625, - "learning_rate": 0.00010187210206816953, - "loss": 0.5328, - "step": 5384 - }, - { - "epoch": 4.867208672086721, - "grad_norm": 0.380859375, - "learning_rate": 0.00010166851335155913, - "loss": 0.5068, - "step": 5388 - }, - { - "epoch": 4.870822041553749, - "grad_norm": 0.380859375, - "learning_rate": 0.00010146502393461637, - "loss": 0.4827, - "step": 5392 - }, - { - "epoch": 4.874435411020777, - "grad_norm": 0.365234375, - "learning_rate": 0.00010126163423542121, - "loss": 0.5177, - "step": 5396 - }, - { - "epoch": 4.878048780487805, - "grad_norm": 0.376953125, - "learning_rate": 0.00010105834467184889, - "loss": 0.4967, - "step": 5400 - }, - { - "epoch": 4.881662149954833, - "grad_norm": 0.365234375, - "learning_rate": 0.00010085515566156888, - "loss": 0.516, - "step": 5404 - }, - { - "epoch": 4.885275519421861, - "grad_norm": 0.3515625, - "learning_rate": 0.000100652067622044, - "loss": 0.4888, - "step": 5408 - }, - { - "epoch": 4.888888888888889, - "grad_norm": 0.35546875, - "learning_rate": 0.00010044908097052976, - "loss": 0.5017, - "step": 5412 - }, - { - "epoch": 4.892502258355917, - "grad_norm": 0.365234375, - "learning_rate": 0.00010024619612407317, - "loss": 0.4858, - "step": 5416 - }, - { - "epoch": 4.896115627822945, - "grad_norm": 0.3828125, - "learning_rate": 0.00010004341349951228, - "loss": 0.4965, - "step": 5420 - }, - { - "epoch": 4.899728997289973, - "grad_norm": 0.369140625, - "learning_rate": 9.984073351347498e-05, - "loss": 0.4791, - "step": 5424 - }, - { - "epoch": 4.903342366757001, - "grad_norm": 0.392578125, - "learning_rate": 9.963815658237829e-05, - "loss": 0.5123, - "step": 5428 - }, - { - "epoch": 4.906955736224029, - "grad_norm": 0.388671875, - "learning_rate": 9.943568312242756e-05, - "loss": 0.5253, - "step": 5432 - }, - { - "epoch": 4.9105691056910565, - "grad_norm": 0.353515625, - "learning_rate": 9.923331354961551e-05, - "loss": 0.4842, - "step": 5436 - }, - { - "epoch": 4.914182475158085, - "grad_norm": 0.34765625, - "learning_rate": 9.903104827972147e-05, - "loss": 0.5174, - "step": 5440 - }, - { - "epoch": 4.917795844625113, - "grad_norm": 0.388671875, - "learning_rate": 9.882888772831041e-05, - "loss": 0.5506, - "step": 5444 - }, - { - "epoch": 4.921409214092141, - "grad_norm": 0.40234375, - "learning_rate": 9.862683231073215e-05, - "loss": 0.4965, - "step": 5448 - }, - { - "epoch": 4.925022583559169, - "grad_norm": 0.359375, - "learning_rate": 9.842488244212057e-05, - "loss": 0.5504, - "step": 5452 - }, - { - "epoch": 4.9286359530261965, - "grad_norm": 0.3671875, - "learning_rate": 9.822303853739265e-05, - "loss": 0.4471, - "step": 5456 - }, - { - "epoch": 4.932249322493225, - "grad_norm": 0.3671875, - "learning_rate": 9.802130101124765e-05, - "loss": 0.4667, - "step": 5460 - }, - { - "epoch": 4.935862691960253, - "grad_norm": 0.376953125, - "learning_rate": 9.781967027816631e-05, - "loss": 0.4884, - "step": 5464 - }, - { - "epoch": 4.939476061427281, - "grad_norm": 0.380859375, - "learning_rate": 9.761814675240995e-05, - "loss": 0.4699, - "step": 5468 - }, - { - "epoch": 4.943089430894309, - "grad_norm": 0.37109375, - "learning_rate": 9.741673084801959e-05, - "loss": 0.4987, - "step": 5472 - }, - { - "epoch": 4.9467028003613365, - "grad_norm": 0.375, - "learning_rate": 9.721542297881511e-05, - "loss": 0.5054, - "step": 5476 - }, - { - "epoch": 4.950316169828365, - "grad_norm": 0.38671875, - "learning_rate": 9.701422355839458e-05, - "loss": 0.516, - "step": 5480 - }, - { - "epoch": 4.953929539295393, - "grad_norm": 0.3671875, - "learning_rate": 9.681313300013305e-05, - "loss": 0.5119, - "step": 5484 - }, - { - "epoch": 4.957542908762421, - "grad_norm": 0.37109375, - "learning_rate": 9.661215171718207e-05, - "loss": 0.4833, - "step": 5488 - }, - { - "epoch": 4.961156278229449, - "grad_norm": 0.3515625, - "learning_rate": 9.641128012246858e-05, - "loss": 0.5013, - "step": 5492 - }, - { - "epoch": 4.964769647696477, - "grad_norm": 0.357421875, - "learning_rate": 9.621051862869423e-05, - "loss": 0.4904, - "step": 5496 - }, - { - "epoch": 4.968383017163505, - "grad_norm": 0.380859375, - "learning_rate": 9.600986764833441e-05, - "loss": 0.4939, - "step": 5500 - }, - { - "epoch": 4.971996386630533, - "grad_norm": 0.376953125, - "learning_rate": 9.580932759363744e-05, - "loss": 0.4751, - "step": 5504 - }, - { - "epoch": 4.975609756097561, - "grad_norm": 0.37890625, - "learning_rate": 9.560889887662382e-05, - "loss": 0.5234, - "step": 5508 - }, - { - "epoch": 4.979223125564589, - "grad_norm": 0.35546875, - "learning_rate": 9.540858190908521e-05, - "loss": 0.5123, - "step": 5512 - }, - { - "epoch": 4.982836495031617, - "grad_norm": 0.392578125, - "learning_rate": 9.520837710258374e-05, - "loss": 0.4779, - "step": 5516 - }, - { - "epoch": 4.9864498644986455, - "grad_norm": 0.36328125, - "learning_rate": 9.500828486845106e-05, - "loss": 0.4851, - "step": 5520 - }, - { - "epoch": 4.990063233965673, - "grad_norm": 0.357421875, - "learning_rate": 9.480830561778751e-05, - "loss": 0.4492, - "step": 5524 - }, - { - "epoch": 4.993676603432701, - "grad_norm": 0.34375, - "learning_rate": 9.460843976146142e-05, - "loss": 0.457, - "step": 5528 - }, - { - "epoch": 4.997289972899729, - "grad_norm": 0.39453125, - "learning_rate": 9.440868771010797e-05, - "loss": 0.4986, - "step": 5532 - }, - { - "epoch": 5.000903342366757, - "grad_norm": 0.349609375, - "learning_rate": 9.420904987412866e-05, - "loss": 0.5165, - "step": 5536 - }, - { - "epoch": 5.004516711833785, - "grad_norm": 0.359375, - "learning_rate": 9.400952666369025e-05, - "loss": 0.4546, - "step": 5540 - }, - { - "epoch": 5.008130081300813, - "grad_norm": 0.361328125, - "learning_rate": 9.38101184887241e-05, - "loss": 0.5119, - "step": 5544 - }, - { - "epoch": 5.011743450767841, - "grad_norm": 0.365234375, - "learning_rate": 9.361082575892508e-05, - "loss": 0.4755, - "step": 5548 - }, - { - "epoch": 5.015356820234869, - "grad_norm": 0.3671875, - "learning_rate": 9.341164888375093e-05, - "loss": 0.458, - "step": 5552 - }, - { - "epoch": 5.018970189701897, - "grad_norm": 0.359375, - "learning_rate": 9.321258827242145e-05, - "loss": 0.4431, - "step": 5556 - }, - { - "epoch": 5.022583559168925, - "grad_norm": 0.37890625, - "learning_rate": 9.301364433391738e-05, - "loss": 0.4425, - "step": 5560 - }, - { - "epoch": 5.026196928635953, - "grad_norm": 0.33203125, - "learning_rate": 9.281481747697994e-05, - "loss": 0.4147, - "step": 5564 - }, - { - "epoch": 5.029810298102981, - "grad_norm": 0.392578125, - "learning_rate": 9.261610811010969e-05, - "loss": 0.4241, - "step": 5568 - }, - { - "epoch": 5.033423667570009, - "grad_norm": 0.357421875, - "learning_rate": 9.241751664156588e-05, - "loss": 0.4338, - "step": 5572 - }, - { - "epoch": 5.037037037037037, - "grad_norm": 0.359375, - "learning_rate": 9.221904347936543e-05, - "loss": 0.4574, - "step": 5576 - }, - { - "epoch": 5.040650406504065, - "grad_norm": 0.369140625, - "learning_rate": 9.202068903128222e-05, - "loss": 0.4555, - "step": 5580 - }, - { - "epoch": 5.044263775971093, - "grad_norm": 0.3984375, - "learning_rate": 9.182245370484633e-05, - "loss": 0.4773, - "step": 5584 - }, - { - "epoch": 5.047877145438121, - "grad_norm": 0.341796875, - "learning_rate": 9.162433790734292e-05, - "loss": 0.4972, - "step": 5588 - }, - { - "epoch": 5.051490514905149, - "grad_norm": 0.37109375, - "learning_rate": 9.142634204581177e-05, - "loss": 0.4566, - "step": 5592 - }, - { - "epoch": 5.055103884372177, - "grad_norm": 0.3671875, - "learning_rate": 9.122846652704608e-05, - "loss": 0.4461, - "step": 5596 - }, - { - "epoch": 5.058717253839205, - "grad_norm": 0.36328125, - "learning_rate": 9.103071175759188e-05, - "loss": 0.4297, - "step": 5600 - }, - { - "epoch": 5.062330623306233, - "grad_norm": 0.3671875, - "learning_rate": 9.083307814374715e-05, - "loss": 0.485, - "step": 5604 - }, - { - "epoch": 5.065943992773261, - "grad_norm": 0.37890625, - "learning_rate": 9.063556609156077e-05, - "loss": 0.4577, - "step": 5608 - }, - { - "epoch": 5.069557362240289, - "grad_norm": 0.357421875, - "learning_rate": 9.04381760068321e-05, - "loss": 0.4168, - "step": 5612 - }, - { - "epoch": 5.073170731707317, - "grad_norm": 0.39453125, - "learning_rate": 9.024090829510976e-05, - "loss": 0.4594, - "step": 5616 - }, - { - "epoch": 5.076784101174345, - "grad_norm": 0.34765625, - "learning_rate": 9.004376336169102e-05, - "loss": 0.4259, - "step": 5620 - }, - { - "epoch": 5.0803974706413735, - "grad_norm": 0.38671875, - "learning_rate": 8.984674161162085e-05, - "loss": 0.4593, - "step": 5624 - }, - { - "epoch": 5.084010840108401, - "grad_norm": 0.3671875, - "learning_rate": 8.964984344969111e-05, - "loss": 0.4226, - "step": 5628 - }, - { - "epoch": 5.087624209575429, - "grad_norm": 0.365234375, - "learning_rate": 8.945306928043988e-05, - "loss": 0.4343, - "step": 5632 - }, - { - "epoch": 5.091237579042457, - "grad_norm": 0.376953125, - "learning_rate": 8.925641950815026e-05, - "loss": 0.4335, - "step": 5636 - }, - { - "epoch": 5.094850948509485, - "grad_norm": 0.376953125, - "learning_rate": 8.905989453685003e-05, - "loss": 0.435, - "step": 5640 - }, - { - "epoch": 5.0984643179765134, - "grad_norm": 0.421875, - "learning_rate": 8.88634947703103e-05, - "loss": 0.4573, - "step": 5644 - }, - { - "epoch": 5.102077687443541, - "grad_norm": 0.408203125, - "learning_rate": 8.866722061204522e-05, - "loss": 0.5077, - "step": 5648 - }, - { - "epoch": 5.105691056910569, - "grad_norm": 0.38671875, - "learning_rate": 8.847107246531064e-05, - "loss": 0.4731, - "step": 5652 - }, - { - "epoch": 5.109304426377597, - "grad_norm": 0.38671875, - "learning_rate": 8.827505073310353e-05, - "loss": 0.431, - "step": 5656 - }, - { - "epoch": 5.112917795844625, - "grad_norm": 0.375, - "learning_rate": 8.807915581816132e-05, - "loss": 0.4606, - "step": 5660 - }, - { - "epoch": 5.116531165311653, - "grad_norm": 0.3828125, - "learning_rate": 8.788338812296068e-05, - "loss": 0.4314, - "step": 5664 - }, - { - "epoch": 5.120144534778681, - "grad_norm": 0.39453125, - "learning_rate": 8.768774804971705e-05, - "loss": 0.4581, - "step": 5668 - }, - { - "epoch": 5.123757904245709, - "grad_norm": 0.37890625, - "learning_rate": 8.749223600038354e-05, - "loss": 0.4931, - "step": 5672 - }, - { - "epoch": 5.127371273712737, - "grad_norm": 0.375, - "learning_rate": 8.72968523766503e-05, - "loss": 0.4472, - "step": 5676 - }, - { - "epoch": 5.130984643179765, - "grad_norm": 0.40234375, - "learning_rate": 8.710159757994366e-05, - "loss": 0.4415, - "step": 5680 - }, - { - "epoch": 5.134598012646793, - "grad_norm": 0.388671875, - "learning_rate": 8.690647201142513e-05, - "loss": 0.4492, - "step": 5684 - }, - { - "epoch": 5.138211382113822, - "grad_norm": 0.3828125, - "learning_rate": 8.671147607199096e-05, - "loss": 0.4317, - "step": 5688 - }, - { - "epoch": 5.141824751580849, - "grad_norm": 0.37109375, - "learning_rate": 8.651661016227067e-05, - "loss": 0.4452, - "step": 5692 - }, - { - "epoch": 5.145438121047877, - "grad_norm": 0.35546875, - "learning_rate": 8.632187468262708e-05, - "loss": 0.4734, - "step": 5696 - }, - { - "epoch": 5.149051490514905, - "grad_norm": 0.3671875, - "learning_rate": 8.612727003315476e-05, - "loss": 0.4279, - "step": 5700 - }, - { - "epoch": 5.152664859981933, - "grad_norm": 0.388671875, - "learning_rate": 8.593279661367942e-05, - "loss": 0.4164, - "step": 5704 - }, - { - "epoch": 5.1562782294489615, - "grad_norm": 0.373046875, - "learning_rate": 8.573845482375754e-05, - "loss": 0.4438, - "step": 5708 - }, - { - "epoch": 5.159891598915989, - "grad_norm": 0.37890625, - "learning_rate": 8.554424506267461e-05, - "loss": 0.4514, - "step": 5712 - }, - { - "epoch": 5.163504968383017, - "grad_norm": 0.392578125, - "learning_rate": 8.535016772944534e-05, - "loss": 0.4454, - "step": 5716 - }, - { - "epoch": 5.167118337850045, - "grad_norm": 0.392578125, - "learning_rate": 8.515622322281203e-05, - "loss": 0.4425, - "step": 5720 - }, - { - "epoch": 5.170731707317073, - "grad_norm": 0.37109375, - "learning_rate": 8.496241194124436e-05, - "loss": 0.4466, - "step": 5724 - }, - { - "epoch": 5.1743450767841015, - "grad_norm": 0.369140625, - "learning_rate": 8.476873428293812e-05, - "loss": 0.428, - "step": 5728 - }, - { - "epoch": 5.177958446251129, - "grad_norm": 0.390625, - "learning_rate": 8.457519064581444e-05, - "loss": 0.4494, - "step": 5732 - }, - { - "epoch": 5.181571815718157, - "grad_norm": 0.375, - "learning_rate": 8.438178142751939e-05, - "loss": 0.4485, - "step": 5736 - }, - { - "epoch": 5.185185185185185, - "grad_norm": 0.400390625, - "learning_rate": 8.418850702542265e-05, - "loss": 0.4973, - "step": 5740 - }, - { - "epoch": 5.188798554652213, - "grad_norm": 0.38671875, - "learning_rate": 8.399536783661713e-05, - "loss": 0.4924, - "step": 5744 - }, - { - "epoch": 5.1924119241192415, - "grad_norm": 0.3828125, - "learning_rate": 8.380236425791759e-05, - "loss": 0.4346, - "step": 5748 - }, - { - "epoch": 5.196025293586269, - "grad_norm": 0.390625, - "learning_rate": 8.360949668586055e-05, - "loss": 0.4849, - "step": 5752 - }, - { - "epoch": 5.199638663053297, - "grad_norm": 0.369140625, - "learning_rate": 8.341676551670283e-05, - "loss": 0.4523, - "step": 5756 - }, - { - "epoch": 5.203252032520325, - "grad_norm": 0.400390625, - "learning_rate": 8.322417114642107e-05, - "loss": 0.4236, - "step": 5760 - }, - { - "epoch": 5.206865401987353, - "grad_norm": 0.373046875, - "learning_rate": 8.303171397071106e-05, - "loss": 0.4416, - "step": 5764 - }, - { - "epoch": 5.210478771454381, - "grad_norm": 0.359375, - "learning_rate": 8.283939438498627e-05, - "loss": 0.4522, - "step": 5768 - }, - { - "epoch": 5.21409214092141, - "grad_norm": 0.36328125, - "learning_rate": 8.264721278437795e-05, - "loss": 0.476, - "step": 5772 - }, - { - "epoch": 5.217705510388437, - "grad_norm": 0.388671875, - "learning_rate": 8.245516956373359e-05, - "loss": 0.4237, - "step": 5776 - }, - { - "epoch": 5.221318879855465, - "grad_norm": 0.3828125, - "learning_rate": 8.226326511761635e-05, - "loss": 0.4215, - "step": 5780 - }, - { - "epoch": 5.224932249322493, - "grad_norm": 0.412109375, - "learning_rate": 8.207149984030453e-05, - "loss": 0.4488, - "step": 5784 - }, - { - "epoch": 5.228545618789521, - "grad_norm": 0.3828125, - "learning_rate": 8.187987412579011e-05, - "loss": 0.4941, - "step": 5788 - }, - { - "epoch": 5.23215898825655, - "grad_norm": 0.375, - "learning_rate": 8.168838836777871e-05, - "loss": 0.4512, - "step": 5792 - }, - { - "epoch": 5.235772357723577, - "grad_norm": 0.3671875, - "learning_rate": 8.14970429596881e-05, - "loss": 0.4581, - "step": 5796 - }, - { - "epoch": 5.239385727190605, - "grad_norm": 0.35546875, - "learning_rate": 8.130583829464793e-05, - "loss": 0.4529, - "step": 5800 - }, - { - "epoch": 5.242999096657633, - "grad_norm": 0.373046875, - "learning_rate": 8.11147747654986e-05, - "loss": 0.4779, - "step": 5804 - }, - { - "epoch": 5.246612466124661, - "grad_norm": 0.37109375, - "learning_rate": 8.092385276479038e-05, - "loss": 0.3973, - "step": 5808 - }, - { - "epoch": 5.2502258355916895, - "grad_norm": 0.376953125, - "learning_rate": 8.073307268478303e-05, - "loss": 0.436, - "step": 5812 - }, - { - "epoch": 5.253839205058718, - "grad_norm": 0.37890625, - "learning_rate": 8.054243491744448e-05, - "loss": 0.4576, - "step": 5816 - }, - { - "epoch": 5.257452574525745, - "grad_norm": 0.3828125, - "learning_rate": 8.03519398544506e-05, - "loss": 0.4253, - "step": 5820 - }, - { - "epoch": 5.261065943992773, - "grad_norm": 0.3828125, - "learning_rate": 8.01615878871836e-05, - "loss": 0.4457, - "step": 5824 - }, - { - "epoch": 5.264679313459801, - "grad_norm": 0.390625, - "learning_rate": 7.997137940673212e-05, - "loss": 0.4777, - "step": 5828 - }, - { - "epoch": 5.2682926829268295, - "grad_norm": 0.39453125, - "learning_rate": 7.97813148038898e-05, - "loss": 0.4356, - "step": 5832 - }, - { - "epoch": 5.271906052393858, - "grad_norm": 0.37109375, - "learning_rate": 7.95913944691546e-05, - "loss": 0.4404, - "step": 5836 - }, - { - "epoch": 5.275519421860885, - "grad_norm": 0.37890625, - "learning_rate": 7.940161879272837e-05, - "loss": 0.4555, - "step": 5840 - }, - { - "epoch": 5.279132791327913, - "grad_norm": 0.365234375, - "learning_rate": 7.921198816451531e-05, - "loss": 0.4435, - "step": 5844 - }, - { - "epoch": 5.282746160794941, - "grad_norm": 0.375, - "learning_rate": 7.902250297412209e-05, - "loss": 0.4695, - "step": 5848 - }, - { - "epoch": 5.2863595302619695, - "grad_norm": 0.388671875, - "learning_rate": 7.883316361085618e-05, - "loss": 0.414, - "step": 5852 - }, - { - "epoch": 5.289972899728998, - "grad_norm": 0.36328125, - "learning_rate": 7.864397046372563e-05, - "loss": 0.4476, - "step": 5856 - }, - { - "epoch": 5.293586269196025, - "grad_norm": 0.365234375, - "learning_rate": 7.84549239214382e-05, - "loss": 0.4597, - "step": 5860 - }, - { - "epoch": 5.297199638663053, - "grad_norm": 0.359375, - "learning_rate": 7.826602437240003e-05, - "loss": 0.4677, - "step": 5864 - }, - { - "epoch": 5.300813008130081, - "grad_norm": 0.36328125, - "learning_rate": 7.80772722047157e-05, - "loss": 0.4376, - "step": 5868 - }, - { - "epoch": 5.304426377597109, - "grad_norm": 0.376953125, - "learning_rate": 7.788866780618672e-05, - "loss": 0.4448, - "step": 5872 - }, - { - "epoch": 5.308039747064138, - "grad_norm": 0.38671875, - "learning_rate": 7.770021156431115e-05, - "loss": 0.4528, - "step": 5876 - }, - { - "epoch": 5.311653116531165, - "grad_norm": 0.37890625, - "learning_rate": 7.751190386628261e-05, - "loss": 0.49, - "step": 5880 - }, - { - "epoch": 5.315266485998193, - "grad_norm": 0.384765625, - "learning_rate": 7.732374509898935e-05, - "loss": 0.4535, - "step": 5884 - }, - { - "epoch": 5.318879855465221, - "grad_norm": 0.361328125, - "learning_rate": 7.713573564901393e-05, - "loss": 0.4324, - "step": 5888 - }, - { - "epoch": 5.322493224932249, - "grad_norm": 0.3671875, - "learning_rate": 7.694787590263191e-05, - "loss": 0.4557, - "step": 5892 - }, - { - "epoch": 5.326106594399278, - "grad_norm": 0.34375, - "learning_rate": 7.676016624581143e-05, - "loss": 0.4346, - "step": 5896 - }, - { - "epoch": 5.329719963866305, - "grad_norm": 0.412109375, - "learning_rate": 7.657260706421217e-05, - "loss": 0.4433, - "step": 5900 - }, - { - "epoch": 5.333333333333333, - "grad_norm": 0.380859375, - "learning_rate": 7.638519874318461e-05, - "loss": 0.4558, - "step": 5904 - }, - { - "epoch": 5.336946702800361, - "grad_norm": 0.3828125, - "learning_rate": 7.619794166776943e-05, - "loss": 0.4721, - "step": 5908 - }, - { - "epoch": 5.340560072267389, - "grad_norm": 0.365234375, - "learning_rate": 7.60108362226963e-05, - "loss": 0.4502, - "step": 5912 - }, - { - "epoch": 5.3441734417344176, - "grad_norm": 0.359375, - "learning_rate": 7.582388279238381e-05, - "loss": 0.4422, - "step": 5916 - }, - { - "epoch": 5.347786811201446, - "grad_norm": 0.400390625, - "learning_rate": 7.563708176093765e-05, - "loss": 0.4733, - "step": 5920 - }, - { - "epoch": 5.351400180668473, - "grad_norm": 0.40234375, - "learning_rate": 7.545043351215092e-05, - "loss": 0.489, - "step": 5924 - }, - { - "epoch": 5.355013550135501, - "grad_norm": 0.373046875, - "learning_rate": 7.52639384295025e-05, - "loss": 0.4533, - "step": 5928 - }, - { - "epoch": 5.358626919602529, - "grad_norm": 0.380859375, - "learning_rate": 7.507759689615662e-05, - "loss": 0.4594, - "step": 5932 - }, - { - "epoch": 5.3622402890695575, - "grad_norm": 0.361328125, - "learning_rate": 7.489140929496229e-05, - "loss": 0.4292, - "step": 5936 - }, - { - "epoch": 5.365853658536586, - "grad_norm": 0.3671875, - "learning_rate": 7.470537600845182e-05, - "loss": 0.4571, - "step": 5940 - }, - { - "epoch": 5.369467028003613, - "grad_norm": 0.35546875, - "learning_rate": 7.45194974188409e-05, - "loss": 0.4081, - "step": 5944 - }, - { - "epoch": 5.373080397470641, - "grad_norm": 0.390625, - "learning_rate": 7.433377390802705e-05, - "loss": 0.4749, - "step": 5948 - }, - { - "epoch": 5.376693766937669, - "grad_norm": 0.396484375, - "learning_rate": 7.414820585758949e-05, - "loss": 0.4672, - "step": 5952 - }, - { - "epoch": 5.3803071364046975, - "grad_norm": 0.423828125, - "learning_rate": 7.396279364878781e-05, - "loss": 0.4913, - "step": 5956 - }, - { - "epoch": 5.383920505871726, - "grad_norm": 0.37890625, - "learning_rate": 7.377753766256135e-05, - "loss": 0.42, - "step": 5960 - }, - { - "epoch": 5.387533875338754, - "grad_norm": 0.3828125, - "learning_rate": 7.359243827952877e-05, - "loss": 0.4313, - "step": 5964 - }, - { - "epoch": 5.391147244805781, - "grad_norm": 0.3671875, - "learning_rate": 7.340749587998669e-05, - "loss": 0.4101, - "step": 5968 - }, - { - "epoch": 5.394760614272809, - "grad_norm": 0.3671875, - "learning_rate": 7.322271084390945e-05, - "loss": 0.4171, - "step": 5972 - }, - { - "epoch": 5.3983739837398375, - "grad_norm": 0.4140625, - "learning_rate": 7.303808355094791e-05, - "loss": 0.4458, - "step": 5976 - }, - { - "epoch": 5.401987353206866, - "grad_norm": 0.388671875, - "learning_rate": 7.285361438042887e-05, - "loss": 0.4423, - "step": 5980 - }, - { - "epoch": 5.405600722673894, - "grad_norm": 0.39453125, - "learning_rate": 7.266930371135426e-05, - "loss": 0.46, - "step": 5984 - }, - { - "epoch": 5.409214092140921, - "grad_norm": 0.3671875, - "learning_rate": 7.248515192240033e-05, - "loss": 0.4832, - "step": 5988 - }, - { - "epoch": 5.412827461607949, - "grad_norm": 0.376953125, - "learning_rate": 7.230115939191713e-05, - "loss": 0.4656, - "step": 5992 - }, - { - "epoch": 5.416440831074977, - "grad_norm": 0.384765625, - "learning_rate": 7.211732649792706e-05, - "loss": 0.4732, - "step": 5996 - }, - { - "epoch": 5.420054200542006, - "grad_norm": 0.34765625, - "learning_rate": 7.193365361812499e-05, - "loss": 0.4405, - "step": 6000 - }, - { - "epoch": 5.423667570009034, - "grad_norm": 0.37109375, - "learning_rate": 7.175014112987677e-05, - "loss": 0.4558, - "step": 6004 - }, - { - "epoch": 5.427280939476061, - "grad_norm": 0.400390625, - "learning_rate": 7.156678941021874e-05, - "loss": 0.4448, - "step": 6008 - }, - { - "epoch": 5.430894308943089, - "grad_norm": 0.375, - "learning_rate": 7.138359883585715e-05, - "loss": 0.4381, - "step": 6012 - }, - { - "epoch": 5.434507678410117, - "grad_norm": 0.37109375, - "learning_rate": 7.120056978316676e-05, - "loss": 0.4776, - "step": 6016 - }, - { - "epoch": 5.438121047877146, - "grad_norm": 0.376953125, - "learning_rate": 7.101770262819089e-05, - "loss": 0.4322, - "step": 6020 - }, - { - "epoch": 5.441734417344174, - "grad_norm": 0.3671875, - "learning_rate": 7.083499774663994e-05, - "loss": 0.4651, - "step": 6024 - }, - { - "epoch": 5.445347786811201, - "grad_norm": 0.369140625, - "learning_rate": 7.065245551389112e-05, - "loss": 0.4518, - "step": 6028 - }, - { - "epoch": 5.448961156278229, - "grad_norm": 0.365234375, - "learning_rate": 7.04700763049874e-05, - "loss": 0.4403, - "step": 6032 - }, - { - "epoch": 5.452574525745257, - "grad_norm": 0.361328125, - "learning_rate": 7.02878604946366e-05, - "loss": 0.4851, - "step": 6036 - }, - { - "epoch": 5.4561878952122855, - "grad_norm": 0.41796875, - "learning_rate": 7.010580845721119e-05, - "loss": 0.4262, - "step": 6040 - }, - { - "epoch": 5.459801264679314, - "grad_norm": 0.384765625, - "learning_rate": 6.992392056674687e-05, - "loss": 0.4951, - "step": 6044 - }, - { - "epoch": 5.463414634146342, - "grad_norm": 0.408203125, - "learning_rate": 6.974219719694229e-05, - "loss": 0.4697, - "step": 6048 - }, - { - "epoch": 5.467028003613369, - "grad_norm": 0.3671875, - "learning_rate": 6.956063872115796e-05, - "loss": 0.4828, - "step": 6052 - }, - { - "epoch": 5.470641373080397, - "grad_norm": 0.390625, - "learning_rate": 6.937924551241562e-05, - "loss": 0.4649, - "step": 6056 - }, - { - "epoch": 5.4742547425474255, - "grad_norm": 0.353515625, - "learning_rate": 6.919801794339751e-05, - "loss": 0.4759, - "step": 6060 - }, - { - "epoch": 5.477868112014454, - "grad_norm": 0.39453125, - "learning_rate": 6.901695638644547e-05, - "loss": 0.4503, - "step": 6064 - }, - { - "epoch": 5.481481481481482, - "grad_norm": 0.365234375, - "learning_rate": 6.883606121356046e-05, - "loss": 0.4506, - "step": 6068 - }, - { - "epoch": 5.485094850948509, - "grad_norm": 0.404296875, - "learning_rate": 6.865533279640121e-05, - "loss": 0.4817, - "step": 6072 - }, - { - "epoch": 5.488708220415537, - "grad_norm": 0.365234375, - "learning_rate": 6.84747715062843e-05, - "loss": 0.4343, - "step": 6076 - }, - { - "epoch": 5.4923215898825655, - "grad_norm": 0.380859375, - "learning_rate": 6.829437771418257e-05, - "loss": 0.4502, - "step": 6080 - }, - { - "epoch": 5.495934959349594, - "grad_norm": 0.3671875, - "learning_rate": 6.811415179072501e-05, - "loss": 0.4064, - "step": 6084 - }, - { - "epoch": 5.499548328816622, - "grad_norm": 0.37109375, - "learning_rate": 6.793409410619556e-05, - "loss": 0.4308, - "step": 6088 - }, - { - "epoch": 5.50316169828365, - "grad_norm": 0.388671875, - "learning_rate": 6.775420503053239e-05, - "loss": 0.4258, - "step": 6092 - }, - { - "epoch": 5.506775067750677, - "grad_norm": 0.357421875, - "learning_rate": 6.757448493332752e-05, - "loss": 0.4541, - "step": 6096 - }, - { - "epoch": 5.510388437217705, - "grad_norm": 0.365234375, - "learning_rate": 6.739493418382555e-05, - "loss": 0.4554, - "step": 6100 - }, - { - "epoch": 5.514001806684734, - "grad_norm": 0.41796875, - "learning_rate": 6.721555315092341e-05, - "loss": 0.5069, - "step": 6104 - }, - { - "epoch": 5.517615176151762, - "grad_norm": 0.341796875, - "learning_rate": 6.703634220316914e-05, - "loss": 0.4701, - "step": 6108 - }, - { - "epoch": 5.52122854561879, - "grad_norm": 0.376953125, - "learning_rate": 6.685730170876122e-05, - "loss": 0.4212, - "step": 6112 - }, - { - "epoch": 5.524841915085817, - "grad_norm": 0.38671875, - "learning_rate": 6.667843203554827e-05, - "loss": 0.4479, - "step": 6116 - }, - { - "epoch": 5.528455284552845, - "grad_norm": 0.35546875, - "learning_rate": 6.649973355102765e-05, - "loss": 0.4272, - "step": 6120 - }, - { - "epoch": 5.532068654019874, - "grad_norm": 0.384765625, - "learning_rate": 6.632120662234521e-05, - "loss": 0.4513, - "step": 6124 - }, - { - "epoch": 5.535682023486902, - "grad_norm": 0.39453125, - "learning_rate": 6.614285161629417e-05, - "loss": 0.4464, - "step": 6128 - }, - { - "epoch": 5.53929539295393, - "grad_norm": 0.376953125, - "learning_rate": 6.596466889931462e-05, - "loss": 0.4526, - "step": 6132 - }, - { - "epoch": 5.542908762420957, - "grad_norm": 0.369140625, - "learning_rate": 6.578665883749262e-05, - "loss": 0.455, - "step": 6136 - }, - { - "epoch": 5.546522131887985, - "grad_norm": 0.359375, - "learning_rate": 6.560882179655946e-05, - "loss": 0.4683, - "step": 6140 - }, - { - "epoch": 5.5501355013550135, - "grad_norm": 0.361328125, - "learning_rate": 6.543115814189112e-05, - "loss": 0.4441, - "step": 6144 - }, - { - "epoch": 5.553748870822042, - "grad_norm": 0.3828125, - "learning_rate": 6.52536682385072e-05, - "loss": 0.445, - "step": 6148 - }, - { - "epoch": 5.55736224028907, - "grad_norm": 0.390625, - "learning_rate": 6.507635245107036e-05, - "loss": 0.4254, - "step": 6152 - }, - { - "epoch": 5.560975609756097, - "grad_norm": 0.376953125, - "learning_rate": 6.489921114388547e-05, - "loss": 0.4385, - "step": 6156 - }, - { - "epoch": 5.564588979223125, - "grad_norm": 0.396484375, - "learning_rate": 6.472224468089909e-05, - "loss": 0.4733, - "step": 6160 - }, - { - "epoch": 5.5682023486901535, - "grad_norm": 0.376953125, - "learning_rate": 6.454545342569842e-05, - "loss": 0.4438, - "step": 6164 - }, - { - "epoch": 5.571815718157182, - "grad_norm": 0.3984375, - "learning_rate": 6.436883774151057e-05, - "loss": 0.4355, - "step": 6168 - }, - { - "epoch": 5.57542908762421, - "grad_norm": 0.396484375, - "learning_rate": 6.419239799120222e-05, - "loss": 0.486, - "step": 6172 - }, - { - "epoch": 5.579042457091237, - "grad_norm": 0.388671875, - "learning_rate": 6.401613453727834e-05, - "loss": 0.4683, - "step": 6176 - }, - { - "epoch": 5.582655826558265, - "grad_norm": 0.373046875, - "learning_rate": 6.384004774188184e-05, - "loss": 0.4502, - "step": 6180 - }, - { - "epoch": 5.5862691960252935, - "grad_norm": 0.388671875, - "learning_rate": 6.366413796679266e-05, - "loss": 0.4511, - "step": 6184 - }, - { - "epoch": 5.589882565492322, - "grad_norm": 0.40234375, - "learning_rate": 6.348840557342684e-05, - "loss": 0.4497, - "step": 6188 - }, - { - "epoch": 5.59349593495935, - "grad_norm": 0.369140625, - "learning_rate": 6.331285092283627e-05, - "loss": 0.4689, - "step": 6192 - }, - { - "epoch": 5.597109304426377, - "grad_norm": 0.388671875, - "learning_rate": 6.313747437570746e-05, - "loss": 0.4535, - "step": 6196 - }, - { - "epoch": 5.600722673893405, - "grad_norm": 0.376953125, - "learning_rate": 6.296227629236113e-05, - "loss": 0.4724, - "step": 6200 - }, - { - "epoch": 5.6043360433604335, - "grad_norm": 0.3984375, - "learning_rate": 6.278725703275124e-05, - "loss": 0.4687, - "step": 6204 - }, - { - "epoch": 5.607949412827462, - "grad_norm": 0.373046875, - "learning_rate": 6.261241695646437e-05, - "loss": 0.4474, - "step": 6208 - }, - { - "epoch": 5.61156278229449, - "grad_norm": 0.37890625, - "learning_rate": 6.243775642271895e-05, - "loss": 0.4471, - "step": 6212 - }, - { - "epoch": 5.615176151761518, - "grad_norm": 0.376953125, - "learning_rate": 6.226327579036451e-05, - "loss": 0.4865, - "step": 6216 - }, - { - "epoch": 5.618789521228545, - "grad_norm": 0.3828125, - "learning_rate": 6.208897541788109e-05, - "loss": 0.4483, - "step": 6220 - }, - { - "epoch": 5.622402890695573, - "grad_norm": 0.380859375, - "learning_rate": 6.191485566337825e-05, - "loss": 0.4523, - "step": 6224 - }, - { - "epoch": 5.626016260162602, - "grad_norm": 0.396484375, - "learning_rate": 6.174091688459448e-05, - "loss": 0.4679, - "step": 6228 - }, - { - "epoch": 5.62962962962963, - "grad_norm": 0.37890625, - "learning_rate": 6.156715943889641e-05, - "loss": 0.4603, - "step": 6232 - }, - { - "epoch": 5.633242999096658, - "grad_norm": 0.41796875, - "learning_rate": 6.139358368327826e-05, - "loss": 0.451, - "step": 6236 - }, - { - "epoch": 5.636856368563686, - "grad_norm": 0.408203125, - "learning_rate": 6.122018997436087e-05, - "loss": 0.4728, - "step": 6240 - }, - { - "epoch": 5.640469738030713, - "grad_norm": 0.396484375, - "learning_rate": 6.104697866839089e-05, - "loss": 0.4546, - "step": 6244 - }, - { - "epoch": 5.644083107497742, - "grad_norm": 0.42578125, - "learning_rate": 6.0873950121240505e-05, - "loss": 0.4323, - "step": 6248 - }, - { - "epoch": 5.64769647696477, - "grad_norm": 0.3828125, - "learning_rate": 6.070110468840617e-05, - "loss": 0.4531, - "step": 6252 - }, - { - "epoch": 5.651309846431798, - "grad_norm": 0.3828125, - "learning_rate": 6.052844272500836e-05, - "loss": 0.4546, - "step": 6256 - }, - { - "epoch": 5.654923215898826, - "grad_norm": 0.400390625, - "learning_rate": 6.0355964585790454e-05, - "loss": 0.449, - "step": 6260 - }, - { - "epoch": 5.658536585365853, - "grad_norm": 0.423828125, - "learning_rate": 6.018367062511799e-05, - "loss": 0.4686, - "step": 6264 - }, - { - "epoch": 5.6621499548328815, - "grad_norm": 0.375, - "learning_rate": 6.001156119697846e-05, - "loss": 0.4225, - "step": 6268 - }, - { - "epoch": 5.66576332429991, - "grad_norm": 0.380859375, - "learning_rate": 5.983963665497993e-05, - "loss": 0.4427, - "step": 6272 - }, - { - "epoch": 5.669376693766938, - "grad_norm": 0.39453125, - "learning_rate": 5.966789735235082e-05, - "loss": 0.4798, - "step": 6276 - }, - { - "epoch": 5.672990063233966, - "grad_norm": 0.37890625, - "learning_rate": 5.949634364193883e-05, - "loss": 0.4689, - "step": 6280 - }, - { - "epoch": 5.676603432700993, - "grad_norm": 0.376953125, - "learning_rate": 5.9324975876210376e-05, - "loss": 0.4478, - "step": 6284 - }, - { - "epoch": 5.6802168021680215, - "grad_norm": 0.36328125, - "learning_rate": 5.915379440724986e-05, - "loss": 0.5014, - "step": 6288 - }, - { - "epoch": 5.68383017163505, - "grad_norm": 0.4375, - "learning_rate": 5.898279958675886e-05, - "loss": 0.4768, - "step": 6292 - }, - { - "epoch": 5.687443541102078, - "grad_norm": 0.396484375, - "learning_rate": 5.8811991766055636e-05, - "loss": 0.4688, - "step": 6296 - }, - { - "epoch": 5.691056910569106, - "grad_norm": 0.3671875, - "learning_rate": 5.864137129607409e-05, - "loss": 0.4298, - "step": 6300 - }, - { - "epoch": 5.694670280036133, - "grad_norm": 0.388671875, - "learning_rate": 5.847093852736326e-05, - "loss": 0.4591, - "step": 6304 - }, - { - "epoch": 5.6982836495031615, - "grad_norm": 0.357421875, - "learning_rate": 5.8300693810086485e-05, - "loss": 0.4219, - "step": 6308 - }, - { - "epoch": 5.70189701897019, - "grad_norm": 0.36328125, - "learning_rate": 5.813063749402089e-05, - "loss": 0.4736, - "step": 6312 - }, - { - "epoch": 5.705510388437218, - "grad_norm": 0.40625, - "learning_rate": 5.796076992855645e-05, - "loss": 0.4435, - "step": 6316 - }, - { - "epoch": 5.709123757904246, - "grad_norm": 0.359375, - "learning_rate": 5.7791091462695134e-05, - "loss": 0.4688, - "step": 6320 - }, - { - "epoch": 5.712737127371273, - "grad_norm": 0.384765625, - "learning_rate": 5.7621602445050744e-05, - "loss": 0.4473, - "step": 6324 - }, - { - "epoch": 5.716350496838301, - "grad_norm": 0.39453125, - "learning_rate": 5.745230322384757e-05, - "loss": 0.4365, - "step": 6328 - }, - { - "epoch": 5.71996386630533, - "grad_norm": 0.384765625, - "learning_rate": 5.728319414692019e-05, - "loss": 0.45, - "step": 6332 - }, - { - "epoch": 5.723577235772358, - "grad_norm": 0.390625, - "learning_rate": 5.7114275561712394e-05, - "loss": 0.4337, - "step": 6336 - }, - { - "epoch": 5.727190605239386, - "grad_norm": 0.369140625, - "learning_rate": 5.694554781527643e-05, - "loss": 0.4329, - "step": 6340 - }, - { - "epoch": 5.730803974706414, - "grad_norm": 0.361328125, - "learning_rate": 5.677701125427278e-05, - "loss": 0.459, - "step": 6344 - }, - { - "epoch": 5.734417344173441, - "grad_norm": 0.416015625, - "learning_rate": 5.66086662249689e-05, - "loss": 0.4699, - "step": 6348 - }, - { - "epoch": 5.73803071364047, - "grad_norm": 0.375, - "learning_rate": 5.6440513073238885e-05, - "loss": 0.4946, - "step": 6352 - }, - { - "epoch": 5.741644083107498, - "grad_norm": 0.369140625, - "learning_rate": 5.627255214456249e-05, - "loss": 0.4601, - "step": 6356 - }, - { - "epoch": 5.745257452574526, - "grad_norm": 0.37890625, - "learning_rate": 5.610478378402457e-05, - "loss": 0.4391, - "step": 6360 - }, - { - "epoch": 5.748870822041554, - "grad_norm": 0.380859375, - "learning_rate": 5.593720833631434e-05, - "loss": 0.4241, - "step": 6364 - }, - { - "epoch": 5.752484191508581, - "grad_norm": 0.375, - "learning_rate": 5.576982614572464e-05, - "loss": 0.4315, - "step": 6368 - }, - { - "epoch": 5.7560975609756095, - "grad_norm": 0.376953125, - "learning_rate": 5.560263755615137e-05, - "loss": 0.4534, - "step": 6372 - }, - { - "epoch": 5.759710930442638, - "grad_norm": 0.373046875, - "learning_rate": 5.5435642911092534e-05, - "loss": 0.4628, - "step": 6376 - }, - { - "epoch": 5.763324299909666, - "grad_norm": 0.388671875, - "learning_rate": 5.526884255364771e-05, - "loss": 0.4114, - "step": 6380 - }, - { - "epoch": 5.766937669376694, - "grad_norm": 0.421875, - "learning_rate": 5.5102236826517234e-05, - "loss": 0.495, - "step": 6384 - }, - { - "epoch": 5.770551038843722, - "grad_norm": 0.36328125, - "learning_rate": 5.493582607200176e-05, - "loss": 0.4702, - "step": 6388 - }, - { - "epoch": 5.7741644083107495, - "grad_norm": 0.38671875, - "learning_rate": 5.4769610632001164e-05, - "loss": 0.478, - "step": 6392 - }, - { - "epoch": 5.777777777777778, - "grad_norm": 0.375, - "learning_rate": 5.460359084801411e-05, - "loss": 0.4456, - "step": 6396 - }, - { - "epoch": 5.781391147244806, - "grad_norm": 0.392578125, - "learning_rate": 5.4437767061137285e-05, - "loss": 0.4366, - "step": 6400 - }, - { - "epoch": 5.785004516711834, - "grad_norm": 0.376953125, - "learning_rate": 5.42721396120646e-05, - "loss": 0.4836, - "step": 6404 - }, - { - "epoch": 5.788617886178862, - "grad_norm": 0.376953125, - "learning_rate": 5.41067088410868e-05, - "loss": 0.4784, - "step": 6408 - }, - { - "epoch": 5.7922312556458895, - "grad_norm": 0.357421875, - "learning_rate": 5.394147508809036e-05, - "loss": 0.4618, - "step": 6412 - }, - { - "epoch": 5.795844625112918, - "grad_norm": 0.396484375, - "learning_rate": 5.377643869255687e-05, - "loss": 0.4266, - "step": 6416 - }, - { - "epoch": 5.799457994579946, - "grad_norm": 0.380859375, - "learning_rate": 5.3611599993562726e-05, - "loss": 0.4439, - "step": 6420 - }, - { - "epoch": 5.803071364046974, - "grad_norm": 0.365234375, - "learning_rate": 5.344695932977791e-05, - "loss": 0.4716, - "step": 6424 - }, - { - "epoch": 5.806684733514002, - "grad_norm": 0.3671875, - "learning_rate": 5.328251703946573e-05, - "loss": 0.4725, - "step": 6428 - }, - { - "epoch": 5.8102981029810294, - "grad_norm": 0.3828125, - "learning_rate": 5.311827346048174e-05, - "loss": 0.4721, - "step": 6432 - }, - { - "epoch": 5.813911472448058, - "grad_norm": 0.38671875, - "learning_rate": 5.2954228930273306e-05, - "loss": 0.4933, - "step": 6436 - }, - { - "epoch": 5.817524841915086, - "grad_norm": 0.376953125, - "learning_rate": 5.279038378587884e-05, - "loss": 0.4656, - "step": 6440 - }, - { - "epoch": 5.821138211382114, - "grad_norm": 0.375, - "learning_rate": 5.262673836392707e-05, - "loss": 0.4642, - "step": 6444 - }, - { - "epoch": 5.824751580849142, - "grad_norm": 0.390625, - "learning_rate": 5.2463293000636485e-05, - "loss": 0.4859, - "step": 6448 - }, - { - "epoch": 5.828364950316169, - "grad_norm": 0.373046875, - "learning_rate": 5.230004803181444e-05, - "loss": 0.477, - "step": 6452 - }, - { - "epoch": 5.831978319783198, - "grad_norm": 0.3671875, - "learning_rate": 5.213700379285659e-05, - "loss": 0.4692, - "step": 6456 - }, - { - "epoch": 5.835591689250226, - "grad_norm": 0.38671875, - "learning_rate": 5.197416061874615e-05, - "loss": 0.4841, - "step": 6460 - }, - { - "epoch": 5.839205058717254, - "grad_norm": 0.388671875, - "learning_rate": 5.181151884405336e-05, - "loss": 0.4608, - "step": 6464 - }, - { - "epoch": 5.842818428184282, - "grad_norm": 0.39453125, - "learning_rate": 5.1649078802934575e-05, - "loss": 0.4724, - "step": 6468 - }, - { - "epoch": 5.846431797651309, - "grad_norm": 0.361328125, - "learning_rate": 5.1486840829131664e-05, - "loss": 0.4707, - "step": 6472 - }, - { - "epoch": 5.8500451671183376, - "grad_norm": 0.37109375, - "learning_rate": 5.132480525597137e-05, - "loss": 0.4186, - "step": 6476 - }, - { - "epoch": 5.853658536585366, - "grad_norm": 0.37890625, - "learning_rate": 5.116297241636455e-05, - "loss": 0.4941, - "step": 6480 - }, - { - "epoch": 5.857271906052394, - "grad_norm": 0.400390625, - "learning_rate": 5.100134264280566e-05, - "loss": 0.4667, - "step": 6484 - }, - { - "epoch": 5.860885275519422, - "grad_norm": 0.390625, - "learning_rate": 5.0839916267371863e-05, - "loss": 0.466, - "step": 6488 - }, - { - "epoch": 5.86449864498645, - "grad_norm": 0.369140625, - "learning_rate": 5.067869362172227e-05, - "loss": 0.4296, - "step": 6492 - }, - { - "epoch": 5.8681120144534775, - "grad_norm": 0.3984375, - "learning_rate": 5.051767503709776e-05, - "loss": 0.4593, - "step": 6496 - }, - { - "epoch": 5.871725383920506, - "grad_norm": 0.361328125, - "learning_rate": 5.035686084431961e-05, - "loss": 0.4522, - "step": 6500 - }, - { - "epoch": 5.875338753387534, - "grad_norm": 0.37109375, - "learning_rate": 5.0196251373789485e-05, - "loss": 0.4357, - "step": 6504 - }, - { - "epoch": 5.878952122854562, - "grad_norm": 0.369140625, - "learning_rate": 5.003584695548819e-05, - "loss": 0.467, - "step": 6508 - }, - { - "epoch": 5.88256549232159, - "grad_norm": 0.396484375, - "learning_rate": 4.987564791897532e-05, - "loss": 0.4651, - "step": 6512 - }, - { - "epoch": 5.886178861788618, - "grad_norm": 0.3828125, - "learning_rate": 4.971565459338854e-05, - "loss": 0.4648, - "step": 6516 - }, - { - "epoch": 5.889792231255646, - "grad_norm": 0.380859375, - "learning_rate": 4.9555867307442745e-05, - "loss": 0.4459, - "step": 6520 - }, - { - "epoch": 5.893405600722674, - "grad_norm": 0.380859375, - "learning_rate": 4.939628638942975e-05, - "loss": 0.4776, - "step": 6524 - }, - { - "epoch": 5.897018970189702, - "grad_norm": 0.3828125, - "learning_rate": 4.9236912167217154e-05, - "loss": 0.4891, - "step": 6528 - }, - { - "epoch": 5.90063233965673, - "grad_norm": 0.396484375, - "learning_rate": 4.9077744968247956e-05, - "loss": 0.4506, - "step": 6532 - }, - { - "epoch": 5.904245709123758, - "grad_norm": 0.37109375, - "learning_rate": 4.891878511953977e-05, - "loss": 0.4284, - "step": 6536 - }, - { - "epoch": 5.907859078590786, - "grad_norm": 0.376953125, - "learning_rate": 4.876003294768437e-05, - "loss": 0.4613, - "step": 6540 - }, - { - "epoch": 5.911472448057814, - "grad_norm": 0.37890625, - "learning_rate": 4.860148877884667e-05, - "loss": 0.4556, - "step": 6544 - }, - { - "epoch": 5.915085817524842, - "grad_norm": 0.40234375, - "learning_rate": 4.844315293876428e-05, - "loss": 0.4644, - "step": 6548 - }, - { - "epoch": 5.91869918699187, - "grad_norm": 0.388671875, - "learning_rate": 4.8285025752746794e-05, - "loss": 0.473, - "step": 6552 - }, - { - "epoch": 5.922312556458898, - "grad_norm": 0.38671875, - "learning_rate": 4.812710754567507e-05, - "loss": 0.4223, - "step": 6556 - }, - { - "epoch": 5.925925925925926, - "grad_norm": 0.40625, - "learning_rate": 4.7969398642000745e-05, - "loss": 0.4756, - "step": 6560 - }, - { - "epoch": 5.929539295392954, - "grad_norm": 0.396484375, - "learning_rate": 4.781189936574534e-05, - "loss": 0.4198, - "step": 6564 - }, - { - "epoch": 5.933152664859982, - "grad_norm": 0.37890625, - "learning_rate": 4.7654610040499535e-05, - "loss": 0.4755, - "step": 6568 - }, - { - "epoch": 5.93676603432701, - "grad_norm": 0.400390625, - "learning_rate": 4.749753098942295e-05, - "loss": 0.4607, - "step": 6572 - }, - { - "epoch": 5.940379403794038, - "grad_norm": 0.390625, - "learning_rate": 4.734066253524295e-05, - "loss": 0.4254, - "step": 6576 - }, - { - "epoch": 5.943992773261066, - "grad_norm": 0.384765625, - "learning_rate": 4.718400500025437e-05, - "loss": 0.4586, - "step": 6580 - }, - { - "epoch": 5.947606142728094, - "grad_norm": 0.3828125, - "learning_rate": 4.702755870631862e-05, - "loss": 0.4384, - "step": 6584 - }, - { - "epoch": 5.951219512195122, - "grad_norm": 0.3671875, - "learning_rate": 4.68713239748631e-05, - "loss": 0.3966, - "step": 6588 - }, - { - "epoch": 5.95483288166215, - "grad_norm": 0.384765625, - "learning_rate": 4.6715301126880574e-05, - "loss": 0.4167, - "step": 6592 - }, - { - "epoch": 5.958446251129178, - "grad_norm": 0.36328125, - "learning_rate": 4.6559490482928415e-05, - "loss": 0.4627, - "step": 6596 - }, - { - "epoch": 5.9620596205962055, - "grad_norm": 0.376953125, - "learning_rate": 4.640389236312816e-05, - "loss": 0.4451, - "step": 6600 - }, - { - "epoch": 5.965672990063234, - "grad_norm": 0.375, - "learning_rate": 4.6248507087164544e-05, - "loss": 0.4344, - "step": 6604 - }, - { - "epoch": 5.969286359530262, - "grad_norm": 0.3828125, - "learning_rate": 4.609333497428508e-05, - "loss": 0.4447, - "step": 6608 - }, - { - "epoch": 5.97289972899729, - "grad_norm": 0.376953125, - "learning_rate": 4.593837634329928e-05, - "loss": 0.4399, - "step": 6612 - }, - { - "epoch": 5.976513098464318, - "grad_norm": 0.388671875, - "learning_rate": 4.578363151257815e-05, - "loss": 0.459, - "step": 6616 - }, - { - "epoch": 5.9801264679313455, - "grad_norm": 0.390625, - "learning_rate": 4.5629100800053315e-05, - "loss": 0.452, - "step": 6620 - }, - { - "epoch": 5.983739837398374, - "grad_norm": 0.369140625, - "learning_rate": 4.547478452321654e-05, - "loss": 0.4198, - "step": 6624 - }, - { - "epoch": 5.987353206865402, - "grad_norm": 0.373046875, - "learning_rate": 4.532068299911899e-05, - "loss": 0.4881, - "step": 6628 - }, - { - "epoch": 5.99096657633243, - "grad_norm": 0.404296875, - "learning_rate": 4.5166796544370584e-05, - "loss": 0.4498, - "step": 6632 - }, - { - "epoch": 5.994579945799458, - "grad_norm": 0.384765625, - "learning_rate": 4.501312547513953e-05, - "loss": 0.4678, - "step": 6636 - }, - { - "epoch": 5.998193315266486, - "grad_norm": 0.373046875, - "learning_rate": 4.485967010715131e-05, - "loss": 0.4275, - "step": 6640 - }, - { - "epoch": 6.001806684733514, - "grad_norm": 0.359375, - "learning_rate": 4.4706430755688363e-05, - "loss": 0.4462, - "step": 6644 - }, - { - "epoch": 6.005420054200542, - "grad_norm": 0.353515625, - "learning_rate": 4.455340773558923e-05, - "loss": 0.4047, - "step": 6648 - }, - { - "epoch": 6.00903342366757, - "grad_norm": 0.36328125, - "learning_rate": 4.4400601361248e-05, - "loss": 0.4123, - "step": 6652 - }, - { - "epoch": 6.012646793134598, - "grad_norm": 0.357421875, - "learning_rate": 4.424801194661377e-05, - "loss": 0.4104, - "step": 6656 - }, - { - "epoch": 6.016260162601626, - "grad_norm": 0.396484375, - "learning_rate": 4.409563980518975e-05, - "loss": 0.4238, - "step": 6660 - }, - { - "epoch": 6.019873532068654, - "grad_norm": 0.3828125, - "learning_rate": 4.394348525003279e-05, - "loss": 0.4041, - "step": 6664 - }, - { - "epoch": 6.023486901535682, - "grad_norm": 0.361328125, - "learning_rate": 4.379154859375268e-05, - "loss": 0.431, - "step": 6668 - }, - { - "epoch": 6.02710027100271, - "grad_norm": 0.36328125, - "learning_rate": 4.363983014851151e-05, - "loss": 0.3976, - "step": 6672 - }, - { - "epoch": 6.030713640469738, - "grad_norm": 0.36328125, - "learning_rate": 4.3488330226023183e-05, - "loss": 0.4231, - "step": 6676 - }, - { - "epoch": 6.034327009936766, - "grad_norm": 0.36328125, - "learning_rate": 4.333704913755244e-05, - "loss": 0.4579, - "step": 6680 - }, - { - "epoch": 6.0379403794037945, - "grad_norm": 0.380859375, - "learning_rate": 4.318598719391455e-05, - "loss": 0.4168, - "step": 6684 - }, - { - "epoch": 6.041553748870822, - "grad_norm": 0.412109375, - "learning_rate": 4.303514470547442e-05, - "loss": 0.4644, - "step": 6688 - }, - { - "epoch": 6.04516711833785, - "grad_norm": 0.3671875, - "learning_rate": 4.288452198214625e-05, - "loss": 0.4276, - "step": 6692 - }, - { - "epoch": 6.048780487804878, - "grad_norm": 0.396484375, - "learning_rate": 4.273411933339255e-05, - "loss": 0.4901, - "step": 6696 - }, - { - "epoch": 6.052393857271906, - "grad_norm": 0.392578125, - "learning_rate": 4.258393706822376e-05, - "loss": 0.388, - "step": 6700 - }, - { - "epoch": 6.056007226738934, - "grad_norm": 0.3515625, - "learning_rate": 4.243397549519749e-05, - "loss": 0.4287, - "step": 6704 - }, - { - "epoch": 6.059620596205962, - "grad_norm": 0.37109375, - "learning_rate": 4.228423492241793e-05, - "loss": 0.4281, - "step": 6708 - }, - { - "epoch": 6.06323396567299, - "grad_norm": 0.38671875, - "learning_rate": 4.2134715657535315e-05, - "loss": 0.4374, - "step": 6712 - }, - { - "epoch": 6.066847335140018, - "grad_norm": 0.373046875, - "learning_rate": 4.198541800774502e-05, - "loss": 0.4366, - "step": 6716 - }, - { - "epoch": 6.070460704607046, - "grad_norm": 0.380859375, - "learning_rate": 4.183634227978724e-05, - "loss": 0.4594, - "step": 6720 - }, - { - "epoch": 6.074074074074074, - "grad_norm": 0.3671875, - "learning_rate": 4.168748877994613e-05, - "loss": 0.452, - "step": 6724 - }, - { - "epoch": 6.077687443541102, - "grad_norm": 0.3984375, - "learning_rate": 4.1538857814049266e-05, - "loss": 0.4504, - "step": 6728 - }, - { - "epoch": 6.08130081300813, - "grad_norm": 0.390625, - "learning_rate": 4.139044968746713e-05, - "loss": 0.4236, - "step": 6732 - }, - { - "epoch": 6.084914182475158, - "grad_norm": 0.375, - "learning_rate": 4.124226470511228e-05, - "loss": 0.4384, - "step": 6736 - }, - { - "epoch": 6.088527551942186, - "grad_norm": 0.35546875, - "learning_rate": 4.109430317143876e-05, - "loss": 0.4464, - "step": 6740 - }, - { - "epoch": 6.092140921409214, - "grad_norm": 0.376953125, - "learning_rate": 4.0946565390441636e-05, - "loss": 0.4661, - "step": 6744 - }, - { - "epoch": 6.095754290876242, - "grad_norm": 0.361328125, - "learning_rate": 4.079905166565616e-05, - "loss": 0.4124, - "step": 6748 - }, - { - "epoch": 6.09936766034327, - "grad_norm": 0.390625, - "learning_rate": 4.065176230015737e-05, - "loss": 0.3999, - "step": 6752 - }, - { - "epoch": 6.102981029810298, - "grad_norm": 0.41015625, - "learning_rate": 4.0504697596559244e-05, - "loss": 0.4859, - "step": 6756 - }, - { - "epoch": 6.106594399277326, - "grad_norm": 0.38671875, - "learning_rate": 4.035785785701422e-05, - "loss": 0.4529, - "step": 6760 - }, - { - "epoch": 6.110207768744354, - "grad_norm": 0.4140625, - "learning_rate": 4.021124338321249e-05, - "loss": 0.4509, - "step": 6764 - }, - { - "epoch": 6.1138211382113825, - "grad_norm": 0.400390625, - "learning_rate": 4.0064854476381525e-05, - "loss": 0.4201, - "step": 6768 - }, - { - "epoch": 6.11743450767841, - "grad_norm": 0.390625, - "learning_rate": 3.991869143728526e-05, - "loss": 0.4354, - "step": 6772 - }, - { - "epoch": 6.121047877145438, - "grad_norm": 0.357421875, - "learning_rate": 3.977275456622363e-05, - "loss": 0.4321, - "step": 6776 - }, - { - "epoch": 6.124661246612466, - "grad_norm": 0.396484375, - "learning_rate": 3.962704416303184e-05, - "loss": 0.4153, - "step": 6780 - }, - { - "epoch": 6.128274616079494, - "grad_norm": 0.34375, - "learning_rate": 3.948156052707982e-05, - "loss": 0.4049, - "step": 6784 - }, - { - "epoch": 6.1318879855465225, - "grad_norm": 0.392578125, - "learning_rate": 3.933630395727167e-05, - "loss": 0.4081, - "step": 6788 - }, - { - "epoch": 6.13550135501355, - "grad_norm": 0.369140625, - "learning_rate": 3.91912747520449e-05, - "loss": 0.4504, - "step": 6792 - }, - { - "epoch": 6.139114724480578, - "grad_norm": 0.37109375, - "learning_rate": 3.9046473209369874e-05, - "loss": 0.4187, - "step": 6796 - }, - { - "epoch": 6.142728093947606, - "grad_norm": 0.38671875, - "learning_rate": 3.890189962674926e-05, - "loss": 0.4668, - "step": 6800 - }, - { - "epoch": 6.146341463414634, - "grad_norm": 0.33984375, - "learning_rate": 3.8757554301217266e-05, - "loss": 0.4381, - "step": 6804 - }, - { - "epoch": 6.1499548328816624, - "grad_norm": 0.3984375, - "learning_rate": 3.8613437529339335e-05, - "loss": 0.4509, - "step": 6808 - }, - { - "epoch": 6.15356820234869, - "grad_norm": 0.3671875, - "learning_rate": 3.8469549607211156e-05, - "loss": 0.4341, - "step": 6812 - }, - { - "epoch": 6.157181571815718, - "grad_norm": 0.35546875, - "learning_rate": 3.832589083045827e-05, - "loss": 0.4367, - "step": 6816 - }, - { - "epoch": 6.160794941282746, - "grad_norm": 0.373046875, - "learning_rate": 3.818246149423548e-05, - "loss": 0.4402, - "step": 6820 - }, - { - "epoch": 6.164408310749774, - "grad_norm": 0.40625, - "learning_rate": 3.8039261893226106e-05, - "loss": 0.4786, - "step": 6824 - }, - { - "epoch": 6.168021680216802, - "grad_norm": 0.384765625, - "learning_rate": 3.789629232164157e-05, - "loss": 0.4196, - "step": 6828 - }, - { - "epoch": 6.171635049683831, - "grad_norm": 0.390625, - "learning_rate": 3.775355307322063e-05, - "loss": 0.3895, - "step": 6832 - }, - { - "epoch": 6.175248419150858, - "grad_norm": 0.392578125, - "learning_rate": 3.76110444412288e-05, - "loss": 0.3994, - "step": 6836 - }, - { - "epoch": 6.178861788617886, - "grad_norm": 0.373046875, - "learning_rate": 3.7468766718457794e-05, - "loss": 0.4655, - "step": 6840 - }, - { - "epoch": 6.182475158084914, - "grad_norm": 0.388671875, - "learning_rate": 3.732672019722501e-05, - "loss": 0.4477, - "step": 6844 - }, - { - "epoch": 6.186088527551942, - "grad_norm": 0.3828125, - "learning_rate": 3.718490516937271e-05, - "loss": 0.4314, - "step": 6848 - }, - { - "epoch": 6.1897018970189706, - "grad_norm": 0.38671875, - "learning_rate": 3.7043321926267576e-05, - "loss": 0.4109, - "step": 6852 - }, - { - "epoch": 6.193315266485998, - "grad_norm": 0.375, - "learning_rate": 3.6901970758800064e-05, - "loss": 0.4238, - "step": 6856 - }, - { - "epoch": 6.196928635953026, - "grad_norm": 0.359375, - "learning_rate": 3.676085195738384e-05, - "loss": 0.4011, - "step": 6860 - }, - { - "epoch": 6.200542005420054, - "grad_norm": 0.359375, - "learning_rate": 3.661996581195518e-05, - "loss": 0.4163, - "step": 6864 - }, - { - "epoch": 6.204155374887082, - "grad_norm": 0.373046875, - "learning_rate": 3.647931261197233e-05, - "loss": 0.4219, - "step": 6868 - }, - { - "epoch": 6.2077687443541105, - "grad_norm": 0.396484375, - "learning_rate": 3.633889264641493e-05, - "loss": 0.4499, - "step": 6872 - }, - { - "epoch": 6.211382113821138, - "grad_norm": 0.3984375, - "learning_rate": 3.6198706203783414e-05, - "loss": 0.4499, - "step": 6876 - }, - { - "epoch": 6.214995483288166, - "grad_norm": 0.404296875, - "learning_rate": 3.6058753572098446e-05, - "loss": 0.4586, - "step": 6880 - }, - { - "epoch": 6.218608852755194, - "grad_norm": 0.384765625, - "learning_rate": 3.5919035038900376e-05, - "loss": 0.4048, - "step": 6884 - }, - { - "epoch": 6.222222222222222, - "grad_norm": 0.380859375, - "learning_rate": 3.577955089124848e-05, - "loss": 0.4284, - "step": 6888 - }, - { - "epoch": 6.2258355916892505, - "grad_norm": 0.359375, - "learning_rate": 3.564030141572055e-05, - "loss": 0.414, - "step": 6892 - }, - { - "epoch": 6.229448961156278, - "grad_norm": 0.357421875, - "learning_rate": 3.5501286898412204e-05, - "loss": 0.4433, - "step": 6896 - }, - { - "epoch": 6.233062330623306, - "grad_norm": 0.376953125, - "learning_rate": 3.5362507624936265e-05, - "loss": 0.4141, - "step": 6900 - }, - { - "epoch": 6.236675700090334, - "grad_norm": 0.3671875, - "learning_rate": 3.522396388042238e-05, - "loss": 0.4694, - "step": 6904 - }, - { - "epoch": 6.240289069557362, - "grad_norm": 0.361328125, - "learning_rate": 3.50856559495162e-05, - "loss": 0.4377, - "step": 6908 - }, - { - "epoch": 6.2439024390243905, - "grad_norm": 0.40625, - "learning_rate": 3.494758411637884e-05, - "loss": 0.4518, - "step": 6912 - }, - { - "epoch": 6.247515808491419, - "grad_norm": 0.3828125, - "learning_rate": 3.480974866468638e-05, - "loss": 0.4469, - "step": 6916 - }, - { - "epoch": 6.251129177958446, - "grad_norm": 0.373046875, - "learning_rate": 3.4672149877629346e-05, - "loss": 0.4501, - "step": 6920 - }, - { - "epoch": 6.254742547425474, - "grad_norm": 0.359375, - "learning_rate": 3.453478803791184e-05, - "loss": 0.3961, - "step": 6924 - }, - { - "epoch": 6.258355916892502, - "grad_norm": 0.3828125, - "learning_rate": 3.439766342775127e-05, - "loss": 0.4447, - "step": 6928 - }, - { - "epoch": 6.26196928635953, - "grad_norm": 0.412109375, - "learning_rate": 3.42607763288776e-05, - "loss": 0.4497, - "step": 6932 - }, - { - "epoch": 6.265582655826559, - "grad_norm": 0.380859375, - "learning_rate": 3.412412702253277e-05, - "loss": 0.4258, - "step": 6936 - }, - { - "epoch": 6.269196025293586, - "grad_norm": 0.3671875, - "learning_rate": 3.39877157894703e-05, - "loss": 0.4462, - "step": 6940 - }, - { - "epoch": 6.272809394760614, - "grad_norm": 0.369140625, - "learning_rate": 3.385154290995443e-05, - "loss": 0.4711, - "step": 6944 - }, - { - "epoch": 6.276422764227642, - "grad_norm": 0.3671875, - "learning_rate": 3.371560866375976e-05, - "loss": 0.4123, - "step": 6948 - }, - { - "epoch": 6.28003613369467, - "grad_norm": 0.40234375, - "learning_rate": 3.3579913330170586e-05, - "loss": 0.4543, - "step": 6952 - }, - { - "epoch": 6.283649503161699, - "grad_norm": 0.37890625, - "learning_rate": 3.34444571879803e-05, - "loss": 0.4503, - "step": 6956 - }, - { - "epoch": 6.287262872628727, - "grad_norm": 0.36328125, - "learning_rate": 3.330924051549101e-05, - "loss": 0.4416, - "step": 6960 - }, - { - "epoch": 6.290876242095754, - "grad_norm": 0.373046875, - "learning_rate": 3.317426359051268e-05, - "loss": 0.4267, - "step": 6964 - }, - { - "epoch": 6.294489611562782, - "grad_norm": 0.376953125, - "learning_rate": 3.303952669036274e-05, - "loss": 0.4078, - "step": 6968 - }, - { - "epoch": 6.29810298102981, - "grad_norm": 0.3984375, - "learning_rate": 3.290503009186545e-05, - "loss": 0.4495, - "step": 6972 - }, - { - "epoch": 6.3017163504968385, - "grad_norm": 0.38671875, - "learning_rate": 3.27707740713514e-05, - "loss": 0.4247, - "step": 6976 - }, - { - "epoch": 6.305329719963867, - "grad_norm": 0.390625, - "learning_rate": 3.263675890465692e-05, - "loss": 0.4255, - "step": 6980 - }, - { - "epoch": 6.308943089430894, - "grad_norm": 0.369140625, - "learning_rate": 3.250298486712345e-05, - "loss": 0.448, - "step": 6984 - }, - { - "epoch": 6.312556458897922, - "grad_norm": 0.37109375, - "learning_rate": 3.2369452233596994e-05, - "loss": 0.3959, - "step": 6988 - }, - { - "epoch": 6.31616982836495, - "grad_norm": 0.400390625, - "learning_rate": 3.22361612784276e-05, - "loss": 0.4513, - "step": 6992 - }, - { - "epoch": 6.3197831978319785, - "grad_norm": 0.37109375, - "learning_rate": 3.210311227546884e-05, - "loss": 0.4498, - "step": 6996 - }, - { - "epoch": 6.323396567299007, - "grad_norm": 0.388671875, - "learning_rate": 3.197030549807708e-05, - "loss": 0.491, - "step": 7000 - }, - { - "epoch": 6.327009936766034, - "grad_norm": 0.37109375, - "learning_rate": 3.183774121911109e-05, - "loss": 0.436, - "step": 7004 - }, - { - "epoch": 6.330623306233062, - "grad_norm": 0.3984375, - "learning_rate": 3.1705419710931377e-05, - "loss": 0.438, - "step": 7008 - }, - { - "epoch": 6.33423667570009, - "grad_norm": 0.3515625, - "learning_rate": 3.1573341245399645e-05, - "loss": 0.4001, - "step": 7012 - }, - { - "epoch": 6.3378500451671185, - "grad_norm": 0.38671875, - "learning_rate": 3.1441506093878346e-05, - "loss": 0.4078, - "step": 7016 - }, - { - "epoch": 6.341463414634147, - "grad_norm": 0.36328125, - "learning_rate": 3.1309914527229934e-05, - "loss": 0.4194, - "step": 7020 - }, - { - "epoch": 6.345076784101174, - "grad_norm": 0.361328125, - "learning_rate": 3.117856681581645e-05, - "loss": 0.4423, - "step": 7024 - }, - { - "epoch": 6.348690153568202, - "grad_norm": 0.3671875, - "learning_rate": 3.10474632294989e-05, - "loss": 0.4444, - "step": 7028 - }, - { - "epoch": 6.35230352303523, - "grad_norm": 0.388671875, - "learning_rate": 3.091660403763669e-05, - "loss": 0.4364, - "step": 7032 - }, - { - "epoch": 6.355916892502258, - "grad_norm": 0.44140625, - "learning_rate": 3.078598950908724e-05, - "loss": 0.4639, - "step": 7036 - }, - { - "epoch": 6.359530261969287, - "grad_norm": 0.3515625, - "learning_rate": 3.0655619912205136e-05, - "loss": 0.425, - "step": 7040 - }, - { - "epoch": 6.363143631436314, - "grad_norm": 0.39453125, - "learning_rate": 3.052549551484184e-05, - "loss": 0.4536, - "step": 7044 - }, - { - "epoch": 6.366757000903342, - "grad_norm": 0.40234375, - "learning_rate": 3.0395616584344962e-05, - "loss": 0.4423, - "step": 7048 - }, - { - "epoch": 6.37037037037037, - "grad_norm": 0.376953125, - "learning_rate": 3.026598338755783e-05, - "loss": 0.4083, - "step": 7052 - }, - { - "epoch": 6.373983739837398, - "grad_norm": 0.369140625, - "learning_rate": 3.013659619081893e-05, - "loss": 0.4333, - "step": 7056 - }, - { - "epoch": 6.377597109304427, - "grad_norm": 0.404296875, - "learning_rate": 3.00074552599613e-05, - "loss": 0.4283, - "step": 7060 - }, - { - "epoch": 6.381210478771455, - "grad_norm": 0.36328125, - "learning_rate": 2.9878560860311967e-05, - "loss": 0.4207, - "step": 7064 - }, - { - "epoch": 6.384823848238482, - "grad_norm": 0.392578125, - "learning_rate": 2.9749913256691448e-05, - "loss": 0.4735, - "step": 7068 - }, - { - "epoch": 6.38843721770551, - "grad_norm": 0.421875, - "learning_rate": 2.9621512713413348e-05, - "loss": 0.4453, - "step": 7072 - }, - { - "epoch": 6.392050587172538, - "grad_norm": 0.419921875, - "learning_rate": 2.9493359494283507e-05, - "loss": 0.4303, - "step": 7076 - }, - { - "epoch": 6.3956639566395665, - "grad_norm": 0.384765625, - "learning_rate": 2.93654538625997e-05, - "loss": 0.4261, - "step": 7080 - }, - { - "epoch": 6.399277326106595, - "grad_norm": 0.373046875, - "learning_rate": 2.9237796081150973e-05, - "loss": 0.4499, - "step": 7084 - }, - { - "epoch": 6.402890695573622, - "grad_norm": 0.392578125, - "learning_rate": 2.9110386412217146e-05, - "loss": 0.4481, - "step": 7088 - }, - { - "epoch": 6.40650406504065, - "grad_norm": 0.365234375, - "learning_rate": 2.898322511756841e-05, - "loss": 0.4148, - "step": 7092 - }, - { - "epoch": 6.410117434507678, - "grad_norm": 0.376953125, - "learning_rate": 2.8856312458464524e-05, - "loss": 0.4011, - "step": 7096 - }, - { - "epoch": 6.4137308039747065, - "grad_norm": 0.37109375, - "learning_rate": 2.8729648695654415e-05, - "loss": 0.4418, - "step": 7100 - }, - { - "epoch": 6.417344173441735, - "grad_norm": 0.375, - "learning_rate": 2.86032340893757e-05, - "loss": 0.5009, - "step": 7104 - }, - { - "epoch": 6.420957542908763, - "grad_norm": 0.3828125, - "learning_rate": 2.847706889935401e-05, - "loss": 0.4655, - "step": 7108 - }, - { - "epoch": 6.42457091237579, - "grad_norm": 0.3828125, - "learning_rate": 2.835115338480269e-05, - "loss": 0.4362, - "step": 7112 - }, - { - "epoch": 6.428184281842818, - "grad_norm": 0.388671875, - "learning_rate": 2.8225487804421925e-05, - "loss": 0.4247, - "step": 7116 - }, - { - "epoch": 6.4317976513098465, - "grad_norm": 0.384765625, - "learning_rate": 2.810007241639853e-05, - "loss": 0.4774, - "step": 7120 - }, - { - "epoch": 6.435411020776875, - "grad_norm": 0.388671875, - "learning_rate": 2.797490747840518e-05, - "loss": 0.4086, - "step": 7124 - }, - { - "epoch": 6.439024390243903, - "grad_norm": 0.359375, - "learning_rate": 2.784999324760012e-05, - "loss": 0.4817, - "step": 7128 - }, - { - "epoch": 6.44263775971093, - "grad_norm": 0.390625, - "learning_rate": 2.7725329980626414e-05, - "loss": 0.3906, - "step": 7132 - }, - { - "epoch": 6.446251129177958, - "grad_norm": 0.37109375, - "learning_rate": 2.7600917933611493e-05, - "loss": 0.4507, - "step": 7136 - }, - { - "epoch": 6.4498644986449865, - "grad_norm": 0.40625, - "learning_rate": 2.747675736216669e-05, - "loss": 0.4406, - "step": 7140 - }, - { - "epoch": 6.453477868112015, - "grad_norm": 0.408203125, - "learning_rate": 2.7352848521386616e-05, - "loss": 0.4542, - "step": 7144 - }, - { - "epoch": 6.457091237579043, - "grad_norm": 0.37890625, - "learning_rate": 2.7229191665848797e-05, - "loss": 0.4176, - "step": 7148 - }, - { - "epoch": 6.46070460704607, - "grad_norm": 0.390625, - "learning_rate": 2.7105787049612927e-05, - "loss": 0.3961, - "step": 7152 - }, - { - "epoch": 6.464317976513098, - "grad_norm": 0.373046875, - "learning_rate": 2.6982634926220487e-05, - "loss": 0.4227, - "step": 7156 - }, - { - "epoch": 6.467931345980126, - "grad_norm": 0.36328125, - "learning_rate": 2.6859735548694234e-05, - "loss": 0.4482, - "step": 7160 - }, - { - "epoch": 6.471544715447155, - "grad_norm": 0.400390625, - "learning_rate": 2.6737089169537566e-05, - "loss": 0.4689, - "step": 7164 - }, - { - "epoch": 6.475158084914183, - "grad_norm": 0.36328125, - "learning_rate": 2.6614696040734225e-05, - "loss": 0.419, - "step": 7168 - }, - { - "epoch": 6.47877145438121, - "grad_norm": 0.365234375, - "learning_rate": 2.64925564137475e-05, - "loss": 0.4143, - "step": 7172 - }, - { - "epoch": 6.482384823848238, - "grad_norm": 0.357421875, - "learning_rate": 2.6370670539519915e-05, - "loss": 0.4473, - "step": 7176 - }, - { - "epoch": 6.485998193315266, - "grad_norm": 0.37109375, - "learning_rate": 2.624903866847262e-05, - "loss": 0.4058, - "step": 7180 - }, - { - "epoch": 6.489611562782295, - "grad_norm": 0.40625, - "learning_rate": 2.6127661050504865e-05, - "loss": 0.4485, - "step": 7184 - }, - { - "epoch": 6.493224932249323, - "grad_norm": 0.3671875, - "learning_rate": 2.6006537934993636e-05, - "loss": 0.4406, - "step": 7188 - }, - { - "epoch": 6.496838301716351, - "grad_norm": 0.359375, - "learning_rate": 2.588566957079294e-05, - "loss": 0.4612, - "step": 7192 - }, - { - "epoch": 6.500451671183378, - "grad_norm": 0.39453125, - "learning_rate": 2.576505620623339e-05, - "loss": 0.4491, - "step": 7196 - }, - { - "epoch": 6.504065040650406, - "grad_norm": 0.388671875, - "learning_rate": 2.564469808912169e-05, - "loss": 0.3919, - "step": 7200 - }, - { - "epoch": 6.5076784101174345, - "grad_norm": 0.349609375, - "learning_rate": 2.5524595466740178e-05, - "loss": 0.4153, - "step": 7204 - }, - { - "epoch": 6.511291779584463, - "grad_norm": 0.388671875, - "learning_rate": 2.5404748585846203e-05, - "loss": 0.4253, - "step": 7208 - }, - { - "epoch": 6.514905149051491, - "grad_norm": 0.390625, - "learning_rate": 2.528515769267168e-05, - "loss": 0.4433, - "step": 7212 - }, - { - "epoch": 6.518518518518518, - "grad_norm": 0.373046875, - "learning_rate": 2.5165823032922638e-05, - "loss": 0.4563, - "step": 7216 - }, - { - "epoch": 6.522131887985546, - "grad_norm": 0.375, - "learning_rate": 2.504674485177855e-05, - "loss": 0.4117, - "step": 7220 - }, - { - "epoch": 6.5257452574525745, - "grad_norm": 0.37109375, - "learning_rate": 2.492792339389213e-05, - "loss": 0.4138, - "step": 7224 - }, - { - "epoch": 6.529358626919603, - "grad_norm": 0.390625, - "learning_rate": 2.480935890338845e-05, - "loss": 0.4567, - "step": 7228 - }, - { - "epoch": 6.532971996386631, - "grad_norm": 0.375, - "learning_rate": 2.4691051623864717e-05, - "loss": 0.4305, - "step": 7232 - }, - { - "epoch": 6.536585365853659, - "grad_norm": 0.3828125, - "learning_rate": 2.4573001798389664e-05, - "loss": 0.469, - "step": 7236 - }, - { - "epoch": 6.540198735320686, - "grad_norm": 0.349609375, - "learning_rate": 2.4455209669503045e-05, - "loss": 0.4539, - "step": 7240 - }, - { - "epoch": 6.5438121047877145, - "grad_norm": 0.376953125, - "learning_rate": 2.433767547921529e-05, - "loss": 0.4723, - "step": 7244 - }, - { - "epoch": 6.547425474254743, - "grad_norm": 0.37890625, - "learning_rate": 2.4220399469006717e-05, - "loss": 0.4098, - "step": 7248 - }, - { - "epoch": 6.551038843721771, - "grad_norm": 0.369140625, - "learning_rate": 2.4103381879827273e-05, - "loss": 0.444, - "step": 7252 - }, - { - "epoch": 6.554652213188799, - "grad_norm": 0.37109375, - "learning_rate": 2.398662295209595e-05, - "loss": 0.4071, - "step": 7256 - }, - { - "epoch": 6.558265582655826, - "grad_norm": 0.380859375, - "learning_rate": 2.387012292570032e-05, - "loss": 0.421, - "step": 7260 - }, - { - "epoch": 6.561878952122854, - "grad_norm": 0.390625, - "learning_rate": 2.3753882039996042e-05, - "loss": 0.4232, - "step": 7264 - }, - { - "epoch": 6.565492321589883, - "grad_norm": 0.369140625, - "learning_rate": 2.3637900533806333e-05, - "loss": 0.3635, - "step": 7268 - }, - { - "epoch": 6.569105691056911, - "grad_norm": 0.384765625, - "learning_rate": 2.352217864542149e-05, - "loss": 0.4244, - "step": 7272 - }, - { - "epoch": 6.572719060523939, - "grad_norm": 0.3828125, - "learning_rate": 2.3406716612598386e-05, - "loss": 0.4546, - "step": 7276 - }, - { - "epoch": 6.576332429990966, - "grad_norm": 0.392578125, - "learning_rate": 2.3291514672560142e-05, - "loss": 0.445, - "step": 7280 - }, - { - "epoch": 6.579945799457994, - "grad_norm": 0.400390625, - "learning_rate": 2.317657306199536e-05, - "loss": 0.4254, - "step": 7284 - }, - { - "epoch": 6.583559168925023, - "grad_norm": 0.3984375, - "learning_rate": 2.3061892017057815e-05, - "loss": 0.3915, - "step": 7288 - }, - { - "epoch": 6.587172538392051, - "grad_norm": 0.408203125, - "learning_rate": 2.2947471773365954e-05, - "loss": 0.4793, - "step": 7292 - }, - { - "epoch": 6.590785907859079, - "grad_norm": 0.380859375, - "learning_rate": 2.283331256600238e-05, - "loss": 0.4429, - "step": 7296 - }, - { - "epoch": 6.594399277326106, - "grad_norm": 0.390625, - "learning_rate": 2.2719414629513443e-05, - "loss": 0.4268, - "step": 7300 - }, - { - "epoch": 6.598012646793134, - "grad_norm": 0.369140625, - "learning_rate": 2.2605778197908604e-05, - "loss": 0.423, - "step": 7304 - }, - { - "epoch": 6.6016260162601625, - "grad_norm": 0.3828125, - "learning_rate": 2.24924035046601e-05, - "loss": 0.4656, - "step": 7308 - }, - { - "epoch": 6.605239385727191, - "grad_norm": 0.3671875, - "learning_rate": 2.2379290782702425e-05, - "loss": 0.4556, - "step": 7312 - }, - { - "epoch": 6.608852755194219, - "grad_norm": 0.380859375, - "learning_rate": 2.2266440264431747e-05, - "loss": 0.4469, - "step": 7316 - }, - { - "epoch": 6.612466124661246, - "grad_norm": 0.3984375, - "learning_rate": 2.2153852181705677e-05, - "loss": 0.4537, - "step": 7320 - }, - { - "epoch": 6.616079494128274, - "grad_norm": 0.3671875, - "learning_rate": 2.2041526765842522e-05, - "loss": 0.4403, - "step": 7324 - }, - { - "epoch": 6.6196928635953025, - "grad_norm": 0.3984375, - "learning_rate": 2.1929464247620953e-05, - "loss": 0.4521, - "step": 7328 - }, - { - "epoch": 6.623306233062331, - "grad_norm": 0.361328125, - "learning_rate": 2.1817664857279506e-05, - "loss": 0.4094, - "step": 7332 - }, - { - "epoch": 6.626919602529359, - "grad_norm": 0.375, - "learning_rate": 2.1706128824516083e-05, - "loss": 0.4411, - "step": 7336 - }, - { - "epoch": 6.630532971996386, - "grad_norm": 0.37109375, - "learning_rate": 2.1594856378487584e-05, - "loss": 0.4586, - "step": 7340 - }, - { - "epoch": 6.634146341463414, - "grad_norm": 0.41015625, - "learning_rate": 2.1483847747809286e-05, - "loss": 0.4866, - "step": 7344 - }, - { - "epoch": 6.6377597109304425, - "grad_norm": 0.373046875, - "learning_rate": 2.1373103160554426e-05, - "loss": 0.4324, - "step": 7348 - }, - { - "epoch": 6.641373080397471, - "grad_norm": 0.380859375, - "learning_rate": 2.126262284425379e-05, - "loss": 0.4169, - "step": 7352 - }, - { - "epoch": 6.644986449864499, - "grad_norm": 0.35546875, - "learning_rate": 2.1152407025895217e-05, - "loss": 0.4381, - "step": 7356 - }, - { - "epoch": 6.648599819331527, - "grad_norm": 0.3828125, - "learning_rate": 2.104245593192309e-05, - "loss": 0.432, - "step": 7360 - }, - { - "epoch": 6.652213188798554, - "grad_norm": 0.361328125, - "learning_rate": 2.0932769788237906e-05, - "loss": 0.4295, - "step": 7364 - }, - { - "epoch": 6.6558265582655824, - "grad_norm": 0.39453125, - "learning_rate": 2.0823348820195836e-05, - "loss": 0.4128, - "step": 7368 - }, - { - "epoch": 6.659439927732611, - "grad_norm": 0.412109375, - "learning_rate": 2.0714193252608125e-05, - "loss": 0.4225, - "step": 7372 - }, - { - "epoch": 6.663053297199639, - "grad_norm": 0.35546875, - "learning_rate": 2.0605303309740934e-05, - "loss": 0.4306, - "step": 7376 - }, - { - "epoch": 6.666666666666667, - "grad_norm": 0.38671875, - "learning_rate": 2.049667921531452e-05, - "loss": 0.4465, - "step": 7380 - }, - { - "epoch": 6.670280036133695, - "grad_norm": 0.408203125, - "learning_rate": 2.0388321192502982e-05, - "loss": 0.4366, - "step": 7384 - }, - { - "epoch": 6.673893405600722, - "grad_norm": 0.365234375, - "learning_rate": 2.02802294639338e-05, - "loss": 0.4409, - "step": 7388 - }, - { - "epoch": 6.677506775067751, - "grad_norm": 0.396484375, - "learning_rate": 2.0172404251687246e-05, - "loss": 0.4679, - "step": 7392 - }, - { - "epoch": 6.681120144534779, - "grad_norm": 0.37109375, - "learning_rate": 2.0064845777296163e-05, - "loss": 0.442, - "step": 7396 - }, - { - "epoch": 6.684733514001807, - "grad_norm": 0.39453125, - "learning_rate": 1.9957554261745247e-05, - "loss": 0.4684, - "step": 7400 - }, - { - "epoch": 6.688346883468835, - "grad_norm": 0.357421875, - "learning_rate": 1.9850529925470748e-05, - "loss": 0.4214, - "step": 7404 - }, - { - "epoch": 6.691960252935862, - "grad_norm": 0.3828125, - "learning_rate": 1.9743772988359997e-05, - "loss": 0.4291, - "step": 7408 - }, - { - "epoch": 6.695573622402891, - "grad_norm": 0.380859375, - "learning_rate": 1.9637283669750902e-05, - "loss": 0.4313, - "step": 7412 - }, - { - "epoch": 6.699186991869919, - "grad_norm": 0.404296875, - "learning_rate": 1.953106218843159e-05, - "loss": 0.4548, - "step": 7416 - }, - { - "epoch": 6.702800361336947, - "grad_norm": 0.36328125, - "learning_rate": 1.942510876263987e-05, - "loss": 0.3669, - "step": 7420 - }, - { - "epoch": 6.706413730803975, - "grad_norm": 0.41796875, - "learning_rate": 1.931942361006279e-05, - "loss": 0.426, - "step": 7424 - }, - { - "epoch": 6.710027100271002, - "grad_norm": 0.369140625, - "learning_rate": 1.921400694783623e-05, - "loss": 0.451, - "step": 7428 - }, - { - "epoch": 6.7136404697380305, - "grad_norm": 0.369140625, - "learning_rate": 1.9108858992544525e-05, - "loss": 0.4351, - "step": 7432 - }, - { - "epoch": 6.717253839205059, - "grad_norm": 0.38671875, - "learning_rate": 1.900397996021981e-05, - "loss": 0.4537, - "step": 7436 - }, - { - "epoch": 6.720867208672087, - "grad_norm": 0.392578125, - "learning_rate": 1.8899370066341794e-05, - "loss": 0.4404, - "step": 7440 - }, - { - "epoch": 6.724480578139115, - "grad_norm": 0.37890625, - "learning_rate": 1.879502952583715e-05, - "loss": 0.3894, - "step": 7444 - }, - { - "epoch": 6.728093947606142, - "grad_norm": 0.39453125, - "learning_rate": 1.86909585530792e-05, - "loss": 0.4797, - "step": 7448 - }, - { - "epoch": 6.7317073170731705, - "grad_norm": 0.361328125, - "learning_rate": 1.8587157361887467e-05, - "loss": 0.4323, - "step": 7452 - }, - { - "epoch": 6.735320686540199, - "grad_norm": 0.375, - "learning_rate": 1.848362616552711e-05, - "loss": 0.4711, - "step": 7456 - }, - { - "epoch": 6.738934056007227, - "grad_norm": 0.39453125, - "learning_rate": 1.8380365176708595e-05, - "loss": 0.4414, - "step": 7460 - }, - { - "epoch": 6.742547425474255, - "grad_norm": 0.38671875, - "learning_rate": 1.827737460758723e-05, - "loss": 0.4656, - "step": 7464 - }, - { - "epoch": 6.746160794941282, - "grad_norm": 0.353515625, - "learning_rate": 1.817465466976274e-05, - "loss": 0.3938, - "step": 7468 - }, - { - "epoch": 6.7497741644083105, - "grad_norm": 0.384765625, - "learning_rate": 1.8072205574278868e-05, - "loss": 0.4171, - "step": 7472 - }, - { - "epoch": 6.753387533875339, - "grad_norm": 0.38671875, - "learning_rate": 1.7970027531622817e-05, - "loss": 0.4572, - "step": 7476 - }, - { - "epoch": 6.757000903342367, - "grad_norm": 0.345703125, - "learning_rate": 1.786812075172495e-05, - "loss": 0.4105, - "step": 7480 - }, - { - "epoch": 6.760614272809395, - "grad_norm": 0.3828125, - "learning_rate": 1.776648544395826e-05, - "loss": 0.4421, - "step": 7484 - }, - { - "epoch": 6.764227642276423, - "grad_norm": 0.384765625, - "learning_rate": 1.7665121817137996e-05, - "loss": 0.4284, - "step": 7488 - }, - { - "epoch": 6.76784101174345, - "grad_norm": 0.373046875, - "learning_rate": 1.7564030079521312e-05, - "loss": 0.4323, - "step": 7492 - }, - { - "epoch": 6.771454381210479, - "grad_norm": 0.369140625, - "learning_rate": 1.746321043880664e-05, - "loss": 0.4626, - "step": 7496 - }, - { - "epoch": 6.775067750677507, - "grad_norm": 0.38671875, - "learning_rate": 1.7362663102133418e-05, - "loss": 0.4493, - "step": 7500 - }, - { - "epoch": 6.778681120144535, - "grad_norm": 0.400390625, - "learning_rate": 1.726238827608158e-05, - "loss": 0.4453, - "step": 7504 - }, - { - "epoch": 6.782294489611563, - "grad_norm": 0.384765625, - "learning_rate": 1.7162386166671278e-05, - "loss": 0.4169, - "step": 7508 - }, - { - "epoch": 6.78590785907859, - "grad_norm": 0.380859375, - "learning_rate": 1.7062656979362243e-05, - "loss": 0.413, - "step": 7512 - }, - { - "epoch": 6.789521228545619, - "grad_norm": 0.3828125, - "learning_rate": 1.6963200919053527e-05, - "loss": 0.413, - "step": 7516 - }, - { - "epoch": 6.793134598012647, - "grad_norm": 0.388671875, - "learning_rate": 1.6864018190083005e-05, - "loss": 0.4069, - "step": 7520 - }, - { - "epoch": 6.796747967479675, - "grad_norm": 0.380859375, - "learning_rate": 1.676510899622696e-05, - "loss": 0.4229, - "step": 7524 - }, - { - "epoch": 6.800361336946703, - "grad_norm": 0.392578125, - "learning_rate": 1.6666473540699772e-05, - "loss": 0.4883, - "step": 7528 - }, - { - "epoch": 6.803974706413731, - "grad_norm": 0.37109375, - "learning_rate": 1.6568112026153312e-05, - "loss": 0.4513, - "step": 7532 - }, - { - "epoch": 6.8075880758807585, - "grad_norm": 0.39453125, - "learning_rate": 1.6470024654676666e-05, - "loss": 0.44, - "step": 7536 - }, - { - "epoch": 6.811201445347787, - "grad_norm": 0.36328125, - "learning_rate": 1.637221162779568e-05, - "loss": 0.4285, - "step": 7540 - }, - { - "epoch": 6.814814814814815, - "grad_norm": 0.400390625, - "learning_rate": 1.627467314647248e-05, - "loss": 0.424, - "step": 7544 - }, - { - "epoch": 6.818428184281843, - "grad_norm": 0.392578125, - "learning_rate": 1.6177409411105262e-05, - "loss": 0.4215, - "step": 7548 - }, - { - "epoch": 6.822041553748871, - "grad_norm": 0.359375, - "learning_rate": 1.6080420621527634e-05, - "loss": 0.3964, - "step": 7552 - }, - { - "epoch": 6.8256549232158985, - "grad_norm": 0.396484375, - "learning_rate": 1.5983706977008327e-05, - "loss": 0.435, - "step": 7556 - }, - { - "epoch": 6.829268292682927, - "grad_norm": 0.390625, - "learning_rate": 1.5887268676250775e-05, - "loss": 0.4214, - "step": 7560 - }, - { - "epoch": 6.832881662149955, - "grad_norm": 0.400390625, - "learning_rate": 1.5791105917392704e-05, - "loss": 0.4442, - "step": 7564 - }, - { - "epoch": 6.836495031616983, - "grad_norm": 0.375, - "learning_rate": 1.5695218898005775e-05, - "loss": 0.42, - "step": 7568 - }, - { - "epoch": 6.840108401084011, - "grad_norm": 0.39453125, - "learning_rate": 1.5599607815095043e-05, - "loss": 0.4371, - "step": 7572 - }, - { - "epoch": 6.8437217705510385, - "grad_norm": 0.396484375, - "learning_rate": 1.5504272865098717e-05, - "loss": 0.4163, - "step": 7576 - }, - { - "epoch": 6.847335140018067, - "grad_norm": 0.37890625, - "learning_rate": 1.5409214243887558e-05, - "loss": 0.4216, - "step": 7580 - }, - { - "epoch": 6.850948509485095, - "grad_norm": 0.384765625, - "learning_rate": 1.531443214676476e-05, - "loss": 0.4511, - "step": 7584 - }, - { - "epoch": 6.854561878952123, - "grad_norm": 0.3828125, - "learning_rate": 1.5219926768465269e-05, - "loss": 0.4452, - "step": 7588 - }, - { - "epoch": 6.858175248419151, - "grad_norm": 0.37890625, - "learning_rate": 1.5125698303155514e-05, - "loss": 0.4575, - "step": 7592 - }, - { - "epoch": 6.861788617886178, - "grad_norm": 0.37109375, - "learning_rate": 1.5031746944433005e-05, - "loss": 0.4174, - "step": 7596 - }, - { - "epoch": 6.865401987353207, - "grad_norm": 0.3828125, - "learning_rate": 1.493807288532588e-05, - "loss": 0.4515, - "step": 7600 - }, - { - "epoch": 6.869015356820235, - "grad_norm": 0.365234375, - "learning_rate": 1.4844676318292659e-05, - "loss": 0.4567, - "step": 7604 - }, - { - "epoch": 6.872628726287263, - "grad_norm": 0.373046875, - "learning_rate": 1.4751557435221629e-05, - "loss": 0.4487, - "step": 7608 - }, - { - "epoch": 6.876242095754291, - "grad_norm": 0.38671875, - "learning_rate": 1.4658716427430584e-05, - "loss": 0.4181, - "step": 7612 - }, - { - "epoch": 6.879855465221318, - "grad_norm": 0.39453125, - "learning_rate": 1.456615348566646e-05, - "loss": 0.447, - "step": 7616 - }, - { - "epoch": 6.883468834688347, - "grad_norm": 0.37109375, - "learning_rate": 1.4473868800104799e-05, - "loss": 0.4173, - "step": 7620 - }, - { - "epoch": 6.887082204155375, - "grad_norm": 0.400390625, - "learning_rate": 1.4381862560349582e-05, - "loss": 0.4065, - "step": 7624 - }, - { - "epoch": 6.890695573622403, - "grad_norm": 0.40234375, - "learning_rate": 1.4290134955432587e-05, - "loss": 0.4809, - "step": 7628 - }, - { - "epoch": 6.894308943089431, - "grad_norm": 0.38671875, - "learning_rate": 1.4198686173813157e-05, - "loss": 0.448, - "step": 7632 - }, - { - "epoch": 6.897922312556459, - "grad_norm": 0.380859375, - "learning_rate": 1.4107516403377817e-05, - "loss": 0.4446, - "step": 7636 - }, - { - "epoch": 6.9015356820234866, - "grad_norm": 0.376953125, - "learning_rate": 1.4016625831439764e-05, - "loss": 0.4578, - "step": 7640 - }, - { - "epoch": 6.905149051490515, - "grad_norm": 0.396484375, - "learning_rate": 1.3926014644738693e-05, - "loss": 0.4206, - "step": 7644 - }, - { - "epoch": 6.908762420957543, - "grad_norm": 0.380859375, - "learning_rate": 1.383568302944018e-05, - "loss": 0.4116, - "step": 7648 - }, - { - "epoch": 6.912375790424571, - "grad_norm": 0.384765625, - "learning_rate": 1.374563117113544e-05, - "loss": 0.4477, - "step": 7652 - }, - { - "epoch": 6.915989159891599, - "grad_norm": 0.384765625, - "learning_rate": 1.365585925484089e-05, - "loss": 0.4452, - "step": 7656 - }, - { - "epoch": 6.919602529358627, - "grad_norm": 0.388671875, - "learning_rate": 1.3566367464997856e-05, - "loss": 0.4501, - "step": 7660 - }, - { - "epoch": 6.923215898825655, - "grad_norm": 0.3671875, - "learning_rate": 1.3477155985472065e-05, - "loss": 0.4455, - "step": 7664 - }, - { - "epoch": 6.926829268292683, - "grad_norm": 0.3828125, - "learning_rate": 1.3388224999553365e-05, - "loss": 0.4419, - "step": 7668 - }, - { - "epoch": 6.930442637759711, - "grad_norm": 0.361328125, - "learning_rate": 1.32995746899553e-05, - "loss": 0.4685, - "step": 7672 - }, - { - "epoch": 6.934056007226739, - "grad_norm": 0.36328125, - "learning_rate": 1.3211205238814709e-05, - "loss": 0.4227, - "step": 7676 - }, - { - "epoch": 6.937669376693767, - "grad_norm": 0.388671875, - "learning_rate": 1.312311682769151e-05, - "loss": 0.4738, - "step": 7680 - }, - { - "epoch": 6.941282746160795, - "grad_norm": 0.396484375, - "learning_rate": 1.3035309637568081e-05, - "loss": 0.4349, - "step": 7684 - }, - { - "epoch": 6.944896115627823, - "grad_norm": 0.41796875, - "learning_rate": 1.2947783848849098e-05, - "loss": 0.4168, - "step": 7688 - }, - { - "epoch": 6.948509485094851, - "grad_norm": 0.36328125, - "learning_rate": 1.2860539641361028e-05, - "loss": 0.4453, - "step": 7692 - }, - { - "epoch": 6.952122854561879, - "grad_norm": 0.396484375, - "learning_rate": 1.2773577194351824e-05, - "loss": 0.4207, - "step": 7696 - }, - { - "epoch": 6.955736224028907, - "grad_norm": 0.375, - "learning_rate": 1.2686896686490633e-05, - "loss": 0.3943, - "step": 7700 - }, - { - "epoch": 6.959349593495935, - "grad_norm": 0.3671875, - "learning_rate": 1.2600498295867211e-05, - "loss": 0.4065, - "step": 7704 - }, - { - "epoch": 6.962962962962963, - "grad_norm": 0.388671875, - "learning_rate": 1.2514382199991767e-05, - "loss": 0.4228, - "step": 7708 - }, - { - "epoch": 6.966576332429991, - "grad_norm": 0.40234375, - "learning_rate": 1.2428548575794506e-05, - "loss": 0.4366, - "step": 7712 - }, - { - "epoch": 6.970189701897019, - "grad_norm": 0.3515625, - "learning_rate": 1.2342997599625259e-05, - "loss": 0.4551, - "step": 7716 - }, - { - "epoch": 6.973803071364047, - "grad_norm": 0.390625, - "learning_rate": 1.2257729447253178e-05, - "loss": 0.4836, - "step": 7720 - }, - { - "epoch": 6.977416440831075, - "grad_norm": 0.375, - "learning_rate": 1.2172744293866328e-05, - "loss": 0.4464, - "step": 7724 - }, - { - "epoch": 6.981029810298103, - "grad_norm": 0.359375, - "learning_rate": 1.2088042314071316e-05, - "loss": 0.4342, - "step": 7728 - }, - { - "epoch": 6.984643179765131, - "grad_norm": 0.369140625, - "learning_rate": 1.2003623681892926e-05, - "loss": 0.4355, - "step": 7732 - }, - { - "epoch": 6.988256549232159, - "grad_norm": 0.412109375, - "learning_rate": 1.1919488570773899e-05, - "loss": 0.4295, - "step": 7736 - }, - { - "epoch": 6.991869918699187, - "grad_norm": 0.3828125, - "learning_rate": 1.1835637153574378e-05, - "loss": 0.457, - "step": 7740 - }, - { - "epoch": 6.995483288166215, - "grad_norm": 0.396484375, - "learning_rate": 1.1752069602571623e-05, - "loss": 0.4613, - "step": 7744 - }, - { - "epoch": 6.999096657633243, - "grad_norm": 0.400390625, - "learning_rate": 1.1668786089459741e-05, - "loss": 0.4617, - "step": 7748 - }, - { - "epoch": 7.002710027100271, - "grad_norm": 0.375, - "learning_rate": 1.158578678534921e-05, - "loss": 0.4494, - "step": 7752 - }, - { - "epoch": 7.006323396567299, - "grad_norm": 0.37890625, - "learning_rate": 1.1503071860766655e-05, - "loss": 0.4519, - "step": 7756 - }, - { - "epoch": 7.009936766034327, - "grad_norm": 0.40625, - "learning_rate": 1.142064148565437e-05, - "loss": 0.4355, - "step": 7760 - }, - { - "epoch": 7.013550135501355, - "grad_norm": 0.365234375, - "learning_rate": 1.1338495829370031e-05, - "loss": 0.3998, - "step": 7764 - }, - { - "epoch": 7.017163504968383, - "grad_norm": 0.3828125, - "learning_rate": 1.1256635060686369e-05, - "loss": 0.4348, - "step": 7768 - }, - { - "epoch": 7.020776874435411, - "grad_norm": 0.36328125, - "learning_rate": 1.1175059347790766e-05, - "loss": 0.4116, - "step": 7772 - }, - { - "epoch": 7.024390243902439, - "grad_norm": 0.3828125, - "learning_rate": 1.1093768858284985e-05, - "loss": 0.399, - "step": 7776 - }, - { - "epoch": 7.028003613369467, - "grad_norm": 0.37890625, - "learning_rate": 1.1012763759184761e-05, - "loss": 0.4358, - "step": 7780 - }, - { - "epoch": 7.031616982836495, - "grad_norm": 0.396484375, - "learning_rate": 1.093204421691945e-05, - "loss": 0.4335, - "step": 7784 - }, - { - "epoch": 7.035230352303523, - "grad_norm": 0.369140625, - "learning_rate": 1.0851610397331756e-05, - "loss": 0.4048, - "step": 7788 - }, - { - "epoch": 7.038843721770551, - "grad_norm": 0.376953125, - "learning_rate": 1.0771462465677306e-05, - "loss": 0.4309, - "step": 7792 - }, - { - "epoch": 7.042457091237579, - "grad_norm": 0.376953125, - "learning_rate": 1.0691600586624427e-05, - "loss": 0.4309, - "step": 7796 - }, - { - "epoch": 7.046070460704607, - "grad_norm": 0.3515625, - "learning_rate": 1.0612024924253687e-05, - "loss": 0.4151, - "step": 7800 - }, - { - "epoch": 7.049683830171635, - "grad_norm": 0.39453125, - "learning_rate": 1.0532735642057588e-05, - "loss": 0.4171, - "step": 7804 - }, - { - "epoch": 7.053297199638663, - "grad_norm": 0.3828125, - "learning_rate": 1.045373290294027e-05, - "loss": 0.4623, - "step": 7808 - }, - { - "epoch": 7.056910569105691, - "grad_norm": 0.36328125, - "learning_rate": 1.0375016869217196e-05, - "loss": 0.4726, - "step": 7812 - }, - { - "epoch": 7.060523938572719, - "grad_norm": 0.388671875, - "learning_rate": 1.0296587702614695e-05, - "loss": 0.4241, - "step": 7816 - }, - { - "epoch": 7.064137308039747, - "grad_norm": 0.361328125, - "learning_rate": 1.021844556426979e-05, - "loss": 0.4337, - "step": 7820 - }, - { - "epoch": 7.067750677506775, - "grad_norm": 0.37890625, - "learning_rate": 1.0140590614729721e-05, - "loss": 0.4324, - "step": 7824 - }, - { - "epoch": 7.0713640469738035, - "grad_norm": 0.37890625, - "learning_rate": 1.0063023013951705e-05, - "loss": 0.4542, - "step": 7828 - }, - { - "epoch": 7.074977416440831, - "grad_norm": 0.39453125, - "learning_rate": 9.985742921302658e-06, - "loss": 0.4335, - "step": 7832 - }, - { - "epoch": 7.078590785907859, - "grad_norm": 0.380859375, - "learning_rate": 9.908750495558693e-06, - "loss": 0.4224, - "step": 7836 - }, - { - "epoch": 7.082204155374887, - "grad_norm": 0.388671875, - "learning_rate": 9.832045894904922e-06, - "loss": 0.4485, - "step": 7840 - }, - { - "epoch": 7.085817524841915, - "grad_norm": 0.373046875, - "learning_rate": 9.75562927693515e-06, - "loss": 0.4516, - "step": 7844 - }, - { - "epoch": 7.0894308943089435, - "grad_norm": 0.390625, - "learning_rate": 9.679500798651424e-06, - "loss": 0.4164, - "step": 7848 - }, - { - "epoch": 7.093044263775971, - "grad_norm": 0.390625, - "learning_rate": 9.60366061646391e-06, - "loss": 0.4383, - "step": 7852 - }, - { - "epoch": 7.096657633242999, - "grad_norm": 0.376953125, - "learning_rate": 9.52810888619035e-06, - "loss": 0.4478, - "step": 7856 - }, - { - "epoch": 7.100271002710027, - "grad_norm": 0.3515625, - "learning_rate": 9.452845763055894e-06, - "loss": 0.3991, - "step": 7860 - }, - { - "epoch": 7.103884372177055, - "grad_norm": 0.365234375, - "learning_rate": 9.3778714016927e-06, - "loss": 0.4196, - "step": 7864 - }, - { - "epoch": 7.107497741644083, - "grad_norm": 0.357421875, - "learning_rate": 9.303185956139664e-06, - "loss": 0.409, - "step": 7868 - }, - { - "epoch": 7.111111111111111, - "grad_norm": 0.3984375, - "learning_rate": 9.228789579842127e-06, - "loss": 0.4081, - "step": 7872 - }, - { - "epoch": 7.114724480578139, - "grad_norm": 0.390625, - "learning_rate": 9.154682425651488e-06, - "loss": 0.4448, - "step": 7876 - }, - { - "epoch": 7.118337850045167, - "grad_norm": 0.39453125, - "learning_rate": 9.08086464582492e-06, - "loss": 0.4208, - "step": 7880 - }, - { - "epoch": 7.121951219512195, - "grad_norm": 0.388671875, - "learning_rate": 9.007336392025038e-06, - "loss": 0.4266, - "step": 7884 - }, - { - "epoch": 7.125564588979223, - "grad_norm": 0.376953125, - "learning_rate": 8.93409781531967e-06, - "loss": 0.4273, - "step": 7888 - }, - { - "epoch": 7.129177958446251, - "grad_norm": 0.388671875, - "learning_rate": 8.861149066181445e-06, - "loss": 0.4212, - "step": 7892 - }, - { - "epoch": 7.132791327913279, - "grad_norm": 0.35546875, - "learning_rate": 8.788490294487543e-06, - "loss": 0.3971, - "step": 7896 - }, - { - "epoch": 7.136404697380307, - "grad_norm": 0.37890625, - "learning_rate": 8.716121649519354e-06, - "loss": 0.4149, - "step": 7900 - }, - { - "epoch": 7.140018066847335, - "grad_norm": 0.376953125, - "learning_rate": 8.64404327996217e-06, - "loss": 0.4063, - "step": 7904 - }, - { - "epoch": 7.143631436314363, - "grad_norm": 0.373046875, - "learning_rate": 8.57225533390497e-06, - "loss": 0.4607, - "step": 7908 - }, - { - "epoch": 7.1472448057813915, - "grad_norm": 0.3828125, - "learning_rate": 8.500757958839954e-06, - "loss": 0.4243, - "step": 7912 - }, - { - "epoch": 7.150858175248419, - "grad_norm": 0.373046875, - "learning_rate": 8.429551301662386e-06, - "loss": 0.4444, - "step": 7916 - }, - { - "epoch": 7.154471544715447, - "grad_norm": 0.404296875, - "learning_rate": 8.35863550867022e-06, - "loss": 0.4668, - "step": 7920 - }, - { - "epoch": 7.158084914182475, - "grad_norm": 0.3984375, - "learning_rate": 8.288010725563759e-06, - "loss": 0.4559, - "step": 7924 - }, - { - "epoch": 7.161698283649503, - "grad_norm": 0.365234375, - "learning_rate": 8.217677097445535e-06, - "loss": 0.426, - "step": 7928 - }, - { - "epoch": 7.1653116531165315, - "grad_norm": 0.384765625, - "learning_rate": 8.147634768819788e-06, - "loss": 0.449, - "step": 7932 - }, - { - "epoch": 7.168925022583559, - "grad_norm": 0.37890625, - "learning_rate": 8.077883883592256e-06, - "loss": 0.4111, - "step": 7936 - }, - { - "epoch": 7.172538392050587, - "grad_norm": 0.404296875, - "learning_rate": 8.008424585069972e-06, - "loss": 0.4177, - "step": 7940 - }, - { - "epoch": 7.176151761517615, - "grad_norm": 0.392578125, - "learning_rate": 7.939257015960782e-06, - "loss": 0.4485, - "step": 7944 - }, - { - "epoch": 7.179765130984643, - "grad_norm": 0.36328125, - "learning_rate": 7.870381318373293e-06, - "loss": 0.4079, - "step": 7948 - }, - { - "epoch": 7.1833785004516715, - "grad_norm": 0.376953125, - "learning_rate": 7.801797633816325e-06, - "loss": 0.423, - "step": 7952 - }, - { - "epoch": 7.186991869918699, - "grad_norm": 0.376953125, - "learning_rate": 7.733506103198784e-06, - "loss": 0.4961, - "step": 7956 - }, - { - "epoch": 7.190605239385727, - "grad_norm": 0.376953125, - "learning_rate": 7.665506866829347e-06, - "loss": 0.4292, - "step": 7960 - }, - { - "epoch": 7.194218608852755, - "grad_norm": 0.380859375, - "learning_rate": 7.597800064416143e-06, - "loss": 0.4689, - "step": 7964 - }, - { - "epoch": 7.197831978319783, - "grad_norm": 0.376953125, - "learning_rate": 7.530385835066499e-06, - "loss": 0.4048, - "step": 7968 - }, - { - "epoch": 7.2014453477868114, - "grad_norm": 0.388671875, - "learning_rate": 7.463264317286571e-06, - "loss": 0.4369, - "step": 7972 - }, - { - "epoch": 7.20505871725384, - "grad_norm": 0.3828125, - "learning_rate": 7.396435648981214e-06, - "loss": 0.4248, - "step": 7976 - }, - { - "epoch": 7.208672086720867, - "grad_norm": 0.38671875, - "learning_rate": 7.329899967453495e-06, - "loss": 0.4861, - "step": 7980 - }, - { - "epoch": 7.212285456187895, - "grad_norm": 0.3828125, - "learning_rate": 7.263657409404666e-06, - "loss": 0.4476, - "step": 7984 - }, - { - "epoch": 7.215898825654923, - "grad_norm": 0.353515625, - "learning_rate": 7.197708110933642e-06, - "loss": 0.4006, - "step": 7988 - }, - { - "epoch": 7.219512195121951, - "grad_norm": 0.375, - "learning_rate": 7.1320522075368205e-06, - "loss": 0.4444, - "step": 7992 - }, - { - "epoch": 7.22312556458898, - "grad_norm": 0.380859375, - "learning_rate": 7.0666898341078305e-06, - "loss": 0.4436, - "step": 7996 - }, - { - "epoch": 7.226738934056007, - "grad_norm": 0.357421875, - "learning_rate": 7.001621124937218e-06, - "loss": 0.4231, - "step": 8000 - }, - { - "epoch": 7.230352303523035, - "grad_norm": 0.37890625, - "learning_rate": 6.936846213712211e-06, - "loss": 0.437, - "step": 8004 - }, - { - "epoch": 7.233965672990063, - "grad_norm": 0.3671875, - "learning_rate": 6.872365233516369e-06, - "loss": 0.4139, - "step": 8008 - }, - { - "epoch": 7.237579042457091, - "grad_norm": 0.380859375, - "learning_rate": 6.8081783168293876e-06, - "loss": 0.4395, - "step": 8012 - }, - { - "epoch": 7.2411924119241196, - "grad_norm": 0.361328125, - "learning_rate": 6.744285595526761e-06, - "loss": 0.4561, - "step": 8016 - }, - { - "epoch": 7.244805781391147, - "grad_norm": 0.412109375, - "learning_rate": 6.680687200879531e-06, - "loss": 0.4554, - "step": 8020 - }, - { - "epoch": 7.248419150858175, - "grad_norm": 0.3515625, - "learning_rate": 6.617383263554127e-06, - "loss": 0.4205, - "step": 8024 - }, - { - "epoch": 7.252032520325203, - "grad_norm": 0.42578125, - "learning_rate": 6.554373913611893e-06, - "loss": 0.4264, - "step": 8028 - }, - { - "epoch": 7.255645889792231, - "grad_norm": 0.361328125, - "learning_rate": 6.49165928050896e-06, - "loss": 0.429, - "step": 8032 - }, - { - "epoch": 7.2592592592592595, - "grad_norm": 0.384765625, - "learning_rate": 6.429239493095939e-06, - "loss": 0.4016, - "step": 8036 - }, - { - "epoch": 7.262872628726287, - "grad_norm": 0.357421875, - "learning_rate": 6.367114679617713e-06, - "loss": 0.4297, - "step": 8040 - }, - { - "epoch": 7.266485998193315, - "grad_norm": 0.38671875, - "learning_rate": 6.305284967713081e-06, - "loss": 0.4781, - "step": 8044 - }, - { - "epoch": 7.270099367660343, - "grad_norm": 0.396484375, - "learning_rate": 6.243750484414528e-06, - "loss": 0.4424, - "step": 8048 - }, - { - "epoch": 7.273712737127371, - "grad_norm": 0.369140625, - "learning_rate": 6.182511356148023e-06, - "loss": 0.4723, - "step": 8052 - }, - { - "epoch": 7.2773261065943995, - "grad_norm": 0.37109375, - "learning_rate": 6.121567708732639e-06, - "loss": 0.4389, - "step": 8056 - }, - { - "epoch": 7.280939476061428, - "grad_norm": 0.3984375, - "learning_rate": 6.0609196673804814e-06, - "loss": 0.4257, - "step": 8060 - }, - { - "epoch": 7.284552845528455, - "grad_norm": 0.357421875, - "learning_rate": 6.0005673566961784e-06, - "loss": 0.4441, - "step": 8064 - }, - { - "epoch": 7.288166214995483, - "grad_norm": 0.375, - "learning_rate": 5.940510900676926e-06, - "loss": 0.4201, - "step": 8068 - }, - { - "epoch": 7.291779584462511, - "grad_norm": 0.38671875, - "learning_rate": 5.8807504227118884e-06, - "loss": 0.4438, - "step": 8072 - }, - { - "epoch": 7.2953929539295395, - "grad_norm": 0.396484375, - "learning_rate": 5.821286045582235e-06, - "loss": 0.4535, - "step": 8076 - }, - { - "epoch": 7.299006323396568, - "grad_norm": 0.3828125, - "learning_rate": 5.762117891460788e-06, - "loss": 0.4425, - "step": 8080 - }, - { - "epoch": 7.302619692863595, - "grad_norm": 0.392578125, - "learning_rate": 5.703246081911739e-06, - "loss": 0.4208, - "step": 8084 - }, - { - "epoch": 7.306233062330623, - "grad_norm": 0.390625, - "learning_rate": 5.644670737890417e-06, - "loss": 0.4197, - "step": 8088 - }, - { - "epoch": 7.309846431797651, - "grad_norm": 0.384765625, - "learning_rate": 5.586391979743038e-06, - "loss": 0.4257, - "step": 8092 - }, - { - "epoch": 7.313459801264679, - "grad_norm": 0.365234375, - "learning_rate": 5.528409927206473e-06, - "loss": 0.4205, - "step": 8096 - }, - { - "epoch": 7.317073170731708, - "grad_norm": 0.365234375, - "learning_rate": 5.47072469940803e-06, - "loss": 0.4079, - "step": 8100 - }, - { - "epoch": 7.320686540198736, - "grad_norm": 0.3828125, - "learning_rate": 5.413336414865138e-06, - "loss": 0.4147, - "step": 8104 - }, - { - "epoch": 7.324299909665763, - "grad_norm": 0.404296875, - "learning_rate": 5.356245191485148e-06, - "loss": 0.4364, - "step": 8108 - }, - { - "epoch": 7.327913279132791, - "grad_norm": 0.3671875, - "learning_rate": 5.299451146565048e-06, - "loss": 0.3964, - "step": 8112 - }, - { - "epoch": 7.331526648599819, - "grad_norm": 0.37890625, - "learning_rate": 5.242954396791349e-06, - "loss": 0.4233, - "step": 8116 - }, - { - "epoch": 7.335140018066848, - "grad_norm": 0.3671875, - "learning_rate": 5.1867550582396666e-06, - "loss": 0.4263, - "step": 8120 - }, - { - "epoch": 7.338753387533876, - "grad_norm": 0.416015625, - "learning_rate": 5.130853246374589e-06, - "loss": 0.456, - "step": 8124 - }, - { - "epoch": 7.342366757000903, - "grad_norm": 0.376953125, - "learning_rate": 5.075249076049459e-06, - "loss": 0.4169, - "step": 8128 - }, - { - "epoch": 7.345980126467931, - "grad_norm": 0.3828125, - "learning_rate": 5.0199426615060264e-06, - "loss": 0.4212, - "step": 8132 - }, - { - "epoch": 7.349593495934959, - "grad_norm": 0.38671875, - "learning_rate": 4.964934116374364e-06, - "loss": 0.4339, - "step": 8136 - }, - { - "epoch": 7.3532068654019875, - "grad_norm": 0.373046875, - "learning_rate": 4.910223553672482e-06, - "loss": 0.4498, - "step": 8140 - }, - { - "epoch": 7.356820234869016, - "grad_norm": 0.40234375, - "learning_rate": 4.855811085806266e-06, - "loss": 0.4583, - "step": 8144 - }, - { - "epoch": 7.360433604336043, - "grad_norm": 0.3984375, - "learning_rate": 4.80169682456904e-06, - "loss": 0.4343, - "step": 8148 - }, - { - "epoch": 7.364046973803071, - "grad_norm": 0.390625, - "learning_rate": 4.747880881141502e-06, - "loss": 0.4173, - "step": 8152 - }, - { - "epoch": 7.367660343270099, - "grad_norm": 0.396484375, - "learning_rate": 4.694363366091441e-06, - "loss": 0.4739, - "step": 8156 - }, - { - "epoch": 7.3712737127371275, - "grad_norm": 0.369140625, - "learning_rate": 4.641144389373519e-06, - "loss": 0.4291, - "step": 8160 - }, - { - "epoch": 7.374887082204156, - "grad_norm": 0.369140625, - "learning_rate": 4.5882240603290045e-06, - "loss": 0.4318, - "step": 8164 - }, - { - "epoch": 7.378500451671183, - "grad_norm": 0.3984375, - "learning_rate": 4.535602487685558e-06, - "loss": 0.4472, - "step": 8168 - }, - { - "epoch": 7.382113821138211, - "grad_norm": 0.400390625, - "learning_rate": 4.483279779557114e-06, - "loss": 0.4248, - "step": 8172 - }, - { - "epoch": 7.385727190605239, - "grad_norm": 0.384765625, - "learning_rate": 4.431256043443515e-06, - "loss": 0.4148, - "step": 8176 - }, - { - "epoch": 7.3893405600722675, - "grad_norm": 0.388671875, - "learning_rate": 4.379531386230328e-06, - "loss": 0.3603, - "step": 8180 - }, - { - "epoch": 7.392953929539296, - "grad_norm": 0.388671875, - "learning_rate": 4.328105914188695e-06, - "loss": 0.4299, - "step": 8184 - }, - { - "epoch": 7.396567299006323, - "grad_norm": 0.392578125, - "learning_rate": 4.276979732975017e-06, - "loss": 0.4313, - "step": 8188 - }, - { - "epoch": 7.400180668473351, - "grad_norm": 0.3515625, - "learning_rate": 4.2261529476308345e-06, - "loss": 0.417, - "step": 8192 - }, - { - "epoch": 7.403794037940379, - "grad_norm": 0.39453125, - "learning_rate": 4.175625662582566e-06, - "loss": 0.4687, - "step": 8196 - }, - { - "epoch": 7.407407407407407, - "grad_norm": 0.396484375, - "learning_rate": 4.1253979816412206e-06, - "loss": 0.4499, - "step": 8200 - }, - { - "epoch": 7.411020776874436, - "grad_norm": 0.36328125, - "learning_rate": 4.075470008002335e-06, - "loss": 0.4295, - "step": 8204 - }, - { - "epoch": 7.414634146341464, - "grad_norm": 0.376953125, - "learning_rate": 4.0258418442456195e-06, - "loss": 0.4032, - "step": 8208 - }, - { - "epoch": 7.418247515808491, - "grad_norm": 0.375, - "learning_rate": 3.976513592334863e-06, - "loss": 0.3908, - "step": 8212 - }, - { - "epoch": 7.421860885275519, - "grad_norm": 0.40234375, - "learning_rate": 3.9274853536175935e-06, - "loss": 0.4193, - "step": 8216 - }, - { - "epoch": 7.425474254742547, - "grad_norm": 0.388671875, - "learning_rate": 3.878757228825086e-06, - "loss": 0.4653, - "step": 8220 - }, - { - "epoch": 7.429087624209576, - "grad_norm": 0.37109375, - "learning_rate": 3.830329318071823e-06, - "loss": 0.416, - "step": 8224 - }, - { - "epoch": 7.432700993676604, - "grad_norm": 0.392578125, - "learning_rate": 3.7822017208556154e-06, - "loss": 0.4302, - "step": 8228 - }, - { - "epoch": 7.436314363143631, - "grad_norm": 0.384765625, - "learning_rate": 3.73437453605725e-06, - "loss": 0.4451, - "step": 8232 - }, - { - "epoch": 7.439927732610659, - "grad_norm": 0.388671875, - "learning_rate": 3.6868478619402564e-06, - "loss": 0.4333, - "step": 8236 - }, - { - "epoch": 7.443541102077687, - "grad_norm": 0.376953125, - "learning_rate": 3.6396217961507936e-06, - "loss": 0.4261, - "step": 8240 - }, - { - "epoch": 7.4471544715447155, - "grad_norm": 0.369140625, - "learning_rate": 3.592696435717346e-06, - "loss": 0.4147, - "step": 8244 - }, - { - "epoch": 7.450767841011744, - "grad_norm": 0.3828125, - "learning_rate": 3.5460718770506434e-06, - "loss": 0.4278, - "step": 8248 - }, - { - "epoch": 7.454381210478772, - "grad_norm": 0.41796875, - "learning_rate": 3.4997482159433764e-06, - "loss": 0.4272, - "step": 8252 - }, - { - "epoch": 7.457994579945799, - "grad_norm": 0.384765625, - "learning_rate": 3.45372554757003e-06, - "loss": 0.4055, - "step": 8256 - }, - { - "epoch": 7.461607949412827, - "grad_norm": 0.359375, - "learning_rate": 3.4080039664866497e-06, - "loss": 0.4364, - "step": 8260 - }, - { - "epoch": 7.4652213188798555, - "grad_norm": 0.404296875, - "learning_rate": 3.3625835666307445e-06, - "loss": 0.4313, - "step": 8264 - }, - { - "epoch": 7.468834688346884, - "grad_norm": 0.37109375, - "learning_rate": 3.317464441320983e-06, - "loss": 0.4039, - "step": 8268 - }, - { - "epoch": 7.472448057813912, - "grad_norm": 0.365234375, - "learning_rate": 3.2726466832570464e-06, - "loss": 0.4671, - "step": 8272 - }, - { - "epoch": 7.476061427280939, - "grad_norm": 0.3984375, - "learning_rate": 3.228130384519495e-06, - "loss": 0.4394, - "step": 8276 - }, - { - "epoch": 7.479674796747967, - "grad_norm": 0.404296875, - "learning_rate": 3.1839156365694506e-06, - "loss": 0.4598, - "step": 8280 - }, - { - "epoch": 7.4832881662149955, - "grad_norm": 0.392578125, - "learning_rate": 3.1400025302485145e-06, - "loss": 0.4358, - "step": 8284 - }, - { - "epoch": 7.486901535682024, - "grad_norm": 0.34375, - "learning_rate": 3.0963911557785824e-06, - "loss": 0.4136, - "step": 8288 - }, - { - "epoch": 7.490514905149052, - "grad_norm": 0.3671875, - "learning_rate": 3.053081602761548e-06, - "loss": 0.3874, - "step": 8292 - }, - { - "epoch": 7.494128274616079, - "grad_norm": 0.357421875, - "learning_rate": 3.0100739601793154e-06, - "loss": 0.4364, - "step": 8296 - }, - { - "epoch": 7.497741644083107, - "grad_norm": 0.376953125, - "learning_rate": 2.96736831639337e-06, - "loss": 0.4315, - "step": 8300 - }, - { - "epoch": 7.5013550135501355, - "grad_norm": 0.3671875, - "learning_rate": 2.92496475914476e-06, - "loss": 0.4645, - "step": 8304 - }, - { - "epoch": 7.504968383017164, - "grad_norm": 0.37890625, - "learning_rate": 2.8828633755539623e-06, - "loss": 0.4215, - "step": 8308 - }, - { - "epoch": 7.508581752484192, - "grad_norm": 0.3671875, - "learning_rate": 2.8410642521205007e-06, - "loss": 0.4524, - "step": 8312 - }, - { - "epoch": 7.512195121951219, - "grad_norm": 0.388671875, - "learning_rate": 2.7995674747230136e-06, - "loss": 0.4665, - "step": 8316 - }, - { - "epoch": 7.515808491418247, - "grad_norm": 0.37109375, - "learning_rate": 2.758373128618818e-06, - "loss": 0.4556, - "step": 8320 - }, - { - "epoch": 7.519421860885275, - "grad_norm": 0.3984375, - "learning_rate": 2.717481298443963e-06, - "loss": 0.4349, - "step": 8324 - }, - { - "epoch": 7.523035230352304, - "grad_norm": 0.375, - "learning_rate": 2.6768920682129604e-06, - "loss": 0.4189, - "step": 8328 - }, - { - "epoch": 7.526648599819332, - "grad_norm": 0.359375, - "learning_rate": 2.636605521318569e-06, - "loss": 0.4243, - "step": 8332 - }, - { - "epoch": 7.530261969286359, - "grad_norm": 0.369140625, - "learning_rate": 2.5966217405316956e-06, - "loss": 0.4516, - "step": 8336 - }, - { - "epoch": 7.533875338753387, - "grad_norm": 0.37890625, - "learning_rate": 2.5569408080011624e-06, - "loss": 0.4545, - "step": 8340 - }, - { - "epoch": 7.537488708220415, - "grad_norm": 0.376953125, - "learning_rate": 2.5175628052536532e-06, - "loss": 0.4271, - "step": 8344 - }, - { - "epoch": 7.541102077687444, - "grad_norm": 0.396484375, - "learning_rate": 2.478487813193386e-06, - "loss": 0.4578, - "step": 8348 - }, - { - "epoch": 7.544715447154472, - "grad_norm": 0.388671875, - "learning_rate": 2.439715912102075e-06, - "loss": 0.4533, - "step": 8352 - }, - { - "epoch": 7.5483288166215, - "grad_norm": 0.375, - "learning_rate": 2.4012471816386827e-06, - "loss": 0.4591, - "step": 8356 - }, - { - "epoch": 7.551942186088527, - "grad_norm": 0.380859375, - "learning_rate": 2.3630817008393033e-06, - "loss": 0.445, - "step": 8360 - }, - { - "epoch": 7.555555555555555, - "grad_norm": 0.431640625, - "learning_rate": 2.3252195481170133e-06, - "loss": 0.4664, - "step": 8364 - }, - { - "epoch": 7.5591689250225835, - "grad_norm": 0.412109375, - "learning_rate": 2.2876608012616537e-06, - "loss": 0.393, - "step": 8368 - }, - { - "epoch": 7.562782294489612, - "grad_norm": 0.404296875, - "learning_rate": 2.2504055374397144e-06, - "loss": 0.4321, - "step": 8372 - }, - { - "epoch": 7.56639566395664, - "grad_norm": 0.35546875, - "learning_rate": 2.2134538331941677e-06, - "loss": 0.4305, - "step": 8376 - }, - { - "epoch": 7.570009033423668, - "grad_norm": 0.375, - "learning_rate": 2.1768057644442673e-06, - "loss": 0.4251, - "step": 8380 - }, - { - "epoch": 7.573622402890695, - "grad_norm": 0.361328125, - "learning_rate": 2.1404614064855e-06, - "loss": 0.3862, - "step": 8384 - }, - { - "epoch": 7.5772357723577235, - "grad_norm": 0.37890625, - "learning_rate": 2.104420833989301e-06, - "loss": 0.4189, - "step": 8388 - }, - { - "epoch": 7.580849141824752, - "grad_norm": 0.37109375, - "learning_rate": 2.0686841210030223e-06, - "loss": 0.4428, - "step": 8392 - }, - { - "epoch": 7.58446251129178, - "grad_norm": 0.380859375, - "learning_rate": 2.033251340949632e-06, - "loss": 0.3944, - "step": 8396 - }, - { - "epoch": 7.588075880758808, - "grad_norm": 0.384765625, - "learning_rate": 1.9981225666277466e-06, - "loss": 0.4303, - "step": 8400 - }, - { - "epoch": 7.591689250225835, - "grad_norm": 0.384765625, - "learning_rate": 1.9632978702113335e-06, - "loss": 0.4411, - "step": 8404 - }, - { - "epoch": 7.5953026196928635, - "grad_norm": 0.359375, - "learning_rate": 1.9287773232496084e-06, - "loss": 0.4177, - "step": 8408 - }, - { - "epoch": 7.598915989159892, - "grad_norm": 0.37890625, - "learning_rate": 1.8945609966669383e-06, - "loss": 0.4171, - "step": 8412 - }, - { - "epoch": 7.60252935862692, - "grad_norm": 0.376953125, - "learning_rate": 1.8606489607625897e-06, - "loss": 0.4203, - "step": 8416 - }, - { - "epoch": 7.606142728093948, - "grad_norm": 0.384765625, - "learning_rate": 1.827041285210712e-06, - "loss": 0.4233, - "step": 8420 - }, - { - "epoch": 7.609756097560975, - "grad_norm": 0.3828125, - "learning_rate": 1.7937380390601054e-06, - "loss": 0.4779, - "step": 8424 - }, - { - "epoch": 7.613369467028003, - "grad_norm": 0.40234375, - "learning_rate": 1.7607392907340868e-06, - "loss": 0.4273, - "step": 8428 - }, - { - "epoch": 7.616982836495032, - "grad_norm": 0.38671875, - "learning_rate": 1.7280451080303902e-06, - "loss": 0.4323, - "step": 8432 - }, - { - "epoch": 7.62059620596206, - "grad_norm": 0.373046875, - "learning_rate": 1.6956555581209674e-06, - "loss": 0.4783, - "step": 8436 - }, - { - "epoch": 7.624209575429088, - "grad_norm": 0.38671875, - "learning_rate": 1.6635707075519533e-06, - "loss": 0.4164, - "step": 8440 - }, - { - "epoch": 7.627822944896115, - "grad_norm": 0.373046875, - "learning_rate": 1.6317906222433675e-06, - "loss": 0.4485, - "step": 8444 - }, - { - "epoch": 7.631436314363143, - "grad_norm": 0.3515625, - "learning_rate": 1.60031536748918e-06, - "loss": 0.4338, - "step": 8448 - }, - { - "epoch": 7.635049683830172, - "grad_norm": 0.40625, - "learning_rate": 1.5691450079569622e-06, - "loss": 0.4312, - "step": 8452 - }, - { - "epoch": 7.6386630532972, - "grad_norm": 0.365234375, - "learning_rate": 1.5382796076879201e-06, - "loss": 0.4701, - "step": 8456 - }, - { - "epoch": 7.642276422764228, - "grad_norm": 0.388671875, - "learning_rate": 1.5077192300967268e-06, - "loss": 0.4259, - "step": 8460 - }, - { - "epoch": 7.645889792231255, - "grad_norm": 0.380859375, - "learning_rate": 1.477463937971307e-06, - "loss": 0.4558, - "step": 8464 - }, - { - "epoch": 7.649503161698283, - "grad_norm": 0.37109375, - "learning_rate": 1.4475137934728543e-06, - "loss": 0.3784, - "step": 8468 - }, - { - "epoch": 7.6531165311653115, - "grad_norm": 0.3671875, - "learning_rate": 1.4178688581354957e-06, - "loss": 0.4166, - "step": 8472 - }, - { - "epoch": 7.65672990063234, - "grad_norm": 0.361328125, - "learning_rate": 1.3885291928664278e-06, - "loss": 0.4005, - "step": 8476 - }, - { - "epoch": 7.660343270099368, - "grad_norm": 0.34765625, - "learning_rate": 1.359494857945548e-06, - "loss": 0.4405, - "step": 8480 - }, - { - "epoch": 7.663956639566395, - "grad_norm": 0.369140625, - "learning_rate": 1.3307659130255056e-06, - "loss": 0.4088, - "step": 8484 - }, - { - "epoch": 7.667570009033423, - "grad_norm": 0.380859375, - "learning_rate": 1.3023424171314523e-06, - "loss": 0.45, - "step": 8488 - }, - { - "epoch": 7.6711833785004515, - "grad_norm": 0.365234375, - "learning_rate": 1.2742244286610081e-06, - "loss": 0.4228, - "step": 8492 - }, - { - "epoch": 7.67479674796748, - "grad_norm": 0.349609375, - "learning_rate": 1.2464120053841454e-06, - "loss": 0.4276, - "step": 8496 - }, - { - "epoch": 7.678410117434508, - "grad_norm": 0.388671875, - "learning_rate": 1.2189052044429547e-06, - "loss": 0.3931, - "step": 8500 + "epoch": 7.0, + "step": 3661, + "total_flos": 2.3920283694378516e+18, + "train_loss": 0.561195495056344, + "train_runtime": 14692.9794, + "train_samples_per_second": 7.96, + "train_steps_per_second": 0.249 } ], "logging_steps": 4, - "max_steps": 8856, + "max_steps": 3661, "num_input_tokens_seen": 0, - "num_train_epochs": 8, + "num_train_epochs": 7, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { @@ -14901,7 +6440,7 @@ "attributes": {} } }, - "total_flos": 5.560613094763266e+18, + "total_flos": 2.3920283694378516e+18, "train_batch_size": 32, "trial_name": null, "trial_params": null