{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9979522184300342, "eval_steps": 500, "global_step": 1098, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0027303754266211604, "grad_norm": 1.3254656791687012, "learning_rate": 1.818181818181818e-06, "loss": 1.2897, "step": 1 }, { "epoch": 0.005460750853242321, "grad_norm": 1.3267881870269775, "learning_rate": 3.636363636363636e-06, "loss": 1.2823, "step": 2 }, { "epoch": 0.008191126279863481, "grad_norm": 1.2982481718063354, "learning_rate": 5.4545454545454545e-06, "loss": 1.2595, "step": 3 }, { "epoch": 0.010921501706484642, "grad_norm": 1.2894413471221924, "learning_rate": 7.272727272727272e-06, "loss": 1.2653, "step": 4 }, { "epoch": 0.013651877133105802, "grad_norm": 1.2869772911071777, "learning_rate": 9.090909090909091e-06, "loss": 1.2545, "step": 5 }, { "epoch": 0.016382252559726963, "grad_norm": 1.2373387813568115, "learning_rate": 1.0909090909090909e-05, "loss": 1.2194, "step": 6 }, { "epoch": 0.01911262798634812, "grad_norm": 1.20195734500885, "learning_rate": 1.2727272727272727e-05, "loss": 1.2139, "step": 7 }, { "epoch": 0.021843003412969283, "grad_norm": 1.1426103115081787, "learning_rate": 1.4545454545454545e-05, "loss": 1.2315, "step": 8 }, { "epoch": 0.024573378839590442, "grad_norm": 1.0495123863220215, "learning_rate": 1.6363636363636366e-05, "loss": 1.1944, "step": 9 }, { "epoch": 0.027303754266211604, "grad_norm": 0.8776500821113586, "learning_rate": 1.8181818181818182e-05, "loss": 1.183, "step": 10 }, { "epoch": 0.030034129692832763, "grad_norm": 0.8687052130699158, "learning_rate": 2e-05, "loss": 1.1592, "step": 11 }, { "epoch": 0.032764505119453925, "grad_norm": 0.7476271390914917, "learning_rate": 2.1818181818181818e-05, "loss": 1.1483, "step": 12 }, { "epoch": 0.03549488054607509, "grad_norm": 0.6418495774269104, "learning_rate": 2.3636363636363637e-05, "loss": 1.0762, "step": 13 }, { "epoch": 0.03822525597269624, "grad_norm": 0.600390613079071, "learning_rate": 2.5454545454545454e-05, "loss": 1.0492, "step": 14 }, { "epoch": 0.040955631399317405, "grad_norm": 0.5653348565101624, "learning_rate": 2.7272727272727273e-05, "loss": 1.0548, "step": 15 }, { "epoch": 0.04368600682593857, "grad_norm": 0.5357097387313843, "learning_rate": 2.909090909090909e-05, "loss": 1.0273, "step": 16 }, { "epoch": 0.04641638225255973, "grad_norm": 0.4480445683002472, "learning_rate": 3.090909090909091e-05, "loss": 1.0065, "step": 17 }, { "epoch": 0.049146757679180884, "grad_norm": 0.40983352065086365, "learning_rate": 3.272727272727273e-05, "loss": 0.9908, "step": 18 }, { "epoch": 0.05187713310580205, "grad_norm": 0.42159780859947205, "learning_rate": 3.454545454545455e-05, "loss": 0.9888, "step": 19 }, { "epoch": 0.05460750853242321, "grad_norm": 0.41620174050331116, "learning_rate": 3.6363636363636364e-05, "loss": 0.9575, "step": 20 }, { "epoch": 0.05733788395904437, "grad_norm": 0.3804452419281006, "learning_rate": 3.818181818181819e-05, "loss": 0.9413, "step": 21 }, { "epoch": 0.060068259385665526, "grad_norm": 0.37021100521087646, "learning_rate": 4e-05, "loss": 0.9223, "step": 22 }, { "epoch": 0.06279863481228669, "grad_norm": 0.34090206027030945, "learning_rate": 4.181818181818182e-05, "loss": 0.8878, "step": 23 }, { "epoch": 0.06552901023890785, "grad_norm": 0.32232972979545593, "learning_rate": 4.3636363636363636e-05, "loss": 0.8986, "step": 24 }, { "epoch": 0.06825938566552901, "grad_norm": 0.2941684424877167, "learning_rate": 4.545454545454546e-05, "loss": 0.8857, "step": 25 }, { "epoch": 0.07098976109215017, "grad_norm": 0.27072674036026, "learning_rate": 4.7272727272727275e-05, "loss": 0.8736, "step": 26 }, { "epoch": 0.07372013651877134, "grad_norm": 0.2696637511253357, "learning_rate": 4.909090909090909e-05, "loss": 0.8698, "step": 27 }, { "epoch": 0.07645051194539249, "grad_norm": 0.2565267086029053, "learning_rate": 5.090909090909091e-05, "loss": 0.8324, "step": 28 }, { "epoch": 0.07918088737201365, "grad_norm": 0.2474038451910019, "learning_rate": 5.272727272727272e-05, "loss": 0.841, "step": 29 }, { "epoch": 0.08191126279863481, "grad_norm": 0.22430865466594696, "learning_rate": 5.4545454545454546e-05, "loss": 0.8219, "step": 30 }, { "epoch": 0.08464163822525597, "grad_norm": 0.21238166093826294, "learning_rate": 5.636363636363636e-05, "loss": 0.8328, "step": 31 }, { "epoch": 0.08737201365187713, "grad_norm": 0.2210783213376999, "learning_rate": 5.818181818181818e-05, "loss": 0.8187, "step": 32 }, { "epoch": 0.0901023890784983, "grad_norm": 0.24119816720485687, "learning_rate": 6e-05, "loss": 0.8078, "step": 33 }, { "epoch": 0.09283276450511946, "grad_norm": 0.21313577890396118, "learning_rate": 6.181818181818182e-05, "loss": 0.8051, "step": 34 }, { "epoch": 0.09556313993174062, "grad_norm": 0.17824789881706238, "learning_rate": 6.363636363636364e-05, "loss": 0.7841, "step": 35 }, { "epoch": 0.09829351535836177, "grad_norm": 0.18413369357585907, "learning_rate": 6.545454545454546e-05, "loss": 0.7851, "step": 36 }, { "epoch": 0.10102389078498293, "grad_norm": 0.2003067582845688, "learning_rate": 6.727272727272727e-05, "loss": 0.8064, "step": 37 }, { "epoch": 0.1037542662116041, "grad_norm": 0.1989540457725525, "learning_rate": 6.90909090909091e-05, "loss": 0.7841, "step": 38 }, { "epoch": 0.10648464163822526, "grad_norm": 0.14544272422790527, "learning_rate": 7.090909090909092e-05, "loss": 0.7745, "step": 39 }, { "epoch": 0.10921501706484642, "grad_norm": 0.1559988260269165, "learning_rate": 7.272727272727273e-05, "loss": 0.7845, "step": 40 }, { "epoch": 0.11194539249146758, "grad_norm": 0.1705523580312729, "learning_rate": 7.454545454545455e-05, "loss": 0.7876, "step": 41 }, { "epoch": 0.11467576791808874, "grad_norm": 0.1464846283197403, "learning_rate": 7.636363636363637e-05, "loss": 0.7846, "step": 42 }, { "epoch": 0.1174061433447099, "grad_norm": 0.1304199993610382, "learning_rate": 7.818181818181818e-05, "loss": 0.7734, "step": 43 }, { "epoch": 0.12013651877133105, "grad_norm": 0.1516261249780655, "learning_rate": 8e-05, "loss": 0.7836, "step": 44 }, { "epoch": 0.12286689419795221, "grad_norm": 0.1361905336380005, "learning_rate": 8.181818181818183e-05, "loss": 0.7751, "step": 45 }, { "epoch": 0.12559726962457338, "grad_norm": 0.14435631036758423, "learning_rate": 8.363636363636364e-05, "loss": 0.7655, "step": 46 }, { "epoch": 0.12832764505119454, "grad_norm": 0.13407501578330994, "learning_rate": 8.545454545454545e-05, "loss": 0.7729, "step": 47 }, { "epoch": 0.1310580204778157, "grad_norm": 0.13555769622325897, "learning_rate": 8.727272727272727e-05, "loss": 0.7693, "step": 48 }, { "epoch": 0.13378839590443686, "grad_norm": 0.1439952850341797, "learning_rate": 8.90909090909091e-05, "loss": 0.7787, "step": 49 }, { "epoch": 0.13651877133105803, "grad_norm": 0.14037510752677917, "learning_rate": 9.090909090909092e-05, "loss": 0.7757, "step": 50 }, { "epoch": 0.1392491467576792, "grad_norm": 0.11772281676530838, "learning_rate": 9.272727272727273e-05, "loss": 0.7408, "step": 51 }, { "epoch": 0.14197952218430035, "grad_norm": 0.1545950025320053, "learning_rate": 9.454545454545455e-05, "loss": 0.7381, "step": 52 }, { "epoch": 0.1447098976109215, "grad_norm": 0.12565699219703674, "learning_rate": 9.636363636363637e-05, "loss": 0.7769, "step": 53 }, { "epoch": 0.14744027303754267, "grad_norm": 0.15412947535514832, "learning_rate": 9.818181818181818e-05, "loss": 0.7585, "step": 54 }, { "epoch": 0.15017064846416384, "grad_norm": 0.11638892441987991, "learning_rate": 0.0001, "loss": 0.7661, "step": 55 }, { "epoch": 0.15290102389078497, "grad_norm": 0.16432470083236694, "learning_rate": 0.00010181818181818181, "loss": 0.7546, "step": 56 }, { "epoch": 0.15563139931740613, "grad_norm": 0.11529026180505753, "learning_rate": 0.00010363636363636364, "loss": 0.7535, "step": 57 }, { "epoch": 0.1583617747440273, "grad_norm": 0.23582805693149567, "learning_rate": 0.00010545454545454545, "loss": 0.7683, "step": 58 }, { "epoch": 0.16109215017064846, "grad_norm": 0.12393908202648163, "learning_rate": 0.00010727272727272728, "loss": 0.7526, "step": 59 }, { "epoch": 0.16382252559726962, "grad_norm": 0.20981422066688538, "learning_rate": 0.00010909090909090909, "loss": 0.7397, "step": 60 }, { "epoch": 0.16655290102389078, "grad_norm": 0.1519405096769333, "learning_rate": 0.00011090909090909092, "loss": 0.7472, "step": 61 }, { "epoch": 0.16928327645051194, "grad_norm": 0.15360122919082642, "learning_rate": 0.00011272727272727272, "loss": 0.7452, "step": 62 }, { "epoch": 0.1720136518771331, "grad_norm": 0.14730164408683777, "learning_rate": 0.00011454545454545456, "loss": 0.7068, "step": 63 }, { "epoch": 0.17474402730375427, "grad_norm": 0.166826993227005, "learning_rate": 0.00011636363636363636, "loss": 0.7476, "step": 64 }, { "epoch": 0.17747440273037543, "grad_norm": 0.17365337908267975, "learning_rate": 0.0001181818181818182, "loss": 0.7083, "step": 65 }, { "epoch": 0.1802047781569966, "grad_norm": 0.2101927250623703, "learning_rate": 0.00012, "loss": 0.755, "step": 66 }, { "epoch": 0.18293515358361775, "grad_norm": 0.1802549660205841, "learning_rate": 0.00012181818181818183, "loss": 0.7388, "step": 67 }, { "epoch": 0.18566552901023892, "grad_norm": 0.17372193932533264, "learning_rate": 0.00012363636363636364, "loss": 0.7067, "step": 68 }, { "epoch": 0.18839590443686008, "grad_norm": 0.15312258899211884, "learning_rate": 0.00012545454545454546, "loss": 0.7424, "step": 69 }, { "epoch": 0.19112627986348124, "grad_norm": 0.24253840744495392, "learning_rate": 0.00012727272727272728, "loss": 0.7418, "step": 70 }, { "epoch": 0.19385665529010238, "grad_norm": 0.2043231725692749, "learning_rate": 0.0001290909090909091, "loss": 0.7362, "step": 71 }, { "epoch": 0.19658703071672354, "grad_norm": 0.19859246909618378, "learning_rate": 0.00013090909090909093, "loss": 0.7071, "step": 72 }, { "epoch": 0.1993174061433447, "grad_norm": 0.20175667107105255, "learning_rate": 0.00013272727272727275, "loss": 0.7202, "step": 73 }, { "epoch": 0.20204778156996586, "grad_norm": 0.1653033196926117, "learning_rate": 0.00013454545454545455, "loss": 0.7229, "step": 74 }, { "epoch": 0.20477815699658702, "grad_norm": 0.17003491520881653, "learning_rate": 0.00013636363636363637, "loss": 0.7353, "step": 75 }, { "epoch": 0.2075085324232082, "grad_norm": 0.18356764316558838, "learning_rate": 0.0001381818181818182, "loss": 0.7138, "step": 76 }, { "epoch": 0.21023890784982935, "grad_norm": 0.2215511053800583, "learning_rate": 0.00014, "loss": 0.7238, "step": 77 }, { "epoch": 0.2129692832764505, "grad_norm": 0.34184327721595764, "learning_rate": 0.00014181818181818184, "loss": 0.7297, "step": 78 }, { "epoch": 0.21569965870307167, "grad_norm": 0.25543472170829773, "learning_rate": 0.00014363636363636363, "loss": 0.742, "step": 79 }, { "epoch": 0.21843003412969283, "grad_norm": 0.2220849245786667, "learning_rate": 0.00014545454545454546, "loss": 0.6897, "step": 80 }, { "epoch": 0.221160409556314, "grad_norm": 0.25552013516426086, "learning_rate": 0.00014727272727272728, "loss": 0.744, "step": 81 }, { "epoch": 0.22389078498293516, "grad_norm": 0.2621108889579773, "learning_rate": 0.0001490909090909091, "loss": 0.7193, "step": 82 }, { "epoch": 0.22662116040955632, "grad_norm": 0.1840047836303711, "learning_rate": 0.0001509090909090909, "loss": 0.6991, "step": 83 }, { "epoch": 0.22935153583617748, "grad_norm": 0.21538959443569183, "learning_rate": 0.00015272727272727275, "loss": 0.7058, "step": 84 }, { "epoch": 0.23208191126279865, "grad_norm": 0.34613293409347534, "learning_rate": 0.00015454545454545454, "loss": 0.7011, "step": 85 }, { "epoch": 0.2348122866894198, "grad_norm": 0.26010966300964355, "learning_rate": 0.00015636363636363637, "loss": 0.7009, "step": 86 }, { "epoch": 0.23754266211604094, "grad_norm": 0.18031255900859833, "learning_rate": 0.0001581818181818182, "loss": 0.7036, "step": 87 }, { "epoch": 0.2402730375426621, "grad_norm": 0.24567286670207977, "learning_rate": 0.00016, "loss": 0.6921, "step": 88 }, { "epoch": 0.24300341296928327, "grad_norm": 0.19522973895072937, "learning_rate": 0.00016181818181818184, "loss": 0.7109, "step": 89 }, { "epoch": 0.24573378839590443, "grad_norm": 0.2405068725347519, "learning_rate": 0.00016363636363636366, "loss": 0.7134, "step": 90 }, { "epoch": 0.2484641638225256, "grad_norm": 0.15669392049312592, "learning_rate": 0.00016545454545454545, "loss": 0.6966, "step": 91 }, { "epoch": 0.25119453924914675, "grad_norm": 0.23415732383728027, "learning_rate": 0.00016727272727272728, "loss": 0.6771, "step": 92 }, { "epoch": 0.25392491467576794, "grad_norm": 0.1842266321182251, "learning_rate": 0.0001690909090909091, "loss": 0.6895, "step": 93 }, { "epoch": 0.2566552901023891, "grad_norm": 0.21642841398715973, "learning_rate": 0.0001709090909090909, "loss": 0.6913, "step": 94 }, { "epoch": 0.2593856655290102, "grad_norm": 0.26327016949653625, "learning_rate": 0.00017272727272727275, "loss": 0.6856, "step": 95 }, { "epoch": 0.2621160409556314, "grad_norm": 0.20735357701778412, "learning_rate": 0.00017454545454545454, "loss": 0.6769, "step": 96 }, { "epoch": 0.26484641638225254, "grad_norm": 0.3127861022949219, "learning_rate": 0.00017636363636363637, "loss": 0.6518, "step": 97 }, { "epoch": 0.2675767918088737, "grad_norm": 0.34650346636772156, "learning_rate": 0.0001781818181818182, "loss": 0.6937, "step": 98 }, { "epoch": 0.27030716723549486, "grad_norm": 0.252059668302536, "learning_rate": 0.00018, "loss": 0.6589, "step": 99 }, { "epoch": 0.27303754266211605, "grad_norm": 0.1896669715642929, "learning_rate": 0.00018181818181818183, "loss": 0.6717, "step": 100 }, { "epoch": 0.2757679180887372, "grad_norm": 0.2724236249923706, "learning_rate": 0.00018363636363636366, "loss": 0.6671, "step": 101 }, { "epoch": 0.2784982935153584, "grad_norm": 0.1814826875925064, "learning_rate": 0.00018545454545454545, "loss": 0.6564, "step": 102 }, { "epoch": 0.2812286689419795, "grad_norm": 0.24140000343322754, "learning_rate": 0.00018727272727272728, "loss": 0.6708, "step": 103 }, { "epoch": 0.2839590443686007, "grad_norm": 0.19333204627037048, "learning_rate": 0.0001890909090909091, "loss": 0.6628, "step": 104 }, { "epoch": 0.28668941979522183, "grad_norm": 0.21080803871154785, "learning_rate": 0.00019090909090909092, "loss": 0.6457, "step": 105 }, { "epoch": 0.289419795221843, "grad_norm": 0.20848962664604187, "learning_rate": 0.00019272727272727274, "loss": 0.6589, "step": 106 }, { "epoch": 0.29215017064846416, "grad_norm": 0.2381501942873001, "learning_rate": 0.00019454545454545457, "loss": 0.6846, "step": 107 }, { "epoch": 0.29488054607508535, "grad_norm": 0.1734190434217453, "learning_rate": 0.00019636363636363636, "loss": 0.6691, "step": 108 }, { "epoch": 0.2976109215017065, "grad_norm": 0.2187424749135971, "learning_rate": 0.00019818181818181821, "loss": 0.6806, "step": 109 }, { "epoch": 0.3003412969283277, "grad_norm": 0.21607345342636108, "learning_rate": 0.0002, "loss": 0.6588, "step": 110 }, { "epoch": 0.3030716723549488, "grad_norm": 0.1873304545879364, "learning_rate": 0.00019999949446003433, "loss": 0.6513, "step": 111 }, { "epoch": 0.30580204778156994, "grad_norm": 0.21443282067775726, "learning_rate": 0.00019999797784524866, "loss": 0.6704, "step": 112 }, { "epoch": 0.30853242320819113, "grad_norm": 0.18743731081485748, "learning_rate": 0.00019999545017097728, "loss": 0.6346, "step": 113 }, { "epoch": 0.31126279863481227, "grad_norm": 0.18916335701942444, "learning_rate": 0.0001999919114627769, "loss": 0.68, "step": 114 }, { "epoch": 0.31399317406143346, "grad_norm": 0.1925644427537918, "learning_rate": 0.00019998736175642673, "loss": 0.6408, "step": 115 }, { "epoch": 0.3167235494880546, "grad_norm": 0.17722898721694946, "learning_rate": 0.0001999818010979279, "loss": 0.6787, "step": 116 }, { "epoch": 0.3194539249146758, "grad_norm": 0.19374825060367584, "learning_rate": 0.0001999752295435032, "loss": 0.669, "step": 117 }, { "epoch": 0.3221843003412969, "grad_norm": 0.20013949275016785, "learning_rate": 0.00019996764715959618, "loss": 0.653, "step": 118 }, { "epoch": 0.3249146757679181, "grad_norm": 0.18780681490898132, "learning_rate": 0.00019995905402287094, "loss": 0.6557, "step": 119 }, { "epoch": 0.32764505119453924, "grad_norm": 0.1718084216117859, "learning_rate": 0.00019994945022021082, "loss": 0.6272, "step": 120 }, { "epoch": 0.33037542662116043, "grad_norm": 0.1613592952489853, "learning_rate": 0.00019993883584871808, "loss": 0.6515, "step": 121 }, { "epoch": 0.33310580204778156, "grad_norm": 0.1737043410539627, "learning_rate": 0.00019992721101571236, "loss": 0.6134, "step": 122 }, { "epoch": 0.33583617747440275, "grad_norm": 0.16362418234348297, "learning_rate": 0.0001999145758387301, "loss": 0.6448, "step": 123 }, { "epoch": 0.3385665529010239, "grad_norm": 0.19181552529335022, "learning_rate": 0.00019990093044552304, "loss": 0.6497, "step": 124 }, { "epoch": 0.3412969283276451, "grad_norm": 0.15803317725658417, "learning_rate": 0.00019988627497405696, "loss": 0.6116, "step": 125 }, { "epoch": 0.3440273037542662, "grad_norm": 0.2208717167377472, "learning_rate": 0.00019987060957251047, "loss": 0.6459, "step": 126 }, { "epoch": 0.34675767918088735, "grad_norm": 0.20142869651317596, "learning_rate": 0.00019985393439927323, "loss": 0.6589, "step": 127 }, { "epoch": 0.34948805460750854, "grad_norm": 0.17945925891399384, "learning_rate": 0.00019983624962294458, "loss": 0.6252, "step": 128 }, { "epoch": 0.35221843003412967, "grad_norm": 0.22226247191429138, "learning_rate": 0.00019981755542233177, "loss": 0.6379, "step": 129 }, { "epoch": 0.35494880546075086, "grad_norm": 0.1588139832019806, "learning_rate": 0.00019979785198644806, "loss": 0.6408, "step": 130 }, { "epoch": 0.357679180887372, "grad_norm": 0.22639498114585876, "learning_rate": 0.00019977713951451102, "loss": 0.6434, "step": 131 }, { "epoch": 0.3604095563139932, "grad_norm": 0.16015386581420898, "learning_rate": 0.00019975541821594026, "loss": 0.6151, "step": 132 }, { "epoch": 0.3631399317406143, "grad_norm": 0.21671050786972046, "learning_rate": 0.00019973268831035545, "loss": 0.6357, "step": 133 }, { "epoch": 0.3658703071672355, "grad_norm": 0.1871589720249176, "learning_rate": 0.00019970895002757413, "loss": 0.6436, "step": 134 }, { "epoch": 0.36860068259385664, "grad_norm": 0.19527480006217957, "learning_rate": 0.00019968420360760926, "loss": 0.6308, "step": 135 }, { "epoch": 0.37133105802047783, "grad_norm": 0.20158074796199799, "learning_rate": 0.000199658449300667, "loss": 0.6227, "step": 136 }, { "epoch": 0.37406143344709897, "grad_norm": 0.15605966746807098, "learning_rate": 0.00019963168736714392, "loss": 0.615, "step": 137 }, { "epoch": 0.37679180887372016, "grad_norm": 0.22042252123355865, "learning_rate": 0.00019960391807762463, "loss": 0.6263, "step": 138 }, { "epoch": 0.3795221843003413, "grad_norm": 0.16206978261470795, "learning_rate": 0.00019957514171287875, "loss": 0.6182, "step": 139 }, { "epoch": 0.3822525597269625, "grad_norm": 0.2251751869916916, "learning_rate": 0.00019954535856385837, "loss": 0.6376, "step": 140 }, { "epoch": 0.3849829351535836, "grad_norm": 0.16586551070213318, "learning_rate": 0.00019951456893169497, "loss": 0.6285, "step": 141 }, { "epoch": 0.38771331058020475, "grad_norm": 0.27427414059638977, "learning_rate": 0.0001994827731276963, "loss": 0.6397, "step": 142 }, { "epoch": 0.39044368600682594, "grad_norm": 0.21177491545677185, "learning_rate": 0.00019944997147334337, "loss": 0.6034, "step": 143 }, { "epoch": 0.3931740614334471, "grad_norm": 0.25477880239486694, "learning_rate": 0.0001994161643002871, "loss": 0.6199, "step": 144 }, { "epoch": 0.39590443686006827, "grad_norm": 0.23290970921516418, "learning_rate": 0.00019938135195034508, "loss": 0.6201, "step": 145 }, { "epoch": 0.3986348122866894, "grad_norm": 0.19198672473430634, "learning_rate": 0.00019934553477549794, "loss": 0.6213, "step": 146 }, { "epoch": 0.4013651877133106, "grad_norm": 0.1911400705575943, "learning_rate": 0.000199308713137886, "loss": 0.6146, "step": 147 }, { "epoch": 0.4040955631399317, "grad_norm": 0.18605491518974304, "learning_rate": 0.0001992708874098054, "loss": 0.6123, "step": 148 }, { "epoch": 0.4068259385665529, "grad_norm": 0.18028293550014496, "learning_rate": 0.0001992320579737045, "loss": 0.6061, "step": 149 }, { "epoch": 0.40955631399317405, "grad_norm": 0.1961037963628769, "learning_rate": 0.00019919222522217996, "loss": 0.622, "step": 150 }, { "epoch": 0.41228668941979524, "grad_norm": 0.17400594055652618, "learning_rate": 0.00019915138955797272, "loss": 0.6138, "step": 151 }, { "epoch": 0.4150170648464164, "grad_norm": 0.17892149090766907, "learning_rate": 0.00019910955139396396, "loss": 0.6242, "step": 152 }, { "epoch": 0.41774744027303756, "grad_norm": 0.21851663291454315, "learning_rate": 0.000199066711153171, "loss": 0.5913, "step": 153 }, { "epoch": 0.4204778156996587, "grad_norm": 0.1468774825334549, "learning_rate": 0.0001990228692687429, "loss": 0.6025, "step": 154 }, { "epoch": 0.4232081911262799, "grad_norm": 0.1920468658208847, "learning_rate": 0.00019897802618395614, "loss": 0.6127, "step": 155 }, { "epoch": 0.425938566552901, "grad_norm": 0.17375442385673523, "learning_rate": 0.00019893218235221015, "loss": 0.6211, "step": 156 }, { "epoch": 0.4286689419795222, "grad_norm": 0.15414904057979584, "learning_rate": 0.00019888533823702277, "loss": 0.6183, "step": 157 }, { "epoch": 0.43139931740614335, "grad_norm": 0.2245103418827057, "learning_rate": 0.0001988374943120254, "loss": 0.6248, "step": 158 }, { "epoch": 0.4341296928327645, "grad_norm": 0.17193332314491272, "learning_rate": 0.00019878865106095835, "loss": 0.5969, "step": 159 }, { "epoch": 0.43686006825938567, "grad_norm": 0.13767646253108978, "learning_rate": 0.00019873880897766598, "loss": 0.5943, "step": 160 }, { "epoch": 0.4395904436860068, "grad_norm": 0.1449906826019287, "learning_rate": 0.00019868796856609152, "loss": 0.573, "step": 161 }, { "epoch": 0.442320819112628, "grad_norm": 0.1392473578453064, "learning_rate": 0.00019863613034027224, "loss": 0.5926, "step": 162 }, { "epoch": 0.44505119453924913, "grad_norm": 0.1772463023662567, "learning_rate": 0.00019858329482433403, "loss": 0.6007, "step": 163 }, { "epoch": 0.4477815699658703, "grad_norm": 0.13768768310546875, "learning_rate": 0.0001985294625524861, "loss": 0.5901, "step": 164 }, { "epoch": 0.45051194539249145, "grad_norm": 0.1631435751914978, "learning_rate": 0.00019847463406901588, "loss": 0.5907, "step": 165 }, { "epoch": 0.45324232081911264, "grad_norm": 0.14781758189201355, "learning_rate": 0.00019841880992828306, "loss": 0.5903, "step": 166 }, { "epoch": 0.4559726962457338, "grad_norm": 0.13440802693367004, "learning_rate": 0.00019836199069471437, "loss": 0.5884, "step": 167 }, { "epoch": 0.45870307167235497, "grad_norm": 0.1414463371038437, "learning_rate": 0.00019830417694279766, "loss": 0.598, "step": 168 }, { "epoch": 0.4614334470989761, "grad_norm": 0.13185666501522064, "learning_rate": 0.0001982453692570762, "loss": 0.621, "step": 169 }, { "epoch": 0.4641638225255973, "grad_norm": 0.14422471821308136, "learning_rate": 0.00019818556823214268, "loss": 0.6065, "step": 170 }, { "epoch": 0.4668941979522184, "grad_norm": 0.13765788078308105, "learning_rate": 0.00019812477447263326, "loss": 0.6073, "step": 171 }, { "epoch": 0.4696245733788396, "grad_norm": 0.16028070449829102, "learning_rate": 0.0001980629885932214, "loss": 0.5767, "step": 172 }, { "epoch": 0.47235494880546075, "grad_norm": 0.14638394117355347, "learning_rate": 0.00019800021121861182, "loss": 0.5971, "step": 173 }, { "epoch": 0.4750853242320819, "grad_norm": 0.14843404293060303, "learning_rate": 0.0001979364429835339, "loss": 0.5894, "step": 174 }, { "epoch": 0.4778156996587031, "grad_norm": 0.13411492109298706, "learning_rate": 0.00019787168453273544, "loss": 0.5757, "step": 175 }, { "epoch": 0.4805460750853242, "grad_norm": 0.14304684102535248, "learning_rate": 0.0001978059365209762, "loss": 0.5846, "step": 176 }, { "epoch": 0.4832764505119454, "grad_norm": 0.13569754362106323, "learning_rate": 0.00019773919961302113, "loss": 0.5872, "step": 177 }, { "epoch": 0.48600682593856653, "grad_norm": 0.14318887889385223, "learning_rate": 0.00019767147448363366, "loss": 0.5804, "step": 178 }, { "epoch": 0.4887372013651877, "grad_norm": 0.1457952857017517, "learning_rate": 0.00019760276181756903, "loss": 0.5973, "step": 179 }, { "epoch": 0.49146757679180886, "grad_norm": 0.13820476830005646, "learning_rate": 0.00019753306230956718, "loss": 0.569, "step": 180 }, { "epoch": 0.49419795221843005, "grad_norm": 0.19338561594486237, "learning_rate": 0.00019746237666434587, "loss": 0.5723, "step": 181 }, { "epoch": 0.4969283276450512, "grad_norm": 0.17352697253227234, "learning_rate": 0.00019739070559659347, "loss": 0.578, "step": 182 }, { "epoch": 0.49965870307167237, "grad_norm": 0.15502339601516724, "learning_rate": 0.00019731804983096177, "loss": 0.5953, "step": 183 }, { "epoch": 0.5023890784982935, "grad_norm": 0.18948784470558167, "learning_rate": 0.00019724441010205863, "loss": 0.5883, "step": 184 }, { "epoch": 0.5051194539249146, "grad_norm": 0.17587606608867645, "learning_rate": 0.00019716978715444056, "loss": 0.5723, "step": 185 }, { "epoch": 0.5078498293515359, "grad_norm": 0.1599951833486557, "learning_rate": 0.0001970941817426052, "loss": 0.5799, "step": 186 }, { "epoch": 0.510580204778157, "grad_norm": 0.1717846840620041, "learning_rate": 0.00019701759463098374, "loss": 0.5543, "step": 187 }, { "epoch": 0.5133105802047782, "grad_norm": 0.14032602310180664, "learning_rate": 0.00019694002659393305, "loss": 0.5845, "step": 188 }, { "epoch": 0.5160409556313993, "grad_norm": 0.17668449878692627, "learning_rate": 0.000196861478415728, "loss": 0.6026, "step": 189 }, { "epoch": 0.5187713310580204, "grad_norm": 0.17806965112686157, "learning_rate": 0.00019678195089055346, "loss": 0.5681, "step": 190 }, { "epoch": 0.5215017064846417, "grad_norm": 0.13321803510189056, "learning_rate": 0.00019670144482249627, "loss": 0.5586, "step": 191 }, { "epoch": 0.5242320819112628, "grad_norm": 0.14684653282165527, "learning_rate": 0.00019661996102553718, "loss": 0.5589, "step": 192 }, { "epoch": 0.5269624573378839, "grad_norm": 0.1308140754699707, "learning_rate": 0.0001965375003235424, "loss": 0.568, "step": 193 }, { "epoch": 0.5296928327645051, "grad_norm": 0.17461615800857544, "learning_rate": 0.00019645406355025565, "loss": 0.5757, "step": 194 }, { "epoch": 0.5324232081911263, "grad_norm": 0.15591022372245789, "learning_rate": 0.0001963696515492893, "loss": 0.5946, "step": 195 }, { "epoch": 0.5351535836177475, "grad_norm": 0.14174342155456543, "learning_rate": 0.00019628426517411625, "loss": 0.5839, "step": 196 }, { "epoch": 0.5378839590443686, "grad_norm": 0.15242989361286163, "learning_rate": 0.0001961979052880609, "loss": 0.5567, "step": 197 }, { "epoch": 0.5406143344709897, "grad_norm": 0.16651766002178192, "learning_rate": 0.00019611057276429085, "loss": 0.5593, "step": 198 }, { "epoch": 0.543344709897611, "grad_norm": 0.14858382940292358, "learning_rate": 0.00019602226848580763, "loss": 0.5848, "step": 199 }, { "epoch": 0.5460750853242321, "grad_norm": 0.14774656295776367, "learning_rate": 0.00019593299334543808, "loss": 0.563, "step": 200 }, { "epoch": 0.5488054607508532, "grad_norm": 0.13993892073631287, "learning_rate": 0.0001958427482458253, "loss": 0.5742, "step": 201 }, { "epoch": 0.5515358361774744, "grad_norm": 0.15201717615127563, "learning_rate": 0.0001957515340994193, "loss": 0.5726, "step": 202 }, { "epoch": 0.5542662116040956, "grad_norm": 0.1567879170179367, "learning_rate": 0.00019565935182846802, "loss": 0.5707, "step": 203 }, { "epoch": 0.5569965870307167, "grad_norm": 0.13955365121364594, "learning_rate": 0.00019556620236500793, "loss": 0.5339, "step": 204 }, { "epoch": 0.5597269624573379, "grad_norm": 0.1425381898880005, "learning_rate": 0.00019547208665085457, "loss": 0.5698, "step": 205 }, { "epoch": 0.562457337883959, "grad_norm": 0.14695167541503906, "learning_rate": 0.00019537700563759304, "loss": 0.578, "step": 206 }, { "epoch": 0.5651877133105802, "grad_norm": 0.15581448376178741, "learning_rate": 0.00019528096028656832, "loss": 0.5552, "step": 207 }, { "epoch": 0.5679180887372014, "grad_norm": 0.13141174614429474, "learning_rate": 0.00019518395156887576, "loss": 0.5598, "step": 208 }, { "epoch": 0.5706484641638225, "grad_norm": 0.16678418219089508, "learning_rate": 0.00019508598046535095, "loss": 0.5485, "step": 209 }, { "epoch": 0.5733788395904437, "grad_norm": 0.1717272400856018, "learning_rate": 0.00019498704796656018, "loss": 0.5849, "step": 210 }, { "epoch": 0.5761092150170648, "grad_norm": 0.14453086256980896, "learning_rate": 0.00019488715507278998, "loss": 0.5757, "step": 211 }, { "epoch": 0.578839590443686, "grad_norm": 0.16329538822174072, "learning_rate": 0.0001947863027940374, "loss": 0.5632, "step": 212 }, { "epoch": 0.5815699658703072, "grad_norm": 0.15865112841129303, "learning_rate": 0.00019468449214999955, "loss": 0.5728, "step": 213 }, { "epoch": 0.5843003412969283, "grad_norm": 0.15316785871982574, "learning_rate": 0.00019458172417006347, "loss": 0.5556, "step": 214 }, { "epoch": 0.5870307167235495, "grad_norm": 0.16666734218597412, "learning_rate": 0.00019447799989329555, "loss": 0.5759, "step": 215 }, { "epoch": 0.5897610921501707, "grad_norm": 0.16525249183177948, "learning_rate": 0.00019437332036843118, "loss": 0.5667, "step": 216 }, { "epoch": 0.5924914675767918, "grad_norm": 0.14022761583328247, "learning_rate": 0.00019426768665386398, "loss": 0.5611, "step": 217 }, { "epoch": 0.595221843003413, "grad_norm": 0.15930500626564026, "learning_rate": 0.00019416109981763526, "loss": 0.5414, "step": 218 }, { "epoch": 0.5979522184300341, "grad_norm": 0.141464963555336, "learning_rate": 0.00019405356093742313, "loss": 0.5363, "step": 219 }, { "epoch": 0.6006825938566553, "grad_norm": 0.1541200429201126, "learning_rate": 0.0001939450711005316, "loss": 0.5487, "step": 220 }, { "epoch": 0.6034129692832765, "grad_norm": 0.13717712461948395, "learning_rate": 0.00019383563140387965, "loss": 0.5564, "step": 221 }, { "epoch": 0.6061433447098976, "grad_norm": 0.14139863848686218, "learning_rate": 0.00019372524295399013, "loss": 0.5592, "step": 222 }, { "epoch": 0.6088737201365187, "grad_norm": 0.13494791090488434, "learning_rate": 0.00019361390686697846, "loss": 0.5452, "step": 223 }, { "epoch": 0.6116040955631399, "grad_norm": 0.1512797623872757, "learning_rate": 0.0001935016242685415, "loss": 0.5595, "step": 224 }, { "epoch": 0.6143344709897611, "grad_norm": 0.1422545164823532, "learning_rate": 0.00019338839629394605, "loss": 0.5602, "step": 225 }, { "epoch": 0.6170648464163823, "grad_norm": 0.14444862306118011, "learning_rate": 0.00019327422408801744, "loss": 0.5574, "step": 226 }, { "epoch": 0.6197952218430034, "grad_norm": 0.15434536337852478, "learning_rate": 0.0001931591088051279, "loss": 0.5467, "step": 227 }, { "epoch": 0.6225255972696245, "grad_norm": 0.1420368254184723, "learning_rate": 0.000193043051609185, "loss": 0.5519, "step": 228 }, { "epoch": 0.6252559726962458, "grad_norm": 0.12724490463733673, "learning_rate": 0.00019292605367361978, "loss": 0.5779, "step": 229 }, { "epoch": 0.6279863481228669, "grad_norm": 0.13820625841617584, "learning_rate": 0.00019280811618137484, "loss": 0.5468, "step": 230 }, { "epoch": 0.630716723549488, "grad_norm": 0.1629246473312378, "learning_rate": 0.00019268924032489248, "loss": 0.5721, "step": 231 }, { "epoch": 0.6334470989761092, "grad_norm": 0.193836510181427, "learning_rate": 0.00019256942730610268, "loss": 0.5392, "step": 232 }, { "epoch": 0.6361774744027304, "grad_norm": 0.18511579930782318, "learning_rate": 0.0001924486783364108, "loss": 0.5718, "step": 233 }, { "epoch": 0.6389078498293516, "grad_norm": 0.14278100430965424, "learning_rate": 0.00019232699463668542, "loss": 0.5589, "step": 234 }, { "epoch": 0.6416382252559727, "grad_norm": 0.14693580567836761, "learning_rate": 0.00019220437743724605, "loss": 0.515, "step": 235 }, { "epoch": 0.6443686006825938, "grad_norm": 0.20018483698368073, "learning_rate": 0.00019208082797785055, "loss": 0.5484, "step": 236 }, { "epoch": 0.647098976109215, "grad_norm": 0.1597984880208969, "learning_rate": 0.00019195634750768275, "loss": 0.539, "step": 237 }, { "epoch": 0.6498293515358362, "grad_norm": 0.14270828664302826, "learning_rate": 0.00019183093728533966, "loss": 0.5496, "step": 238 }, { "epoch": 0.6525597269624573, "grad_norm": 0.1788954734802246, "learning_rate": 0.0001917045985788189, "loss": 0.5629, "step": 239 }, { "epoch": 0.6552901023890785, "grad_norm": 0.15234531462192535, "learning_rate": 0.00019157733266550575, "loss": 0.5454, "step": 240 }, { "epoch": 0.6580204778156996, "grad_norm": 0.17805363237857819, "learning_rate": 0.00019144914083216034, "loss": 0.5599, "step": 241 }, { "epoch": 0.6607508532423209, "grad_norm": 0.18313759565353394, "learning_rate": 0.00019132002437490458, "loss": 0.5544, "step": 242 }, { "epoch": 0.663481228668942, "grad_norm": 0.16873271763324738, "learning_rate": 0.00019118998459920902, "loss": 0.5445, "step": 243 }, { "epoch": 0.6662116040955631, "grad_norm": 0.23297423124313354, "learning_rate": 0.00019105902281987976, "loss": 0.5623, "step": 244 }, { "epoch": 0.6689419795221843, "grad_norm": 0.13279114663600922, "learning_rate": 0.00019092714036104508, "loss": 0.5354, "step": 245 }, { "epoch": 0.6716723549488055, "grad_norm": 0.18942788243293762, "learning_rate": 0.00019079433855614201, "loss": 0.5574, "step": 246 }, { "epoch": 0.6744027303754266, "grad_norm": 0.15293289721012115, "learning_rate": 0.00019066061874790303, "loss": 0.5473, "step": 247 }, { "epoch": 0.6771331058020478, "grad_norm": 0.1669221818447113, "learning_rate": 0.00019052598228834217, "loss": 0.5272, "step": 248 }, { "epoch": 0.6798634812286689, "grad_norm": 0.16368111968040466, "learning_rate": 0.00019039043053874175, "loss": 0.5387, "step": 249 }, { "epoch": 0.6825938566552902, "grad_norm": 0.15945561230182648, "learning_rate": 0.00019025396486963827, "loss": 0.548, "step": 250 }, { "epoch": 0.6853242320819113, "grad_norm": 0.1538572758436203, "learning_rate": 0.00019011658666080874, "loss": 0.5419, "step": 251 }, { "epoch": 0.6880546075085324, "grad_norm": 0.13356320559978485, "learning_rate": 0.00018997829730125663, "loss": 0.5271, "step": 252 }, { "epoch": 0.6907849829351536, "grad_norm": 0.1389850527048111, "learning_rate": 0.0001898390981891979, "loss": 0.5489, "step": 253 }, { "epoch": 0.6935153583617747, "grad_norm": 0.14726200699806213, "learning_rate": 0.00018969899073204686, "loss": 0.554, "step": 254 }, { "epoch": 0.6962457337883959, "grad_norm": 0.13865283131599426, "learning_rate": 0.0001895579763464019, "loss": 0.5486, "step": 255 }, { "epoch": 0.6989761092150171, "grad_norm": 0.15216147899627686, "learning_rate": 0.00018941605645803115, "loss": 0.5156, "step": 256 }, { "epoch": 0.7017064846416382, "grad_norm": 0.13091793656349182, "learning_rate": 0.00018927323250185815, "loss": 0.5359, "step": 257 }, { "epoch": 0.7044368600682593, "grad_norm": 0.14097946882247925, "learning_rate": 0.0001891295059219472, "loss": 0.5367, "step": 258 }, { "epoch": 0.7071672354948806, "grad_norm": 0.1434548795223236, "learning_rate": 0.00018898487817148898, "loss": 0.5395, "step": 259 }, { "epoch": 0.7098976109215017, "grad_norm": 0.15296564996242523, "learning_rate": 0.0001888393507127856, "loss": 0.5552, "step": 260 }, { "epoch": 0.7126279863481229, "grad_norm": 0.15522688627243042, "learning_rate": 0.000188692925017236, "loss": 0.5433, "step": 261 }, { "epoch": 0.715358361774744, "grad_norm": 0.1461726576089859, "learning_rate": 0.000188545602565321, "loss": 0.5291, "step": 262 }, { "epoch": 0.7180887372013652, "grad_norm": 0.14327403903007507, "learning_rate": 0.00018839738484658836, "loss": 0.5502, "step": 263 }, { "epoch": 0.7208191126279864, "grad_norm": 0.1613275706768036, "learning_rate": 0.00018824827335963765, "loss": 0.5565, "step": 264 }, { "epoch": 0.7235494880546075, "grad_norm": 0.2040315866470337, "learning_rate": 0.00018809826961210525, "loss": 0.5324, "step": 265 }, { "epoch": 0.7262798634812286, "grad_norm": 0.13186219334602356, "learning_rate": 0.0001879473751206489, "loss": 0.5395, "step": 266 }, { "epoch": 0.7290102389078499, "grad_norm": 0.15462934970855713, "learning_rate": 0.00018779559141093258, "loss": 0.527, "step": 267 }, { "epoch": 0.731740614334471, "grad_norm": 0.155660942196846, "learning_rate": 0.0001876429200176108, "loss": 0.5429, "step": 268 }, { "epoch": 0.7344709897610922, "grad_norm": 0.15280231833457947, "learning_rate": 0.00018748936248431353, "loss": 0.5571, "step": 269 }, { "epoch": 0.7372013651877133, "grad_norm": 0.19389967620372772, "learning_rate": 0.00018733492036363005, "loss": 0.5325, "step": 270 }, { "epoch": 0.7399317406143344, "grad_norm": 0.14486828446388245, "learning_rate": 0.0001871795952170937, "loss": 0.5229, "step": 271 }, { "epoch": 0.7426621160409557, "grad_norm": 0.18223008513450623, "learning_rate": 0.00018702338861516587, "loss": 0.538, "step": 272 }, { "epoch": 0.7453924914675768, "grad_norm": 0.1434670090675354, "learning_rate": 0.00018686630213722016, "loss": 0.5353, "step": 273 }, { "epoch": 0.7481228668941979, "grad_norm": 0.1796412169933319, "learning_rate": 0.0001867083373715264, "loss": 0.5221, "step": 274 }, { "epoch": 0.7508532423208191, "grad_norm": 0.14365307986736298, "learning_rate": 0.00018654949591523465, "loss": 0.5211, "step": 275 }, { "epoch": 0.7535836177474403, "grad_norm": 0.15224212408065796, "learning_rate": 0.000186389779374359, "loss": 0.5353, "step": 276 }, { "epoch": 0.7563139931740614, "grad_norm": 0.1557937115430832, "learning_rate": 0.00018622918936376132, "loss": 0.5359, "step": 277 }, { "epoch": 0.7590443686006826, "grad_norm": 0.13806480169296265, "learning_rate": 0.00018606772750713504, "loss": 0.5404, "step": 278 }, { "epoch": 0.7617747440273037, "grad_norm": 0.14786981046199799, "learning_rate": 0.00018590539543698854, "loss": 0.5377, "step": 279 }, { "epoch": 0.764505119453925, "grad_norm": 0.14674220979213715, "learning_rate": 0.00018574219479462878, "loss": 0.531, "step": 280 }, { "epoch": 0.7672354948805461, "grad_norm": 0.146319180727005, "learning_rate": 0.00018557812723014476, "loss": 0.5244, "step": 281 }, { "epoch": 0.7699658703071672, "grad_norm": 0.15404731035232544, "learning_rate": 0.00018541319440239066, "loss": 0.5419, "step": 282 }, { "epoch": 0.7726962457337884, "grad_norm": 0.13534784317016602, "learning_rate": 0.00018524739797896923, "loss": 0.5162, "step": 283 }, { "epoch": 0.7754266211604095, "grad_norm": 0.14462386071681976, "learning_rate": 0.0001850807396362148, "loss": 0.5338, "step": 284 }, { "epoch": 0.7781569965870307, "grad_norm": 0.126676544547081, "learning_rate": 0.00018491322105917645, "loss": 0.5291, "step": 285 }, { "epoch": 0.7808873720136519, "grad_norm": 0.1487646847963333, "learning_rate": 0.0001847448439416009, "loss": 0.5235, "step": 286 }, { "epoch": 0.783617747440273, "grad_norm": 0.14197687804698944, "learning_rate": 0.00018457560998591538, "loss": 0.518, "step": 287 }, { "epoch": 0.7863481228668942, "grad_norm": 0.13646024465560913, "learning_rate": 0.00018440552090321047, "loss": 0.5425, "step": 288 }, { "epoch": 0.7890784982935154, "grad_norm": 0.15339186787605286, "learning_rate": 0.00018423457841322273, "loss": 0.5203, "step": 289 }, { "epoch": 0.7918088737201365, "grad_norm": 0.14754898846149445, "learning_rate": 0.00018406278424431736, "loss": 0.5259, "step": 290 }, { "epoch": 0.7945392491467577, "grad_norm": 0.1315944641828537, "learning_rate": 0.00018389014013347078, "loss": 0.5248, "step": 291 }, { "epoch": 0.7972696245733788, "grad_norm": 0.14218468964099884, "learning_rate": 0.00018371664782625287, "loss": 0.5328, "step": 292 }, { "epoch": 0.8, "grad_norm": 0.1456083059310913, "learning_rate": 0.00018354230907680958, "loss": 0.5352, "step": 293 }, { "epoch": 0.8027303754266212, "grad_norm": 0.14852645993232727, "learning_rate": 0.00018336712564784503, "loss": 0.5379, "step": 294 }, { "epoch": 0.8054607508532423, "grad_norm": 0.17426247894763947, "learning_rate": 0.0001831910993106037, "loss": 0.5358, "step": 295 }, { "epoch": 0.8081911262798634, "grad_norm": 0.14848141372203827, "learning_rate": 0.0001830142318448525, "loss": 0.5377, "step": 296 }, { "epoch": 0.8109215017064847, "grad_norm": 0.14610137045383453, "learning_rate": 0.000182836525038863, "loss": 0.5331, "step": 297 }, { "epoch": 0.8136518771331058, "grad_norm": 0.1565829962491989, "learning_rate": 0.00018265798068939294, "loss": 0.5162, "step": 298 }, { "epoch": 0.816382252559727, "grad_norm": 0.1328669935464859, "learning_rate": 0.0001824786006016685, "loss": 0.5045, "step": 299 }, { "epoch": 0.8191126279863481, "grad_norm": 0.14329080283641815, "learning_rate": 0.00018229838658936564, "loss": 0.5083, "step": 300 }, { "epoch": 0.8218430034129692, "grad_norm": 0.13927890360355377, "learning_rate": 0.00018211734047459216, "loss": 0.5221, "step": 301 }, { "epoch": 0.8245733788395905, "grad_norm": 0.14164070785045624, "learning_rate": 0.00018193546408786898, "loss": 0.5243, "step": 302 }, { "epoch": 0.8273037542662116, "grad_norm": 0.14735499024391174, "learning_rate": 0.00018175275926811174, "loss": 0.5136, "step": 303 }, { "epoch": 0.8300341296928327, "grad_norm": 0.1411520093679428, "learning_rate": 0.00018156922786261216, "loss": 0.5164, "step": 304 }, { "epoch": 0.8327645051194539, "grad_norm": 0.14603470265865326, "learning_rate": 0.0001813848717270195, "loss": 0.5306, "step": 305 }, { "epoch": 0.8354948805460751, "grad_norm": 0.13602910935878754, "learning_rate": 0.00018119969272532166, "loss": 0.5014, "step": 306 }, { "epoch": 0.8382252559726963, "grad_norm": 0.15259377658367157, "learning_rate": 0.00018101369272982632, "loss": 0.5309, "step": 307 }, { "epoch": 0.8409556313993174, "grad_norm": 0.13250286877155304, "learning_rate": 0.00018082687362114212, "loss": 0.5034, "step": 308 }, { "epoch": 0.8436860068259385, "grad_norm": 0.15984928607940674, "learning_rate": 0.00018063923728815957, "loss": 0.5167, "step": 309 }, { "epoch": 0.8464163822525598, "grad_norm": 0.14640702307224274, "learning_rate": 0.00018045078562803203, "loss": 0.5361, "step": 310 }, { "epoch": 0.8491467576791809, "grad_norm": 0.1568063646554947, "learning_rate": 0.00018026152054615634, "loss": 0.5221, "step": 311 }, { "epoch": 0.851877133105802, "grad_norm": 0.14193738996982574, "learning_rate": 0.0001800714439561538, "loss": 0.5151, "step": 312 }, { "epoch": 0.8546075085324232, "grad_norm": 0.15847285091876984, "learning_rate": 0.00017988055777985067, "loss": 0.4923, "step": 313 }, { "epoch": 0.8573378839590444, "grad_norm": 0.16404108703136444, "learning_rate": 0.00017968886394725874, "loss": 0.5114, "step": 314 }, { "epoch": 0.8600682593856656, "grad_norm": 0.22749370336532593, "learning_rate": 0.00017949636439655592, "loss": 0.5225, "step": 315 }, { "epoch": 0.8627986348122867, "grad_norm": 0.17752817273139954, "learning_rate": 0.00017930306107406653, "loss": 0.5138, "step": 316 }, { "epoch": 0.8655290102389078, "grad_norm": 0.16554813086986542, "learning_rate": 0.00017910895593424163, "loss": 0.5087, "step": 317 }, { "epoch": 0.868259385665529, "grad_norm": 0.1888076812028885, "learning_rate": 0.00017891405093963938, "loss": 0.5163, "step": 318 }, { "epoch": 0.8709897610921502, "grad_norm": 0.15817560255527496, "learning_rate": 0.00017871834806090501, "loss": 0.4918, "step": 319 }, { "epoch": 0.8737201365187713, "grad_norm": 0.15167172253131866, "learning_rate": 0.00017852184927675112, "loss": 0.5248, "step": 320 }, { "epoch": 0.8764505119453925, "grad_norm": 0.15847422182559967, "learning_rate": 0.00017832455657393746, "loss": 0.5205, "step": 321 }, { "epoch": 0.8791808873720136, "grad_norm": 0.1436643898487091, "learning_rate": 0.00017812647194725094, "loss": 0.4869, "step": 322 }, { "epoch": 0.8819112627986349, "grad_norm": 0.1533607393503189, "learning_rate": 0.00017792759739948546, "loss": 0.5141, "step": 323 }, { "epoch": 0.884641638225256, "grad_norm": 0.1520746648311615, "learning_rate": 0.00017772793494142167, "loss": 0.5158, "step": 324 }, { "epoch": 0.8873720136518771, "grad_norm": 0.1398312896490097, "learning_rate": 0.0001775274865918066, "loss": 0.5126, "step": 325 }, { "epoch": 0.8901023890784983, "grad_norm": 0.14314454793930054, "learning_rate": 0.00017732625437733335, "loss": 0.5187, "step": 326 }, { "epoch": 0.8928327645051195, "grad_norm": 0.1434595286846161, "learning_rate": 0.00017712424033262042, "loss": 0.5018, "step": 327 }, { "epoch": 0.8955631399317406, "grad_norm": 0.13712714612483978, "learning_rate": 0.00017692144650019125, "loss": 0.5044, "step": 328 }, { "epoch": 0.8982935153583618, "grad_norm": 0.13560262322425842, "learning_rate": 0.00017671787493045356, "loss": 0.5153, "step": 329 }, { "epoch": 0.9010238907849829, "grad_norm": 0.15035240352153778, "learning_rate": 0.0001765135276816787, "loss": 0.5023, "step": 330 }, { "epoch": 0.903754266211604, "grad_norm": 0.14878690242767334, "learning_rate": 0.00017630840681998066, "loss": 0.4916, "step": 331 }, { "epoch": 0.9064846416382253, "grad_norm": 0.14930297434329987, "learning_rate": 0.00017610251441929533, "loss": 0.5092, "step": 332 }, { "epoch": 0.9092150170648464, "grad_norm": 0.14058218896389008, "learning_rate": 0.0001758958525613594, "loss": 0.5091, "step": 333 }, { "epoch": 0.9119453924914676, "grad_norm": 0.18505944311618805, "learning_rate": 0.00017568842333568952, "loss": 0.5047, "step": 334 }, { "epoch": 0.9146757679180887, "grad_norm": 0.16390664875507355, "learning_rate": 0.0001754802288395609, "loss": 0.5115, "step": 335 }, { "epoch": 0.9174061433447099, "grad_norm": 0.16126009821891785, "learning_rate": 0.00017527127117798635, "loss": 0.5017, "step": 336 }, { "epoch": 0.9201365187713311, "grad_norm": 0.16674400866031647, "learning_rate": 0.0001750615524636948, "loss": 0.5193, "step": 337 }, { "epoch": 0.9228668941979522, "grad_norm": 0.13656651973724365, "learning_rate": 0.00017485107481711012, "loss": 0.4909, "step": 338 }, { "epoch": 0.9255972696245733, "grad_norm": 0.15961140394210815, "learning_rate": 0.00017463984036632954, "loss": 0.5145, "step": 339 }, { "epoch": 0.9283276450511946, "grad_norm": 0.13122031092643738, "learning_rate": 0.00017442785124710227, "loss": 0.4854, "step": 340 }, { "epoch": 0.9310580204778157, "grad_norm": 0.14256154000759125, "learning_rate": 0.0001742151096028076, "loss": 0.4852, "step": 341 }, { "epoch": 0.9337883959044369, "grad_norm": 0.14853668212890625, "learning_rate": 0.00017400161758443375, "loss": 0.5202, "step": 342 }, { "epoch": 0.936518771331058, "grad_norm": 0.16885364055633545, "learning_rate": 0.00017378737735055562, "loss": 0.5136, "step": 343 }, { "epoch": 0.9392491467576792, "grad_norm": 0.1398458182811737, "learning_rate": 0.00017357239106731317, "loss": 0.4908, "step": 344 }, { "epoch": 0.9419795221843004, "grad_norm": 0.1871606856584549, "learning_rate": 0.00017335666090838965, "loss": 0.5076, "step": 345 }, { "epoch": 0.9447098976109215, "grad_norm": 0.17375800013542175, "learning_rate": 0.00017314018905498931, "loss": 0.5032, "step": 346 }, { "epoch": 0.9474402730375426, "grad_norm": 0.14859388768672943, "learning_rate": 0.00017292297769581571, "loss": 0.5119, "step": 347 }, { "epoch": 0.9501706484641638, "grad_norm": 0.18141716718673706, "learning_rate": 0.00017270502902704926, "loss": 0.4982, "step": 348 }, { "epoch": 0.952901023890785, "grad_norm": 0.17015349864959717, "learning_rate": 0.00017248634525232523, "loss": 0.4835, "step": 349 }, { "epoch": 0.9556313993174061, "grad_norm": 0.14770372211933136, "learning_rate": 0.00017226692858271134, "loss": 0.4851, "step": 350 }, { "epoch": 0.9583617747440273, "grad_norm": 0.19386035203933716, "learning_rate": 0.00017204678123668556, "loss": 0.5172, "step": 351 }, { "epoch": 0.9610921501706484, "grad_norm": 0.15225833654403687, "learning_rate": 0.00017182590544011347, "loss": 0.4972, "step": 352 }, { "epoch": 0.9638225255972697, "grad_norm": 0.1522843837738037, "learning_rate": 0.00017160430342622596, "loss": 0.493, "step": 353 }, { "epoch": 0.9665529010238908, "grad_norm": 0.13483871519565582, "learning_rate": 0.00017138197743559654, "loss": 0.4882, "step": 354 }, { "epoch": 0.9692832764505119, "grad_norm": 0.14176194369792938, "learning_rate": 0.00017115892971611863, "loss": 0.4921, "step": 355 }, { "epoch": 0.9720136518771331, "grad_norm": 0.15088023245334625, "learning_rate": 0.00017093516252298296, "loss": 0.5073, "step": 356 }, { "epoch": 0.9747440273037543, "grad_norm": 0.14758853614330292, "learning_rate": 0.00017071067811865476, "loss": 0.5106, "step": 357 }, { "epoch": 0.9774744027303754, "grad_norm": 0.14779959619045258, "learning_rate": 0.00017048547877285077, "loss": 0.5066, "step": 358 }, { "epoch": 0.9802047781569966, "grad_norm": 0.14550894498825073, "learning_rate": 0.00017025956676251636, "loss": 0.484, "step": 359 }, { "epoch": 0.9829351535836177, "grad_norm": 0.14111794531345367, "learning_rate": 0.00017003294437180255, "loss": 0.4719, "step": 360 }, { "epoch": 0.985665529010239, "grad_norm": 0.15354299545288086, "learning_rate": 0.00016980561389204284, "loss": 0.4874, "step": 361 }, { "epoch": 0.9883959044368601, "grad_norm": 0.14522601664066315, "learning_rate": 0.0001695775776217301, "loss": 0.4722, "step": 362 }, { "epoch": 0.9911262798634812, "grad_norm": 0.143354594707489, "learning_rate": 0.00016934883786649333, "loss": 0.5038, "step": 363 }, { "epoch": 0.9938566552901024, "grad_norm": 0.14298783242702484, "learning_rate": 0.0001691193969390742, "loss": 0.5007, "step": 364 }, { "epoch": 0.9965870307167235, "grad_norm": 0.14418889582157135, "learning_rate": 0.00016888925715930394, "loss": 0.5054, "step": 365 }, { "epoch": 0.9993174061433447, "grad_norm": 0.16697633266448975, "learning_rate": 0.0001686584208540797, "loss": 0.4948, "step": 366 }, { "epoch": 1.0020477815699658, "grad_norm": 0.4087926745414734, "learning_rate": 0.000168426890357341, "loss": 0.8762, "step": 367 }, { "epoch": 1.004778156996587, "grad_norm": 0.17126424610614777, "learning_rate": 0.00016819466801004621, "loss": 0.4964, "step": 368 }, { "epoch": 1.0075085324232083, "grad_norm": 0.18887566030025482, "learning_rate": 0.00016796175616014893, "loss": 0.4842, "step": 369 }, { "epoch": 1.0102389078498293, "grad_norm": 0.16404148936271667, "learning_rate": 0.00016772815716257412, "loss": 0.4761, "step": 370 }, { "epoch": 1.0129692832764505, "grad_norm": 0.2001708298921585, "learning_rate": 0.00016749387337919433, "loss": 0.5005, "step": 371 }, { "epoch": 1.0156996587030718, "grad_norm": 0.16301579773426056, "learning_rate": 0.0001672589071788059, "loss": 0.497, "step": 372 }, { "epoch": 1.0184300341296928, "grad_norm": 0.15777342021465302, "learning_rate": 0.0001670232609371049, "loss": 0.4771, "step": 373 }, { "epoch": 1.021160409556314, "grad_norm": 0.14179396629333496, "learning_rate": 0.00016678693703666325, "loss": 0.4491, "step": 374 }, { "epoch": 1.023890784982935, "grad_norm": 0.16155582666397095, "learning_rate": 0.00016654993786690444, "loss": 0.4846, "step": 375 }, { "epoch": 1.0266211604095563, "grad_norm": 0.15554013848304749, "learning_rate": 0.00016631226582407952, "loss": 0.4663, "step": 376 }, { "epoch": 1.0293515358361776, "grad_norm": 0.15946047008037567, "learning_rate": 0.00016607392331124282, "loss": 0.4978, "step": 377 }, { "epoch": 1.0320819112627986, "grad_norm": 0.14758449792861938, "learning_rate": 0.00016583491273822765, "loss": 0.4868, "step": 378 }, { "epoch": 1.0348122866894198, "grad_norm": 0.16146820783615112, "learning_rate": 0.0001655952365216219, "loss": 0.5056, "step": 379 }, { "epoch": 1.0375426621160408, "grad_norm": 0.14648115634918213, "learning_rate": 0.0001653548970847438, "loss": 0.5028, "step": 380 }, { "epoch": 1.040273037542662, "grad_norm": 0.14611080288887024, "learning_rate": 0.00016511389685761708, "loss": 0.4975, "step": 381 }, { "epoch": 1.0430034129692833, "grad_norm": 0.13934309780597687, "learning_rate": 0.00016487223827694672, "loss": 0.4881, "step": 382 }, { "epoch": 1.0457337883959044, "grad_norm": 0.1680765300989151, "learning_rate": 0.00016462992378609407, "loss": 0.4947, "step": 383 }, { "epoch": 1.0484641638225256, "grad_norm": 0.16694432497024536, "learning_rate": 0.00016438695583505242, "loss": 0.5, "step": 384 }, { "epoch": 1.0511945392491469, "grad_norm": 0.13027995824813843, "learning_rate": 0.00016414333688042186, "loss": 0.4783, "step": 385 }, { "epoch": 1.0539249146757679, "grad_norm": 0.15040864050388336, "learning_rate": 0.0001638990693853848, "loss": 0.4743, "step": 386 }, { "epoch": 1.0566552901023891, "grad_norm": 0.17924848198890686, "learning_rate": 0.00016365415581968084, "loss": 0.493, "step": 387 }, { "epoch": 1.0593856655290101, "grad_norm": 0.14788490533828735, "learning_rate": 0.0001634085986595819, "loss": 0.4932, "step": 388 }, { "epoch": 1.0621160409556314, "grad_norm": 0.1745985448360443, "learning_rate": 0.00016316240038786718, "loss": 0.5192, "step": 389 }, { "epoch": 1.0648464163822526, "grad_norm": 0.1356726437807083, "learning_rate": 0.00016291556349379795, "loss": 0.4829, "step": 390 }, { "epoch": 1.0675767918088737, "grad_norm": 0.1533443182706833, "learning_rate": 0.00016266809047309251, "loss": 0.4881, "step": 391 }, { "epoch": 1.070307167235495, "grad_norm": 0.16772933304309845, "learning_rate": 0.00016241998382790095, "loss": 0.4967, "step": 392 }, { "epoch": 1.073037542662116, "grad_norm": 0.1441749781370163, "learning_rate": 0.0001621712460667797, "loss": 0.4867, "step": 393 }, { "epoch": 1.0757679180887372, "grad_norm": 0.15063367784023285, "learning_rate": 0.00016192187970466644, "loss": 0.4916, "step": 394 }, { "epoch": 1.0784982935153584, "grad_norm": 0.14568567276000977, "learning_rate": 0.00016167188726285434, "loss": 0.488, "step": 395 }, { "epoch": 1.0812286689419794, "grad_norm": 0.14870832860469818, "learning_rate": 0.0001614212712689668, "loss": 0.4867, "step": 396 }, { "epoch": 1.0839590443686007, "grad_norm": 0.15368396043777466, "learning_rate": 0.00016117003425693188, "loss": 0.4766, "step": 397 }, { "epoch": 1.086689419795222, "grad_norm": 0.163841113448143, "learning_rate": 0.00016091817876695655, "loss": 0.4803, "step": 398 }, { "epoch": 1.089419795221843, "grad_norm": 0.17485839128494263, "learning_rate": 0.0001606657073455012, "loss": 0.4875, "step": 399 }, { "epoch": 1.0921501706484642, "grad_norm": 0.14765095710754395, "learning_rate": 0.00016041262254525362, "loss": 0.5058, "step": 400 }, { "epoch": 1.0948805460750852, "grad_norm": 0.1618352234363556, "learning_rate": 0.0001601589269251035, "loss": 0.4903, "step": 401 }, { "epoch": 1.0976109215017065, "grad_norm": 0.1456744223833084, "learning_rate": 0.0001599046230501163, "loss": 0.4822, "step": 402 }, { "epoch": 1.1003412969283277, "grad_norm": 0.15738914906978607, "learning_rate": 0.00015964971349150746, "loss": 0.4944, "step": 403 }, { "epoch": 1.1030716723549487, "grad_norm": 0.1475927084684372, "learning_rate": 0.0001593942008266164, "loss": 0.4971, "step": 404 }, { "epoch": 1.10580204778157, "grad_norm": 0.14080214500427246, "learning_rate": 0.00015913808763888039, "loss": 0.4864, "step": 405 }, { "epoch": 1.108532423208191, "grad_norm": 0.16853410005569458, "learning_rate": 0.00015888137651780845, "loss": 0.4891, "step": 406 }, { "epoch": 1.1112627986348123, "grad_norm": 0.13720810413360596, "learning_rate": 0.00015862407005895522, "loss": 0.4822, "step": 407 }, { "epoch": 1.1139931740614335, "grad_norm": 0.16895629465579987, "learning_rate": 0.00015836617086389468, "loss": 0.4598, "step": 408 }, { "epoch": 1.1167235494880545, "grad_norm": 0.15242214500904083, "learning_rate": 0.00015810768154019385, "loss": 0.4761, "step": 409 }, { "epoch": 1.1194539249146758, "grad_norm": 0.1633676141500473, "learning_rate": 0.00015784860470138633, "loss": 0.4912, "step": 410 }, { "epoch": 1.122184300341297, "grad_norm": 0.16565294563770294, "learning_rate": 0.00015758894296694615, "loss": 0.4836, "step": 411 }, { "epoch": 1.124914675767918, "grad_norm": 0.160204216837883, "learning_rate": 0.00015732869896226094, "loss": 0.4694, "step": 412 }, { "epoch": 1.1276450511945393, "grad_norm": 0.19067788124084473, "learning_rate": 0.00015706787531860557, "loss": 0.4853, "step": 413 }, { "epoch": 1.1303754266211605, "grad_norm": 0.14547327160835266, "learning_rate": 0.00015680647467311557, "loss": 0.4709, "step": 414 }, { "epoch": 1.1331058020477816, "grad_norm": 0.18047383427619934, "learning_rate": 0.0001565444996687605, "loss": 0.4738, "step": 415 }, { "epoch": 1.1358361774744028, "grad_norm": 0.16099834442138672, "learning_rate": 0.00015628195295431697, "loss": 0.4805, "step": 416 }, { "epoch": 1.1385665529010238, "grad_norm": 0.1426517814397812, "learning_rate": 0.00015601883718434207, "loss": 0.4573, "step": 417 }, { "epoch": 1.141296928327645, "grad_norm": 0.14879460632801056, "learning_rate": 0.00015575515501914668, "loss": 0.4778, "step": 418 }, { "epoch": 1.144027303754266, "grad_norm": 0.1393750160932541, "learning_rate": 0.0001554909091247682, "loss": 0.447, "step": 419 }, { "epoch": 1.1467576791808873, "grad_norm": 0.17599152028560638, "learning_rate": 0.00015522610217294375, "loss": 0.4969, "step": 420 }, { "epoch": 1.1494880546075086, "grad_norm": 0.17399606108665466, "learning_rate": 0.0001549607368410834, "loss": 0.4933, "step": 421 }, { "epoch": 1.1522184300341296, "grad_norm": 0.16555064916610718, "learning_rate": 0.00015469481581224272, "loss": 0.475, "step": 422 }, { "epoch": 1.1549488054607508, "grad_norm": 0.20848453044891357, "learning_rate": 0.00015442834177509582, "loss": 0.4872, "step": 423 }, { "epoch": 1.157679180887372, "grad_norm": 0.15129271149635315, "learning_rate": 0.00015416131742390827, "loss": 0.4963, "step": 424 }, { "epoch": 1.1604095563139931, "grad_norm": 0.1703735888004303, "learning_rate": 0.00015389374545850973, "loss": 0.479, "step": 425 }, { "epoch": 1.1631399317406144, "grad_norm": 0.152608722448349, "learning_rate": 0.00015362562858426654, "loss": 0.4831, "step": 426 }, { "epoch": 1.1658703071672356, "grad_norm": 0.14749537408351898, "learning_rate": 0.0001533569695120547, "loss": 0.4784, "step": 427 }, { "epoch": 1.1686006825938566, "grad_norm": 0.15642118453979492, "learning_rate": 0.0001530877709582321, "loss": 0.4679, "step": 428 }, { "epoch": 1.1713310580204779, "grad_norm": 0.13505250215530396, "learning_rate": 0.00015281803564461135, "loss": 0.4779, "step": 429 }, { "epoch": 1.174061433447099, "grad_norm": 0.14518040418624878, "learning_rate": 0.00015254776629843205, "loss": 0.448, "step": 430 }, { "epoch": 1.1767918088737201, "grad_norm": 0.12947289645671844, "learning_rate": 0.0001522769656523333, "loss": 0.4735, "step": 431 }, { "epoch": 1.1795221843003414, "grad_norm": 0.15066657960414886, "learning_rate": 0.00015200563644432612, "loss": 0.4791, "step": 432 }, { "epoch": 1.1822525597269624, "grad_norm": 0.13076473772525787, "learning_rate": 0.00015173378141776568, "loss": 0.4582, "step": 433 }, { "epoch": 1.1849829351535837, "grad_norm": 0.15804897248744965, "learning_rate": 0.00015146140332132358, "loss": 0.482, "step": 434 }, { "epoch": 1.1877133105802047, "grad_norm": 0.1330975741147995, "learning_rate": 0.00015118850490896012, "loss": 0.4736, "step": 435 }, { "epoch": 1.190443686006826, "grad_norm": 0.16358092427253723, "learning_rate": 0.00015091508893989633, "loss": 0.4992, "step": 436 }, { "epoch": 1.1931740614334472, "grad_norm": 0.14177009463310242, "learning_rate": 0.00015064115817858622, "loss": 0.4646, "step": 437 }, { "epoch": 1.1959044368600682, "grad_norm": 0.1569090485572815, "learning_rate": 0.00015036671539468878, "loss": 0.4683, "step": 438 }, { "epoch": 1.1986348122866894, "grad_norm": 0.15897628664970398, "learning_rate": 0.00015009176336303986, "loss": 0.4829, "step": 439 }, { "epoch": 1.2013651877133107, "grad_norm": 0.1482827216386795, "learning_rate": 0.00014981630486362435, "loss": 0.4552, "step": 440 }, { "epoch": 1.2040955631399317, "grad_norm": 0.1546843945980072, "learning_rate": 0.00014954034268154778, "loss": 0.4702, "step": 441 }, { "epoch": 1.206825938566553, "grad_norm": 0.1493782103061676, "learning_rate": 0.00014926387960700842, "loss": 0.4708, "step": 442 }, { "epoch": 1.209556313993174, "grad_norm": 0.15456125140190125, "learning_rate": 0.00014898691843526899, "loss": 0.4738, "step": 443 }, { "epoch": 1.2122866894197952, "grad_norm": 0.17920009791851044, "learning_rate": 0.00014870946196662822, "loss": 0.4616, "step": 444 }, { "epoch": 1.2150170648464165, "grad_norm": 0.15904481709003448, "learning_rate": 0.00014843151300639282, "loss": 0.4791, "step": 445 }, { "epoch": 1.2177474402730375, "grad_norm": 0.18129463493824005, "learning_rate": 0.00014815307436484898, "loss": 0.4789, "step": 446 }, { "epoch": 1.2204778156996587, "grad_norm": 0.14939218759536743, "learning_rate": 0.00014787414885723385, "loss": 0.4774, "step": 447 }, { "epoch": 1.2232081911262798, "grad_norm": 0.17625784873962402, "learning_rate": 0.00014759473930370736, "loss": 0.4785, "step": 448 }, { "epoch": 1.225938566552901, "grad_norm": 0.158066987991333, "learning_rate": 0.0001473148485293234, "loss": 0.4588, "step": 449 }, { "epoch": 1.2286689419795223, "grad_norm": 0.16107094287872314, "learning_rate": 0.00014703447936400134, "loss": 0.4734, "step": 450 }, { "epoch": 1.2313993174061433, "grad_norm": 0.1387171894311905, "learning_rate": 0.00014675363464249763, "loss": 0.4733, "step": 451 }, { "epoch": 1.2341296928327645, "grad_norm": 0.1561625450849533, "learning_rate": 0.00014647231720437686, "loss": 0.4502, "step": 452 }, { "epoch": 1.2368600682593858, "grad_norm": 0.1419583112001419, "learning_rate": 0.00014619052989398322, "loss": 0.4674, "step": 453 }, { "epoch": 1.2395904436860068, "grad_norm": 0.1578853279352188, "learning_rate": 0.00014590827556041158, "loss": 0.4635, "step": 454 }, { "epoch": 1.242320819112628, "grad_norm": 0.1325417309999466, "learning_rate": 0.00014562555705747892, "loss": 0.4557, "step": 455 }, { "epoch": 1.245051194539249, "grad_norm": 0.17075014114379883, "learning_rate": 0.00014534237724369534, "loss": 0.4928, "step": 456 }, { "epoch": 1.2477815699658703, "grad_norm": 0.136972576379776, "learning_rate": 0.00014505873898223496, "loss": 0.4699, "step": 457 }, { "epoch": 1.2505119453924913, "grad_norm": 0.16119800508022308, "learning_rate": 0.00014477464514090743, "loss": 0.4807, "step": 458 }, { "epoch": 1.2532423208191126, "grad_norm": 0.13321495056152344, "learning_rate": 0.00014449009859212857, "loss": 0.4806, "step": 459 }, { "epoch": 1.2559726962457338, "grad_norm": 0.14326980710029602, "learning_rate": 0.00014420510221289137, "loss": 0.4491, "step": 460 }, { "epoch": 1.2587030716723548, "grad_norm": 0.13098248839378357, "learning_rate": 0.00014391965888473703, "loss": 0.4653, "step": 461 }, { "epoch": 1.261433447098976, "grad_norm": 0.14561250805854797, "learning_rate": 0.00014363377149372584, "loss": 0.479, "step": 462 }, { "epoch": 1.2641638225255973, "grad_norm": 0.1377183198928833, "learning_rate": 0.0001433474429304077, "loss": 0.4607, "step": 463 }, { "epoch": 1.2668941979522184, "grad_norm": 0.14758490025997162, "learning_rate": 0.0001430606760897934, "loss": 0.472, "step": 464 }, { "epoch": 1.2696245733788396, "grad_norm": 0.15359081327915192, "learning_rate": 0.0001427734738713248, "loss": 0.469, "step": 465 }, { "epoch": 1.2723549488054609, "grad_norm": 0.16317234933376312, "learning_rate": 0.00014248583917884594, "loss": 0.4749, "step": 466 }, { "epoch": 1.2750853242320819, "grad_norm": 0.1310819685459137, "learning_rate": 0.00014219777492057348, "loss": 0.4722, "step": 467 }, { "epoch": 1.2778156996587031, "grad_norm": 0.14236976206302643, "learning_rate": 0.0001419092840090673, "loss": 0.4704, "step": 468 }, { "epoch": 1.2805460750853244, "grad_norm": 0.1274275928735733, "learning_rate": 0.00014162036936120115, "loss": 0.4687, "step": 469 }, { "epoch": 1.2832764505119454, "grad_norm": 0.13622865080833435, "learning_rate": 0.00014133103389813302, "loss": 0.4649, "step": 470 }, { "epoch": 1.2860068259385666, "grad_norm": 0.14750082790851593, "learning_rate": 0.0001410412805452757, "loss": 0.4654, "step": 471 }, { "epoch": 1.2887372013651877, "grad_norm": 0.14838138222694397, "learning_rate": 0.0001407511122322672, "loss": 0.4626, "step": 472 }, { "epoch": 1.291467576791809, "grad_norm": 0.1439974308013916, "learning_rate": 0.00014046053189294112, "loss": 0.4499, "step": 473 }, { "epoch": 1.29419795221843, "grad_norm": 0.14686186611652374, "learning_rate": 0.00014016954246529696, "loss": 0.4676, "step": 474 }, { "epoch": 1.2969283276450512, "grad_norm": 0.15870115160942078, "learning_rate": 0.00013987814689147041, "loss": 0.4448, "step": 475 }, { "epoch": 1.2996587030716724, "grad_norm": 0.16382190585136414, "learning_rate": 0.0001395863481177036, "loss": 0.4476, "step": 476 }, { "epoch": 1.3023890784982934, "grad_norm": 0.21564428508281708, "learning_rate": 0.00013929414909431544, "loss": 0.4751, "step": 477 }, { "epoch": 1.3051194539249147, "grad_norm": 0.15450774133205414, "learning_rate": 0.00013900155277567157, "loss": 0.4776, "step": 478 }, { "epoch": 1.307849829351536, "grad_norm": 0.15922100841999054, "learning_rate": 0.00013870856212015468, "loss": 0.4693, "step": 479 }, { "epoch": 1.310580204778157, "grad_norm": 0.13661661744117737, "learning_rate": 0.00013841518009013445, "loss": 0.454, "step": 480 }, { "epoch": 1.3133105802047782, "grad_norm": 0.1507856249809265, "learning_rate": 0.00013812140965193773, "loss": 0.4701, "step": 481 }, { "epoch": 1.3160409556313994, "grad_norm": 0.14306284487247467, "learning_rate": 0.00013782725377581848, "loss": 0.4842, "step": 482 }, { "epoch": 1.3187713310580205, "grad_norm": 0.15984225273132324, "learning_rate": 0.00013753271543592773, "loss": 0.4628, "step": 483 }, { "epoch": 1.3215017064846417, "grad_norm": 0.15242429077625275, "learning_rate": 0.00013723779761028347, "loss": 0.4708, "step": 484 }, { "epoch": 1.3242320819112627, "grad_norm": 0.14520719647407532, "learning_rate": 0.0001369425032807407, "loss": 0.4608, "step": 485 }, { "epoch": 1.326962457337884, "grad_norm": 0.16152748465538025, "learning_rate": 0.00013664683543296112, "loss": 0.4729, "step": 486 }, { "epoch": 1.329692832764505, "grad_norm": 0.15082891285419464, "learning_rate": 0.00013635079705638298, "loss": 0.4593, "step": 487 }, { "epoch": 1.3324232081911263, "grad_norm": 0.16038447618484497, "learning_rate": 0.00013605439114419094, "loss": 0.4718, "step": 488 }, { "epoch": 1.3351535836177475, "grad_norm": 0.1532922238111496, "learning_rate": 0.00013575762069328566, "loss": 0.4758, "step": 489 }, { "epoch": 1.3378839590443685, "grad_norm": 0.12969861924648285, "learning_rate": 0.00013546048870425356, "loss": 0.4582, "step": 490 }, { "epoch": 1.3406143344709898, "grad_norm": 0.13830237090587616, "learning_rate": 0.00013516299818133664, "loss": 0.4585, "step": 491 }, { "epoch": 1.343344709897611, "grad_norm": 0.148755744099617, "learning_rate": 0.00013486515213240188, "loss": 0.4803, "step": 492 }, { "epoch": 1.346075085324232, "grad_norm": 0.13623669743537903, "learning_rate": 0.0001345669535689108, "loss": 0.4562, "step": 493 }, { "epoch": 1.3488054607508533, "grad_norm": 0.14996616542339325, "learning_rate": 0.00013426840550588933, "loss": 0.4546, "step": 494 }, { "epoch": 1.3515358361774745, "grad_norm": 0.1468917429447174, "learning_rate": 0.000133969510961897, "loss": 0.4712, "step": 495 }, { "epoch": 1.3542662116040955, "grad_norm": 0.14525148272514343, "learning_rate": 0.0001336702729589965, "loss": 0.4614, "step": 496 }, { "epoch": 1.3569965870307168, "grad_norm": 0.13571806252002716, "learning_rate": 0.00013337069452272333, "loss": 0.4601, "step": 497 }, { "epoch": 1.3597269624573378, "grad_norm": 0.15235814452171326, "learning_rate": 0.00013307077868205487, "loss": 0.4785, "step": 498 }, { "epoch": 1.362457337883959, "grad_norm": 0.14220909774303436, "learning_rate": 0.00013277052846937996, "loss": 0.4561, "step": 499 }, { "epoch": 1.36518771331058, "grad_norm": 0.14699751138687134, "learning_rate": 0.00013246994692046836, "loss": 0.446, "step": 500 }, { "epoch": 1.3679180887372013, "grad_norm": 0.1636335849761963, "learning_rate": 0.00013216903707443967, "loss": 0.4614, "step": 501 }, { "epoch": 1.3706484641638226, "grad_norm": 0.1447010040283203, "learning_rate": 0.00013186780197373306, "loss": 0.4573, "step": 502 }, { "epoch": 1.3733788395904436, "grad_norm": 0.17758530378341675, "learning_rate": 0.0001315662446640761, "loss": 0.4701, "step": 503 }, { "epoch": 1.3761092150170648, "grad_norm": 0.1493985950946808, "learning_rate": 0.00013126436819445422, "loss": 0.4671, "step": 504 }, { "epoch": 1.378839590443686, "grad_norm": 0.16412951052188873, "learning_rate": 0.0001309621756170799, "loss": 0.4705, "step": 505 }, { "epoch": 1.3815699658703071, "grad_norm": 0.14819127321243286, "learning_rate": 0.00013065966998736155, "loss": 0.4579, "step": 506 }, { "epoch": 1.3843003412969284, "grad_norm": 0.1500328630208969, "learning_rate": 0.00013035685436387298, "loss": 0.4484, "step": 507 }, { "epoch": 1.3870307167235496, "grad_norm": 0.14388103783130646, "learning_rate": 0.0001300537318083221, "loss": 0.4325, "step": 508 }, { "epoch": 1.3897610921501706, "grad_norm": 0.17138421535491943, "learning_rate": 0.00012975030538552032, "loss": 0.4717, "step": 509 }, { "epoch": 1.3924914675767919, "grad_norm": 0.15119241178035736, "learning_rate": 0.00012944657816335123, "loss": 0.4279, "step": 510 }, { "epoch": 1.395221843003413, "grad_norm": 0.15165849030017853, "learning_rate": 0.00012914255321273986, "loss": 0.4716, "step": 511 }, { "epoch": 1.3979522184300341, "grad_norm": 0.14222781360149384, "learning_rate": 0.0001288382336076215, "loss": 0.4393, "step": 512 }, { "epoch": 1.4006825938566552, "grad_norm": 0.1435043066740036, "learning_rate": 0.00012853362242491053, "loss": 0.4589, "step": 513 }, { "epoch": 1.4034129692832764, "grad_norm": 0.15017302334308624, "learning_rate": 0.00012822872274446958, "loss": 0.4492, "step": 514 }, { "epoch": 1.4061433447098977, "grad_norm": 0.14196786284446716, "learning_rate": 0.00012792353764907804, "loss": 0.4588, "step": 515 }, { "epoch": 1.4088737201365187, "grad_norm": 0.144223153591156, "learning_rate": 0.0001276180702244012, "loss": 0.4678, "step": 516 }, { "epoch": 1.41160409556314, "grad_norm": 0.12913116812705994, "learning_rate": 0.0001273123235589589, "loss": 0.4515, "step": 517 }, { "epoch": 1.4143344709897612, "grad_norm": 0.1370343565940857, "learning_rate": 0.00012700630074409427, "loss": 0.444, "step": 518 }, { "epoch": 1.4170648464163822, "grad_norm": 0.15820138156414032, "learning_rate": 0.00012670000487394266, "loss": 0.4693, "step": 519 }, { "epoch": 1.4197952218430034, "grad_norm": 0.13780242204666138, "learning_rate": 0.0001263934390454001, "loss": 0.4621, "step": 520 }, { "epoch": 1.4225255972696247, "grad_norm": 0.13959269225597382, "learning_rate": 0.00012608660635809207, "loss": 0.448, "step": 521 }, { "epoch": 1.4252559726962457, "grad_norm": 0.14923076331615448, "learning_rate": 0.00012577950991434248, "loss": 0.4694, "step": 522 }, { "epoch": 1.427986348122867, "grad_norm": 0.14546504616737366, "learning_rate": 0.00012547215281914168, "loss": 0.4679, "step": 523 }, { "epoch": 1.430716723549488, "grad_norm": 0.13985708355903625, "learning_rate": 0.00012516453818011566, "loss": 0.469, "step": 524 }, { "epoch": 1.4334470989761092, "grad_norm": 0.15217062830924988, "learning_rate": 0.00012485666910749428, "loss": 0.4698, "step": 525 }, { "epoch": 1.4361774744027302, "grad_norm": 0.1293148696422577, "learning_rate": 0.00012454854871407994, "loss": 0.4448, "step": 526 }, { "epoch": 1.4389078498293515, "grad_norm": 0.16416317224502563, "learning_rate": 0.0001242401801152161, "loss": 0.4589, "step": 527 }, { "epoch": 1.4416382252559727, "grad_norm": 0.14671452343463898, "learning_rate": 0.0001239315664287558, "loss": 0.4498, "step": 528 }, { "epoch": 1.4443686006825938, "grad_norm": 0.17271259427070618, "learning_rate": 0.00012362271077503008, "loss": 0.4492, "step": 529 }, { "epoch": 1.447098976109215, "grad_norm": 0.15541909635066986, "learning_rate": 0.00012331361627681645, "loss": 0.4281, "step": 530 }, { "epoch": 1.4498293515358363, "grad_norm": 0.14913444221019745, "learning_rate": 0.00012300428605930736, "loss": 0.4334, "step": 531 }, { "epoch": 1.4525597269624573, "grad_norm": 0.167875275015831, "learning_rate": 0.00012269472325007858, "loss": 0.4736, "step": 532 }, { "epoch": 1.4552901023890785, "grad_norm": 0.15073426067829132, "learning_rate": 0.00012238493097905756, "loss": 0.4744, "step": 533 }, { "epoch": 1.4580204778156998, "grad_norm": 0.16593150794506073, "learning_rate": 0.00012207491237849172, "loss": 0.455, "step": 534 }, { "epoch": 1.4607508532423208, "grad_norm": 0.13389617204666138, "learning_rate": 0.00012176467058291699, "loss": 0.4394, "step": 535 }, { "epoch": 1.463481228668942, "grad_norm": 0.16609057784080505, "learning_rate": 0.00012145420872912585, "loss": 0.4562, "step": 536 }, { "epoch": 1.466211604095563, "grad_norm": 0.1394118070602417, "learning_rate": 0.00012114352995613582, "loss": 0.4451, "step": 537 }, { "epoch": 1.4689419795221843, "grad_norm": 0.17529746890068054, "learning_rate": 0.00012083263740515765, "loss": 0.4522, "step": 538 }, { "epoch": 1.4716723549488053, "grad_norm": 0.15724115073680878, "learning_rate": 0.00012052153421956342, "loss": 0.4556, "step": 539 }, { "epoch": 1.4744027303754266, "grad_norm": 0.17186792194843292, "learning_rate": 0.00012021022354485514, "loss": 0.4546, "step": 540 }, { "epoch": 1.4771331058020478, "grad_norm": 0.1301499456167221, "learning_rate": 0.00011989870852863254, "loss": 0.4431, "step": 541 }, { "epoch": 1.4798634812286688, "grad_norm": 0.1480223387479782, "learning_rate": 0.00011958699232056134, "loss": 0.4491, "step": 542 }, { "epoch": 1.48259385665529, "grad_norm": 0.13150086998939514, "learning_rate": 0.00011927507807234168, "loss": 0.4568, "step": 543 }, { "epoch": 1.4853242320819113, "grad_norm": 0.15769197046756744, "learning_rate": 0.00011896296893767587, "loss": 0.4501, "step": 544 }, { "epoch": 1.4880546075085324, "grad_norm": 0.13996848464012146, "learning_rate": 0.0001186506680722367, "loss": 0.4673, "step": 545 }, { "epoch": 1.4907849829351536, "grad_norm": 0.16406555473804474, "learning_rate": 0.00011833817863363564, "loss": 0.4634, "step": 546 }, { "epoch": 1.4935153583617748, "grad_norm": 0.14007951319217682, "learning_rate": 0.0001180255037813906, "loss": 0.4466, "step": 547 }, { "epoch": 1.4962457337883959, "grad_norm": 0.15525664389133453, "learning_rate": 0.00011771264667689427, "loss": 0.4323, "step": 548 }, { "epoch": 1.4989761092150171, "grad_norm": 0.143234983086586, "learning_rate": 0.00011739961048338213, "loss": 0.4395, "step": 549 }, { "epoch": 1.5017064846416384, "grad_norm": 0.15597446262836456, "learning_rate": 0.00011708639836590023, "loss": 0.4546, "step": 550 }, { "epoch": 1.5044368600682594, "grad_norm": 0.13816912472248077, "learning_rate": 0.00011677301349127348, "loss": 0.4586, "step": 551 }, { "epoch": 1.5071672354948804, "grad_norm": 0.16299140453338623, "learning_rate": 0.00011645945902807341, "loss": 0.4465, "step": 552 }, { "epoch": 1.5098976109215017, "grad_norm": 0.15032370388507843, "learning_rate": 0.00011614573814658629, "loss": 0.4579, "step": 553 }, { "epoch": 1.512627986348123, "grad_norm": 0.158245250582695, "learning_rate": 0.00011583185401878101, "loss": 0.4462, "step": 554 }, { "epoch": 1.515358361774744, "grad_norm": 0.16943717002868652, "learning_rate": 0.00011551780981827698, "loss": 0.4572, "step": 555 }, { "epoch": 1.5180887372013652, "grad_norm": 0.14559145271778107, "learning_rate": 0.00011520360872031209, "loss": 0.4693, "step": 556 }, { "epoch": 1.5208191126279864, "grad_norm": 0.16138285398483276, "learning_rate": 0.00011488925390171059, "loss": 0.4623, "step": 557 }, { "epoch": 1.5235494880546074, "grad_norm": 0.14859908819198608, "learning_rate": 0.00011457474854085096, "loss": 0.4684, "step": 558 }, { "epoch": 1.5262798634812287, "grad_norm": 0.15695518255233765, "learning_rate": 0.00011426009581763377, "loss": 0.4619, "step": 559 }, { "epoch": 1.52901023890785, "grad_norm": 0.14711041748523712, "learning_rate": 0.00011394529891344958, "loss": 0.4556, "step": 560 }, { "epoch": 1.531740614334471, "grad_norm": 0.15271785855293274, "learning_rate": 0.0001136303610111467, "loss": 0.4557, "step": 561 }, { "epoch": 1.5344709897610922, "grad_norm": 0.1541603058576584, "learning_rate": 0.00011331528529499909, "loss": 0.4644, "step": 562 }, { "epoch": 1.5372013651877134, "grad_norm": 0.1686154007911682, "learning_rate": 0.00011300007495067401, "loss": 0.4446, "step": 563 }, { "epoch": 1.5399317406143345, "grad_norm": 0.14042454957962036, "learning_rate": 0.00011268473316520007, "loss": 0.4275, "step": 564 }, { "epoch": 1.5426621160409555, "grad_norm": 0.14575007557868958, "learning_rate": 0.00011236926312693479, "loss": 0.4579, "step": 565 }, { "epoch": 1.545392491467577, "grad_norm": 0.16124123334884644, "learning_rate": 0.0001120536680255323, "loss": 0.441, "step": 566 }, { "epoch": 1.548122866894198, "grad_norm": 0.14471302926540375, "learning_rate": 0.00011173795105191145, "loss": 0.4543, "step": 567 }, { "epoch": 1.550853242320819, "grad_norm": 0.162650004029274, "learning_rate": 0.00011142211539822318, "loss": 0.4483, "step": 568 }, { "epoch": 1.5535836177474402, "grad_norm": 0.14518044888973236, "learning_rate": 0.00011110616425781833, "loss": 0.4386, "step": 569 }, { "epoch": 1.5563139931740615, "grad_norm": 0.1545732617378235, "learning_rate": 0.00011079010082521557, "loss": 0.4505, "step": 570 }, { "epoch": 1.5590443686006825, "grad_norm": 0.12941716611385345, "learning_rate": 0.00011047392829606876, "loss": 0.4288, "step": 571 }, { "epoch": 1.5617747440273038, "grad_norm": 0.15107029676437378, "learning_rate": 0.0001101576498671349, "loss": 0.4275, "step": 572 }, { "epoch": 1.564505119453925, "grad_norm": 0.13776972889900208, "learning_rate": 0.00010984126873624179, "loss": 0.432, "step": 573 }, { "epoch": 1.567235494880546, "grad_norm": 0.15370745956897736, "learning_rate": 0.00010952478810225548, "loss": 0.4523, "step": 574 }, { "epoch": 1.5699658703071673, "grad_norm": 0.14505314826965332, "learning_rate": 0.00010920821116504816, "loss": 0.4444, "step": 575 }, { "epoch": 1.5726962457337885, "grad_norm": 0.13053090870380402, "learning_rate": 0.0001088915411254657, "loss": 0.4492, "step": 576 }, { "epoch": 1.5754266211604095, "grad_norm": 0.15613074600696564, "learning_rate": 0.00010857478118529533, "loss": 0.465, "step": 577 }, { "epoch": 1.5781569965870306, "grad_norm": 0.164808988571167, "learning_rate": 0.00010825793454723325, "loss": 0.4498, "step": 578 }, { "epoch": 1.580887372013652, "grad_norm": 0.14633600413799286, "learning_rate": 0.0001079410044148522, "loss": 0.4416, "step": 579 }, { "epoch": 1.583617747440273, "grad_norm": 0.15268942713737488, "learning_rate": 0.00010762399399256917, "loss": 0.4431, "step": 580 }, { "epoch": 1.586348122866894, "grad_norm": 0.16290055215358734, "learning_rate": 0.00010730690648561292, "loss": 0.465, "step": 581 }, { "epoch": 1.5890784982935153, "grad_norm": 0.13567085564136505, "learning_rate": 0.00010698974509999158, "loss": 0.4397, "step": 582 }, { "epoch": 1.5918088737201366, "grad_norm": 0.14896200597286224, "learning_rate": 0.00010667251304246029, "loss": 0.4458, "step": 583 }, { "epoch": 1.5945392491467576, "grad_norm": 0.14076146483421326, "learning_rate": 0.00010635521352048872, "loss": 0.4476, "step": 584 }, { "epoch": 1.5972696245733788, "grad_norm": 0.13334687054157257, "learning_rate": 0.00010603784974222861, "loss": 0.4283, "step": 585 }, { "epoch": 1.6, "grad_norm": 0.16997142136096954, "learning_rate": 0.00010572042491648149, "loss": 0.4472, "step": 586 }, { "epoch": 1.6027303754266211, "grad_norm": 0.15508471429347992, "learning_rate": 0.00010540294225266607, "loss": 0.4518, "step": 587 }, { "epoch": 1.6054607508532424, "grad_norm": 0.1571933478116989, "learning_rate": 0.0001050854049607858, "loss": 0.4493, "step": 588 }, { "epoch": 1.6081911262798636, "grad_norm": 0.15027360618114471, "learning_rate": 0.00010476781625139656, "loss": 0.4562, "step": 589 }, { "epoch": 1.6109215017064846, "grad_norm": 0.14502452313899994, "learning_rate": 0.00010445017933557404, "loss": 0.4476, "step": 590 }, { "epoch": 1.6136518771331056, "grad_norm": 0.1403171718120575, "learning_rate": 0.00010413249742488131, "loss": 0.4154, "step": 591 }, { "epoch": 1.6163822525597271, "grad_norm": 0.14815428853034973, "learning_rate": 0.00010381477373133652, "loss": 0.4501, "step": 592 }, { "epoch": 1.6191126279863481, "grad_norm": 0.16107513010501862, "learning_rate": 0.00010349701146738007, "loss": 0.4507, "step": 593 }, { "epoch": 1.6218430034129692, "grad_norm": 0.14111128449440002, "learning_rate": 0.00010317921384584244, "loss": 0.4483, "step": 594 }, { "epoch": 1.6245733788395904, "grad_norm": 0.1680098921060562, "learning_rate": 0.0001028613840799117, "loss": 0.4584, "step": 595 }, { "epoch": 1.6273037542662117, "grad_norm": 0.13612088561058044, "learning_rate": 0.00010254352538310075, "loss": 0.4389, "step": 596 }, { "epoch": 1.6300341296928327, "grad_norm": 0.16298632323741913, "learning_rate": 0.00010222564096921505, "loss": 0.452, "step": 597 }, { "epoch": 1.632764505119454, "grad_norm": 0.14744146168231964, "learning_rate": 0.00010190773405232024, "loss": 0.4588, "step": 598 }, { "epoch": 1.6354948805460752, "grad_norm": 0.1594633013010025, "learning_rate": 0.00010158980784670927, "loss": 0.4403, "step": 599 }, { "epoch": 1.6382252559726962, "grad_norm": 0.15123943984508514, "learning_rate": 0.00010127186556687019, "loss": 0.4395, "step": 600 }, { "epoch": 1.6409556313993174, "grad_norm": 0.1518649309873581, "learning_rate": 0.00010095391042745361, "loss": 0.4265, "step": 601 }, { "epoch": 1.6436860068259387, "grad_norm": 0.1611323356628418, "learning_rate": 0.00010063594564324012, "loss": 0.444, "step": 602 }, { "epoch": 1.6464163822525597, "grad_norm": 0.14762264490127563, "learning_rate": 0.00010031797442910789, "loss": 0.4409, "step": 603 }, { "epoch": 1.6491467576791807, "grad_norm": 0.1546734869480133, "learning_rate": 0.0001, "loss": 0.4472, "step": 604 }, { "epoch": 1.6518771331058022, "grad_norm": 0.1580485701560974, "learning_rate": 9.968202557089212e-05, "loss": 0.4497, "step": 605 }, { "epoch": 1.6546075085324232, "grad_norm": 0.151153564453125, "learning_rate": 9.93640543567599e-05, "loss": 0.4495, "step": 606 }, { "epoch": 1.6573378839590442, "grad_norm": 0.14238281548023224, "learning_rate": 9.904608957254642e-05, "loss": 0.4481, "step": 607 }, { "epoch": 1.6600682593856655, "grad_norm": 0.13984693586826324, "learning_rate": 9.872813443312984e-05, "loss": 0.4262, "step": 608 }, { "epoch": 1.6627986348122867, "grad_norm": 0.14657770097255707, "learning_rate": 9.84101921532908e-05, "loss": 0.4554, "step": 609 }, { "epoch": 1.6655290102389078, "grad_norm": 0.15702606737613678, "learning_rate": 9.809226594767978e-05, "loss": 0.4553, "step": 610 }, { "epoch": 1.668259385665529, "grad_norm": 0.15596133470535278, "learning_rate": 9.777435903078494e-05, "loss": 0.4353, "step": 611 }, { "epoch": 1.6709897610921502, "grad_norm": 0.16193975508213043, "learning_rate": 9.745647461689931e-05, "loss": 0.4313, "step": 612 }, { "epoch": 1.6737201365187713, "grad_norm": 0.15690681338310242, "learning_rate": 9.713861592008833e-05, "loss": 0.4498, "step": 613 }, { "epoch": 1.6764505119453925, "grad_norm": 0.15277935564517975, "learning_rate": 9.682078615415754e-05, "loss": 0.4358, "step": 614 }, { "epoch": 1.6791808873720138, "grad_norm": 0.1447397917509079, "learning_rate": 9.650298853261997e-05, "loss": 0.4264, "step": 615 }, { "epoch": 1.6819112627986348, "grad_norm": 0.14292937517166138, "learning_rate": 9.61852262686635e-05, "loss": 0.4521, "step": 616 }, { "epoch": 1.6846416382252558, "grad_norm": 0.15041732788085938, "learning_rate": 9.586750257511867e-05, "loss": 0.445, "step": 617 }, { "epoch": 1.6873720136518773, "grad_norm": 0.14610610902309418, "learning_rate": 9.5549820664426e-05, "loss": 0.4493, "step": 618 }, { "epoch": 1.6901023890784983, "grad_norm": 0.1494503617286682, "learning_rate": 9.523218374860348e-05, "loss": 0.4473, "step": 619 }, { "epoch": 1.6928327645051193, "grad_norm": 0.14970283210277557, "learning_rate": 9.491459503921421e-05, "loss": 0.453, "step": 620 }, { "epoch": 1.6955631399317406, "grad_norm": 0.15402431786060333, "learning_rate": 9.459705774733396e-05, "loss": 0.4435, "step": 621 }, { "epoch": 1.6982935153583618, "grad_norm": 0.1583550125360489, "learning_rate": 9.427957508351852e-05, "loss": 0.4379, "step": 622 }, { "epoch": 1.7010238907849828, "grad_norm": 0.15820656716823578, "learning_rate": 9.396215025777139e-05, "loss": 0.4416, "step": 623 }, { "epoch": 1.703754266211604, "grad_norm": 0.14503423869609833, "learning_rate": 9.364478647951133e-05, "loss": 0.434, "step": 624 }, { "epoch": 1.7064846416382253, "grad_norm": 0.1509648561477661, "learning_rate": 9.332748695753973e-05, "loss": 0.4516, "step": 625 }, { "epoch": 1.7092150170648464, "grad_norm": 0.15958918631076813, "learning_rate": 9.301025490000841e-05, "loss": 0.4504, "step": 626 }, { "epoch": 1.7119453924914676, "grad_norm": 0.1522430032491684, "learning_rate": 9.269309351438711e-05, "loss": 0.4339, "step": 627 }, { "epoch": 1.7146757679180888, "grad_norm": 0.14348183572292328, "learning_rate": 9.237600600743085e-05, "loss": 0.4236, "step": 628 }, { "epoch": 1.7174061433447099, "grad_norm": 0.1558932512998581, "learning_rate": 9.20589955851478e-05, "loss": 0.4514, "step": 629 }, { "epoch": 1.7201365187713311, "grad_norm": 0.13524165749549866, "learning_rate": 9.174206545276677e-05, "loss": 0.4426, "step": 630 }, { "epoch": 1.7228668941979524, "grad_norm": 0.1464926302433014, "learning_rate": 9.142521881470469e-05, "loss": 0.4397, "step": 631 }, { "epoch": 1.7255972696245734, "grad_norm": 0.14485323429107666, "learning_rate": 9.11084588745343e-05, "loss": 0.4497, "step": 632 }, { "epoch": 1.7283276450511944, "grad_norm": 0.139760822057724, "learning_rate": 9.07917888349519e-05, "loss": 0.4465, "step": 633 }, { "epoch": 1.7310580204778157, "grad_norm": 0.16193385422229767, "learning_rate": 9.047521189774455e-05, "loss": 0.4377, "step": 634 }, { "epoch": 1.733788395904437, "grad_norm": 0.17076647281646729, "learning_rate": 9.015873126375822e-05, "loss": 0.4632, "step": 635 }, { "epoch": 1.736518771331058, "grad_norm": 0.14881980419158936, "learning_rate": 8.984235013286511e-05, "loss": 0.4401, "step": 636 }, { "epoch": 1.7392491467576792, "grad_norm": 0.15158087015151978, "learning_rate": 8.952607170393125e-05, "loss": 0.4292, "step": 637 }, { "epoch": 1.7419795221843004, "grad_norm": 0.14353424310684204, "learning_rate": 8.920989917478447e-05, "loss": 0.4439, "step": 638 }, { "epoch": 1.7447098976109214, "grad_norm": 0.13718026876449585, "learning_rate": 8.88938357421817e-05, "loss": 0.4267, "step": 639 }, { "epoch": 1.7474402730375427, "grad_norm": 0.15788930654525757, "learning_rate": 8.857788460177686e-05, "loss": 0.4241, "step": 640 }, { "epoch": 1.750170648464164, "grad_norm": 0.14324265718460083, "learning_rate": 8.826204894808855e-05, "loss": 0.4472, "step": 641 }, { "epoch": 1.752901023890785, "grad_norm": 0.15020030736923218, "learning_rate": 8.79463319744677e-05, "loss": 0.4339, "step": 642 }, { "epoch": 1.7556313993174062, "grad_norm": 0.13715054094791412, "learning_rate": 8.763073687306524e-05, "loss": 0.4429, "step": 643 }, { "epoch": 1.7583617747440274, "grad_norm": 0.16862636804580688, "learning_rate": 8.731526683479992e-05, "loss": 0.4509, "step": 644 }, { "epoch": 1.7610921501706485, "grad_norm": 0.14781633019447327, "learning_rate": 8.6999925049326e-05, "loss": 0.4319, "step": 645 }, { "epoch": 1.7638225255972695, "grad_norm": 0.15873977541923523, "learning_rate": 8.668471470500095e-05, "loss": 0.4486, "step": 646 }, { "epoch": 1.7665529010238907, "grad_norm": 0.1477411389350891, "learning_rate": 8.63696389888533e-05, "loss": 0.4461, "step": 647 }, { "epoch": 1.769283276450512, "grad_norm": 0.1804722100496292, "learning_rate": 8.605470108655045e-05, "loss": 0.4402, "step": 648 }, { "epoch": 1.772013651877133, "grad_norm": 0.15354932844638824, "learning_rate": 8.573990418236625e-05, "loss": 0.4564, "step": 649 }, { "epoch": 1.7747440273037542, "grad_norm": 0.1494126319885254, "learning_rate": 8.542525145914905e-05, "loss": 0.4377, "step": 650 }, { "epoch": 1.7774744027303755, "grad_norm": 0.14122453331947327, "learning_rate": 8.511074609828944e-05, "loss": 0.4361, "step": 651 }, { "epoch": 1.7802047781569965, "grad_norm": 0.16938751935958862, "learning_rate": 8.479639127968792e-05, "loss": 0.4593, "step": 652 }, { "epoch": 1.7829351535836178, "grad_norm": 0.1502314805984497, "learning_rate": 8.448219018172303e-05, "loss": 0.4336, "step": 653 }, { "epoch": 1.785665529010239, "grad_norm": 0.180609330534935, "learning_rate": 8.4168145981219e-05, "loss": 0.45, "step": 654 }, { "epoch": 1.78839590443686, "grad_norm": 0.1395808309316635, "learning_rate": 8.385426185341374e-05, "loss": 0.4188, "step": 655 }, { "epoch": 1.7911262798634813, "grad_norm": 0.1596853882074356, "learning_rate": 8.35405409719266e-05, "loss": 0.452, "step": 656 }, { "epoch": 1.7938566552901025, "grad_norm": 0.15330305695533752, "learning_rate": 8.322698650872656e-05, "loss": 0.4355, "step": 657 }, { "epoch": 1.7965870307167235, "grad_norm": 0.1545482873916626, "learning_rate": 8.291360163409978e-05, "loss": 0.4366, "step": 658 }, { "epoch": 1.7993174061433446, "grad_norm": 0.13950030505657196, "learning_rate": 8.260038951661787e-05, "loss": 0.4169, "step": 659 }, { "epoch": 1.802047781569966, "grad_norm": 0.14191307127475739, "learning_rate": 8.228735332310575e-05, "loss": 0.4471, "step": 660 }, { "epoch": 1.804778156996587, "grad_norm": 0.14557993412017822, "learning_rate": 8.197449621860943e-05, "loss": 0.4028, "step": 661 }, { "epoch": 1.807508532423208, "grad_norm": 0.13985979557037354, "learning_rate": 8.16618213663644e-05, "loss": 0.4293, "step": 662 }, { "epoch": 1.8102389078498293, "grad_norm": 0.1420183628797531, "learning_rate": 8.134933192776333e-05, "loss": 0.4313, "step": 663 }, { "epoch": 1.8129692832764506, "grad_norm": 0.1498919129371643, "learning_rate": 8.103703106232416e-05, "loss": 0.4315, "step": 664 }, { "epoch": 1.8156996587030716, "grad_norm": 0.1570868194103241, "learning_rate": 8.072492192765833e-05, "loss": 0.4348, "step": 665 }, { "epoch": 1.8184300341296928, "grad_norm": 0.15423277020454407, "learning_rate": 8.041300767943867e-05, "loss": 0.44, "step": 666 }, { "epoch": 1.821160409556314, "grad_norm": 0.13805197179317474, "learning_rate": 8.010129147136749e-05, "loss": 0.4317, "step": 667 }, { "epoch": 1.823890784982935, "grad_norm": 0.14628642797470093, "learning_rate": 7.978977645514487e-05, "loss": 0.4379, "step": 668 }, { "epoch": 1.8266211604095564, "grad_norm": 0.15710268914699554, "learning_rate": 7.947846578043659e-05, "loss": 0.4557, "step": 669 }, { "epoch": 1.8293515358361776, "grad_norm": 0.15317128598690033, "learning_rate": 7.916736259484239e-05, "loss": 0.4506, "step": 670 }, { "epoch": 1.8320819112627986, "grad_norm": 0.15919502079486847, "learning_rate": 7.88564700438642e-05, "loss": 0.435, "step": 671 }, { "epoch": 1.8348122866894196, "grad_norm": 0.1551041305065155, "learning_rate": 7.854579127087417e-05, "loss": 0.4283, "step": 672 }, { "epoch": 1.8375426621160411, "grad_norm": 0.16782739758491516, "learning_rate": 7.823532941708303e-05, "loss": 0.4507, "step": 673 }, { "epoch": 1.8402730375426621, "grad_norm": 0.15251149237155914, "learning_rate": 7.792508762150833e-05, "loss": 0.4255, "step": 674 }, { "epoch": 1.8430034129692832, "grad_norm": 0.17279517650604248, "learning_rate": 7.761506902094248e-05, "loss": 0.4255, "step": 675 }, { "epoch": 1.8457337883959044, "grad_norm": 0.13385094702243805, "learning_rate": 7.730527674992143e-05, "loss": 0.4229, "step": 676 }, { "epoch": 1.8484641638225257, "grad_norm": 0.15888231992721558, "learning_rate": 7.699571394069269e-05, "loss": 0.4335, "step": 677 }, { "epoch": 1.8511945392491467, "grad_norm": 0.1549587994813919, "learning_rate": 7.668638372318359e-05, "loss": 0.4083, "step": 678 }, { "epoch": 1.853924914675768, "grad_norm": 0.16143332421779633, "learning_rate": 7.637728922496996e-05, "loss": 0.4367, "step": 679 }, { "epoch": 1.8566552901023892, "grad_norm": 0.13735996186733246, "learning_rate": 7.606843357124426e-05, "loss": 0.4296, "step": 680 }, { "epoch": 1.8593856655290102, "grad_norm": 0.14317500591278076, "learning_rate": 7.575981988478392e-05, "loss": 0.4419, "step": 681 }, { "epoch": 1.8621160409556314, "grad_norm": 0.14451129734516144, "learning_rate": 7.54514512859201e-05, "loss": 0.4389, "step": 682 }, { "epoch": 1.8648464163822527, "grad_norm": 0.14233650267124176, "learning_rate": 7.514333089250577e-05, "loss": 0.4258, "step": 683 }, { "epoch": 1.8675767918088737, "grad_norm": 0.14275044202804565, "learning_rate": 7.483546181988436e-05, "loss": 0.4456, "step": 684 }, { "epoch": 1.8703071672354947, "grad_norm": 0.15278606116771698, "learning_rate": 7.452784718085833e-05, "loss": 0.4344, "step": 685 }, { "epoch": 1.8730375426621162, "grad_norm": 0.13797658681869507, "learning_rate": 7.422049008565757e-05, "loss": 0.4392, "step": 686 }, { "epoch": 1.8757679180887372, "grad_norm": 0.1518598347902298, "learning_rate": 7.391339364190794e-05, "loss": 0.431, "step": 687 }, { "epoch": 1.8784982935153582, "grad_norm": 0.1477964073419571, "learning_rate": 7.360656095459995e-05, "loss": 0.4449, "step": 688 }, { "epoch": 1.8812286689419795, "grad_norm": 0.14295299351215363, "learning_rate": 7.329999512605738e-05, "loss": 0.4384, "step": 689 }, { "epoch": 1.8839590443686007, "grad_norm": 0.1487056165933609, "learning_rate": 7.299369925590574e-05, "loss": 0.4311, "step": 690 }, { "epoch": 1.8866894197952218, "grad_norm": 0.14108945429325104, "learning_rate": 7.268767644104112e-05, "loss": 0.4239, "step": 691 }, { "epoch": 1.889419795221843, "grad_norm": 0.15047363936901093, "learning_rate": 7.238192977559884e-05, "loss": 0.4427, "step": 692 }, { "epoch": 1.8921501706484642, "grad_norm": 0.14634403586387634, "learning_rate": 7.2076462350922e-05, "loss": 0.4416, "step": 693 }, { "epoch": 1.8948805460750853, "grad_norm": 0.15272392332553864, "learning_rate": 7.177127725553045e-05, "loss": 0.43, "step": 694 }, { "epoch": 1.8976109215017065, "grad_norm": 0.15047992765903473, "learning_rate": 7.146637757508949e-05, "loss": 0.4436, "step": 695 }, { "epoch": 1.9003412969283278, "grad_norm": 0.14413118362426758, "learning_rate": 7.116176639237852e-05, "loss": 0.4169, "step": 696 }, { "epoch": 1.9030716723549488, "grad_norm": 0.1437167525291443, "learning_rate": 7.085744678726013e-05, "loss": 0.4389, "step": 697 }, { "epoch": 1.9058020477815698, "grad_norm": 0.14662359654903412, "learning_rate": 7.05534218366488e-05, "loss": 0.435, "step": 698 }, { "epoch": 1.9085324232081913, "grad_norm": 0.14521794021129608, "learning_rate": 7.024969461447972e-05, "loss": 0.4505, "step": 699 }, { "epoch": 1.9112627986348123, "grad_norm": 0.14155706763267517, "learning_rate": 6.994626819167789e-05, "loss": 0.4301, "step": 700 }, { "epoch": 1.9139931740614333, "grad_norm": 0.14663158357143402, "learning_rate": 6.964314563612708e-05, "loss": 0.4262, "step": 701 }, { "epoch": 1.9167235494880546, "grad_norm": 0.15034128725528717, "learning_rate": 6.934033001263847e-05, "loss": 0.4424, "step": 702 }, { "epoch": 1.9194539249146758, "grad_norm": 0.15093255043029785, "learning_rate": 6.903782438292015e-05, "loss": 0.4425, "step": 703 }, { "epoch": 1.9221843003412968, "grad_norm": 0.1556250900030136, "learning_rate": 6.873563180554583e-05, "loss": 0.4225, "step": 704 }, { "epoch": 1.924914675767918, "grad_norm": 0.16173475980758667, "learning_rate": 6.843375533592395e-05, "loss": 0.4228, "step": 705 }, { "epoch": 1.9276450511945393, "grad_norm": 0.1532420516014099, "learning_rate": 6.813219802626698e-05, "loss": 0.432, "step": 706 }, { "epoch": 1.9303754266211604, "grad_norm": 0.1596469283103943, "learning_rate": 6.783096292556035e-05, "loss": 0.4232, "step": 707 }, { "epoch": 1.9331058020477816, "grad_norm": 0.16538076102733612, "learning_rate": 6.753005307953167e-05, "loss": 0.45, "step": 708 }, { "epoch": 1.9358361774744028, "grad_norm": 0.15187640488147736, "learning_rate": 6.722947153062003e-05, "loss": 0.4442, "step": 709 }, { "epoch": 1.9385665529010239, "grad_norm": 0.14827731251716614, "learning_rate": 6.692922131794517e-05, "loss": 0.414, "step": 710 }, { "epoch": 1.9412969283276449, "grad_norm": 0.16438645124435425, "learning_rate": 6.662930547727668e-05, "loss": 0.419, "step": 711 }, { "epoch": 1.9440273037542664, "grad_norm": 0.15135832130908966, "learning_rate": 6.632972704100349e-05, "loss": 0.4155, "step": 712 }, { "epoch": 1.9467576791808874, "grad_norm": 0.15094083547592163, "learning_rate": 6.603048903810305e-05, "loss": 0.4258, "step": 713 }, { "epoch": 1.9494880546075084, "grad_norm": 0.14968033134937286, "learning_rate": 6.57315944941107e-05, "loss": 0.4395, "step": 714 }, { "epoch": 1.9522184300341296, "grad_norm": 0.16519851982593536, "learning_rate": 6.54330464310892e-05, "loss": 0.4406, "step": 715 }, { "epoch": 1.954948805460751, "grad_norm": 0.13855180144309998, "learning_rate": 6.513484786759818e-05, "loss": 0.43, "step": 716 }, { "epoch": 1.957679180887372, "grad_norm": 0.1570328176021576, "learning_rate": 6.483700181866337e-05, "loss": 0.4288, "step": 717 }, { "epoch": 1.9604095563139932, "grad_norm": 0.14928270876407623, "learning_rate": 6.453951129574644e-05, "loss": 0.4224, "step": 718 }, { "epoch": 1.9631399317406144, "grad_norm": 0.16348999738693237, "learning_rate": 6.42423793067144e-05, "loss": 0.4379, "step": 719 }, { "epoch": 1.9658703071672354, "grad_norm": 0.14947615563869476, "learning_rate": 6.39456088558091e-05, "loss": 0.445, "step": 720 }, { "epoch": 1.9686006825938567, "grad_norm": 0.14767783880233765, "learning_rate": 6.3649202943617e-05, "loss": 0.4388, "step": 721 }, { "epoch": 1.971331058020478, "grad_norm": 0.14223739504814148, "learning_rate": 6.33531645670389e-05, "loss": 0.433, "step": 722 }, { "epoch": 1.974061433447099, "grad_norm": 0.1595824956893921, "learning_rate": 6.305749671925931e-05, "loss": 0.4325, "step": 723 }, { "epoch": 1.9767918088737202, "grad_norm": 0.14971914887428284, "learning_rate": 6.276220238971652e-05, "loss": 0.4336, "step": 724 }, { "epoch": 1.9795221843003414, "grad_norm": 0.15426860749721527, "learning_rate": 6.24672845640723e-05, "loss": 0.433, "step": 725 }, { "epoch": 1.9822525597269625, "grad_norm": 0.1470557302236557, "learning_rate": 6.217274622418153e-05, "loss": 0.435, "step": 726 }, { "epoch": 1.9849829351535835, "grad_norm": 0.1564924120903015, "learning_rate": 6.187859034806224e-05, "loss": 0.4371, "step": 727 }, { "epoch": 1.9877133105802047, "grad_norm": 0.16420651972293854, "learning_rate": 6.158481990986557e-05, "loss": 0.4478, "step": 728 }, { "epoch": 1.990443686006826, "grad_norm": 0.15158484876155853, "learning_rate": 6.129143787984533e-05, "loss": 0.4267, "step": 729 }, { "epoch": 1.993174061433447, "grad_norm": 0.15700684487819672, "learning_rate": 6.099844722432843e-05, "loss": 0.4293, "step": 730 }, { "epoch": 1.9959044368600682, "grad_norm": 0.14848262071609497, "learning_rate": 6.070585090568459e-05, "loss": 0.4339, "step": 731 }, { "epoch": 1.9986348122866895, "grad_norm": 0.15596534311771393, "learning_rate": 6.0413651882296406e-05, "loss": 0.4245, "step": 732 }, { "epoch": 2.0013651877133105, "grad_norm": 0.3338797092437744, "learning_rate": 6.012185310852962e-05, "loss": 0.7156, "step": 733 }, { "epoch": 2.0040955631399315, "grad_norm": 0.1896699219942093, "learning_rate": 5.983045753470308e-05, "loss": 0.4228, "step": 734 }, { "epoch": 2.006825938566553, "grad_norm": 0.20619529485702515, "learning_rate": 5.953946810705888e-05, "loss": 0.4244, "step": 735 }, { "epoch": 2.009556313993174, "grad_norm": 0.17412033677101135, "learning_rate": 5.924888776773281e-05, "loss": 0.4186, "step": 736 }, { "epoch": 2.012286689419795, "grad_norm": 0.1896408498287201, "learning_rate": 5.8958719454724346e-05, "loss": 0.4259, "step": 737 }, { "epoch": 2.0150170648464165, "grad_norm": 0.16549214720726013, "learning_rate": 5.8668966101867005e-05, "loss": 0.3967, "step": 738 }, { "epoch": 2.0177474402730375, "grad_norm": 0.16784432530403137, "learning_rate": 5.837963063879884e-05, "loss": 0.4347, "step": 739 }, { "epoch": 2.0204778156996586, "grad_norm": 0.18551002442836761, "learning_rate": 5.809071599093272e-05, "loss": 0.4407, "step": 740 }, { "epoch": 2.02320819112628, "grad_norm": 0.17308658361434937, "learning_rate": 5.780222507942654e-05, "loss": 0.4343, "step": 741 }, { "epoch": 2.025938566552901, "grad_norm": 0.17182452976703644, "learning_rate": 5.751416082115408e-05, "loss": 0.422, "step": 742 }, { "epoch": 2.028668941979522, "grad_norm": 0.17356833815574646, "learning_rate": 5.722652612867523e-05, "loss": 0.4168, "step": 743 }, { "epoch": 2.0313993174061435, "grad_norm": 0.17314977943897247, "learning_rate": 5.6939323910206645e-05, "loss": 0.4285, "step": 744 }, { "epoch": 2.0341296928327646, "grad_norm": 0.1868155151605606, "learning_rate": 5.6652557069592304e-05, "loss": 0.4211, "step": 745 }, { "epoch": 2.0368600682593856, "grad_norm": 0.1524539738893509, "learning_rate": 5.63662285062742e-05, "loss": 0.4233, "step": 746 }, { "epoch": 2.0395904436860066, "grad_norm": 0.17929619550704956, "learning_rate": 5.608034111526298e-05, "loss": 0.4305, "step": 747 }, { "epoch": 2.042320819112628, "grad_norm": 0.1604132503271103, "learning_rate": 5.579489778710867e-05, "loss": 0.4214, "step": 748 }, { "epoch": 2.045051194539249, "grad_norm": 0.16710160672664642, "learning_rate": 5.550990140787147e-05, "loss": 0.4369, "step": 749 }, { "epoch": 2.04778156996587, "grad_norm": 0.16114738583564758, "learning_rate": 5.522535485909257e-05, "loss": 0.4216, "step": 750 }, { "epoch": 2.0505119453924916, "grad_norm": 0.16698378324508667, "learning_rate": 5.494126101776505e-05, "loss": 0.4323, "step": 751 }, { "epoch": 2.0532423208191126, "grad_norm": 0.174140065908432, "learning_rate": 5.4657622756304704e-05, "loss": 0.4135, "step": 752 }, { "epoch": 2.0559726962457336, "grad_norm": 0.1496962457895279, "learning_rate": 5.437444294252107e-05, "loss": 0.438, "step": 753 }, { "epoch": 2.058703071672355, "grad_norm": 0.1660911738872528, "learning_rate": 5.409172443958843e-05, "loss": 0.4262, "step": 754 }, { "epoch": 2.061433447098976, "grad_norm": 0.18081265687942505, "learning_rate": 5.380947010601681e-05, "loss": 0.4172, "step": 755 }, { "epoch": 2.064163822525597, "grad_norm": 0.14415475726127625, "learning_rate": 5.3527682795623146e-05, "loss": 0.4181, "step": 756 }, { "epoch": 2.0668941979522186, "grad_norm": 0.20684713125228882, "learning_rate": 5.324636535750238e-05, "loss": 0.4291, "step": 757 }, { "epoch": 2.0696245733788396, "grad_norm": 0.17085103690624237, "learning_rate": 5.296552063599868e-05, "loss": 0.4372, "step": 758 }, { "epoch": 2.0723549488054607, "grad_norm": 0.17061397433280945, "learning_rate": 5.2685151470676653e-05, "loss": 0.4247, "step": 759 }, { "epoch": 2.0750853242320817, "grad_norm": 0.17692053318023682, "learning_rate": 5.240526069629265e-05, "loss": 0.4261, "step": 760 }, { "epoch": 2.077815699658703, "grad_norm": 0.15353117883205414, "learning_rate": 5.212585114276614e-05, "loss": 0.4272, "step": 761 }, { "epoch": 2.080546075085324, "grad_norm": 0.17618128657341003, "learning_rate": 5.1846925635151045e-05, "loss": 0.4206, "step": 762 }, { "epoch": 2.083276450511945, "grad_norm": 0.154897078871727, "learning_rate": 5.156848699360719e-05, "loss": 0.4086, "step": 763 }, { "epoch": 2.0860068259385667, "grad_norm": 0.15779747068881989, "learning_rate": 5.129053803337181e-05, "loss": 0.4073, "step": 764 }, { "epoch": 2.0887372013651877, "grad_norm": 0.186599463224411, "learning_rate": 5.101308156473104e-05, "loss": 0.4204, "step": 765 }, { "epoch": 2.0914675767918087, "grad_norm": 0.15039357542991638, "learning_rate": 5.073612039299157e-05, "loss": 0.4277, "step": 766 }, { "epoch": 2.09419795221843, "grad_norm": 0.16809961199760437, "learning_rate": 5.0459657318452224e-05, "loss": 0.4312, "step": 767 }, { "epoch": 2.096928327645051, "grad_norm": 0.18085776269435883, "learning_rate": 5.0183695136375664e-05, "loss": 0.4233, "step": 768 }, { "epoch": 2.0996587030716722, "grad_norm": 0.15629561245441437, "learning_rate": 4.9908236636960126e-05, "loss": 0.4252, "step": 769 }, { "epoch": 2.1023890784982937, "grad_norm": 0.18560980260372162, "learning_rate": 4.963328460531127e-05, "loss": 0.4112, "step": 770 }, { "epoch": 2.1051194539249147, "grad_norm": 0.16012516617774963, "learning_rate": 4.935884182141377e-05, "loss": 0.4108, "step": 771 }, { "epoch": 2.1078498293515358, "grad_norm": 0.1524171382188797, "learning_rate": 4.908491106010368e-05, "loss": 0.4183, "step": 772 }, { "epoch": 2.1105802047781568, "grad_norm": 0.173212930560112, "learning_rate": 4.8811495091039926e-05, "loss": 0.4058, "step": 773 }, { "epoch": 2.1133105802047782, "grad_norm": 0.1540430635213852, "learning_rate": 4.8538596678676406e-05, "loss": 0.4016, "step": 774 }, { "epoch": 2.1160409556313993, "grad_norm": 0.1477975845336914, "learning_rate": 4.826621858223431e-05, "loss": 0.3975, "step": 775 }, { "epoch": 2.1187713310580203, "grad_norm": 0.17888864874839783, "learning_rate": 4.79943635556739e-05, "loss": 0.4208, "step": 776 }, { "epoch": 2.1215017064846418, "grad_norm": 0.15199348330497742, "learning_rate": 4.7723034347666696e-05, "loss": 0.4304, "step": 777 }, { "epoch": 2.124232081911263, "grad_norm": 0.15406261384487152, "learning_rate": 4.745223370156797e-05, "loss": 0.439, "step": 778 }, { "epoch": 2.126962457337884, "grad_norm": 0.1821894645690918, "learning_rate": 4.71819643553887e-05, "loss": 0.4298, "step": 779 }, { "epoch": 2.1296928327645053, "grad_norm": 0.15180355310440063, "learning_rate": 4.691222904176791e-05, "loss": 0.4136, "step": 780 }, { "epoch": 2.1324232081911263, "grad_norm": 0.17140239477157593, "learning_rate": 4.6643030487945326e-05, "loss": 0.4292, "step": 781 }, { "epoch": 2.1351535836177473, "grad_norm": 0.15253609418869019, "learning_rate": 4.6374371415733496e-05, "loss": 0.4393, "step": 782 }, { "epoch": 2.137883959044369, "grad_norm": 0.15631216764450073, "learning_rate": 4.6106254541490325e-05, "loss": 0.4162, "step": 783 }, { "epoch": 2.14061433447099, "grad_norm": 0.18031221628189087, "learning_rate": 4.583868257609171e-05, "loss": 0.415, "step": 784 }, { "epoch": 2.143344709897611, "grad_norm": 0.15720027685165405, "learning_rate": 4.55716582249042e-05, "loss": 0.4288, "step": 785 }, { "epoch": 2.146075085324232, "grad_norm": 0.15618009865283966, "learning_rate": 4.530518418775733e-05, "loss": 0.4238, "step": 786 }, { "epoch": 2.1488054607508533, "grad_norm": 0.16991287469863892, "learning_rate": 4.50392631589166e-05, "loss": 0.4321, "step": 787 }, { "epoch": 2.1515358361774743, "grad_norm": 0.16138002276420593, "learning_rate": 4.477389782705628e-05, "loss": 0.4172, "step": 788 }, { "epoch": 2.1542662116040954, "grad_norm": 0.15910767018795013, "learning_rate": 4.450909087523186e-05, "loss": 0.4149, "step": 789 }, { "epoch": 2.156996587030717, "grad_norm": 0.15495507419109344, "learning_rate": 4.424484498085335e-05, "loss": 0.4351, "step": 790 }, { "epoch": 2.159726962457338, "grad_norm": 0.15659591555595398, "learning_rate": 4.398116281565794e-05, "loss": 0.4254, "step": 791 }, { "epoch": 2.162457337883959, "grad_norm": 0.16147974133491516, "learning_rate": 4.371804704568309e-05, "loss": 0.4265, "step": 792 }, { "epoch": 2.1651877133105804, "grad_norm": 0.14304347336292267, "learning_rate": 4.345550033123954e-05, "loss": 0.4211, "step": 793 }, { "epoch": 2.1679180887372014, "grad_norm": 0.15479592978954315, "learning_rate": 4.3193525326884435e-05, "loss": 0.4002, "step": 794 }, { "epoch": 2.1706484641638224, "grad_norm": 0.1610931009054184, "learning_rate": 4.293212468139447e-05, "loss": 0.4156, "step": 795 }, { "epoch": 2.173378839590444, "grad_norm": 0.16558706760406494, "learning_rate": 4.267130103773911e-05, "loss": 0.4285, "step": 796 }, { "epoch": 2.176109215017065, "grad_norm": 0.16455373167991638, "learning_rate": 4.241105703305388e-05, "loss": 0.4058, "step": 797 }, { "epoch": 2.178839590443686, "grad_norm": 0.15886934101581573, "learning_rate": 4.215139529861367e-05, "loss": 0.4311, "step": 798 }, { "epoch": 2.181569965870307, "grad_norm": 0.15277110040187836, "learning_rate": 4.189231845980618e-05, "loss": 0.4176, "step": 799 }, { "epoch": 2.1843003412969284, "grad_norm": 0.15656784176826477, "learning_rate": 4.163382913610533e-05, "loss": 0.4205, "step": 800 }, { "epoch": 2.1870307167235494, "grad_norm": 0.1564100980758667, "learning_rate": 4.1375929941044786e-05, "loss": 0.4313, "step": 801 }, { "epoch": 2.1897610921501705, "grad_norm": 0.14734816551208496, "learning_rate": 4.111862348219158e-05, "loss": 0.4258, "step": 802 }, { "epoch": 2.192491467576792, "grad_norm": 0.15179724991321564, "learning_rate": 4.086191236111964e-05, "loss": 0.4372, "step": 803 }, { "epoch": 2.195221843003413, "grad_norm": 0.16015468537807465, "learning_rate": 4.060579917338362e-05, "loss": 0.4104, "step": 804 }, { "epoch": 2.197952218430034, "grad_norm": 0.1574854701757431, "learning_rate": 4.0350286508492554e-05, "loss": 0.4068, "step": 805 }, { "epoch": 2.2006825938566554, "grad_norm": 0.14426739513874054, "learning_rate": 4.009537694988372e-05, "loss": 0.3904, "step": 806 }, { "epoch": 2.2034129692832765, "grad_norm": 0.14949829876422882, "learning_rate": 3.9841073074896517e-05, "loss": 0.3934, "step": 807 }, { "epoch": 2.2061433447098975, "grad_norm": 0.1610165387392044, "learning_rate": 3.958737745474638e-05, "loss": 0.4207, "step": 808 }, { "epoch": 2.208873720136519, "grad_norm": 0.15804022550582886, "learning_rate": 3.933429265449882e-05, "loss": 0.3968, "step": 809 }, { "epoch": 2.21160409556314, "grad_norm": 0.15507763624191284, "learning_rate": 3.9081821233043436e-05, "loss": 0.4322, "step": 810 }, { "epoch": 2.214334470989761, "grad_norm": 0.16331470012664795, "learning_rate": 3.8829965743068174e-05, "loss": 0.4317, "step": 811 }, { "epoch": 2.217064846416382, "grad_norm": 0.14913159608840942, "learning_rate": 3.857872873103322e-05, "loss": 0.4098, "step": 812 }, { "epoch": 2.2197952218430035, "grad_norm": 0.15193897485733032, "learning_rate": 3.832811273714569e-05, "loss": 0.4319, "step": 813 }, { "epoch": 2.2225255972696245, "grad_norm": 0.15656188130378723, "learning_rate": 3.807812029533362e-05, "loss": 0.3962, "step": 814 }, { "epoch": 2.2252559726962455, "grad_norm": 0.1457897126674652, "learning_rate": 3.7828753933220295e-05, "loss": 0.4044, "step": 815 }, { "epoch": 2.227986348122867, "grad_norm": 0.15471549332141876, "learning_rate": 3.758001617209906e-05, "loss": 0.4251, "step": 816 }, { "epoch": 2.230716723549488, "grad_norm": 0.16151392459869385, "learning_rate": 3.733190952690753e-05, "loss": 0.4278, "step": 817 }, { "epoch": 2.233447098976109, "grad_norm": 0.15231560170650482, "learning_rate": 3.708443650620206e-05, "loss": 0.4286, "step": 818 }, { "epoch": 2.2361774744027305, "grad_norm": 0.13981011509895325, "learning_rate": 3.683759961213282e-05, "loss": 0.4127, "step": 819 }, { "epoch": 2.2389078498293515, "grad_norm": 0.16484162211418152, "learning_rate": 3.6591401340418116e-05, "loss": 0.4399, "step": 820 }, { "epoch": 2.2416382252559726, "grad_norm": 0.15228329598903656, "learning_rate": 3.634584418031915e-05, "loss": 0.4247, "step": 821 }, { "epoch": 2.244368600682594, "grad_norm": 0.14780929684638977, "learning_rate": 3.6100930614615205e-05, "loss": 0.4324, "step": 822 }, { "epoch": 2.247098976109215, "grad_norm": 0.1611994057893753, "learning_rate": 3.585666311957817e-05, "loss": 0.4263, "step": 823 }, { "epoch": 2.249829351535836, "grad_norm": 0.16382110118865967, "learning_rate": 3.561304416494762e-05, "loss": 0.4332, "step": 824 }, { "epoch": 2.252559726962457, "grad_norm": 0.15646643936634064, "learning_rate": 3.53700762139059e-05, "loss": 0.4132, "step": 825 }, { "epoch": 2.2552901023890786, "grad_norm": 0.16182062029838562, "learning_rate": 3.512776172305331e-05, "loss": 0.4199, "step": 826 }, { "epoch": 2.2580204778156996, "grad_norm": 0.15407824516296387, "learning_rate": 3.4886103142382945e-05, "loss": 0.4087, "step": 827 }, { "epoch": 2.260750853242321, "grad_norm": 0.1593010425567627, "learning_rate": 3.46451029152562e-05, "loss": 0.416, "step": 828 }, { "epoch": 2.263481228668942, "grad_norm": 0.15628038346767426, "learning_rate": 3.440476347837811e-05, "loss": 0.4296, "step": 829 }, { "epoch": 2.266211604095563, "grad_norm": 0.1571022868156433, "learning_rate": 3.41650872617724e-05, "loss": 0.419, "step": 830 }, { "epoch": 2.268941979522184, "grad_norm": 0.15472716093063354, "learning_rate": 3.392607668875718e-05, "loss": 0.4151, "step": 831 }, { "epoch": 2.2716723549488056, "grad_norm": 0.1419110745191574, "learning_rate": 3.36877341759205e-05, "loss": 0.3991, "step": 832 }, { "epoch": 2.2744027303754266, "grad_norm": 0.15272633731365204, "learning_rate": 3.345006213309557e-05, "loss": 0.3965, "step": 833 }, { "epoch": 2.2771331058020476, "grad_norm": 0.15313783288002014, "learning_rate": 3.321306296333673e-05, "loss": 0.4022, "step": 834 }, { "epoch": 2.279863481228669, "grad_norm": 0.14547322690486908, "learning_rate": 3.29767390628951e-05, "loss": 0.4127, "step": 835 }, { "epoch": 2.28259385665529, "grad_norm": 0.15116067230701447, "learning_rate": 3.274109282119413e-05, "loss": 0.4086, "step": 836 }, { "epoch": 2.285324232081911, "grad_norm": 0.1611739993095398, "learning_rate": 3.250612662080567e-05, "loss": 0.4261, "step": 837 }, { "epoch": 2.288054607508532, "grad_norm": 0.1671733260154724, "learning_rate": 3.227184283742591e-05, "loss": 0.4244, "step": 838 }, { "epoch": 2.2907849829351536, "grad_norm": 0.16232764720916748, "learning_rate": 3.2038243839851075e-05, "loss": 0.4118, "step": 839 }, { "epoch": 2.2935153583617747, "grad_norm": 0.1595815271139145, "learning_rate": 3.180533198995379e-05, "loss": 0.4333, "step": 840 }, { "epoch": 2.296245733788396, "grad_norm": 0.14766521751880646, "learning_rate": 3.1573109642659024e-05, "loss": 0.4066, "step": 841 }, { "epoch": 2.298976109215017, "grad_norm": 0.16081617772579193, "learning_rate": 3.134157914592032e-05, "loss": 0.4032, "step": 842 }, { "epoch": 2.301706484641638, "grad_norm": 0.15681862831115723, "learning_rate": 3.111074284069606e-05, "loss": 0.4256, "step": 843 }, { "epoch": 2.304436860068259, "grad_norm": 0.15217512845993042, "learning_rate": 3.088060306092582e-05, "loss": 0.4142, "step": 844 }, { "epoch": 2.3071672354948807, "grad_norm": 0.15541419386863708, "learning_rate": 3.065116213350671e-05, "loss": 0.4246, "step": 845 }, { "epoch": 2.3098976109215017, "grad_norm": 0.16863110661506653, "learning_rate": 3.042242237826991e-05, "loss": 0.4167, "step": 846 }, { "epoch": 2.3126279863481227, "grad_norm": 0.15719062089920044, "learning_rate": 3.0194386107957173e-05, "loss": 0.4272, "step": 847 }, { "epoch": 2.315358361774744, "grad_norm": 0.14599116146564484, "learning_rate": 2.9967055628197472e-05, "loss": 0.3826, "step": 848 }, { "epoch": 2.318088737201365, "grad_norm": 0.16562673449516296, "learning_rate": 2.974043323748367e-05, "loss": 0.4246, "step": 849 }, { "epoch": 2.3208191126279862, "grad_norm": 0.16425776481628418, "learning_rate": 2.951452122714926e-05, "loss": 0.4182, "step": 850 }, { "epoch": 2.3235494880546073, "grad_norm": 0.15654055774211884, "learning_rate": 2.9289321881345254e-05, "loss": 0.4349, "step": 851 }, { "epoch": 2.3262798634812287, "grad_norm": 0.15405891835689545, "learning_rate": 2.9064837477017048e-05, "loss": 0.4045, "step": 852 }, { "epoch": 2.3290102389078498, "grad_norm": 0.16419056057929993, "learning_rate": 2.88410702838814e-05, "loss": 0.4174, "step": 853 }, { "epoch": 2.331740614334471, "grad_norm": 0.16045036911964417, "learning_rate": 2.861802256440348e-05, "loss": 0.4185, "step": 854 }, { "epoch": 2.3344709897610922, "grad_norm": 0.15399502217769623, "learning_rate": 2.8395696573774032e-05, "loss": 0.4228, "step": 855 }, { "epoch": 2.3372013651877133, "grad_norm": 0.1503557562828064, "learning_rate": 2.8174094559886534e-05, "loss": 0.4146, "step": 856 }, { "epoch": 2.3399317406143343, "grad_norm": 0.16562052071094513, "learning_rate": 2.7953218763314458e-05, "loss": 0.4127, "step": 857 }, { "epoch": 2.3426621160409558, "grad_norm": 0.15860068798065186, "learning_rate": 2.773307141728867e-05, "loss": 0.4221, "step": 858 }, { "epoch": 2.345392491467577, "grad_norm": 0.16427016258239746, "learning_rate": 2.7513654747674788e-05, "loss": 0.4202, "step": 859 }, { "epoch": 2.348122866894198, "grad_norm": 0.15462426841259003, "learning_rate": 2.729497097295075e-05, "loss": 0.3986, "step": 860 }, { "epoch": 2.3508532423208193, "grad_norm": 0.150539368391037, "learning_rate": 2.7077022304184295e-05, "loss": 0.4112, "step": 861 }, { "epoch": 2.3535836177474403, "grad_norm": 0.14942197501659393, "learning_rate": 2.685981094501069e-05, "loss": 0.4173, "step": 862 }, { "epoch": 2.3563139931740613, "grad_norm": 0.15172016620635986, "learning_rate": 2.6643339091610377e-05, "loss": 0.4273, "step": 863 }, { "epoch": 2.359044368600683, "grad_norm": 0.15123523771762848, "learning_rate": 2.6427608932686843e-05, "loss": 0.4104, "step": 864 }, { "epoch": 2.361774744027304, "grad_norm": 0.15563499927520752, "learning_rate": 2.621262264944444e-05, "loss": 0.4215, "step": 865 }, { "epoch": 2.364505119453925, "grad_norm": 0.15889500081539154, "learning_rate": 2.599838241556626e-05, "loss": 0.4058, "step": 866 }, { "epoch": 2.3672354948805463, "grad_norm": 0.1514395922422409, "learning_rate": 2.5784890397192398e-05, "loss": 0.4058, "step": 867 }, { "epoch": 2.3699658703071673, "grad_norm": 0.15626998245716095, "learning_rate": 2.5572148752897795e-05, "loss": 0.4092, "step": 868 }, { "epoch": 2.3726962457337883, "grad_norm": 0.15478669106960297, "learning_rate": 2.5360159633670457e-05, "loss": 0.424, "step": 869 }, { "epoch": 2.3754266211604094, "grad_norm": 0.14791764318943024, "learning_rate": 2.514892518288988e-05, "loss": 0.4294, "step": 870 }, { "epoch": 2.378156996587031, "grad_norm": 0.15729525685310364, "learning_rate": 2.4938447536305243e-05, "loss": 0.4204, "step": 871 }, { "epoch": 2.380887372013652, "grad_norm": 0.1549883335828781, "learning_rate": 2.472872882201368e-05, "loss": 0.4302, "step": 872 }, { "epoch": 2.383617747440273, "grad_norm": 0.15608322620391846, "learning_rate": 2.451977116043911e-05, "loss": 0.4208, "step": 873 }, { "epoch": 2.3863481228668944, "grad_norm": 0.15836332738399506, "learning_rate": 2.431157666431052e-05, "loss": 0.4141, "step": 874 }, { "epoch": 2.3890784982935154, "grad_norm": 0.15775950253009796, "learning_rate": 2.410414743864059e-05, "loss": 0.4142, "step": 875 }, { "epoch": 2.3918088737201364, "grad_norm": 0.1490509808063507, "learning_rate": 2.3897485580704682e-05, "loss": 0.4079, "step": 876 }, { "epoch": 2.394539249146758, "grad_norm": 0.16001944243907928, "learning_rate": 2.3691593180019366e-05, "loss": 0.4276, "step": 877 }, { "epoch": 2.397269624573379, "grad_norm": 0.14967067539691925, "learning_rate": 2.3486472318321307e-05, "loss": 0.4045, "step": 878 }, { "epoch": 2.4, "grad_norm": 0.14551271498203278, "learning_rate": 2.3282125069546433e-05, "loss": 0.4169, "step": 879 }, { "epoch": 2.4027303754266214, "grad_norm": 0.15170224010944366, "learning_rate": 2.3078553499808797e-05, "loss": 0.4166, "step": 880 }, { "epoch": 2.4054607508532424, "grad_norm": 0.15381450951099396, "learning_rate": 2.2875759667379614e-05, "loss": 0.4046, "step": 881 }, { "epoch": 2.4081911262798634, "grad_norm": 0.15133820474147797, "learning_rate": 2.267374562266662e-05, "loss": 0.4053, "step": 882 }, { "epoch": 2.4109215017064844, "grad_norm": 0.15135996043682098, "learning_rate": 2.2472513408193384e-05, "loss": 0.4145, "step": 883 }, { "epoch": 2.413651877133106, "grad_norm": 0.1551310420036316, "learning_rate": 2.227206505857834e-05, "loss": 0.4107, "step": 884 }, { "epoch": 2.416382252559727, "grad_norm": 0.15202271938323975, "learning_rate": 2.207240260051453e-05, "loss": 0.409, "step": 885 }, { "epoch": 2.419112627986348, "grad_norm": 0.15912394225597382, "learning_rate": 2.1873528052749092e-05, "loss": 0.4293, "step": 886 }, { "epoch": 2.4218430034129694, "grad_norm": 0.164555624127388, "learning_rate": 2.167544342606256e-05, "loss": 0.4153, "step": 887 }, { "epoch": 2.4245733788395905, "grad_norm": 0.1536960005760193, "learning_rate": 2.1478150723248857e-05, "loss": 0.4161, "step": 888 }, { "epoch": 2.4273037542662115, "grad_norm": 0.1572561115026474, "learning_rate": 2.1281651939094992e-05, "loss": 0.3975, "step": 889 }, { "epoch": 2.430034129692833, "grad_norm": 0.14872194826602936, "learning_rate": 2.1085949060360654e-05, "loss": 0.4028, "step": 890 }, { "epoch": 2.432764505119454, "grad_norm": 0.15553632378578186, "learning_rate": 2.089104406575837e-05, "loss": 0.4113, "step": 891 }, { "epoch": 2.435494880546075, "grad_norm": 0.15172426402568817, "learning_rate": 2.0696938925933506e-05, "loss": 0.407, "step": 892 }, { "epoch": 2.4382252559726965, "grad_norm": 0.1533356010913849, "learning_rate": 2.0503635603444094e-05, "loss": 0.4225, "step": 893 }, { "epoch": 2.4409556313993175, "grad_norm": 0.15870912373065948, "learning_rate": 2.0311136052741277e-05, "loss": 0.4176, "step": 894 }, { "epoch": 2.4436860068259385, "grad_norm": 0.15616737306118011, "learning_rate": 2.0119442220149353e-05, "loss": 0.4158, "step": 895 }, { "epoch": 2.4464163822525595, "grad_norm": 0.15833789110183716, "learning_rate": 1.9928556043846214e-05, "loss": 0.403, "step": 896 }, { "epoch": 2.449146757679181, "grad_norm": 0.1554342359304428, "learning_rate": 1.9738479453843682e-05, "loss": 0.396, "step": 897 }, { "epoch": 2.451877133105802, "grad_norm": 0.17515774071216583, "learning_rate": 1.9549214371968004e-05, "loss": 0.4096, "step": 898 }, { "epoch": 2.454607508532423, "grad_norm": 0.15816594660282135, "learning_rate": 1.936076271184044e-05, "loss": 0.4232, "step": 899 }, { "epoch": 2.4573378839590445, "grad_norm": 0.1528582125902176, "learning_rate": 1.9173126378857907e-05, "loss": 0.4145, "step": 900 }, { "epoch": 2.4600682593856655, "grad_norm": 0.16006483137607574, "learning_rate": 1.898630727017371e-05, "loss": 0.4201, "step": 901 }, { "epoch": 2.4627986348122866, "grad_norm": 0.15996922552585602, "learning_rate": 1.8800307274678364e-05, "loss": 0.4056, "step": 902 }, { "epoch": 2.465529010238908, "grad_norm": 0.1555391103029251, "learning_rate": 1.861512827298051e-05, "loss": 0.4315, "step": 903 }, { "epoch": 2.468259385665529, "grad_norm": 0.15501669049263, "learning_rate": 1.8430772137387853e-05, "loss": 0.4159, "step": 904 }, { "epoch": 2.47098976109215, "grad_norm": 0.1523975431919098, "learning_rate": 1.8247240731888294e-05, "loss": 0.4004, "step": 905 }, { "epoch": 2.4737201365187715, "grad_norm": 0.1715194135904312, "learning_rate": 1.806453591213103e-05, "loss": 0.4124, "step": 906 }, { "epoch": 2.4764505119453926, "grad_norm": 0.15396980941295624, "learning_rate": 1.788265952540784e-05, "loss": 0.4094, "step": 907 }, { "epoch": 2.4791808873720136, "grad_norm": 0.1634356677532196, "learning_rate": 1.7701613410634365e-05, "loss": 0.4257, "step": 908 }, { "epoch": 2.4819112627986346, "grad_norm": 0.1548430621623993, "learning_rate": 1.752139939833154e-05, "loss": 0.3942, "step": 909 }, { "epoch": 2.484641638225256, "grad_norm": 0.15945452451705933, "learning_rate": 1.734201931060706e-05, "loss": 0.3979, "step": 910 }, { "epoch": 2.487372013651877, "grad_norm": 0.16635702550411224, "learning_rate": 1.7163474961137028e-05, "loss": 0.422, "step": 911 }, { "epoch": 2.490102389078498, "grad_norm": 0.16245630383491516, "learning_rate": 1.6985768155147496e-05, "loss": 0.4126, "step": 912 }, { "epoch": 2.4928327645051196, "grad_norm": 0.14662671089172363, "learning_rate": 1.6808900689396336e-05, "loss": 0.4062, "step": 913 }, { "epoch": 2.4955631399317406, "grad_norm": 0.1555013507604599, "learning_rate": 1.663287435215498e-05, "loss": 0.4101, "step": 914 }, { "epoch": 2.4982935153583616, "grad_norm": 0.14830157160758972, "learning_rate": 1.645769092319045e-05, "loss": 0.3843, "step": 915 }, { "epoch": 2.5010238907849827, "grad_norm": 0.16135641932487488, "learning_rate": 1.6283352173747145e-05, "loss": 0.4229, "step": 916 }, { "epoch": 2.503754266211604, "grad_norm": 0.15229038894176483, "learning_rate": 1.6109859866529255e-05, "loss": 0.4209, "step": 917 }, { "epoch": 2.506484641638225, "grad_norm": 0.15836934745311737, "learning_rate": 1.5937215755682665e-05, "loss": 0.422, "step": 918 }, { "epoch": 2.5092150170648466, "grad_norm": 0.1639019101858139, "learning_rate": 1.5765421586777284e-05, "loss": 0.4206, "step": 919 }, { "epoch": 2.5119453924914676, "grad_norm": 0.15451960265636444, "learning_rate": 1.5594479096789537e-05, "loss": 0.411, "step": 920 }, { "epoch": 2.5146757679180887, "grad_norm": 0.16513267159461975, "learning_rate": 1.5424390014084644e-05, "loss": 0.4324, "step": 921 }, { "epoch": 2.5174061433447097, "grad_norm": 0.15432654321193695, "learning_rate": 1.5255156058399122e-05, "loss": 0.4074, "step": 922 }, { "epoch": 2.520136518771331, "grad_norm": 0.16064870357513428, "learning_rate": 1.5086778940823543e-05, "loss": 0.417, "step": 923 }, { "epoch": 2.522866894197952, "grad_norm": 0.16009055078029633, "learning_rate": 1.4919260363785215e-05, "loss": 0.4128, "step": 924 }, { "epoch": 2.5255972696245736, "grad_norm": 0.1598517745733261, "learning_rate": 1.4752602021030792e-05, "loss": 0.4191, "step": 925 }, { "epoch": 2.5283276450511947, "grad_norm": 0.15252196788787842, "learning_rate": 1.4586805597609331e-05, "loss": 0.4124, "step": 926 }, { "epoch": 2.5310580204778157, "grad_norm": 0.1643335521221161, "learning_rate": 1.442187276985526e-05, "loss": 0.4207, "step": 927 }, { "epoch": 2.5337883959044367, "grad_norm": 0.15445098280906677, "learning_rate": 1.4257805205371234e-05, "loss": 0.3993, "step": 928 }, { "epoch": 2.536518771331058, "grad_norm": 0.15789660811424255, "learning_rate": 1.4094604563011472e-05, "loss": 0.4103, "step": 929 }, { "epoch": 2.539249146757679, "grad_norm": 0.15002034604549408, "learning_rate": 1.3932272492864984e-05, "loss": 0.4042, "step": 930 }, { "epoch": 2.5419795221843002, "grad_norm": 0.16790151596069336, "learning_rate": 1.3770810636238684e-05, "loss": 0.4245, "step": 931 }, { "epoch": 2.5447098976109217, "grad_norm": 0.1551153063774109, "learning_rate": 1.3610220625641002e-05, "loss": 0.4145, "step": 932 }, { "epoch": 2.5474402730375427, "grad_norm": 0.15363937616348267, "learning_rate": 1.3450504084765381e-05, "loss": 0.384, "step": 933 }, { "epoch": 2.5501706484641637, "grad_norm": 0.15322524309158325, "learning_rate": 1.3291662628473633e-05, "loss": 0.4042, "step": 934 }, { "epoch": 2.5529010238907848, "grad_norm": 0.1649988293647766, "learning_rate": 1.313369786277987e-05, "loss": 0.4236, "step": 935 }, { "epoch": 2.5556313993174062, "grad_norm": 0.150667205452919, "learning_rate": 1.2976611384834148e-05, "loss": 0.3908, "step": 936 }, { "epoch": 2.5583617747440273, "grad_norm": 0.16129009425640106, "learning_rate": 1.2820404782906315e-05, "loss": 0.4167, "step": 937 }, { "epoch": 2.5610921501706487, "grad_norm": 0.16965742409229279, "learning_rate": 1.2665079636369969e-05, "loss": 0.4145, "step": 938 }, { "epoch": 2.5638225255972698, "grad_norm": 0.14878158271312714, "learning_rate": 1.2510637515686496e-05, "loss": 0.4236, "step": 939 }, { "epoch": 2.5665529010238908, "grad_norm": 0.14919213950634003, "learning_rate": 1.2357079982389197e-05, "loss": 0.396, "step": 940 }, { "epoch": 2.569283276450512, "grad_norm": 0.1563798040151596, "learning_rate": 1.2204408589067462e-05, "loss": 0.3992, "step": 941 }, { "epoch": 2.5720136518771333, "grad_norm": 0.15581347048282623, "learning_rate": 1.2052624879351104e-05, "loss": 0.4261, "step": 942 }, { "epoch": 2.5747440273037543, "grad_norm": 0.15438248217105865, "learning_rate": 1.190173038789476e-05, "loss": 0.4013, "step": 943 }, { "epoch": 2.5774744027303753, "grad_norm": 0.15199199318885803, "learning_rate": 1.1751726640362349e-05, "loss": 0.4089, "step": 944 }, { "epoch": 2.580204778156997, "grad_norm": 0.1516939401626587, "learning_rate": 1.1602615153411667e-05, "loss": 0.4008, "step": 945 }, { "epoch": 2.582935153583618, "grad_norm": 0.15474575757980347, "learning_rate": 1.1454397434679021e-05, "loss": 0.4115, "step": 946 }, { "epoch": 2.585665529010239, "grad_norm": 0.15419447422027588, "learning_rate": 1.1307074982764022e-05, "loss": 0.4187, "step": 947 }, { "epoch": 2.58839590443686, "grad_norm": 0.15413175523281097, "learning_rate": 1.116064928721442e-05, "loss": 0.4191, "step": 948 }, { "epoch": 2.5911262798634813, "grad_norm": 0.1559099704027176, "learning_rate": 1.1015121828511032e-05, "loss": 0.4136, "step": 949 }, { "epoch": 2.5938566552901023, "grad_norm": 0.16210560500621796, "learning_rate": 1.0870494078052796e-05, "loss": 0.4204, "step": 950 }, { "epoch": 2.596587030716724, "grad_norm": 0.15410131216049194, "learning_rate": 1.0726767498141877e-05, "loss": 0.4098, "step": 951 }, { "epoch": 2.599317406143345, "grad_norm": 0.14466793835163116, "learning_rate": 1.0583943541968856e-05, "loss": 0.3832, "step": 952 }, { "epoch": 2.602047781569966, "grad_norm": 0.1512717753648758, "learning_rate": 1.044202365359811e-05, "loss": 0.4132, "step": 953 }, { "epoch": 2.604778156996587, "grad_norm": 0.1530720293521881, "learning_rate": 1.0301009267953143e-05, "loss": 0.4165, "step": 954 }, { "epoch": 2.6075085324232083, "grad_norm": 0.16781674325466156, "learning_rate": 1.0160901810802115e-05, "loss": 0.4203, "step": 955 }, { "epoch": 2.6102389078498294, "grad_norm": 0.14876051247119904, "learning_rate": 1.0021702698743407e-05, "loss": 0.4168, "step": 956 }, { "epoch": 2.6129692832764504, "grad_norm": 0.15021638572216034, "learning_rate": 9.883413339191294e-06, "loss": 0.4173, "step": 957 }, { "epoch": 2.615699658703072, "grad_norm": 0.159826397895813, "learning_rate": 9.746035130361742e-06, "loss": 0.4279, "step": 958 }, { "epoch": 2.618430034129693, "grad_norm": 0.156574085354805, "learning_rate": 9.609569461258262e-06, "loss": 0.4277, "step": 959 }, { "epoch": 2.621160409556314, "grad_norm": 0.157151460647583, "learning_rate": 9.474017711657834e-06, "loss": 0.412, "step": 960 }, { "epoch": 2.623890784982935, "grad_norm": 0.14887213706970215, "learning_rate": 9.339381252097e-06, "loss": 0.4012, "step": 961 }, { "epoch": 2.6266211604095564, "grad_norm": 0.15450581908226013, "learning_rate": 9.205661443857994e-06, "loss": 0.4077, "step": 962 }, { "epoch": 2.6293515358361774, "grad_norm": 0.14593878388404846, "learning_rate": 9.072859638954955e-06, "loss": 0.4064, "step": 963 }, { "epoch": 2.632081911262799, "grad_norm": 0.156602144241333, "learning_rate": 8.940977180120247e-06, "loss": 0.4267, "step": 964 }, { "epoch": 2.63481228668942, "grad_norm": 0.16739366948604584, "learning_rate": 8.810015400790994e-06, "loss": 0.4176, "step": 965 }, { "epoch": 2.637542662116041, "grad_norm": 0.15772853791713715, "learning_rate": 8.67997562509546e-06, "loss": 0.424, "step": 966 }, { "epoch": 2.640273037542662, "grad_norm": 0.15860068798065186, "learning_rate": 8.550859167839664e-06, "loss": 0.4167, "step": 967 }, { "epoch": 2.6430034129692834, "grad_norm": 0.15044620633125305, "learning_rate": 8.422667334494249e-06, "loss": 0.3916, "step": 968 }, { "epoch": 2.6457337883959045, "grad_norm": 0.150175541639328, "learning_rate": 8.295401421181125e-06, "loss": 0.3953, "step": 969 }, { "epoch": 2.6484641638225255, "grad_norm": 0.14696063101291656, "learning_rate": 8.169062714660346e-06, "loss": 0.4115, "step": 970 }, { "epoch": 2.651194539249147, "grad_norm": 0.1482568085193634, "learning_rate": 8.043652492317256e-06, "loss": 0.4017, "step": 971 }, { "epoch": 2.653924914675768, "grad_norm": 0.15316608548164368, "learning_rate": 7.919172022149456e-06, "loss": 0.4176, "step": 972 }, { "epoch": 2.656655290102389, "grad_norm": 0.15325787663459778, "learning_rate": 7.795622562753957e-06, "loss": 0.413, "step": 973 }, { "epoch": 2.65938566552901, "grad_norm": 0.16979162395000458, "learning_rate": 7.673005363314579e-06, "loss": 0.4244, "step": 974 }, { "epoch": 2.6621160409556315, "grad_norm": 0.16061224043369293, "learning_rate": 7.551321663589228e-06, "loss": 0.4082, "step": 975 }, { "epoch": 2.6648464163822525, "grad_norm": 0.1511377990245819, "learning_rate": 7.430572693897342e-06, "loss": 0.4047, "step": 976 }, { "epoch": 2.667576791808874, "grad_norm": 0.1549064815044403, "learning_rate": 7.310759675107515e-06, "loss": 0.4181, "step": 977 }, { "epoch": 2.670307167235495, "grad_norm": 0.15855662524700165, "learning_rate": 7.191883818625189e-06, "loss": 0.4242, "step": 978 }, { "epoch": 2.673037542662116, "grad_norm": 0.16046655178070068, "learning_rate": 7.073946326380243e-06, "loss": 0.4077, "step": 979 }, { "epoch": 2.675767918088737, "grad_norm": 0.1561538577079773, "learning_rate": 6.956948390814977e-06, "loss": 0.4117, "step": 980 }, { "epoch": 2.6784982935153585, "grad_norm": 0.16078175604343414, "learning_rate": 6.840891194872112e-06, "loss": 0.4342, "step": 981 }, { "epoch": 2.6812286689419795, "grad_norm": 0.15352275967597961, "learning_rate": 6.725775911982601e-06, "loss": 0.402, "step": 982 }, { "epoch": 2.6839590443686006, "grad_norm": 0.15391647815704346, "learning_rate": 6.6116037060539704e-06, "loss": 0.4095, "step": 983 }, { "epoch": 2.686689419795222, "grad_norm": 0.15556836128234863, "learning_rate": 6.498375731458528e-06, "loss": 0.4063, "step": 984 }, { "epoch": 2.689419795221843, "grad_norm": 0.15222905576229095, "learning_rate": 6.386093133021554e-06, "loss": 0.4139, "step": 985 }, { "epoch": 2.692150170648464, "grad_norm": 0.15189246833324432, "learning_rate": 6.274757046009871e-06, "loss": 0.4195, "step": 986 }, { "epoch": 2.694880546075085, "grad_norm": 0.15635477006435394, "learning_rate": 6.164368596120351e-06, "loss": 0.4137, "step": 987 }, { "epoch": 2.6976109215017066, "grad_norm": 0.15748678147792816, "learning_rate": 6.054928899468426e-06, "loss": 0.396, "step": 988 }, { "epoch": 2.7003412969283276, "grad_norm": 0.1585109978914261, "learning_rate": 5.946439062576903e-06, "loss": 0.4111, "step": 989 }, { "epoch": 2.703071672354949, "grad_norm": 0.15509194135665894, "learning_rate": 5.83890018236476e-06, "loss": 0.4259, "step": 990 }, { "epoch": 2.70580204778157, "grad_norm": 0.15443935990333557, "learning_rate": 5.732313346136031e-06, "loss": 0.4138, "step": 991 }, { "epoch": 2.708532423208191, "grad_norm": 0.1619240939617157, "learning_rate": 5.626679631568832e-06, "loss": 0.4091, "step": 992 }, { "epoch": 2.711262798634812, "grad_norm": 0.1597377061843872, "learning_rate": 5.522000106704439e-06, "loss": 0.4193, "step": 993 }, { "epoch": 2.7139931740614336, "grad_norm": 0.15020039677619934, "learning_rate": 5.418275829936537e-06, "loss": 0.4216, "step": 994 }, { "epoch": 2.7167235494880546, "grad_norm": 0.15186108648777008, "learning_rate": 5.315507850000456e-06, "loss": 0.4057, "step": 995 }, { "epoch": 2.7194539249146756, "grad_norm": 0.1630185842514038, "learning_rate": 5.2136972059626314e-06, "loss": 0.4141, "step": 996 }, { "epoch": 2.722184300341297, "grad_norm": 0.1610775589942932, "learning_rate": 5.112844927210048e-06, "loss": 0.4025, "step": 997 }, { "epoch": 2.724914675767918, "grad_norm": 0.15820352733135223, "learning_rate": 5.012952033439844e-06, "loss": 0.4197, "step": 998 }, { "epoch": 2.727645051194539, "grad_norm": 0.1567496657371521, "learning_rate": 4.914019534649039e-06, "loss": 0.4215, "step": 999 }, { "epoch": 2.73037542662116, "grad_norm": 0.1540801227092743, "learning_rate": 4.816048431124265e-06, "loss": 0.4237, "step": 1000 }, { "epoch": 2.7331058020477816, "grad_norm": 0.15339985489845276, "learning_rate": 4.719039713431694e-06, "loss": 0.4127, "step": 1001 }, { "epoch": 2.7358361774744027, "grad_norm": 0.1545177400112152, "learning_rate": 4.622994362406996e-06, "loss": 0.424, "step": 1002 }, { "epoch": 2.738566552901024, "grad_norm": 0.15001171827316284, "learning_rate": 4.527913349145441e-06, "loss": 0.4199, "step": 1003 }, { "epoch": 2.741296928327645, "grad_norm": 0.15352268517017365, "learning_rate": 4.433797634992077e-06, "loss": 0.3991, "step": 1004 }, { "epoch": 2.744027303754266, "grad_norm": 0.1552933305501938, "learning_rate": 4.340648171531992e-06, "loss": 0.4173, "step": 1005 }, { "epoch": 2.746757679180887, "grad_norm": 0.1541508436203003, "learning_rate": 4.248465900580734e-06, "loss": 0.4231, "step": 1006 }, { "epoch": 2.7494880546075087, "grad_norm": 0.15616647899150848, "learning_rate": 4.1572517541747294e-06, "loss": 0.4295, "step": 1007 }, { "epoch": 2.7522184300341297, "grad_norm": 0.14823675155639648, "learning_rate": 4.0670066545619225e-06, "loss": 0.403, "step": 1008 }, { "epoch": 2.7549488054607507, "grad_norm": 0.15372464060783386, "learning_rate": 3.977731514192385e-06, "loss": 0.4184, "step": 1009 }, { "epoch": 2.757679180887372, "grad_norm": 0.15206997096538544, "learning_rate": 3.889427235709153e-06, "loss": 0.4119, "step": 1010 }, { "epoch": 2.760409556313993, "grad_norm": 0.15120883285999298, "learning_rate": 3.802094711939075e-06, "loss": 0.413, "step": 1011 }, { "epoch": 2.7631399317406142, "grad_norm": 0.16259510815143585, "learning_rate": 3.7157348258837652e-06, "loss": 0.4253, "step": 1012 }, { "epoch": 2.7658703071672353, "grad_norm": 0.15381862223148346, "learning_rate": 3.6303484507106966e-06, "loss": 0.4162, "step": 1013 }, { "epoch": 2.7686006825938567, "grad_norm": 0.15234719216823578, "learning_rate": 3.5459364497443694e-06, "loss": 0.3893, "step": 1014 }, { "epoch": 2.7713310580204777, "grad_norm": 0.1562722623348236, "learning_rate": 3.4624996764575977e-06, "loss": 0.3979, "step": 1015 }, { "epoch": 2.774061433447099, "grad_norm": 0.15057340264320374, "learning_rate": 3.3800389744628404e-06, "loss": 0.3861, "step": 1016 }, { "epoch": 2.7767918088737202, "grad_norm": 0.1512756198644638, "learning_rate": 3.298555177503726e-06, "loss": 0.4135, "step": 1017 }, { "epoch": 2.7795221843003413, "grad_norm": 0.1467510461807251, "learning_rate": 3.2180491094465415e-06, "loss": 0.4133, "step": 1018 }, { "epoch": 2.7822525597269623, "grad_norm": 0.1518513709306717, "learning_rate": 3.1385215842720027e-06, "loss": 0.4032, "step": 1019 }, { "epoch": 2.7849829351535837, "grad_norm": 0.15052153170108795, "learning_rate": 3.059973406066963e-06, "loss": 0.4157, "step": 1020 }, { "epoch": 2.7877133105802048, "grad_norm": 0.15297654271125793, "learning_rate": 2.9824053690162723e-06, "loss": 0.4153, "step": 1021 }, { "epoch": 2.790443686006826, "grad_norm": 0.15268519520759583, "learning_rate": 2.905818257394799e-06, "loss": 0.4083, "step": 1022 }, { "epoch": 2.7931740614334473, "grad_norm": 0.1531766653060913, "learning_rate": 2.8302128455594656e-06, "loss": 0.3943, "step": 1023 }, { "epoch": 2.7959044368600683, "grad_norm": 0.1546778380870819, "learning_rate": 2.7555898979413797e-06, "loss": 0.4091, "step": 1024 }, { "epoch": 2.7986348122866893, "grad_norm": 0.15699811279773712, "learning_rate": 2.6819501690382277e-06, "loss": 0.4249, "step": 1025 }, { "epoch": 2.8013651877133103, "grad_norm": 0.15149515867233276, "learning_rate": 2.609294403406537e-06, "loss": 0.4064, "step": 1026 }, { "epoch": 2.804095563139932, "grad_norm": 0.15525740385055542, "learning_rate": 2.537623335654127e-06, "loss": 0.4086, "step": 1027 }, { "epoch": 2.806825938566553, "grad_norm": 0.15066871047019958, "learning_rate": 2.4669376904328247e-06, "loss": 0.4046, "step": 1028 }, { "epoch": 2.8095563139931743, "grad_norm": 0.1503557562828064, "learning_rate": 2.397238182430994e-06, "loss": 0.4007, "step": 1029 }, { "epoch": 2.8122866894197953, "grad_norm": 0.15836714208126068, "learning_rate": 2.3285255163663532e-06, "loss": 0.4297, "step": 1030 }, { "epoch": 2.8150170648464163, "grad_norm": 0.1532403528690338, "learning_rate": 2.2608003869788786e-06, "loss": 0.4096, "step": 1031 }, { "epoch": 2.8177474402730374, "grad_norm": 0.1478443145751953, "learning_rate": 2.1940634790238e-06, "loss": 0.3819, "step": 1032 }, { "epoch": 2.820477815699659, "grad_norm": 0.1495964378118515, "learning_rate": 2.128315467264552e-06, "loss": 0.4086, "step": 1033 }, { "epoch": 2.82320819112628, "grad_norm": 0.15715493261814117, "learning_rate": 2.063557016466111e-06, "loss": 0.403, "step": 1034 }, { "epoch": 2.825938566552901, "grad_norm": 0.14644889533519745, "learning_rate": 1.999788781388201e-06, "loss": 0.405, "step": 1035 }, { "epoch": 2.8286689419795223, "grad_norm": 0.15347984433174133, "learning_rate": 1.9370114067785994e-06, "loss": 0.4198, "step": 1036 }, { "epoch": 2.8313993174061434, "grad_norm": 0.14934033155441284, "learning_rate": 1.8752255273667752e-06, "loss": 0.4078, "step": 1037 }, { "epoch": 2.8341296928327644, "grad_norm": 0.1513030081987381, "learning_rate": 1.8144317678573497e-06, "loss": 0.4165, "step": 1038 }, { "epoch": 2.8368600682593854, "grad_norm": 0.15821826457977295, "learning_rate": 1.754630742923813e-06, "loss": 0.4213, "step": 1039 }, { "epoch": 2.839590443686007, "grad_norm": 0.1506132036447525, "learning_rate": 1.6958230572023503e-06, "loss": 0.4058, "step": 1040 }, { "epoch": 2.842320819112628, "grad_norm": 0.15292277932167053, "learning_rate": 1.6380093052856483e-06, "loss": 0.4219, "step": 1041 }, { "epoch": 2.8450511945392494, "grad_norm": 0.15926344692707062, "learning_rate": 1.5811900717169538e-06, "loss": 0.4144, "step": 1042 }, { "epoch": 2.8477815699658704, "grad_norm": 0.15735220909118652, "learning_rate": 1.525365930984146e-06, "loss": 0.3986, "step": 1043 }, { "epoch": 2.8505119453924914, "grad_norm": 0.15544278919696808, "learning_rate": 1.4705374475138978e-06, "loss": 0.4151, "step": 1044 }, { "epoch": 2.8532423208191124, "grad_norm": 0.14867156744003296, "learning_rate": 1.416705175666e-06, "loss": 0.4039, "step": 1045 }, { "epoch": 2.855972696245734, "grad_norm": 0.15105663239955902, "learning_rate": 1.3638696597277679e-06, "loss": 0.4022, "step": 1046 }, { "epoch": 2.858703071672355, "grad_norm": 0.15061454474925995, "learning_rate": 1.3120314339084783e-06, "loss": 0.3928, "step": 1047 }, { "epoch": 2.861433447098976, "grad_norm": 0.15541335940361023, "learning_rate": 1.2611910223340407e-06, "loss": 0.4108, "step": 1048 }, { "epoch": 2.8641638225255974, "grad_norm": 0.15430454909801483, "learning_rate": 1.2113489390416566e-06, "loss": 0.4142, "step": 1049 }, { "epoch": 2.8668941979522184, "grad_norm": 0.1592140942811966, "learning_rate": 1.1625056879746133e-06, "loss": 0.4123, "step": 1050 }, { "epoch": 2.8696245733788395, "grad_norm": 0.15730910003185272, "learning_rate": 1.1146617629772315e-06, "loss": 0.4047, "step": 1051 }, { "epoch": 2.8723549488054605, "grad_norm": 0.15714263916015625, "learning_rate": 1.0678176477898372e-06, "loss": 0.4258, "step": 1052 }, { "epoch": 2.875085324232082, "grad_norm": 0.15376383066177368, "learning_rate": 1.0219738160438753e-06, "loss": 0.4098, "step": 1053 }, { "epoch": 2.877815699658703, "grad_norm": 0.15481893718242645, "learning_rate": 9.771307312571254e-07, "loss": 0.408, "step": 1054 }, { "epoch": 2.8805460750853245, "grad_norm": 0.15664780139923096, "learning_rate": 9.332888468290169e-07, "loss": 0.3983, "step": 1055 }, { "epoch": 2.8832764505119455, "grad_norm": 0.15640319883823395, "learning_rate": 8.90448606036054e-07, "loss": 0.4215, "step": 1056 }, { "epoch": 2.8860068259385665, "grad_norm": 0.15267756581306458, "learning_rate": 8.486104420272977e-07, "loss": 0.401, "step": 1057 }, { "epoch": 2.8887372013651875, "grad_norm": 0.15257929265499115, "learning_rate": 8.077747778200473e-07, "loss": 0.4145, "step": 1058 }, { "epoch": 2.891467576791809, "grad_norm": 0.1439686268568039, "learning_rate": 7.679420262954984e-07, "loss": 0.3806, "step": 1059 }, { "epoch": 2.89419795221843, "grad_norm": 0.15835194289684296, "learning_rate": 7.291125901946027e-07, "loss": 0.4156, "step": 1060 }, { "epoch": 2.896928327645051, "grad_norm": 0.15433841943740845, "learning_rate": 6.912868621140045e-07, "loss": 0.4198, "step": 1061 }, { "epoch": 2.8996587030716725, "grad_norm": 0.15369294583797455, "learning_rate": 6.544652245020433e-07, "loss": 0.4086, "step": 1062 }, { "epoch": 2.9023890784982935, "grad_norm": 0.16046328842639923, "learning_rate": 6.18648049654913e-07, "loss": 0.4046, "step": 1063 }, { "epoch": 2.9051194539249146, "grad_norm": 0.15750819444656372, "learning_rate": 5.838356997128869e-07, "loss": 0.4245, "step": 1064 }, { "epoch": 2.9078498293515356, "grad_norm": 0.16013328731060028, "learning_rate": 5.500285266566319e-07, "loss": 0.4128, "step": 1065 }, { "epoch": 2.910580204778157, "grad_norm": 0.15908414125442505, "learning_rate": 5.172268723036999e-07, "loss": 0.4256, "step": 1066 }, { "epoch": 2.913310580204778, "grad_norm": 0.15130603313446045, "learning_rate": 4.854310683050312e-07, "loss": 0.4253, "step": 1067 }, { "epoch": 2.9160409556313995, "grad_norm": 0.1521066129207611, "learning_rate": 4.546414361416229e-07, "loss": 0.4131, "step": 1068 }, { "epoch": 2.9187713310580206, "grad_norm": 0.15544620156288147, "learning_rate": 4.2485828712126583e-07, "loss": 0.4088, "step": 1069 }, { "epoch": 2.9215017064846416, "grad_norm": 0.1541679948568344, "learning_rate": 3.96081922375402e-07, "loss": 0.4083, "step": 1070 }, { "epoch": 2.9242320819112626, "grad_norm": 0.15230417251586914, "learning_rate": 3.6831263285608266e-07, "loss": 0.4067, "step": 1071 }, { "epoch": 2.926962457337884, "grad_norm": 0.15117131173610687, "learning_rate": 3.415506993330153e-07, "loss": 0.4138, "step": 1072 }, { "epoch": 2.929692832764505, "grad_norm": 0.1522316336631775, "learning_rate": 3.1579639239074365e-07, "loss": 0.4052, "step": 1073 }, { "epoch": 2.932423208191126, "grad_norm": 0.1582721471786499, "learning_rate": 2.9104997242590527e-07, "loss": 0.4056, "step": 1074 }, { "epoch": 2.9351535836177476, "grad_norm": 0.1515754610300064, "learning_rate": 2.673116896445671e-07, "loss": 0.4054, "step": 1075 }, { "epoch": 2.9378839590443686, "grad_norm": 0.15097399055957794, "learning_rate": 2.4458178405974975e-07, "loss": 0.4036, "step": 1076 }, { "epoch": 2.9406143344709896, "grad_norm": 0.1505846381187439, "learning_rate": 2.2286048548897376e-07, "loss": 0.4269, "step": 1077 }, { "epoch": 2.9433447098976107, "grad_norm": 0.14722640812397003, "learning_rate": 2.0214801355192824e-07, "loss": 0.3938, "step": 1078 }, { "epoch": 2.946075085324232, "grad_norm": 0.15851718187332153, "learning_rate": 1.824445776682504e-07, "loss": 0.4163, "step": 1079 }, { "epoch": 2.948805460750853, "grad_norm": 0.1539052575826645, "learning_rate": 1.6375037705543826e-07, "loss": 0.4119, "step": 1080 }, { "epoch": 2.9515358361774746, "grad_norm": 0.15709060430526733, "learning_rate": 1.4606560072679687e-07, "loss": 0.4125, "step": 1081 }, { "epoch": 2.9542662116040956, "grad_norm": 0.14519765973091125, "learning_rate": 1.2939042748955077e-07, "loss": 0.399, "step": 1082 }, { "epoch": 2.9569965870307167, "grad_norm": 0.1516365110874176, "learning_rate": 1.1372502594303446e-07, "loss": 0.4181, "step": 1083 }, { "epoch": 2.9597269624573377, "grad_norm": 0.15326349437236786, "learning_rate": 9.906955447697153e-08, "loss": 0.4174, "step": 1084 }, { "epoch": 2.962457337883959, "grad_norm": 0.15291821956634521, "learning_rate": 8.542416126989805e-08, "loss": 0.4166, "step": 1085 }, { "epoch": 2.96518771331058, "grad_norm": 0.14680640399456024, "learning_rate": 7.27889842876417e-08, "loss": 0.4111, "step": 1086 }, { "epoch": 2.967918088737201, "grad_norm": 0.15194863080978394, "learning_rate": 6.116415128194497e-08, "loss": 0.4054, "step": 1087 }, { "epoch": 2.9706484641638227, "grad_norm": 0.160339817404747, "learning_rate": 5.054977978916631e-08, "loss": 0.4133, "step": 1088 }, { "epoch": 2.9733788395904437, "grad_norm": 0.15164430439472198, "learning_rate": 4.094597712908099e-08, "loss": 0.4134, "step": 1089 }, { "epoch": 2.9761092150170647, "grad_norm": 0.14412933588027954, "learning_rate": 3.2352840403804264e-08, "loss": 0.3914, "step": 1090 }, { "epoch": 2.9788395904436857, "grad_norm": 0.15993493795394897, "learning_rate": 2.477045649681431e-08, "loss": 0.4335, "step": 1091 }, { "epoch": 2.981569965870307, "grad_norm": 0.1526769995689392, "learning_rate": 1.81989020720974e-08, "loss": 0.4129, "step": 1092 }, { "epoch": 2.9843003412969282, "grad_norm": 0.14546047151088715, "learning_rate": 1.2638243573293018e-08, "loss": 0.3924, "step": 1093 }, { "epoch": 2.9870307167235497, "grad_norm": 0.150767520070076, "learning_rate": 8.088537223116532e-09, "loss": 0.3982, "step": 1094 }, { "epoch": 2.9897610921501707, "grad_norm": 0.15718306601047516, "learning_rate": 4.549829022748586e-09, "loss": 0.418, "step": 1095 }, { "epoch": 2.9924914675767917, "grad_norm": 0.14539772272109985, "learning_rate": 2.0221547513243897e-09, "loss": 0.4034, "step": 1096 }, { "epoch": 2.9952218430034128, "grad_norm": 0.15158307552337646, "learning_rate": 5.055399656894721e-10, "loss": 0.4038, "step": 1097 }, { "epoch": 2.9979522184300342, "grad_norm": 0.148399218916893, "learning_rate": 0.0, "loss": 0.4121, "step": 1098 } ], "logging_steps": 1, "max_steps": 1098, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.529684799263867e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }