{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 1161, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0008613264427217916, "grad_norm": 599.9192504882812, "learning_rate": 5.714285714285715e-07, "loss": 6.0815, "step": 1 }, { "epoch": 0.0017226528854435831, "grad_norm": 575.8373413085938, "learning_rate": 1.142857142857143e-06, "loss": 5.5586, "step": 2 }, { "epoch": 0.002583979328165375, "grad_norm": 564.647216796875, "learning_rate": 1.7142857142857145e-06, "loss": 5.7547, "step": 3 }, { "epoch": 0.0034453057708871662, "grad_norm": 407.28704833984375, "learning_rate": 2.285714285714286e-06, "loss": 5.438, "step": 4 }, { "epoch": 0.004306632213608958, "grad_norm": 371.2016906738281, "learning_rate": 2.8571428571428573e-06, "loss": 4.8455, "step": 5 }, { "epoch": 0.00516795865633075, "grad_norm": 371.3276672363281, "learning_rate": 3.428571428571429e-06, "loss": 3.7534, "step": 6 }, { "epoch": 0.006029285099052541, "grad_norm": 1765.4417724609375, "learning_rate": 4.000000000000001e-06, "loss": 3.7887, "step": 7 }, { "epoch": 0.0068906115417743325, "grad_norm": 293.5586242675781, "learning_rate": 4.571428571428572e-06, "loss": 3.0103, "step": 8 }, { "epoch": 0.007751937984496124, "grad_norm": 490.7660827636719, "learning_rate": 5.142857142857142e-06, "loss": 2.7897, "step": 9 }, { "epoch": 0.008613264427217916, "grad_norm": 211.81118774414062, "learning_rate": 5.7142857142857145e-06, "loss": 2.951, "step": 10 }, { "epoch": 0.009474590869939707, "grad_norm": 83.9891128540039, "learning_rate": 6.285714285714286e-06, "loss": 2.6668, "step": 11 }, { "epoch": 0.0103359173126615, "grad_norm": 296.01885986328125, "learning_rate": 6.857142857142858e-06, "loss": 2.4551, "step": 12 }, { "epoch": 0.01119724375538329, "grad_norm": 44.15952682495117, "learning_rate": 7.428571428571429e-06, "loss": 2.3069, "step": 13 }, { "epoch": 0.012058570198105082, "grad_norm": 28.404870986938477, "learning_rate": 8.000000000000001e-06, "loss": 2.1574, "step": 14 }, { "epoch": 0.012919896640826873, "grad_norm": 18892.63671875, "learning_rate": 8.571428571428571e-06, "loss": 8.4092, "step": 15 }, { "epoch": 0.013781223083548665, "grad_norm": 13.016464233398438, "learning_rate": 9.142857142857144e-06, "loss": 2.0747, "step": 16 }, { "epoch": 0.014642549526270457, "grad_norm": 24.35378074645996, "learning_rate": 9.714285714285715e-06, "loss": 2.1555, "step": 17 }, { "epoch": 0.015503875968992248, "grad_norm": 129.4594268798828, "learning_rate": 1.0285714285714285e-05, "loss": 2.2513, "step": 18 }, { "epoch": 0.01636520241171404, "grad_norm": 312.46337890625, "learning_rate": 1.0857142857142858e-05, "loss": 2.4386, "step": 19 }, { "epoch": 0.017226528854435832, "grad_norm": 1113.3402099609375, "learning_rate": 1.1428571428571429e-05, "loss": 2.6086, "step": 20 }, { "epoch": 0.01808785529715762, "grad_norm": 758.8408203125, "learning_rate": 1.2e-05, "loss": 2.3477, "step": 21 }, { "epoch": 0.018949181739879414, "grad_norm": 90.13806915283203, "learning_rate": 1.2571428571428572e-05, "loss": 2.0906, "step": 22 }, { "epoch": 0.019810508182601206, "grad_norm": 44.2735710144043, "learning_rate": 1.3142857142857145e-05, "loss": 2.0562, "step": 23 }, { "epoch": 0.020671834625323, "grad_norm": 42.16297149658203, "learning_rate": 1.3714285714285716e-05, "loss": 2.0478, "step": 24 }, { "epoch": 0.02153316106804479, "grad_norm": 35.452392578125, "learning_rate": 1.4285714285714287e-05, "loss": 2.0479, "step": 25 }, { "epoch": 0.02239448751076658, "grad_norm": 23.058853149414062, "learning_rate": 1.4857142857142858e-05, "loss": 1.9681, "step": 26 }, { "epoch": 0.023255813953488372, "grad_norm": 17.171300888061523, "learning_rate": 1.542857142857143e-05, "loss": 1.917, "step": 27 }, { "epoch": 0.024117140396210164, "grad_norm": 12.922857284545898, "learning_rate": 1.6000000000000003e-05, "loss": 1.8992, "step": 28 }, { "epoch": 0.024978466838931956, "grad_norm": 10.135607719421387, "learning_rate": 1.6571428571428574e-05, "loss": 1.8757, "step": 29 }, { "epoch": 0.025839793281653745, "grad_norm": 8.323143005371094, "learning_rate": 1.7142857142857142e-05, "loss": 1.8451, "step": 30 }, { "epoch": 0.026701119724375538, "grad_norm": 5.9443793296813965, "learning_rate": 1.7714285714285717e-05, "loss": 1.7739, "step": 31 }, { "epoch": 0.02756244616709733, "grad_norm": 4.917455196380615, "learning_rate": 1.8285714285714288e-05, "loss": 1.7803, "step": 32 }, { "epoch": 0.028423772609819122, "grad_norm": 5.2614240646362305, "learning_rate": 1.885714285714286e-05, "loss": 1.7771, "step": 33 }, { "epoch": 0.029285099052540915, "grad_norm": 4.4380316734313965, "learning_rate": 1.942857142857143e-05, "loss": 1.7563, "step": 34 }, { "epoch": 0.030146425495262703, "grad_norm": 4.11315393447876, "learning_rate": 2e-05, "loss": 1.7493, "step": 35 }, { "epoch": 0.031007751937984496, "grad_norm": 3.8529300689697266, "learning_rate": 1.9999961078220587e-05, "loss": 1.7646, "step": 36 }, { "epoch": 0.03186907838070629, "grad_norm": 3.429624080657959, "learning_rate": 1.9999844313185335e-05, "loss": 1.7406, "step": 37 }, { "epoch": 0.03273040482342808, "grad_norm": 3.103177785873413, "learning_rate": 1.9999649705803178e-05, "loss": 1.7295, "step": 38 }, { "epoch": 0.03359173126614987, "grad_norm": 2.6700687408447266, "learning_rate": 1.9999377257589012e-05, "loss": 1.6595, "step": 39 }, { "epoch": 0.034453057708871665, "grad_norm": 2.35770583152771, "learning_rate": 1.999902697066367e-05, "loss": 1.6766, "step": 40 }, { "epoch": 0.03531438415159346, "grad_norm": 2.01990008354187, "learning_rate": 1.9998598847753918e-05, "loss": 1.6472, "step": 41 }, { "epoch": 0.03617571059431524, "grad_norm": 1.8652081489562988, "learning_rate": 1.9998092892192403e-05, "loss": 1.6448, "step": 42 }, { "epoch": 0.037037037037037035, "grad_norm": 1.9675177335739136, "learning_rate": 1.999750910791767e-05, "loss": 1.6573, "step": 43 }, { "epoch": 0.03789836347975883, "grad_norm": 2.176922082901001, "learning_rate": 1.9996847499474102e-05, "loss": 1.6364, "step": 44 }, { "epoch": 0.03875968992248062, "grad_norm": 2.259882688522339, "learning_rate": 1.99961080720119e-05, "loss": 1.6164, "step": 45 }, { "epoch": 0.03962101636520241, "grad_norm": 2.2850229740142822, "learning_rate": 1.9995290831287032e-05, "loss": 1.5988, "step": 46 }, { "epoch": 0.040482342807924204, "grad_norm": 2.1443259716033936, "learning_rate": 1.9994395783661177e-05, "loss": 1.5978, "step": 47 }, { "epoch": 0.041343669250646, "grad_norm": 2.1305596828460693, "learning_rate": 1.9993422936101715e-05, "loss": 1.6235, "step": 48 }, { "epoch": 0.04220499569336779, "grad_norm": 2.253634214401245, "learning_rate": 1.9992372296181637e-05, "loss": 1.6344, "step": 49 }, { "epoch": 0.04306632213608958, "grad_norm": 1.944365382194519, "learning_rate": 1.9991243872079495e-05, "loss": 1.5638, "step": 50 }, { "epoch": 0.04392764857881137, "grad_norm": 1.8713358640670776, "learning_rate": 1.9990037672579347e-05, "loss": 1.59, "step": 51 }, { "epoch": 0.04478897502153316, "grad_norm": 1.4943434000015259, "learning_rate": 1.9988753707070675e-05, "loss": 1.6001, "step": 52 }, { "epoch": 0.04565030146425495, "grad_norm": 1.5794744491577148, "learning_rate": 1.9987391985548326e-05, "loss": 1.5618, "step": 53 }, { "epoch": 0.046511627906976744, "grad_norm": 1.5509154796600342, "learning_rate": 1.998595251861243e-05, "loss": 1.5514, "step": 54 }, { "epoch": 0.047372954349698536, "grad_norm": 1.7027164697647095, "learning_rate": 1.9984435317468298e-05, "loss": 1.5591, "step": 55 }, { "epoch": 0.04823428079242033, "grad_norm": 1.851098656654358, "learning_rate": 1.9982840393926374e-05, "loss": 1.6171, "step": 56 }, { "epoch": 0.04909560723514212, "grad_norm": 1.8336509466171265, "learning_rate": 1.9981167760402104e-05, "loss": 1.6013, "step": 57 }, { "epoch": 0.04995693367786391, "grad_norm": 1.5762431621551514, "learning_rate": 1.997941742991587e-05, "loss": 1.5275, "step": 58 }, { "epoch": 0.050818260120585705, "grad_norm": 1.5044102668762207, "learning_rate": 1.997758941609286e-05, "loss": 1.5214, "step": 59 }, { "epoch": 0.05167958656330749, "grad_norm": 1.93209707736969, "learning_rate": 1.9975683733162987e-05, "loss": 1.5819, "step": 60 }, { "epoch": 0.05254091300602928, "grad_norm": 1.545015811920166, "learning_rate": 1.9973700395960765e-05, "loss": 1.5227, "step": 61 }, { "epoch": 0.053402239448751075, "grad_norm": 1.2270804643630981, "learning_rate": 1.9971639419925197e-05, "loss": 1.5119, "step": 62 }, { "epoch": 0.05426356589147287, "grad_norm": 1.7906986474990845, "learning_rate": 1.9969500821099654e-05, "loss": 1.5528, "step": 63 }, { "epoch": 0.05512489233419466, "grad_norm": 1.5916507244110107, "learning_rate": 1.996728461613175e-05, "loss": 1.5367, "step": 64 }, { "epoch": 0.05598621877691645, "grad_norm": 1.413952350616455, "learning_rate": 1.996499082227321e-05, "loss": 1.5247, "step": 65 }, { "epoch": 0.056847545219638244, "grad_norm": 1.3606282472610474, "learning_rate": 1.996261945737975e-05, "loss": 1.459, "step": 66 }, { "epoch": 0.05770887166236004, "grad_norm": 1.327688455581665, "learning_rate": 1.996017053991091e-05, "loss": 1.5062, "step": 67 }, { "epoch": 0.05857019810508183, "grad_norm": 1.2740212678909302, "learning_rate": 1.995764408892994e-05, "loss": 1.478, "step": 68 }, { "epoch": 0.059431524547803614, "grad_norm": 1.1678736209869385, "learning_rate": 1.995504012410363e-05, "loss": 1.4755, "step": 69 }, { "epoch": 0.06029285099052541, "grad_norm": 1.574155569076538, "learning_rate": 1.995235866570217e-05, "loss": 1.5191, "step": 70 }, { "epoch": 0.0611541774332472, "grad_norm": 1.2211532592773438, "learning_rate": 1.9949599734598993e-05, "loss": 1.4561, "step": 71 }, { "epoch": 0.06201550387596899, "grad_norm": 1.313081979751587, "learning_rate": 1.994676335227059e-05, "loss": 1.502, "step": 72 }, { "epoch": 0.06287683031869079, "grad_norm": 1.2082189321517944, "learning_rate": 1.9943849540796375e-05, "loss": 1.4598, "step": 73 }, { "epoch": 0.06373815676141258, "grad_norm": 1.2929805517196655, "learning_rate": 1.9940858322858493e-05, "loss": 1.4993, "step": 74 }, { "epoch": 0.06459948320413436, "grad_norm": 1.199054479598999, "learning_rate": 1.9937789721741654e-05, "loss": 1.4904, "step": 75 }, { "epoch": 0.06546080964685616, "grad_norm": 1.350379467010498, "learning_rate": 1.9934643761332933e-05, "loss": 1.4534, "step": 76 }, { "epoch": 0.06632213608957795, "grad_norm": 1.1363669633865356, "learning_rate": 1.9931420466121613e-05, "loss": 1.453, "step": 77 }, { "epoch": 0.06718346253229975, "grad_norm": 1.1514127254486084, "learning_rate": 1.9928119861198962e-05, "loss": 1.4607, "step": 78 }, { "epoch": 0.06804478897502153, "grad_norm": 1.0154885053634644, "learning_rate": 1.9924741972258076e-05, "loss": 1.4401, "step": 79 }, { "epoch": 0.06890611541774333, "grad_norm": 1.055841326713562, "learning_rate": 1.9921286825593632e-05, "loss": 1.48, "step": 80 }, { "epoch": 0.06976744186046512, "grad_norm": 1.3121042251586914, "learning_rate": 1.9917754448101725e-05, "loss": 1.4263, "step": 81 }, { "epoch": 0.07062876830318691, "grad_norm": 1.102817416191101, "learning_rate": 1.9914144867279644e-05, "loss": 1.4401, "step": 82 }, { "epoch": 0.0714900947459087, "grad_norm": 1.0836132764816284, "learning_rate": 1.9910458111225645e-05, "loss": 1.4866, "step": 83 }, { "epoch": 0.07235142118863049, "grad_norm": 1.0442920923233032, "learning_rate": 1.990669420863875e-05, "loss": 1.4162, "step": 84 }, { "epoch": 0.07321274763135228, "grad_norm": 1.0844299793243408, "learning_rate": 1.9902853188818518e-05, "loss": 1.4707, "step": 85 }, { "epoch": 0.07407407407407407, "grad_norm": 0.993617057800293, "learning_rate": 1.9898935081664814e-05, "loss": 1.4567, "step": 86 }, { "epoch": 0.07493540051679587, "grad_norm": 1.103285312652588, "learning_rate": 1.9894939917677577e-05, "loss": 1.4249, "step": 87 }, { "epoch": 0.07579672695951765, "grad_norm": 1.0927506685256958, "learning_rate": 1.9890867727956587e-05, "loss": 1.4294, "step": 88 }, { "epoch": 0.07665805340223945, "grad_norm": 1.0682389736175537, "learning_rate": 1.988671854420122e-05, "loss": 1.3915, "step": 89 }, { "epoch": 0.07751937984496124, "grad_norm": 1.0784050226211548, "learning_rate": 1.9882492398710192e-05, "loss": 1.4445, "step": 90 }, { "epoch": 0.07838070628768304, "grad_norm": 1.0482838153839111, "learning_rate": 1.987818932438133e-05, "loss": 1.3939, "step": 91 }, { "epoch": 0.07924203273040482, "grad_norm": 1.1594949960708618, "learning_rate": 1.987380935471129e-05, "loss": 1.4543, "step": 92 }, { "epoch": 0.08010335917312661, "grad_norm": 1.0159746408462524, "learning_rate": 1.986935252379532e-05, "loss": 1.4048, "step": 93 }, { "epoch": 0.08096468561584841, "grad_norm": 1.1461023092269897, "learning_rate": 1.9864818866326978e-05, "loss": 1.4314, "step": 94 }, { "epoch": 0.0818260120585702, "grad_norm": 1.1117491722106934, "learning_rate": 1.9860208417597863e-05, "loss": 1.4383, "step": 95 }, { "epoch": 0.082687338501292, "grad_norm": 0.9831107258796692, "learning_rate": 1.9855521213497355e-05, "loss": 1.4245, "step": 96 }, { "epoch": 0.08354866494401378, "grad_norm": 1.0372180938720703, "learning_rate": 1.9850757290512313e-05, "loss": 1.3838, "step": 97 }, { "epoch": 0.08440999138673558, "grad_norm": 0.9056967496871948, "learning_rate": 1.9845916685726808e-05, "loss": 1.4076, "step": 98 }, { "epoch": 0.08527131782945736, "grad_norm": 1.002386450767517, "learning_rate": 1.9840999436821836e-05, "loss": 1.4301, "step": 99 }, { "epoch": 0.08613264427217916, "grad_norm": 0.9989022612571716, "learning_rate": 1.983600558207501e-05, "loss": 1.3993, "step": 100 }, { "epoch": 0.08699397071490095, "grad_norm": 1.0508233308792114, "learning_rate": 1.983093516036027e-05, "loss": 1.4241, "step": 101 }, { "epoch": 0.08785529715762273, "grad_norm": 0.9549940824508667, "learning_rate": 1.9825788211147587e-05, "loss": 1.427, "step": 102 }, { "epoch": 0.08871662360034453, "grad_norm": 1.0347970724105835, "learning_rate": 1.9820564774502644e-05, "loss": 1.4029, "step": 103 }, { "epoch": 0.08957795004306632, "grad_norm": 1.0045099258422852, "learning_rate": 1.981526489108653e-05, "loss": 1.4146, "step": 104 }, { "epoch": 0.09043927648578812, "grad_norm": 1.2777268886566162, "learning_rate": 1.980988860215542e-05, "loss": 1.3599, "step": 105 }, { "epoch": 0.0913006029285099, "grad_norm": 0.9002528190612793, "learning_rate": 1.980443594956027e-05, "loss": 1.4008, "step": 106 }, { "epoch": 0.0921619293712317, "grad_norm": 0.9888056516647339, "learning_rate": 1.9798906975746462e-05, "loss": 1.422, "step": 107 }, { "epoch": 0.09302325581395349, "grad_norm": 0.9847419857978821, "learning_rate": 1.9793301723753494e-05, "loss": 1.3866, "step": 108 }, { "epoch": 0.09388458225667529, "grad_norm": 0.9763849377632141, "learning_rate": 1.9787620237214648e-05, "loss": 1.3883, "step": 109 }, { "epoch": 0.09474590869939707, "grad_norm": 1.1006907224655151, "learning_rate": 1.9781862560356632e-05, "loss": 1.3748, "step": 110 }, { "epoch": 0.09560723514211886, "grad_norm": 1.0140228271484375, "learning_rate": 1.9776028737999256e-05, "loss": 1.3581, "step": 111 }, { "epoch": 0.09646856158484066, "grad_norm": 1.0311439037322998, "learning_rate": 1.9770118815555063e-05, "loss": 1.4253, "step": 112 }, { "epoch": 0.09732988802756244, "grad_norm": 0.9372150301933289, "learning_rate": 1.9764132839029e-05, "loss": 1.4078, "step": 113 }, { "epoch": 0.09819121447028424, "grad_norm": 1.0055251121520996, "learning_rate": 1.9758070855018033e-05, "loss": 1.432, "step": 114 }, { "epoch": 0.09905254091300603, "grad_norm": 0.9005001783370972, "learning_rate": 1.9751932910710808e-05, "loss": 1.3998, "step": 115 }, { "epoch": 0.09991386735572783, "grad_norm": 0.958956778049469, "learning_rate": 1.9745719053887265e-05, "loss": 1.3867, "step": 116 }, { "epoch": 0.10077519379844961, "grad_norm": 1.2123082876205444, "learning_rate": 1.9739429332918276e-05, "loss": 1.382, "step": 117 }, { "epoch": 0.10163652024117141, "grad_norm": 0.858340859413147, "learning_rate": 1.9733063796765267e-05, "loss": 1.3794, "step": 118 }, { "epoch": 0.1024978466838932, "grad_norm": 0.8608000874519348, "learning_rate": 1.972662249497984e-05, "loss": 1.3996, "step": 119 }, { "epoch": 0.10335917312661498, "grad_norm": 0.8888645768165588, "learning_rate": 1.972010547770338e-05, "loss": 1.3874, "step": 120 }, { "epoch": 0.10422049956933678, "grad_norm": 0.9219585061073303, "learning_rate": 1.9713512795666663e-05, "loss": 1.3881, "step": 121 }, { "epoch": 0.10508182601205857, "grad_norm": 0.8192768096923828, "learning_rate": 1.970684450018948e-05, "loss": 1.3827, "step": 122 }, { "epoch": 0.10594315245478036, "grad_norm": 1.0063087940216064, "learning_rate": 1.9700100643180213e-05, "loss": 1.3679, "step": 123 }, { "epoch": 0.10680447889750215, "grad_norm": 0.9152660369873047, "learning_rate": 1.969328127713544e-05, "loss": 1.3637, "step": 124 }, { "epoch": 0.10766580534022395, "grad_norm": 0.9176394939422607, "learning_rate": 1.9686386455139544e-05, "loss": 1.3972, "step": 125 }, { "epoch": 0.10852713178294573, "grad_norm": 0.91953444480896, "learning_rate": 1.9679416230864265e-05, "loss": 1.3833, "step": 126 }, { "epoch": 0.10938845822566753, "grad_norm": 0.967275083065033, "learning_rate": 1.9672370658568306e-05, "loss": 1.3628, "step": 127 }, { "epoch": 0.11024978466838932, "grad_norm": 0.8972539305686951, "learning_rate": 1.966524979309692e-05, "loss": 1.3457, "step": 128 }, { "epoch": 0.1111111111111111, "grad_norm": 0.926163911819458, "learning_rate": 1.9658053689881453e-05, "loss": 1.3278, "step": 129 }, { "epoch": 0.1119724375538329, "grad_norm": 0.9265878200531006, "learning_rate": 1.9650782404938933e-05, "loss": 1.347, "step": 130 }, { "epoch": 0.11283376399655469, "grad_norm": 1.088254451751709, "learning_rate": 1.9643435994871626e-05, "loss": 1.3498, "step": 131 }, { "epoch": 0.11369509043927649, "grad_norm": 0.9928959608078003, "learning_rate": 1.963601451686661e-05, "loss": 1.372, "step": 132 }, { "epoch": 0.11455641688199827, "grad_norm": 0.8130192756652832, "learning_rate": 1.9628518028695307e-05, "loss": 1.3185, "step": 133 }, { "epoch": 0.11541774332472007, "grad_norm": 0.9813799262046814, "learning_rate": 1.9620946588713048e-05, "loss": 1.3723, "step": 134 }, { "epoch": 0.11627906976744186, "grad_norm": 0.8440076112747192, "learning_rate": 1.9613300255858615e-05, "loss": 1.379, "step": 135 }, { "epoch": 0.11714039621016366, "grad_norm": 0.9899678826332092, "learning_rate": 1.960557908965379e-05, "loss": 1.3396, "step": 136 }, { "epoch": 0.11800172265288544, "grad_norm": 0.9470325708389282, "learning_rate": 1.9597783150202873e-05, "loss": 1.3631, "step": 137 }, { "epoch": 0.11886304909560723, "grad_norm": 0.9099454879760742, "learning_rate": 1.9589912498192233e-05, "loss": 1.3623, "step": 138 }, { "epoch": 0.11972437553832903, "grad_norm": 1.1269025802612305, "learning_rate": 1.9581967194889826e-05, "loss": 1.3615, "step": 139 }, { "epoch": 0.12058570198105081, "grad_norm": 0.8688855767250061, "learning_rate": 1.957394730214472e-05, "loss": 1.36, "step": 140 }, { "epoch": 0.12144702842377261, "grad_norm": 0.9251095652580261, "learning_rate": 1.956585288238662e-05, "loss": 1.3274, "step": 141 }, { "epoch": 0.1223083548664944, "grad_norm": 1.2113760709762573, "learning_rate": 1.955768399862536e-05, "loss": 1.3687, "step": 142 }, { "epoch": 0.1231696813092162, "grad_norm": 1.0069940090179443, "learning_rate": 1.9549440714450447e-05, "loss": 1.3602, "step": 143 }, { "epoch": 0.12403100775193798, "grad_norm": 0.987554132938385, "learning_rate": 1.9541123094030528e-05, "loss": 1.3515, "step": 144 }, { "epoch": 0.12489233419465978, "grad_norm": 0.8786775469779968, "learning_rate": 1.9532731202112935e-05, "loss": 1.3849, "step": 145 }, { "epoch": 0.12575366063738158, "grad_norm": 1.0503038167953491, "learning_rate": 1.9524265104023133e-05, "loss": 1.3638, "step": 146 }, { "epoch": 0.12661498708010335, "grad_norm": 0.9434555768966675, "learning_rate": 1.9515724865664242e-05, "loss": 1.3065, "step": 147 }, { "epoch": 0.12747631352282515, "grad_norm": 1.0372234582901, "learning_rate": 1.9507110553516518e-05, "loss": 1.3398, "step": 148 }, { "epoch": 0.12833763996554695, "grad_norm": 0.9711533188819885, "learning_rate": 1.949842223463683e-05, "loss": 1.3428, "step": 149 }, { "epoch": 0.12919896640826872, "grad_norm": 0.8844906091690063, "learning_rate": 1.9489659976658152e-05, "loss": 1.3685, "step": 150 }, { "epoch": 0.13006029285099052, "grad_norm": 0.9219164848327637, "learning_rate": 1.9480823847789007e-05, "loss": 1.3465, "step": 151 }, { "epoch": 0.13092161929371232, "grad_norm": 0.9082310795783997, "learning_rate": 1.947191391681298e-05, "loss": 1.3799, "step": 152 }, { "epoch": 0.13178294573643412, "grad_norm": 0.9203616380691528, "learning_rate": 1.946293025308813e-05, "loss": 1.3636, "step": 153 }, { "epoch": 0.1326442721791559, "grad_norm": 0.8217676877975464, "learning_rate": 1.9453872926546505e-05, "loss": 1.3247, "step": 154 }, { "epoch": 0.1335055986218777, "grad_norm": 0.8964298963546753, "learning_rate": 1.944474200769355e-05, "loss": 1.3357, "step": 155 }, { "epoch": 0.1343669250645995, "grad_norm": 0.9523929357528687, "learning_rate": 1.943553756760759e-05, "loss": 1.3712, "step": 156 }, { "epoch": 0.13522825150732126, "grad_norm": 0.8506253361701965, "learning_rate": 1.9426259677939264e-05, "loss": 1.3522, "step": 157 }, { "epoch": 0.13608957795004306, "grad_norm": 0.9101868867874146, "learning_rate": 1.9416908410910965e-05, "loss": 1.3266, "step": 158 }, { "epoch": 0.13695090439276486, "grad_norm": 0.8775342702865601, "learning_rate": 1.9407483839316284e-05, "loss": 1.3575, "step": 159 }, { "epoch": 0.13781223083548666, "grad_norm": 0.9115650057792664, "learning_rate": 1.939798603651944e-05, "loss": 1.3578, "step": 160 }, { "epoch": 0.13867355727820843, "grad_norm": 0.8985600471496582, "learning_rate": 1.938841507645471e-05, "loss": 1.3059, "step": 161 }, { "epoch": 0.13953488372093023, "grad_norm": 1.053168773651123, "learning_rate": 1.9378771033625855e-05, "loss": 1.3736, "step": 162 }, { "epoch": 0.14039621016365203, "grad_norm": 0.9247081279754639, "learning_rate": 1.9369053983105533e-05, "loss": 1.3268, "step": 163 }, { "epoch": 0.14125753660637383, "grad_norm": 0.937142014503479, "learning_rate": 1.9359264000534726e-05, "loss": 1.3352, "step": 164 }, { "epoch": 0.1421188630490956, "grad_norm": 0.8469811081886292, "learning_rate": 1.934940116212214e-05, "loss": 1.3277, "step": 165 }, { "epoch": 0.1429801894918174, "grad_norm": 0.8988816142082214, "learning_rate": 1.9339465544643623e-05, "loss": 1.3447, "step": 166 }, { "epoch": 0.1438415159345392, "grad_norm": 1.033051609992981, "learning_rate": 1.9329457225441554e-05, "loss": 1.3072, "step": 167 }, { "epoch": 0.14470284237726097, "grad_norm": 0.9636563658714294, "learning_rate": 1.9319376282424255e-05, "loss": 1.3027, "step": 168 }, { "epoch": 0.14556416881998277, "grad_norm": 0.9479746222496033, "learning_rate": 1.9309222794065373e-05, "loss": 1.3319, "step": 169 }, { "epoch": 0.14642549526270457, "grad_norm": 0.9528256058692932, "learning_rate": 1.929899683940327e-05, "loss": 1.3737, "step": 170 }, { "epoch": 0.14728682170542637, "grad_norm": 0.87147057056427, "learning_rate": 1.9288698498040423e-05, "loss": 1.3361, "step": 171 }, { "epoch": 0.14814814814814814, "grad_norm": 0.9613904356956482, "learning_rate": 1.9278327850142783e-05, "loss": 1.3485, "step": 172 }, { "epoch": 0.14900947459086994, "grad_norm": 1.0545318126678467, "learning_rate": 1.9267884976439163e-05, "loss": 1.333, "step": 173 }, { "epoch": 0.14987080103359174, "grad_norm": 0.9187701940536499, "learning_rate": 1.9257369958220612e-05, "loss": 1.3241, "step": 174 }, { "epoch": 0.1507321274763135, "grad_norm": 0.866005539894104, "learning_rate": 1.9246782877339767e-05, "loss": 1.3133, "step": 175 }, { "epoch": 0.1515934539190353, "grad_norm": 0.9420531392097473, "learning_rate": 1.923612381621024e-05, "loss": 1.2784, "step": 176 }, { "epoch": 0.1524547803617571, "grad_norm": 0.9032683968544006, "learning_rate": 1.9225392857805955e-05, "loss": 1.2992, "step": 177 }, { "epoch": 0.1533161068044789, "grad_norm": 0.9758039116859436, "learning_rate": 1.921459008566051e-05, "loss": 1.3013, "step": 178 }, { "epoch": 0.15417743324720068, "grad_norm": 0.9804133176803589, "learning_rate": 1.9203715583866527e-05, "loss": 1.2893, "step": 179 }, { "epoch": 0.15503875968992248, "grad_norm": 0.9305117130279541, "learning_rate": 1.9192769437075e-05, "loss": 1.3307, "step": 180 }, { "epoch": 0.15590008613264428, "grad_norm": 0.9009129405021667, "learning_rate": 1.918175173049463e-05, "loss": 1.3344, "step": 181 }, { "epoch": 0.15676141257536608, "grad_norm": 0.9379671812057495, "learning_rate": 1.9170662549891162e-05, "loss": 1.3434, "step": 182 }, { "epoch": 0.15762273901808785, "grad_norm": 0.8975892066955566, "learning_rate": 1.9159501981586738e-05, "loss": 1.305, "step": 183 }, { "epoch": 0.15848406546080965, "grad_norm": 0.8844696283340454, "learning_rate": 1.9148270112459178e-05, "loss": 1.3373, "step": 184 }, { "epoch": 0.15934539190353145, "grad_norm": 0.9148349761962891, "learning_rate": 1.9136967029941354e-05, "loss": 1.322, "step": 185 }, { "epoch": 0.16020671834625322, "grad_norm": 0.9962027668952942, "learning_rate": 1.9125592822020485e-05, "loss": 1.3065, "step": 186 }, { "epoch": 0.16106804478897502, "grad_norm": 0.9287891983985901, "learning_rate": 1.9114147577237452e-05, "loss": 1.3102, "step": 187 }, { "epoch": 0.16192937123169682, "grad_norm": 0.9015977382659912, "learning_rate": 1.9102631384686116e-05, "loss": 1.3011, "step": 188 }, { "epoch": 0.16279069767441862, "grad_norm": 1.0545498132705688, "learning_rate": 1.909104433401261e-05, "loss": 1.2916, "step": 189 }, { "epoch": 0.1636520241171404, "grad_norm": 0.8966095447540283, "learning_rate": 1.9079386515414667e-05, "loss": 1.2899, "step": 190 }, { "epoch": 0.1645133505598622, "grad_norm": 0.8880913853645325, "learning_rate": 1.9067658019640897e-05, "loss": 1.3187, "step": 191 }, { "epoch": 0.165374677002584, "grad_norm": 1.007127285003662, "learning_rate": 1.9055858937990083e-05, "loss": 1.2909, "step": 192 }, { "epoch": 0.16623600344530576, "grad_norm": 0.8095739483833313, "learning_rate": 1.9043989362310472e-05, "loss": 1.2956, "step": 193 }, { "epoch": 0.16709732988802756, "grad_norm": 0.8505532145500183, "learning_rate": 1.903204938499907e-05, "loss": 1.344, "step": 194 }, { "epoch": 0.16795865633074936, "grad_norm": 0.8656225204467773, "learning_rate": 1.902003909900091e-05, "loss": 1.3279, "step": 195 }, { "epoch": 0.16881998277347116, "grad_norm": 0.8762969374656677, "learning_rate": 1.9007958597808326e-05, "loss": 1.313, "step": 196 }, { "epoch": 0.16968130921619293, "grad_norm": 0.8867192268371582, "learning_rate": 1.8995807975460246e-05, "loss": 1.2943, "step": 197 }, { "epoch": 0.17054263565891473, "grad_norm": 0.9131550192832947, "learning_rate": 1.8983587326541437e-05, "loss": 1.302, "step": 198 }, { "epoch": 0.17140396210163653, "grad_norm": 0.9059438705444336, "learning_rate": 1.8971296746181774e-05, "loss": 1.2927, "step": 199 }, { "epoch": 0.17226528854435832, "grad_norm": 0.912719190120697, "learning_rate": 1.8958936330055516e-05, "loss": 1.2986, "step": 200 }, { "epoch": 0.1731266149870801, "grad_norm": 0.9960724711418152, "learning_rate": 1.894650617438054e-05, "loss": 1.2989, "step": 201 }, { "epoch": 0.1739879414298019, "grad_norm": 0.9200662970542908, "learning_rate": 1.893400637591759e-05, "loss": 1.3312, "step": 202 }, { "epoch": 0.1748492678725237, "grad_norm": 1.040659785270691, "learning_rate": 1.8921437031969557e-05, "loss": 1.3031, "step": 203 }, { "epoch": 0.17571059431524547, "grad_norm": 0.8604618310928345, "learning_rate": 1.8908798240380692e-05, "loss": 1.341, "step": 204 }, { "epoch": 0.17657192075796727, "grad_norm": 0.982661247253418, "learning_rate": 1.8896090099535834e-05, "loss": 1.305, "step": 205 }, { "epoch": 0.17743324720068906, "grad_norm": 0.8687034249305725, "learning_rate": 1.888331270835968e-05, "loss": 1.3114, "step": 206 }, { "epoch": 0.17829457364341086, "grad_norm": 0.9046248197555542, "learning_rate": 1.8870466166315992e-05, "loss": 1.3212, "step": 207 }, { "epoch": 0.17915590008613264, "grad_norm": 0.8828570246696472, "learning_rate": 1.885755057340682e-05, "loss": 1.2929, "step": 208 }, { "epoch": 0.18001722652885443, "grad_norm": 0.8848705291748047, "learning_rate": 1.8844566030171737e-05, "loss": 1.3057, "step": 209 }, { "epoch": 0.18087855297157623, "grad_norm": 1.004079818725586, "learning_rate": 1.8831512637687054e-05, "loss": 1.2909, "step": 210 }, { "epoch": 0.181739879414298, "grad_norm": 0.9026859998703003, "learning_rate": 1.881839049756502e-05, "loss": 1.3081, "step": 211 }, { "epoch": 0.1826012058570198, "grad_norm": 1.0891921520233154, "learning_rate": 1.880519971195304e-05, "loss": 1.3273, "step": 212 }, { "epoch": 0.1834625322997416, "grad_norm": 0.9379571676254272, "learning_rate": 1.879194038353289e-05, "loss": 1.3223, "step": 213 }, { "epoch": 0.1843238587424634, "grad_norm": 0.9415060877799988, "learning_rate": 1.87786126155199e-05, "loss": 1.3464, "step": 214 }, { "epoch": 0.18518518518518517, "grad_norm": 0.9501126408576965, "learning_rate": 1.8765216511662153e-05, "loss": 1.3008, "step": 215 }, { "epoch": 0.18604651162790697, "grad_norm": 0.9133473634719849, "learning_rate": 1.8751752176239693e-05, "loss": 1.2966, "step": 216 }, { "epoch": 0.18690783807062877, "grad_norm": 0.9399188160896301, "learning_rate": 1.87382197140637e-05, "loss": 1.3193, "step": 217 }, { "epoch": 0.18776916451335057, "grad_norm": 0.9107801914215088, "learning_rate": 1.8724619230475675e-05, "loss": 1.3024, "step": 218 }, { "epoch": 0.18863049095607234, "grad_norm": 0.894831120967865, "learning_rate": 1.8710950831346623e-05, "loss": 1.3163, "step": 219 }, { "epoch": 0.18949181739879414, "grad_norm": 1.0155446529388428, "learning_rate": 1.8697214623076222e-05, "loss": 1.2575, "step": 220 }, { "epoch": 0.19035314384151594, "grad_norm": 0.9866533279418945, "learning_rate": 1.8683410712592015e-05, "loss": 1.3209, "step": 221 }, { "epoch": 0.19121447028423771, "grad_norm": 0.8821126222610474, "learning_rate": 1.8669539207348544e-05, "loss": 1.2828, "step": 222 }, { "epoch": 0.1920757967269595, "grad_norm": 0.8987988233566284, "learning_rate": 1.8655600215326547e-05, "loss": 1.2898, "step": 223 }, { "epoch": 0.1929371231696813, "grad_norm": 0.9085987210273743, "learning_rate": 1.8641593845032098e-05, "loss": 1.3065, "step": 224 }, { "epoch": 0.1937984496124031, "grad_norm": 0.9193576574325562, "learning_rate": 1.8627520205495772e-05, "loss": 1.2837, "step": 225 }, { "epoch": 0.19465977605512488, "grad_norm": 0.9095317721366882, "learning_rate": 1.8613379406271784e-05, "loss": 1.3171, "step": 226 }, { "epoch": 0.19552110249784668, "grad_norm": 0.9969485402107239, "learning_rate": 1.8599171557437147e-05, "loss": 1.2819, "step": 227 }, { "epoch": 0.19638242894056848, "grad_norm": 0.9653918743133545, "learning_rate": 1.858489676959081e-05, "loss": 1.2601, "step": 228 }, { "epoch": 0.19724375538329025, "grad_norm": 0.977749764919281, "learning_rate": 1.8570555153852806e-05, "loss": 1.2861, "step": 229 }, { "epoch": 0.19810508182601205, "grad_norm": 0.9468740820884705, "learning_rate": 1.855614682186338e-05, "loss": 1.2869, "step": 230 }, { "epoch": 0.19896640826873385, "grad_norm": 1.1150413751602173, "learning_rate": 1.8541671885782106e-05, "loss": 1.2798, "step": 231 }, { "epoch": 0.19982773471145565, "grad_norm": 0.9200144410133362, "learning_rate": 1.8527130458287047e-05, "loss": 1.2855, "step": 232 }, { "epoch": 0.20068906115417742, "grad_norm": 0.937813401222229, "learning_rate": 1.851252265257384e-05, "loss": 1.2598, "step": 233 }, { "epoch": 0.20155038759689922, "grad_norm": 1.02151620388031, "learning_rate": 1.8497848582354852e-05, "loss": 1.2828, "step": 234 }, { "epoch": 0.20241171403962102, "grad_norm": 0.9327558279037476, "learning_rate": 1.8483108361858263e-05, "loss": 1.2898, "step": 235 }, { "epoch": 0.20327304048234282, "grad_norm": 1.0296434164047241, "learning_rate": 1.8468302105827195e-05, "loss": 1.2869, "step": 236 }, { "epoch": 0.2041343669250646, "grad_norm": 0.9533336758613586, "learning_rate": 1.845342992951882e-05, "loss": 1.2768, "step": 237 }, { "epoch": 0.2049956933677864, "grad_norm": 0.9365238547325134, "learning_rate": 1.8438491948703445e-05, "loss": 1.2783, "step": 238 }, { "epoch": 0.2058570198105082, "grad_norm": 1.104421854019165, "learning_rate": 1.842348827966363e-05, "loss": 1.293, "step": 239 }, { "epoch": 0.20671834625322996, "grad_norm": 0.8812316060066223, "learning_rate": 1.840841903919328e-05, "loss": 1.2856, "step": 240 }, { "epoch": 0.20757967269595176, "grad_norm": 0.9076652526855469, "learning_rate": 1.8393284344596715e-05, "loss": 1.3453, "step": 241 }, { "epoch": 0.20844099913867356, "grad_norm": 0.9731348156929016, "learning_rate": 1.837808431368779e-05, "loss": 1.3084, "step": 242 }, { "epoch": 0.20930232558139536, "grad_norm": 0.8820015788078308, "learning_rate": 1.8362819064788956e-05, "loss": 1.2753, "step": 243 }, { "epoch": 0.21016365202411713, "grad_norm": 1.0030732154846191, "learning_rate": 1.8347488716730343e-05, "loss": 1.302, "step": 244 }, { "epoch": 0.21102497846683893, "grad_norm": 0.9901053309440613, "learning_rate": 1.8332093388848836e-05, "loss": 1.31, "step": 245 }, { "epoch": 0.21188630490956073, "grad_norm": 0.81141597032547, "learning_rate": 1.8316633200987143e-05, "loss": 1.2973, "step": 246 }, { "epoch": 0.2127476313522825, "grad_norm": 0.9569693803787231, "learning_rate": 1.830110827349287e-05, "loss": 1.2805, "step": 247 }, { "epoch": 0.2136089577950043, "grad_norm": 0.9289495348930359, "learning_rate": 1.8285518727217578e-05, "loss": 1.2735, "step": 248 }, { "epoch": 0.2144702842377261, "grad_norm": 0.8315669894218445, "learning_rate": 1.8269864683515847e-05, "loss": 1.2775, "step": 249 }, { "epoch": 0.2153316106804479, "grad_norm": 0.9181628823280334, "learning_rate": 1.8254146264244316e-05, "loss": 1.2328, "step": 250 }, { "epoch": 0.21619293712316967, "grad_norm": 0.8977162837982178, "learning_rate": 1.8238363591760758e-05, "loss": 1.2346, "step": 251 }, { "epoch": 0.21705426356589147, "grad_norm": 0.9173194169998169, "learning_rate": 1.822251678892312e-05, "loss": 1.2653, "step": 252 }, { "epoch": 0.21791559000861327, "grad_norm": 0.9074868559837341, "learning_rate": 1.8206605979088545e-05, "loss": 1.2514, "step": 253 }, { "epoch": 0.21877691645133507, "grad_norm": 0.8458305597305298, "learning_rate": 1.819063128611244e-05, "loss": 1.2885, "step": 254 }, { "epoch": 0.21963824289405684, "grad_norm": 0.914714515209198, "learning_rate": 1.8174592834347503e-05, "loss": 1.2431, "step": 255 }, { "epoch": 0.22049956933677864, "grad_norm": 0.8686874508857727, "learning_rate": 1.815849074864275e-05, "loss": 1.2554, "step": 256 }, { "epoch": 0.22136089577950044, "grad_norm": 0.8784075379371643, "learning_rate": 1.814232515434255e-05, "loss": 1.3005, "step": 257 }, { "epoch": 0.2222222222222222, "grad_norm": 0.919947624206543, "learning_rate": 1.8126096177285637e-05, "loss": 1.2968, "step": 258 }, { "epoch": 0.223083548664944, "grad_norm": 0.9774091839790344, "learning_rate": 1.8109803943804146e-05, "loss": 1.2763, "step": 259 }, { "epoch": 0.2239448751076658, "grad_norm": 0.8449957370758057, "learning_rate": 1.8093448580722617e-05, "loss": 1.2653, "step": 260 }, { "epoch": 0.2248062015503876, "grad_norm": 0.9756072163581848, "learning_rate": 1.8077030215357024e-05, "loss": 1.272, "step": 261 }, { "epoch": 0.22566752799310938, "grad_norm": 0.8839893341064453, "learning_rate": 1.806054897551376e-05, "loss": 1.2743, "step": 262 }, { "epoch": 0.22652885443583118, "grad_norm": 0.8637559413909912, "learning_rate": 1.8044004989488662e-05, "loss": 1.2862, "step": 263 }, { "epoch": 0.22739018087855298, "grad_norm": 0.9832426309585571, "learning_rate": 1.802739838606601e-05, "loss": 1.2886, "step": 264 }, { "epoch": 0.22825150732127478, "grad_norm": 0.9957407116889954, "learning_rate": 1.801072929451751e-05, "loss": 1.2629, "step": 265 }, { "epoch": 0.22911283376399655, "grad_norm": 0.9450539946556091, "learning_rate": 1.7993997844601305e-05, "loss": 1.2606, "step": 266 }, { "epoch": 0.22997416020671835, "grad_norm": 0.9669734239578247, "learning_rate": 1.7977204166560954e-05, "loss": 1.2614, "step": 267 }, { "epoch": 0.23083548664944015, "grad_norm": 0.9341006278991699, "learning_rate": 1.7960348391124422e-05, "loss": 1.2826, "step": 268 }, { "epoch": 0.23169681309216192, "grad_norm": 0.8190209865570068, "learning_rate": 1.7943430649503065e-05, "loss": 1.2575, "step": 269 }, { "epoch": 0.23255813953488372, "grad_norm": 0.9061177968978882, "learning_rate": 1.7926451073390612e-05, "loss": 1.2458, "step": 270 }, { "epoch": 0.23341946597760552, "grad_norm": 0.8741104602813721, "learning_rate": 1.7909409794962115e-05, "loss": 1.2805, "step": 271 }, { "epoch": 0.23428079242032732, "grad_norm": 0.8899810910224915, "learning_rate": 1.7892306946872952e-05, "loss": 1.2581, "step": 272 }, { "epoch": 0.2351421188630491, "grad_norm": 0.9245790243148804, "learning_rate": 1.7875142662257788e-05, "loss": 1.2877, "step": 273 }, { "epoch": 0.2360034453057709, "grad_norm": 0.8939240574836731, "learning_rate": 1.7857917074729513e-05, "loss": 1.2461, "step": 274 }, { "epoch": 0.2368647717484927, "grad_norm": 0.8694809079170227, "learning_rate": 1.7840630318378233e-05, "loss": 1.2458, "step": 275 }, { "epoch": 0.23772609819121446, "grad_norm": 0.8689870238304138, "learning_rate": 1.7823282527770214e-05, "loss": 1.2821, "step": 276 }, { "epoch": 0.23858742463393626, "grad_norm": 0.8938902616500854, "learning_rate": 1.7805873837946833e-05, "loss": 1.2691, "step": 277 }, { "epoch": 0.23944875107665806, "grad_norm": 1.0327783823013306, "learning_rate": 1.778840438442352e-05, "loss": 1.2732, "step": 278 }, { "epoch": 0.24031007751937986, "grad_norm": 0.9891493916511536, "learning_rate": 1.7770874303188727e-05, "loss": 1.253, "step": 279 }, { "epoch": 0.24117140396210163, "grad_norm": 0.9043530821800232, "learning_rate": 1.7753283730702837e-05, "loss": 1.2731, "step": 280 }, { "epoch": 0.24203273040482343, "grad_norm": 0.9076008796691895, "learning_rate": 1.7735632803897135e-05, "loss": 1.2684, "step": 281 }, { "epoch": 0.24289405684754523, "grad_norm": 0.8306509852409363, "learning_rate": 1.7717921660172708e-05, "loss": 1.2577, "step": 282 }, { "epoch": 0.24375538329026702, "grad_norm": 0.8941156268119812, "learning_rate": 1.7700150437399405e-05, "loss": 1.2682, "step": 283 }, { "epoch": 0.2446167097329888, "grad_norm": 0.8855567574501038, "learning_rate": 1.7682319273914755e-05, "loss": 1.2904, "step": 284 }, { "epoch": 0.2454780361757106, "grad_norm": 1.0233161449432373, "learning_rate": 1.766442830852287e-05, "loss": 1.2907, "step": 285 }, { "epoch": 0.2463393626184324, "grad_norm": 1.106757402420044, "learning_rate": 1.76464776804934e-05, "loss": 1.245, "step": 286 }, { "epoch": 0.24720068906115417, "grad_norm": 0.92214435338974, "learning_rate": 1.7628467529560417e-05, "loss": 1.254, "step": 287 }, { "epoch": 0.24806201550387597, "grad_norm": 0.8845908045768738, "learning_rate": 1.7610397995921348e-05, "loss": 1.314, "step": 288 }, { "epoch": 0.24892334194659776, "grad_norm": 0.837216317653656, "learning_rate": 1.759226922023587e-05, "loss": 1.2943, "step": 289 }, { "epoch": 0.24978466838931956, "grad_norm": 0.9128316044807434, "learning_rate": 1.7574081343624827e-05, "loss": 1.2881, "step": 290 }, { "epoch": 0.25064599483204136, "grad_norm": 0.9482645392417908, "learning_rate": 1.7555834507669124e-05, "loss": 1.2846, "step": 291 }, { "epoch": 0.25150732127476316, "grad_norm": 0.9201834797859192, "learning_rate": 1.7537528854408625e-05, "loss": 1.2729, "step": 292 }, { "epoch": 0.2523686477174849, "grad_norm": 0.8601769804954529, "learning_rate": 1.751916452634105e-05, "loss": 1.2601, "step": 293 }, { "epoch": 0.2532299741602067, "grad_norm": 0.8500188589096069, "learning_rate": 1.7500741666420863e-05, "loss": 1.2627, "step": 294 }, { "epoch": 0.2540913006029285, "grad_norm": 0.9467564821243286, "learning_rate": 1.7482260418058167e-05, "loss": 1.2572, "step": 295 }, { "epoch": 0.2549526270456503, "grad_norm": 0.9599546790122986, "learning_rate": 1.7463720925117565e-05, "loss": 1.2445, "step": 296 }, { "epoch": 0.2558139534883721, "grad_norm": 0.8565171957015991, "learning_rate": 1.744512333191708e-05, "loss": 1.2222, "step": 297 }, { "epoch": 0.2566752799310939, "grad_norm": 1.068682074546814, "learning_rate": 1.7426467783226992e-05, "loss": 1.2574, "step": 298 }, { "epoch": 0.2575366063738157, "grad_norm": 0.9868722558021545, "learning_rate": 1.7407754424268727e-05, "loss": 1.2234, "step": 299 }, { "epoch": 0.25839793281653745, "grad_norm": 0.8837171196937561, "learning_rate": 1.7388983400713736e-05, "loss": 1.2489, "step": 300 }, { "epoch": 0.25925925925925924, "grad_norm": 1.0964189767837524, "learning_rate": 1.7370154858682347e-05, "loss": 1.2658, "step": 301 }, { "epoch": 0.26012058570198104, "grad_norm": 0.9035983085632324, "learning_rate": 1.7351268944742626e-05, "loss": 1.261, "step": 302 }, { "epoch": 0.26098191214470284, "grad_norm": 1.021364688873291, "learning_rate": 1.7332325805909256e-05, "loss": 1.2542, "step": 303 }, { "epoch": 0.26184323858742464, "grad_norm": 0.8748368620872498, "learning_rate": 1.7313325589642363e-05, "loss": 1.2155, "step": 304 }, { "epoch": 0.26270456503014644, "grad_norm": 1.0000536441802979, "learning_rate": 1.7294268443846403e-05, "loss": 1.2386, "step": 305 }, { "epoch": 0.26356589147286824, "grad_norm": 0.8522447943687439, "learning_rate": 1.727515451686897e-05, "loss": 1.262, "step": 306 }, { "epoch": 0.26442721791559, "grad_norm": 1.03394615650177, "learning_rate": 1.7255983957499676e-05, "loss": 1.2607, "step": 307 }, { "epoch": 0.2652885443583118, "grad_norm": 0.9327899217605591, "learning_rate": 1.7236756914968985e-05, "loss": 1.239, "step": 308 }, { "epoch": 0.2661498708010336, "grad_norm": 0.8953905701637268, "learning_rate": 1.7217473538947032e-05, "loss": 1.2584, "step": 309 }, { "epoch": 0.2670111972437554, "grad_norm": 0.877118706703186, "learning_rate": 1.719813397954248e-05, "loss": 1.2692, "step": 310 }, { "epoch": 0.2678725236864772, "grad_norm": 0.8814844489097595, "learning_rate": 1.7178738387301342e-05, "loss": 1.258, "step": 311 }, { "epoch": 0.268733850129199, "grad_norm": 0.882537305355072, "learning_rate": 1.7159286913205813e-05, "loss": 1.2923, "step": 312 }, { "epoch": 0.2695951765719208, "grad_norm": 0.9672994613647461, "learning_rate": 1.7139779708673084e-05, "loss": 1.2725, "step": 313 }, { "epoch": 0.2704565030146425, "grad_norm": 0.8873171806335449, "learning_rate": 1.7120216925554185e-05, "loss": 1.2268, "step": 314 }, { "epoch": 0.2713178294573643, "grad_norm": 0.9454041123390198, "learning_rate": 1.7100598716132775e-05, "loss": 1.2358, "step": 315 }, { "epoch": 0.2721791559000861, "grad_norm": 0.943789005279541, "learning_rate": 1.708092523312398e-05, "loss": 1.2138, "step": 316 }, { "epoch": 0.2730404823428079, "grad_norm": 0.8467637300491333, "learning_rate": 1.7061196629673198e-05, "loss": 1.2495, "step": 317 }, { "epoch": 0.2739018087855297, "grad_norm": 0.8850976228713989, "learning_rate": 1.7041413059354893e-05, "loss": 1.2586, "step": 318 }, { "epoch": 0.2747631352282515, "grad_norm": 0.8928930759429932, "learning_rate": 1.7021574676171418e-05, "loss": 1.2279, "step": 319 }, { "epoch": 0.2756244616709733, "grad_norm": 0.9026039242744446, "learning_rate": 1.7001681634551813e-05, "loss": 1.2361, "step": 320 }, { "epoch": 0.27648578811369506, "grad_norm": 0.90481036901474, "learning_rate": 1.6981734089350585e-05, "loss": 1.2308, "step": 321 }, { "epoch": 0.27734711455641686, "grad_norm": 0.8753896951675415, "learning_rate": 1.696173219584653e-05, "loss": 1.2383, "step": 322 }, { "epoch": 0.27820844099913866, "grad_norm": 1.0231022834777832, "learning_rate": 1.6941676109741506e-05, "loss": 1.2472, "step": 323 }, { "epoch": 0.27906976744186046, "grad_norm": 0.8703423142433167, "learning_rate": 1.6921565987159226e-05, "loss": 1.2712, "step": 324 }, { "epoch": 0.27993109388458226, "grad_norm": 0.9238699078559875, "learning_rate": 1.6901401984644034e-05, "loss": 1.2279, "step": 325 }, { "epoch": 0.28079242032730406, "grad_norm": 0.9144941568374634, "learning_rate": 1.6881184259159708e-05, "loss": 1.2416, "step": 326 }, { "epoch": 0.28165374677002586, "grad_norm": 0.8758153915405273, "learning_rate": 1.686091296808822e-05, "loss": 1.2736, "step": 327 }, { "epoch": 0.28251507321274766, "grad_norm": 0.8742189407348633, "learning_rate": 1.6840588269228507e-05, "loss": 1.2876, "step": 328 }, { "epoch": 0.2833763996554694, "grad_norm": 0.9139267206192017, "learning_rate": 1.682021032079526e-05, "loss": 1.2919, "step": 329 }, { "epoch": 0.2842377260981912, "grad_norm": 0.8497806191444397, "learning_rate": 1.6799779281417685e-05, "loss": 1.2275, "step": 330 }, { "epoch": 0.285099052540913, "grad_norm": 0.9016520380973816, "learning_rate": 1.6779295310138264e-05, "loss": 1.2126, "step": 331 }, { "epoch": 0.2859603789836348, "grad_norm": 0.8737862706184387, "learning_rate": 1.6758758566411516e-05, "loss": 1.2641, "step": 332 }, { "epoch": 0.2868217054263566, "grad_norm": 0.9002192616462708, "learning_rate": 1.6738169210102765e-05, "loss": 1.2872, "step": 333 }, { "epoch": 0.2876830318690784, "grad_norm": 0.8395231366157532, "learning_rate": 1.6717527401486882e-05, "loss": 1.2808, "step": 334 }, { "epoch": 0.2885443583118002, "grad_norm": 0.9024813771247864, "learning_rate": 1.669683330124706e-05, "loss": 1.2415, "step": 335 }, { "epoch": 0.28940568475452194, "grad_norm": 0.9167109131813049, "learning_rate": 1.667608707047354e-05, "loss": 1.2375, "step": 336 }, { "epoch": 0.29026701119724374, "grad_norm": 0.8365147113800049, "learning_rate": 1.6655288870662354e-05, "loss": 1.2451, "step": 337 }, { "epoch": 0.29112833763996554, "grad_norm": 0.9034088253974915, "learning_rate": 1.6634438863714108e-05, "loss": 1.2237, "step": 338 }, { "epoch": 0.29198966408268734, "grad_norm": 0.8997372984886169, "learning_rate": 1.661353721193266e-05, "loss": 1.2358, "step": 339 }, { "epoch": 0.29285099052540914, "grad_norm": 0.9689245820045471, "learning_rate": 1.6592584078023915e-05, "loss": 1.2523, "step": 340 }, { "epoch": 0.29371231696813094, "grad_norm": 0.9412267208099365, "learning_rate": 1.657157962509452e-05, "loss": 1.2445, "step": 341 }, { "epoch": 0.29457364341085274, "grad_norm": 0.8464860916137695, "learning_rate": 1.6550524016650617e-05, "loss": 1.2102, "step": 342 }, { "epoch": 0.2954349698535745, "grad_norm": 0.8929284811019897, "learning_rate": 1.652941741659655e-05, "loss": 1.2433, "step": 343 }, { "epoch": 0.2962962962962963, "grad_norm": 1.0461153984069824, "learning_rate": 1.650825998923361e-05, "loss": 1.2289, "step": 344 }, { "epoch": 0.2971576227390181, "grad_norm": 0.8382362127304077, "learning_rate": 1.6487051899258738e-05, "loss": 1.2455, "step": 345 }, { "epoch": 0.2980189491817399, "grad_norm": 0.9788974523544312, "learning_rate": 1.6465793311763255e-05, "loss": 1.2314, "step": 346 }, { "epoch": 0.2988802756244617, "grad_norm": 0.9103128910064697, "learning_rate": 1.6444484392231574e-05, "loss": 1.2331, "step": 347 }, { "epoch": 0.2997416020671835, "grad_norm": 0.9638088345527649, "learning_rate": 1.6423125306539903e-05, "loss": 1.2228, "step": 348 }, { "epoch": 0.3006029285099053, "grad_norm": 0.9347442388534546, "learning_rate": 1.6401716220954968e-05, "loss": 1.227, "step": 349 }, { "epoch": 0.301464254952627, "grad_norm": 0.9683001637458801, "learning_rate": 1.638025730213271e-05, "loss": 1.2253, "step": 350 }, { "epoch": 0.3023255813953488, "grad_norm": 0.9251329898834229, "learning_rate": 1.6358748717116993e-05, "loss": 1.2698, "step": 351 }, { "epoch": 0.3031869078380706, "grad_norm": 0.9530742168426514, "learning_rate": 1.6337190633338294e-05, "loss": 1.2407, "step": 352 }, { "epoch": 0.3040482342807924, "grad_norm": 0.9009522199630737, "learning_rate": 1.631558321861241e-05, "loss": 1.2478, "step": 353 }, { "epoch": 0.3049095607235142, "grad_norm": 1.0106037855148315, "learning_rate": 1.6293926641139154e-05, "loss": 1.2298, "step": 354 }, { "epoch": 0.305770887166236, "grad_norm": 0.8804092407226562, "learning_rate": 1.627222106950102e-05, "loss": 1.2324, "step": 355 }, { "epoch": 0.3066322136089578, "grad_norm": 0.8187827467918396, "learning_rate": 1.625046667266191e-05, "loss": 1.2714, "step": 356 }, { "epoch": 0.30749354005167956, "grad_norm": 1.0687847137451172, "learning_rate": 1.6228663619965787e-05, "loss": 1.2423, "step": 357 }, { "epoch": 0.30835486649440136, "grad_norm": 0.8696728348731995, "learning_rate": 1.620681208113538e-05, "loss": 1.241, "step": 358 }, { "epoch": 0.30921619293712316, "grad_norm": 0.8725491762161255, "learning_rate": 1.6184912226270833e-05, "loss": 1.2741, "step": 359 }, { "epoch": 0.31007751937984496, "grad_norm": 1.0487327575683594, "learning_rate": 1.6162964225848416e-05, "loss": 1.2382, "step": 360 }, { "epoch": 0.31093884582256676, "grad_norm": 0.8088880181312561, "learning_rate": 1.6140968250719177e-05, "loss": 1.2441, "step": 361 }, { "epoch": 0.31180017226528856, "grad_norm": 0.8771573305130005, "learning_rate": 1.611892447210761e-05, "loss": 1.2393, "step": 362 }, { "epoch": 0.31266149870801035, "grad_norm": 0.9969056248664856, "learning_rate": 1.609683306161034e-05, "loss": 1.2319, "step": 363 }, { "epoch": 0.31352282515073215, "grad_norm": 0.9056423902511597, "learning_rate": 1.6074694191194758e-05, "loss": 1.2191, "step": 364 }, { "epoch": 0.3143841515934539, "grad_norm": 0.8870687484741211, "learning_rate": 1.6052508033197713e-05, "loss": 1.249, "step": 365 }, { "epoch": 0.3152454780361757, "grad_norm": 0.9680421352386475, "learning_rate": 1.6030274760324163e-05, "loss": 1.2299, "step": 366 }, { "epoch": 0.3161068044788975, "grad_norm": 0.9495812058448792, "learning_rate": 1.6007994545645807e-05, "loss": 1.2574, "step": 367 }, { "epoch": 0.3169681309216193, "grad_norm": 0.8676847219467163, "learning_rate": 1.598566756259977e-05, "loss": 1.2284, "step": 368 }, { "epoch": 0.3178294573643411, "grad_norm": 0.841946005821228, "learning_rate": 1.596329398498723e-05, "loss": 1.2304, "step": 369 }, { "epoch": 0.3186907838070629, "grad_norm": 1.0011506080627441, "learning_rate": 1.5940873986972078e-05, "loss": 1.2454, "step": 370 }, { "epoch": 0.3195521102497847, "grad_norm": 0.9674389958381653, "learning_rate": 1.5918407743079564e-05, "loss": 1.241, "step": 371 }, { "epoch": 0.32041343669250644, "grad_norm": 1.0047541856765747, "learning_rate": 1.5895895428194915e-05, "loss": 1.2541, "step": 372 }, { "epoch": 0.32127476313522824, "grad_norm": 0.9397938847541809, "learning_rate": 1.5873337217562012e-05, "loss": 1.2182, "step": 373 }, { "epoch": 0.32213608957795004, "grad_norm": 0.8715697526931763, "learning_rate": 1.585073328678199e-05, "loss": 1.243, "step": 374 }, { "epoch": 0.32299741602067183, "grad_norm": 0.9596509337425232, "learning_rate": 1.582808381181189e-05, "loss": 1.2417, "step": 375 }, { "epoch": 0.32385874246339363, "grad_norm": 0.8386455178260803, "learning_rate": 1.5805388968963286e-05, "loss": 1.2625, "step": 376 }, { "epoch": 0.32472006890611543, "grad_norm": 0.897318422794342, "learning_rate": 1.5782648934900915e-05, "loss": 1.2233, "step": 377 }, { "epoch": 0.32558139534883723, "grad_norm": 0.8364303112030029, "learning_rate": 1.575986388664129e-05, "loss": 1.2323, "step": 378 }, { "epoch": 0.326442721791559, "grad_norm": 0.8544413447380066, "learning_rate": 1.5737034001551336e-05, "loss": 1.2373, "step": 379 }, { "epoch": 0.3273040482342808, "grad_norm": 0.9020499587059021, "learning_rate": 1.5714159457347007e-05, "loss": 1.2432, "step": 380 }, { "epoch": 0.3281653746770026, "grad_norm": 0.8884045481681824, "learning_rate": 1.5691240432091892e-05, "loss": 1.2144, "step": 381 }, { "epoch": 0.3290267011197244, "grad_norm": 0.9886582493782043, "learning_rate": 1.566827710419584e-05, "loss": 1.2344, "step": 382 }, { "epoch": 0.3298880275624462, "grad_norm": 0.8105961680412292, "learning_rate": 1.5645269652413574e-05, "loss": 1.2441, "step": 383 }, { "epoch": 0.330749354005168, "grad_norm": 0.9423347115516663, "learning_rate": 1.5622218255843276e-05, "loss": 1.2172, "step": 384 }, { "epoch": 0.33161068044788977, "grad_norm": 0.970775842666626, "learning_rate": 1.559912309392523e-05, "loss": 1.2587, "step": 385 }, { "epoch": 0.3324720068906115, "grad_norm": 0.9015215039253235, "learning_rate": 1.5575984346440393e-05, "loss": 1.2388, "step": 386 }, { "epoch": 0.3333333333333333, "grad_norm": 0.9856666326522827, "learning_rate": 1.5552802193509003e-05, "loss": 1.2181, "step": 387 }, { "epoch": 0.3341946597760551, "grad_norm": 0.9138821959495544, "learning_rate": 1.55295768155892e-05, "loss": 1.2418, "step": 388 }, { "epoch": 0.3350559862187769, "grad_norm": 0.8847649693489075, "learning_rate": 1.5506308393475582e-05, "loss": 1.1956, "step": 389 }, { "epoch": 0.3359173126614987, "grad_norm": 0.847986102104187, "learning_rate": 1.5482997108297834e-05, "loss": 1.2041, "step": 390 }, { "epoch": 0.3367786391042205, "grad_norm": 1.1017706394195557, "learning_rate": 1.545964314151929e-05, "loss": 1.2646, "step": 391 }, { "epoch": 0.3376399655469423, "grad_norm": 0.8774247169494629, "learning_rate": 1.5436246674935543e-05, "loss": 1.2236, "step": 392 }, { "epoch": 0.3385012919896641, "grad_norm": 0.8937813639640808, "learning_rate": 1.5412807890673015e-05, "loss": 1.2192, "step": 393 }, { "epoch": 0.33936261843238585, "grad_norm": 0.8541170358657837, "learning_rate": 1.5389326971187543e-05, "loss": 1.2268, "step": 394 }, { "epoch": 0.34022394487510765, "grad_norm": 0.9567473530769348, "learning_rate": 1.536580409926296e-05, "loss": 1.2489, "step": 395 }, { "epoch": 0.34108527131782945, "grad_norm": 0.8888243436813354, "learning_rate": 1.5342239458009675e-05, "loss": 1.2308, "step": 396 }, { "epoch": 0.34194659776055125, "grad_norm": 0.893013060092926, "learning_rate": 1.5318633230863237e-05, "loss": 1.2237, "step": 397 }, { "epoch": 0.34280792420327305, "grad_norm": 0.874277651309967, "learning_rate": 1.5294985601582922e-05, "loss": 1.186, "step": 398 }, { "epoch": 0.34366925064599485, "grad_norm": 0.8381382822990417, "learning_rate": 1.5271296754250296e-05, "loss": 1.2454, "step": 399 }, { "epoch": 0.34453057708871665, "grad_norm": 0.9372345209121704, "learning_rate": 1.524756687326777e-05, "loss": 1.2233, "step": 400 }, { "epoch": 0.3453919035314384, "grad_norm": 0.869147002696991, "learning_rate": 1.5223796143357188e-05, "loss": 1.2084, "step": 401 }, { "epoch": 0.3462532299741602, "grad_norm": 0.8200699687004089, "learning_rate": 1.5199984749558367e-05, "loss": 1.2355, "step": 402 }, { "epoch": 0.347114556416882, "grad_norm": 0.901074230670929, "learning_rate": 1.5176132877227674e-05, "loss": 1.2653, "step": 403 }, { "epoch": 0.3479758828596038, "grad_norm": 0.871003270149231, "learning_rate": 1.5152240712036573e-05, "loss": 1.2289, "step": 404 }, { "epoch": 0.3488372093023256, "grad_norm": 0.9896201491355896, "learning_rate": 1.5128308439970174e-05, "loss": 1.2388, "step": 405 }, { "epoch": 0.3496985357450474, "grad_norm": 0.9683994054794312, "learning_rate": 1.5104336247325803e-05, "loss": 1.2178, "step": 406 }, { "epoch": 0.3505598621877692, "grad_norm": 0.8784785270690918, "learning_rate": 1.5080324320711542e-05, "loss": 1.2929, "step": 407 }, { "epoch": 0.35142118863049093, "grad_norm": 0.9590929746627808, "learning_rate": 1.505627284704477e-05, "loss": 1.2284, "step": 408 }, { "epoch": 0.35228251507321273, "grad_norm": 0.8345155715942383, "learning_rate": 1.5032182013550719e-05, "loss": 1.2614, "step": 409 }, { "epoch": 0.35314384151593453, "grad_norm": 0.8421818017959595, "learning_rate": 1.5008052007761009e-05, "loss": 1.2258, "step": 410 }, { "epoch": 0.35400516795865633, "grad_norm": 1.0068047046661377, "learning_rate": 1.498388301751219e-05, "loss": 1.2686, "step": 411 }, { "epoch": 0.35486649440137813, "grad_norm": 0.8941269516944885, "learning_rate": 1.495967523094429e-05, "loss": 1.2268, "step": 412 }, { "epoch": 0.35572782084409993, "grad_norm": 0.9728747606277466, "learning_rate": 1.4935428836499333e-05, "loss": 1.2359, "step": 413 }, { "epoch": 0.35658914728682173, "grad_norm": 0.946317732334137, "learning_rate": 1.4911144022919879e-05, "loss": 1.2062, "step": 414 }, { "epoch": 0.35745047372954347, "grad_norm": 0.8988799452781677, "learning_rate": 1.4886820979247561e-05, "loss": 1.2389, "step": 415 }, { "epoch": 0.35831180017226527, "grad_norm": 1.0396331548690796, "learning_rate": 1.4862459894821606e-05, "loss": 1.2247, "step": 416 }, { "epoch": 0.35917312661498707, "grad_norm": 0.8363151550292969, "learning_rate": 1.483806095927737e-05, "loss": 1.1869, "step": 417 }, { "epoch": 0.36003445305770887, "grad_norm": 0.9274395704269409, "learning_rate": 1.481362436254484e-05, "loss": 1.1593, "step": 418 }, { "epoch": 0.36089577950043067, "grad_norm": 1.005424976348877, "learning_rate": 1.4789150294847192e-05, "loss": 1.2132, "step": 419 }, { "epoch": 0.36175710594315247, "grad_norm": 0.9512991309165955, "learning_rate": 1.4764638946699275e-05, "loss": 1.2082, "step": 420 }, { "epoch": 0.36261843238587427, "grad_norm": 0.8919705152511597, "learning_rate": 1.4740090508906147e-05, "loss": 1.1693, "step": 421 }, { "epoch": 0.363479758828596, "grad_norm": 0.8347681760787964, "learning_rate": 1.4715505172561577e-05, "loss": 1.2604, "step": 422 }, { "epoch": 0.3643410852713178, "grad_norm": 0.8956724405288696, "learning_rate": 1.4690883129046585e-05, "loss": 1.2241, "step": 423 }, { "epoch": 0.3652024117140396, "grad_norm": 0.9047289490699768, "learning_rate": 1.466622457002791e-05, "loss": 1.2257, "step": 424 }, { "epoch": 0.3660637381567614, "grad_norm": 1.0073318481445312, "learning_rate": 1.4641529687456558e-05, "loss": 1.2385, "step": 425 }, { "epoch": 0.3669250645994832, "grad_norm": 0.9420737028121948, "learning_rate": 1.4616798673566276e-05, "loss": 1.2037, "step": 426 }, { "epoch": 0.367786391042205, "grad_norm": 0.9980217218399048, "learning_rate": 1.4592031720872086e-05, "loss": 1.2163, "step": 427 }, { "epoch": 0.3686477174849268, "grad_norm": 0.8870885968208313, "learning_rate": 1.4567229022168756e-05, "loss": 1.2356, "step": 428 }, { "epoch": 0.3695090439276486, "grad_norm": 0.9896050095558167, "learning_rate": 1.454239077052932e-05, "loss": 1.2334, "step": 429 }, { "epoch": 0.37037037037037035, "grad_norm": 1.0749104022979736, "learning_rate": 1.4517517159303573e-05, "loss": 1.2024, "step": 430 }, { "epoch": 0.37123169681309215, "grad_norm": 0.8981916904449463, "learning_rate": 1.4492608382116548e-05, "loss": 1.1876, "step": 431 }, { "epoch": 0.37209302325581395, "grad_norm": 0.9399816393852234, "learning_rate": 1.4467664632867042e-05, "loss": 1.2465, "step": 432 }, { "epoch": 0.37295434969853575, "grad_norm": 0.9650474786758423, "learning_rate": 1.4442686105726066e-05, "loss": 1.2525, "step": 433 }, { "epoch": 0.37381567614125755, "grad_norm": 0.9051154255867004, "learning_rate": 1.4417672995135372e-05, "loss": 1.2653, "step": 434 }, { "epoch": 0.37467700258397935, "grad_norm": 0.8648033142089844, "learning_rate": 1.4392625495805913e-05, "loss": 1.2221, "step": 435 }, { "epoch": 0.37553832902670115, "grad_norm": 0.942790687084198, "learning_rate": 1.4367543802716334e-05, "loss": 1.2258, "step": 436 }, { "epoch": 0.3763996554694229, "grad_norm": 0.849907398223877, "learning_rate": 1.4342428111111461e-05, "loss": 1.2307, "step": 437 }, { "epoch": 0.3772609819121447, "grad_norm": 0.9022756218910217, "learning_rate": 1.4317278616500785e-05, "loss": 1.2356, "step": 438 }, { "epoch": 0.3781223083548665, "grad_norm": 1.0210813283920288, "learning_rate": 1.4292095514656907e-05, "loss": 1.2099, "step": 439 }, { "epoch": 0.3789836347975883, "grad_norm": 0.8442511558532715, "learning_rate": 1.4266879001614067e-05, "loss": 1.2509, "step": 440 }, { "epoch": 0.3798449612403101, "grad_norm": 0.8751745223999023, "learning_rate": 1.424162927366657e-05, "loss": 1.2039, "step": 441 }, { "epoch": 0.3807062876830319, "grad_norm": 0.8547453880310059, "learning_rate": 1.4216346527367284e-05, "loss": 1.239, "step": 442 }, { "epoch": 0.3815676141257537, "grad_norm": 0.8485442399978638, "learning_rate": 1.4191030959526106e-05, "loss": 1.2208, "step": 443 }, { "epoch": 0.38242894056847543, "grad_norm": 0.9222948551177979, "learning_rate": 1.4165682767208426e-05, "loss": 1.2407, "step": 444 }, { "epoch": 0.3832902670111972, "grad_norm": 0.8452631235122681, "learning_rate": 1.4140302147733596e-05, "loss": 1.2388, "step": 445 }, { "epoch": 0.384151593453919, "grad_norm": 0.8585206270217896, "learning_rate": 1.4114889298673383e-05, "loss": 1.2138, "step": 446 }, { "epoch": 0.3850129198966408, "grad_norm": 0.8404960036277771, "learning_rate": 1.4089444417850455e-05, "loss": 1.2613, "step": 447 }, { "epoch": 0.3858742463393626, "grad_norm": 0.8335172533988953, "learning_rate": 1.4063967703336814e-05, "loss": 1.2188, "step": 448 }, { "epoch": 0.3867355727820844, "grad_norm": 0.8247851729393005, "learning_rate": 1.403845935345228e-05, "loss": 1.2334, "step": 449 }, { "epoch": 0.3875968992248062, "grad_norm": 0.9984527230262756, "learning_rate": 1.401291956676292e-05, "loss": 1.2272, "step": 450 }, { "epoch": 0.38845822566752797, "grad_norm": 0.8594214916229248, "learning_rate": 1.3987348542079526e-05, "loss": 1.1752, "step": 451 }, { "epoch": 0.38931955211024977, "grad_norm": 0.95940101146698, "learning_rate": 1.396174647845605e-05, "loss": 1.2172, "step": 452 }, { "epoch": 0.39018087855297157, "grad_norm": 0.8622090816497803, "learning_rate": 1.3936113575188074e-05, "loss": 1.206, "step": 453 }, { "epoch": 0.39104220499569337, "grad_norm": 0.8805087804794312, "learning_rate": 1.3910450031811235e-05, "loss": 1.2339, "step": 454 }, { "epoch": 0.39190353143841516, "grad_norm": 0.8449251651763916, "learning_rate": 1.3884756048099688e-05, "loss": 1.2253, "step": 455 }, { "epoch": 0.39276485788113696, "grad_norm": 0.9744880199432373, "learning_rate": 1.3859031824064543e-05, "loss": 1.237, "step": 456 }, { "epoch": 0.39362618432385876, "grad_norm": 0.8672634959220886, "learning_rate": 1.3833277559952323e-05, "loss": 1.2199, "step": 457 }, { "epoch": 0.3944875107665805, "grad_norm": 0.8973323702812195, "learning_rate": 1.380749345624338e-05, "loss": 1.1859, "step": 458 }, { "epoch": 0.3953488372093023, "grad_norm": 0.9631936550140381, "learning_rate": 1.3781679713650349e-05, "loss": 1.2133, "step": 459 }, { "epoch": 0.3962101636520241, "grad_norm": 0.8389044404029846, "learning_rate": 1.3755836533116597e-05, "loss": 1.2095, "step": 460 }, { "epoch": 0.3970714900947459, "grad_norm": 0.8947293162345886, "learning_rate": 1.3729964115814636e-05, "loss": 1.1841, "step": 461 }, { "epoch": 0.3979328165374677, "grad_norm": 0.9099428057670593, "learning_rate": 1.3704062663144569e-05, "loss": 1.2188, "step": 462 }, { "epoch": 0.3987941429801895, "grad_norm": 0.9000226855278015, "learning_rate": 1.3678132376732518e-05, "loss": 1.2673, "step": 463 }, { "epoch": 0.3996554694229113, "grad_norm": 0.9490253925323486, "learning_rate": 1.3652173458429068e-05, "loss": 1.2233, "step": 464 }, { "epoch": 0.4005167958656331, "grad_norm": 0.9906533360481262, "learning_rate": 1.3626186110307673e-05, "loss": 1.2125, "step": 465 }, { "epoch": 0.40137812230835485, "grad_norm": 0.9203736186027527, "learning_rate": 1.3600170534663097e-05, "loss": 1.1732, "step": 466 }, { "epoch": 0.40223944875107664, "grad_norm": 0.8627713918685913, "learning_rate": 1.3574126934009843e-05, "loss": 1.2441, "step": 467 }, { "epoch": 0.40310077519379844, "grad_norm": 0.8728564977645874, "learning_rate": 1.3548055511080568e-05, "loss": 1.1814, "step": 468 }, { "epoch": 0.40396210163652024, "grad_norm": 0.9065693020820618, "learning_rate": 1.3521956468824505e-05, "loss": 1.2281, "step": 469 }, { "epoch": 0.40482342807924204, "grad_norm": 0.860071063041687, "learning_rate": 1.3495830010405884e-05, "loss": 1.2438, "step": 470 }, { "epoch": 0.40568475452196384, "grad_norm": 0.8716009855270386, "learning_rate": 1.346967633920236e-05, "loss": 1.2185, "step": 471 }, { "epoch": 0.40654608096468564, "grad_norm": 1.0129085779190063, "learning_rate": 1.344349565880341e-05, "loss": 1.2253, "step": 472 }, { "epoch": 0.4074074074074074, "grad_norm": 0.8441084623336792, "learning_rate": 1.3417288173008778e-05, "loss": 1.2128, "step": 473 }, { "epoch": 0.4082687338501292, "grad_norm": 0.9617881178855896, "learning_rate": 1.339105408582685e-05, "loss": 1.2085, "step": 474 }, { "epoch": 0.409130060292851, "grad_norm": 0.8780672550201416, "learning_rate": 1.3364793601473105e-05, "loss": 1.1989, "step": 475 }, { "epoch": 0.4099913867355728, "grad_norm": 0.9372087717056274, "learning_rate": 1.3338506924368494e-05, "loss": 1.2343, "step": 476 }, { "epoch": 0.4108527131782946, "grad_norm": 0.8718287348747253, "learning_rate": 1.331219425913787e-05, "loss": 1.2009, "step": 477 }, { "epoch": 0.4117140396210164, "grad_norm": 0.8041454553604126, "learning_rate": 1.3285855810608377e-05, "loss": 1.2006, "step": 478 }, { "epoch": 0.4125753660637382, "grad_norm": 0.8639675378799438, "learning_rate": 1.325949178380788e-05, "loss": 1.2244, "step": 479 }, { "epoch": 0.4134366925064599, "grad_norm": 0.9752419590950012, "learning_rate": 1.3233102383963341e-05, "loss": 1.2374, "step": 480 }, { "epoch": 0.4142980189491817, "grad_norm": 0.9429941177368164, "learning_rate": 1.3206687816499242e-05, "loss": 1.2437, "step": 481 }, { "epoch": 0.4151593453919035, "grad_norm": 0.830390214920044, "learning_rate": 1.3180248287035977e-05, "loss": 1.2173, "step": 482 }, { "epoch": 0.4160206718346253, "grad_norm": 0.8312661051750183, "learning_rate": 1.3153784001388249e-05, "loss": 1.2432, "step": 483 }, { "epoch": 0.4168819982773471, "grad_norm": 0.8787814378738403, "learning_rate": 1.3127295165563476e-05, "loss": 1.2465, "step": 484 }, { "epoch": 0.4177433247200689, "grad_norm": 0.8632099628448486, "learning_rate": 1.3100781985760188e-05, "loss": 1.2347, "step": 485 }, { "epoch": 0.4186046511627907, "grad_norm": 1.0281778573989868, "learning_rate": 1.3074244668366412e-05, "loss": 1.2193, "step": 486 }, { "epoch": 0.41946597760551246, "grad_norm": 0.9415022730827332, "learning_rate": 1.3047683419958062e-05, "loss": 1.2021, "step": 487 }, { "epoch": 0.42032730404823426, "grad_norm": 0.9123542904853821, "learning_rate": 1.3021098447297358e-05, "loss": 1.216, "step": 488 }, { "epoch": 0.42118863049095606, "grad_norm": 0.9520506858825684, "learning_rate": 1.2994489957331183e-05, "loss": 1.2148, "step": 489 }, { "epoch": 0.42204995693367786, "grad_norm": 0.8072666525840759, "learning_rate": 1.2967858157189495e-05, "loss": 1.2404, "step": 490 }, { "epoch": 0.42291128337639966, "grad_norm": 0.9025439620018005, "learning_rate": 1.29412032541837e-05, "loss": 1.2183, "step": 491 }, { "epoch": 0.42377260981912146, "grad_norm": 0.8652177453041077, "learning_rate": 1.2914525455805056e-05, "loss": 1.1991, "step": 492 }, { "epoch": 0.42463393626184326, "grad_norm": 1.0470143556594849, "learning_rate": 1.2887824969723035e-05, "loss": 1.2286, "step": 493 }, { "epoch": 0.425495262704565, "grad_norm": 0.9123655557632446, "learning_rate": 1.2861102003783722e-05, "loss": 1.2405, "step": 494 }, { "epoch": 0.4263565891472868, "grad_norm": 0.8399021625518799, "learning_rate": 1.2834356766008198e-05, "loss": 1.1718, "step": 495 }, { "epoch": 0.4272179155900086, "grad_norm": 1.0004807710647583, "learning_rate": 1.2807589464590908e-05, "loss": 1.1792, "step": 496 }, { "epoch": 0.4280792420327304, "grad_norm": 0.8504483103752136, "learning_rate": 1.2780800307898057e-05, "loss": 1.1962, "step": 497 }, { "epoch": 0.4289405684754522, "grad_norm": 0.9600708484649658, "learning_rate": 1.2753989504465967e-05, "loss": 1.2396, "step": 498 }, { "epoch": 0.429801894918174, "grad_norm": 0.8892411589622498, "learning_rate": 1.2727157262999481e-05, "loss": 1.1949, "step": 499 }, { "epoch": 0.4306632213608958, "grad_norm": 0.9092352986335754, "learning_rate": 1.270030379237031e-05, "loss": 1.2467, "step": 500 }, { "epoch": 0.4315245478036176, "grad_norm": 0.8833507299423218, "learning_rate": 1.2673429301615431e-05, "loss": 1.1803, "step": 501 }, { "epoch": 0.43238587424633934, "grad_norm": 0.8737507462501526, "learning_rate": 1.2646533999935442e-05, "loss": 1.2517, "step": 502 }, { "epoch": 0.43324720068906114, "grad_norm": 0.8569416999816895, "learning_rate": 1.2619618096692942e-05, "loss": 1.2096, "step": 503 }, { "epoch": 0.43410852713178294, "grad_norm": 0.9263505935668945, "learning_rate": 1.25926818014109e-05, "loss": 1.1993, "step": 504 }, { "epoch": 0.43496985357450474, "grad_norm": 1.0186364650726318, "learning_rate": 1.256572532377103e-05, "loss": 1.2233, "step": 505 }, { "epoch": 0.43583118001722654, "grad_norm": 0.83504319190979, "learning_rate": 1.253874887361214e-05, "loss": 1.2251, "step": 506 }, { "epoch": 0.43669250645994834, "grad_norm": 0.9106087684631348, "learning_rate": 1.2511752660928523e-05, "loss": 1.2103, "step": 507 }, { "epoch": 0.43755383290267014, "grad_norm": 0.8864356279373169, "learning_rate": 1.2484736895868306e-05, "loss": 1.2015, "step": 508 }, { "epoch": 0.4384151593453919, "grad_norm": 0.9549910426139832, "learning_rate": 1.2457701788731812e-05, "loss": 1.2394, "step": 509 }, { "epoch": 0.4392764857881137, "grad_norm": 0.9277191758155823, "learning_rate": 1.2430647549969949e-05, "loss": 1.2186, "step": 510 }, { "epoch": 0.4401378122308355, "grad_norm": 0.9281412363052368, "learning_rate": 1.2403574390182529e-05, "loss": 1.2122, "step": 511 }, { "epoch": 0.4409991386735573, "grad_norm": 0.8452510833740234, "learning_rate": 1.2376482520116666e-05, "loss": 1.2203, "step": 512 }, { "epoch": 0.4418604651162791, "grad_norm": 0.9457308650016785, "learning_rate": 1.2349372150665117e-05, "loss": 1.2073, "step": 513 }, { "epoch": 0.4427217915590009, "grad_norm": 0.8949629664421082, "learning_rate": 1.2322243492864651e-05, "loss": 1.2078, "step": 514 }, { "epoch": 0.4435831180017227, "grad_norm": 0.9043914079666138, "learning_rate": 1.2295096757894389e-05, "loss": 1.2159, "step": 515 }, { "epoch": 0.4444444444444444, "grad_norm": 0.8685716390609741, "learning_rate": 1.2267932157074178e-05, "loss": 1.2262, "step": 516 }, { "epoch": 0.4453057708871662, "grad_norm": 0.8798421621322632, "learning_rate": 1.224074990186294e-05, "loss": 1.2551, "step": 517 }, { "epoch": 0.446167097329888, "grad_norm": 0.9253307580947876, "learning_rate": 1.2213550203857025e-05, "loss": 1.2072, "step": 518 }, { "epoch": 0.4470284237726098, "grad_norm": 0.9633350372314453, "learning_rate": 1.2186333274788558e-05, "loss": 1.2122, "step": 519 }, { "epoch": 0.4478897502153316, "grad_norm": 0.8635453581809998, "learning_rate": 1.21590993265238e-05, "loss": 1.2107, "step": 520 }, { "epoch": 0.4487510766580534, "grad_norm": 0.8736230134963989, "learning_rate": 1.2131848571061501e-05, "loss": 1.2315, "step": 521 }, { "epoch": 0.4496124031007752, "grad_norm": 0.8366631269454956, "learning_rate": 1.2104581220531237e-05, "loss": 1.2226, "step": 522 }, { "epoch": 0.45047372954349696, "grad_norm": 0.9264794588088989, "learning_rate": 1.2077297487191771e-05, "loss": 1.2, "step": 523 }, { "epoch": 0.45133505598621876, "grad_norm": 0.8290119171142578, "learning_rate": 1.2049997583429389e-05, "loss": 1.2188, "step": 524 }, { "epoch": 0.45219638242894056, "grad_norm": 0.9655864238739014, "learning_rate": 1.202268172175626e-05, "loss": 1.2326, "step": 525 }, { "epoch": 0.45305770887166236, "grad_norm": 0.8329289555549622, "learning_rate": 1.1995350114808772e-05, "loss": 1.2218, "step": 526 }, { "epoch": 0.45391903531438416, "grad_norm": 0.9534193873405457, "learning_rate": 1.1968002975345882e-05, "loss": 1.2072, "step": 527 }, { "epoch": 0.45478036175710596, "grad_norm": 0.8719503283500671, "learning_rate": 1.194064051624745e-05, "loss": 1.2469, "step": 528 }, { "epoch": 0.45564168819982775, "grad_norm": 0.9249786138534546, "learning_rate": 1.1913262950512605e-05, "loss": 1.2378, "step": 529 }, { "epoch": 0.45650301464254955, "grad_norm": 0.8721954226493835, "learning_rate": 1.1885870491258054e-05, "loss": 1.1767, "step": 530 }, { "epoch": 0.4573643410852713, "grad_norm": 0.9456660747528076, "learning_rate": 1.185846335171645e-05, "loss": 1.1892, "step": 531 }, { "epoch": 0.4582256675279931, "grad_norm": 0.856704592704773, "learning_rate": 1.1831041745234728e-05, "loss": 1.2102, "step": 532 }, { "epoch": 0.4590869939707149, "grad_norm": 0.8333457112312317, "learning_rate": 1.180360588527242e-05, "loss": 1.2255, "step": 533 }, { "epoch": 0.4599483204134367, "grad_norm": 0.9460641741752625, "learning_rate": 1.177615598540003e-05, "loss": 1.1819, "step": 534 }, { "epoch": 0.4608096468561585, "grad_norm": 0.830569326877594, "learning_rate": 1.1748692259297347e-05, "loss": 1.2227, "step": 535 }, { "epoch": 0.4616709732988803, "grad_norm": 0.8055482506752014, "learning_rate": 1.172121492075179e-05, "loss": 1.2011, "step": 536 }, { "epoch": 0.4625322997416021, "grad_norm": 0.9891955256462097, "learning_rate": 1.169372418365674e-05, "loss": 1.1889, "step": 537 }, { "epoch": 0.46339362618432384, "grad_norm": 0.9944434762001038, "learning_rate": 1.1666220262009877e-05, "loss": 1.185, "step": 538 }, { "epoch": 0.46425495262704564, "grad_norm": 0.8905592560768127, "learning_rate": 1.1638703369911517e-05, "loss": 1.2404, "step": 539 }, { "epoch": 0.46511627906976744, "grad_norm": 0.8993557691574097, "learning_rate": 1.161117372156294e-05, "loss": 1.1786, "step": 540 }, { "epoch": 0.46597760551248923, "grad_norm": 0.9197692275047302, "learning_rate": 1.1583631531264723e-05, "loss": 1.2178, "step": 541 }, { "epoch": 0.46683893195521103, "grad_norm": 0.8710423111915588, "learning_rate": 1.1556077013415084e-05, "loss": 1.2422, "step": 542 }, { "epoch": 0.46770025839793283, "grad_norm": 0.8500446081161499, "learning_rate": 1.152851038250819e-05, "loss": 1.2196, "step": 543 }, { "epoch": 0.46856158484065463, "grad_norm": 0.88962721824646, "learning_rate": 1.150093185313251e-05, "loss": 1.1858, "step": 544 }, { "epoch": 0.4694229112833764, "grad_norm": 0.907476544380188, "learning_rate": 1.147334163996913e-05, "loss": 1.1906, "step": 545 }, { "epoch": 0.4702842377260982, "grad_norm": 0.9100911617279053, "learning_rate": 1.1445739957790087e-05, "loss": 1.2185, "step": 546 }, { "epoch": 0.47114556416882, "grad_norm": 0.9100723266601562, "learning_rate": 1.14181270214567e-05, "loss": 1.2539, "step": 547 }, { "epoch": 0.4720068906115418, "grad_norm": 0.9603404998779297, "learning_rate": 1.1390503045917892e-05, "loss": 1.1917, "step": 548 }, { "epoch": 0.4728682170542636, "grad_norm": 0.8922184109687805, "learning_rate": 1.1362868246208519e-05, "loss": 1.1583, "step": 549 }, { "epoch": 0.4737295434969854, "grad_norm": 0.9883030652999878, "learning_rate": 1.1335222837447692e-05, "loss": 1.2329, "step": 550 }, { "epoch": 0.47459086993970717, "grad_norm": 0.9090227484703064, "learning_rate": 1.1307567034837123e-05, "loss": 1.2424, "step": 551 }, { "epoch": 0.4754521963824289, "grad_norm": 0.8161377906799316, "learning_rate": 1.127990105365941e-05, "loss": 1.2166, "step": 552 }, { "epoch": 0.4763135228251507, "grad_norm": 0.9323464632034302, "learning_rate": 1.1252225109276404e-05, "loss": 1.1941, "step": 553 }, { "epoch": 0.4771748492678725, "grad_norm": 0.9303016662597656, "learning_rate": 1.12245394171275e-05, "loss": 1.1916, "step": 554 }, { "epoch": 0.4780361757105943, "grad_norm": 0.9320595264434814, "learning_rate": 1.1196844192727984e-05, "loss": 1.2087, "step": 555 }, { "epoch": 0.4788975021533161, "grad_norm": 0.848597526550293, "learning_rate": 1.1169139651667334e-05, "loss": 1.1857, "step": 556 }, { "epoch": 0.4797588285960379, "grad_norm": 0.9183049201965332, "learning_rate": 1.1141426009607562e-05, "loss": 1.2185, "step": 557 }, { "epoch": 0.4806201550387597, "grad_norm": 0.9825764298439026, "learning_rate": 1.1113703482281515e-05, "loss": 1.2045, "step": 558 }, { "epoch": 0.48148148148148145, "grad_norm": 0.8843861818313599, "learning_rate": 1.1085972285491213e-05, "loss": 1.198, "step": 559 }, { "epoch": 0.48234280792420325, "grad_norm": 0.8294288516044617, "learning_rate": 1.1058232635106167e-05, "loss": 1.1729, "step": 560 }, { "epoch": 0.48320413436692505, "grad_norm": 0.810230553150177, "learning_rate": 1.103048474706168e-05, "loss": 1.2055, "step": 561 }, { "epoch": 0.48406546080964685, "grad_norm": 0.8811335563659668, "learning_rate": 1.1002728837357192e-05, "loss": 1.2168, "step": 562 }, { "epoch": 0.48492678725236865, "grad_norm": 0.9530275464057922, "learning_rate": 1.097496512205458e-05, "loss": 1.2179, "step": 563 }, { "epoch": 0.48578811369509045, "grad_norm": 0.9637301564216614, "learning_rate": 1.0947193817276485e-05, "loss": 1.1692, "step": 564 }, { "epoch": 0.48664944013781225, "grad_norm": 0.8479545712471008, "learning_rate": 1.0919415139204625e-05, "loss": 1.2172, "step": 565 }, { "epoch": 0.48751076658053405, "grad_norm": 0.9883643388748169, "learning_rate": 1.089162930407812e-05, "loss": 1.2697, "step": 566 }, { "epoch": 0.4883720930232558, "grad_norm": 0.8231867551803589, "learning_rate": 1.0863836528191795e-05, "loss": 1.215, "step": 567 }, { "epoch": 0.4892334194659776, "grad_norm": 0.8607746362686157, "learning_rate": 1.0836037027894515e-05, "loss": 1.2322, "step": 568 }, { "epoch": 0.4900947459086994, "grad_norm": 0.85414719581604, "learning_rate": 1.0808231019587472e-05, "loss": 1.2044, "step": 569 }, { "epoch": 0.4909560723514212, "grad_norm": 0.835567831993103, "learning_rate": 1.0780418719722544e-05, "loss": 1.218, "step": 570 }, { "epoch": 0.491817398794143, "grad_norm": 0.841223418712616, "learning_rate": 1.075260034480056e-05, "loss": 1.2253, "step": 571 }, { "epoch": 0.4926787252368648, "grad_norm": 0.8959380984306335, "learning_rate": 1.0724776111369654e-05, "loss": 1.1758, "step": 572 }, { "epoch": 0.4935400516795866, "grad_norm": 0.9385613799095154, "learning_rate": 1.0696946236023566e-05, "loss": 1.2174, "step": 573 }, { "epoch": 0.49440137812230833, "grad_norm": 0.9247633814811707, "learning_rate": 1.0669110935399944e-05, "loss": 1.2104, "step": 574 }, { "epoch": 0.49526270456503013, "grad_norm": 0.8738383650779724, "learning_rate": 1.0641270426178677e-05, "loss": 1.1825, "step": 575 }, { "epoch": 0.49612403100775193, "grad_norm": 1.0910409688949585, "learning_rate": 1.0613424925080194e-05, "loss": 1.1731, "step": 576 }, { "epoch": 0.49698535745047373, "grad_norm": 0.8407815098762512, "learning_rate": 1.058557464886379e-05, "loss": 1.2002, "step": 577 }, { "epoch": 0.49784668389319553, "grad_norm": 0.9195674657821655, "learning_rate": 1.055771981432592e-05, "loss": 1.2013, "step": 578 }, { "epoch": 0.49870801033591733, "grad_norm": 0.8779056072235107, "learning_rate": 1.0529860638298535e-05, "loss": 1.2098, "step": 579 }, { "epoch": 0.49956933677863913, "grad_norm": 0.8528910875320435, "learning_rate": 1.0501997337647372e-05, "loss": 1.1676, "step": 580 }, { "epoch": 0.5004306632213609, "grad_norm": 0.8757944703102112, "learning_rate": 1.0474130129270281e-05, "loss": 1.1858, "step": 581 }, { "epoch": 0.5012919896640827, "grad_norm": 0.8935431241989136, "learning_rate": 1.0446259230095531e-05, "loss": 1.1776, "step": 582 }, { "epoch": 0.5021533161068045, "grad_norm": 0.9777861833572388, "learning_rate": 1.0418384857080118e-05, "loss": 1.1756, "step": 583 }, { "epoch": 0.5030146425495263, "grad_norm": 0.9448710680007935, "learning_rate": 1.039050722720808e-05, "loss": 1.2015, "step": 584 }, { "epoch": 0.5038759689922481, "grad_norm": 0.8520198464393616, "learning_rate": 1.0362626557488811e-05, "loss": 1.1893, "step": 585 }, { "epoch": 0.5047372954349698, "grad_norm": 0.9623661041259766, "learning_rate": 1.0334743064955367e-05, "loss": 1.1882, "step": 586 }, { "epoch": 0.5055986218776917, "grad_norm": 0.8741092681884766, "learning_rate": 1.0306856966662776e-05, "loss": 1.17, "step": 587 }, { "epoch": 0.5064599483204134, "grad_norm": 0.8488198518753052, "learning_rate": 1.027896847968635e-05, "loss": 1.2303, "step": 588 }, { "epoch": 0.5073212747631353, "grad_norm": 0.9547082185745239, "learning_rate": 1.0251077821119998e-05, "loss": 1.2269, "step": 589 }, { "epoch": 0.508182601205857, "grad_norm": 1.0101442337036133, "learning_rate": 1.0223185208074538e-05, "loss": 1.191, "step": 590 }, { "epoch": 0.5090439276485789, "grad_norm": 0.8840098977088928, "learning_rate": 1.0195290857675982e-05, "loss": 1.1732, "step": 591 }, { "epoch": 0.5099052540913006, "grad_norm": 0.7868524789810181, "learning_rate": 1.0167394987063894e-05, "loss": 1.1934, "step": 592 }, { "epoch": 0.5107665805340224, "grad_norm": 0.8917614817619324, "learning_rate": 1.0139497813389654e-05, "loss": 1.2006, "step": 593 }, { "epoch": 0.5116279069767442, "grad_norm": 0.8367542624473572, "learning_rate": 1.0111599553814788e-05, "loss": 1.1675, "step": 594 }, { "epoch": 0.512489233419466, "grad_norm": 0.8534975647926331, "learning_rate": 1.008370042550928e-05, "loss": 1.1979, "step": 595 }, { "epoch": 0.5133505598621878, "grad_norm": 0.8273769617080688, "learning_rate": 1.0055800645649874e-05, "loss": 1.1826, "step": 596 }, { "epoch": 0.5142118863049095, "grad_norm": 0.9318427443504333, "learning_rate": 1.002790043141838e-05, "loss": 1.18, "step": 597 }, { "epoch": 0.5150732127476314, "grad_norm": 0.9464846849441528, "learning_rate": 1e-05, "loss": 1.1945, "step": 598 }, { "epoch": 0.5159345391903531, "grad_norm": 0.893747866153717, "learning_rate": 9.972099568581621e-06, "loss": 1.2057, "step": 599 }, { "epoch": 0.5167958656330749, "grad_norm": 0.9831318259239197, "learning_rate": 9.94419935435013e-06, "loss": 1.2543, "step": 600 }, { "epoch": 0.5176571920757967, "grad_norm": 0.8675307035446167, "learning_rate": 9.916299574490722e-06, "loss": 1.1924, "step": 601 }, { "epoch": 0.5185185185185185, "grad_norm": 0.8235703706741333, "learning_rate": 9.888400446185212e-06, "loss": 1.205, "step": 602 }, { "epoch": 0.5193798449612403, "grad_norm": 0.9362935423851013, "learning_rate": 9.860502186610349e-06, "loss": 1.2045, "step": 603 }, { "epoch": 0.5202411714039621, "grad_norm": 1.0003681182861328, "learning_rate": 9.832605012936107e-06, "loss": 1.1725, "step": 604 }, { "epoch": 0.5211024978466839, "grad_norm": 0.8889096975326538, "learning_rate": 9.80470914232402e-06, "loss": 1.2078, "step": 605 }, { "epoch": 0.5219638242894057, "grad_norm": 0.9049688577651978, "learning_rate": 9.77681479192547e-06, "loss": 1.1603, "step": 606 }, { "epoch": 0.5228251507321274, "grad_norm": 0.8464852571487427, "learning_rate": 9.748922178880005e-06, "loss": 1.2056, "step": 607 }, { "epoch": 0.5236864771748493, "grad_norm": 0.9175320267677307, "learning_rate": 9.721031520313653e-06, "loss": 1.2323, "step": 608 }, { "epoch": 0.524547803617571, "grad_norm": 0.8868247270584106, "learning_rate": 9.693143033337228e-06, "loss": 1.186, "step": 609 }, { "epoch": 0.5254091300602929, "grad_norm": 0.8364291191101074, "learning_rate": 9.665256935044636e-06, "loss": 1.1761, "step": 610 }, { "epoch": 0.5262704565030146, "grad_norm": 0.8297803401947021, "learning_rate": 9.637373442511192e-06, "loss": 1.1775, "step": 611 }, { "epoch": 0.5271317829457365, "grad_norm": 0.9279587268829346, "learning_rate": 9.609492772791924e-06, "loss": 1.1789, "step": 612 }, { "epoch": 0.5279931093884582, "grad_norm": 0.8592475056648254, "learning_rate": 9.581615142919887e-06, "loss": 1.2299, "step": 613 }, { "epoch": 0.52885443583118, "grad_norm": 0.8591251373291016, "learning_rate": 9.55374076990447e-06, "loss": 1.1792, "step": 614 }, { "epoch": 0.5297157622739018, "grad_norm": 0.9070160388946533, "learning_rate": 9.525869870729719e-06, "loss": 1.1989, "step": 615 }, { "epoch": 0.5305770887166236, "grad_norm": 0.8992422223091125, "learning_rate": 9.49800266235263e-06, "loss": 1.1947, "step": 616 }, { "epoch": 0.5314384151593454, "grad_norm": 0.8558951616287231, "learning_rate": 9.470139361701469e-06, "loss": 1.2101, "step": 617 }, { "epoch": 0.5322997416020672, "grad_norm": 0.9297447800636292, "learning_rate": 9.442280185674084e-06, "loss": 1.1918, "step": 618 }, { "epoch": 0.533161068044789, "grad_norm": 0.8094662427902222, "learning_rate": 9.414425351136215e-06, "loss": 1.2102, "step": 619 }, { "epoch": 0.5340223944875108, "grad_norm": 0.8350943326950073, "learning_rate": 9.386575074919806e-06, "loss": 1.1809, "step": 620 }, { "epoch": 0.5348837209302325, "grad_norm": 0.9019107818603516, "learning_rate": 9.358729573821325e-06, "loss": 1.1814, "step": 621 }, { "epoch": 0.5357450473729544, "grad_norm": 0.8287387490272522, "learning_rate": 9.330889064600058e-06, "loss": 1.2297, "step": 622 }, { "epoch": 0.5366063738156761, "grad_norm": 0.8505203723907471, "learning_rate": 9.303053763976436e-06, "loss": 1.2028, "step": 623 }, { "epoch": 0.537467700258398, "grad_norm": 0.8712747097015381, "learning_rate": 9.275223888630348e-06, "loss": 1.1718, "step": 624 }, { "epoch": 0.5383290267011197, "grad_norm": 0.907810389995575, "learning_rate": 9.247399655199444e-06, "loss": 1.2403, "step": 625 }, { "epoch": 0.5391903531438416, "grad_norm": 0.8972229957580566, "learning_rate": 9.219581280277463e-06, "loss": 1.1864, "step": 626 }, { "epoch": 0.5400516795865633, "grad_norm": 0.897445023059845, "learning_rate": 9.191768980412528e-06, "loss": 1.1663, "step": 627 }, { "epoch": 0.540913006029285, "grad_norm": 0.813290536403656, "learning_rate": 9.163962972105488e-06, "loss": 1.2084, "step": 628 }, { "epoch": 0.5417743324720069, "grad_norm": 0.8626779913902283, "learning_rate": 9.136163471808207e-06, "loss": 1.2222, "step": 629 }, { "epoch": 0.5426356589147286, "grad_norm": 0.949863076210022, "learning_rate": 9.108370695921884e-06, "loss": 1.2068, "step": 630 }, { "epoch": 0.5434969853574505, "grad_norm": 1.0227597951889038, "learning_rate": 9.080584860795378e-06, "loss": 1.2174, "step": 631 }, { "epoch": 0.5443583118001722, "grad_norm": 0.8536685705184937, "learning_rate": 9.05280618272352e-06, "loss": 1.1978, "step": 632 }, { "epoch": 0.5452196382428941, "grad_norm": 0.8163442015647888, "learning_rate": 9.025034877945422e-06, "loss": 1.1749, "step": 633 }, { "epoch": 0.5460809646856158, "grad_norm": 0.8537796139717102, "learning_rate": 8.99727116264281e-06, "loss": 1.1979, "step": 634 }, { "epoch": 0.5469422911283376, "grad_norm": 0.8340730667114258, "learning_rate": 8.969515252938323e-06, "loss": 1.1756, "step": 635 }, { "epoch": 0.5478036175710594, "grad_norm": 0.8549203276634216, "learning_rate": 8.941767364893836e-06, "loss": 1.1931, "step": 636 }, { "epoch": 0.5486649440137812, "grad_norm": 0.8984569311141968, "learning_rate": 8.914027714508788e-06, "loss": 1.167, "step": 637 }, { "epoch": 0.549526270456503, "grad_norm": 0.8956097960472107, "learning_rate": 8.88629651771849e-06, "loss": 1.1972, "step": 638 }, { "epoch": 0.5503875968992248, "grad_norm": 0.889928936958313, "learning_rate": 8.85857399039244e-06, "loss": 1.1852, "step": 639 }, { "epoch": 0.5512489233419466, "grad_norm": 0.8526130318641663, "learning_rate": 8.830860348332666e-06, "loss": 1.1825, "step": 640 }, { "epoch": 0.5521102497846684, "grad_norm": 0.9533372521400452, "learning_rate": 8.803155807272019e-06, "loss": 1.208, "step": 641 }, { "epoch": 0.5529715762273901, "grad_norm": 1.0128767490386963, "learning_rate": 8.775460582872502e-06, "loss": 1.2077, "step": 642 }, { "epoch": 0.553832902670112, "grad_norm": 0.9490034580230713, "learning_rate": 8.7477748907236e-06, "loss": 1.2078, "step": 643 }, { "epoch": 0.5546942291128337, "grad_norm": 0.9446279406547546, "learning_rate": 8.720098946340594e-06, "loss": 1.1819, "step": 644 }, { "epoch": 0.5555555555555556, "grad_norm": 0.9068703055381775, "learning_rate": 8.69243296516288e-06, "loss": 1.2096, "step": 645 }, { "epoch": 0.5564168819982773, "grad_norm": 0.8627210259437561, "learning_rate": 8.664777162552308e-06, "loss": 1.1981, "step": 646 }, { "epoch": 0.5572782084409992, "grad_norm": 0.8679218292236328, "learning_rate": 8.637131753791485e-06, "loss": 1.2136, "step": 647 }, { "epoch": 0.5581395348837209, "grad_norm": 0.8559207916259766, "learning_rate": 8.609496954082113e-06, "loss": 1.184, "step": 648 }, { "epoch": 0.5590008613264428, "grad_norm": 0.8170911073684692, "learning_rate": 8.581872978543305e-06, "loss": 1.1849, "step": 649 }, { "epoch": 0.5598621877691645, "grad_norm": 0.9225674867630005, "learning_rate": 8.554260042209918e-06, "loss": 1.1612, "step": 650 }, { "epoch": 0.5607235142118863, "grad_norm": 0.8442004919052124, "learning_rate": 8.526658360030876e-06, "loss": 1.195, "step": 651 }, { "epoch": 0.5615848406546081, "grad_norm": 0.9129129648208618, "learning_rate": 8.499068146867492e-06, "loss": 1.2332, "step": 652 }, { "epoch": 0.5624461670973299, "grad_norm": 0.8818443417549133, "learning_rate": 8.471489617491813e-06, "loss": 1.1633, "step": 653 }, { "epoch": 0.5633074935400517, "grad_norm": 0.8677276968955994, "learning_rate": 8.44392298658492e-06, "loss": 1.2062, "step": 654 }, { "epoch": 0.5641688199827735, "grad_norm": 0.9707540273666382, "learning_rate": 8.41636846873528e-06, "loss": 1.1585, "step": 655 }, { "epoch": 0.5650301464254953, "grad_norm": 0.929904580116272, "learning_rate": 8.388826278437066e-06, "loss": 1.1807, "step": 656 }, { "epoch": 0.5658914728682171, "grad_norm": 0.8509595394134521, "learning_rate": 8.361296630088488e-06, "loss": 1.1799, "step": 657 }, { "epoch": 0.5667527993109388, "grad_norm": 0.9297439455986023, "learning_rate": 8.333779737990124e-06, "loss": 1.2112, "step": 658 }, { "epoch": 0.5676141257536607, "grad_norm": 0.913957953453064, "learning_rate": 8.306275816343262e-06, "loss": 1.1909, "step": 659 }, { "epoch": 0.5684754521963824, "grad_norm": 0.8199489116668701, "learning_rate": 8.278785079248211e-06, "loss": 1.1718, "step": 660 }, { "epoch": 0.5693367786391043, "grad_norm": 0.9394577741622925, "learning_rate": 8.251307740702656e-06, "loss": 1.1927, "step": 661 }, { "epoch": 0.570198105081826, "grad_norm": 1.024397611618042, "learning_rate": 8.223844014599973e-06, "loss": 1.1849, "step": 662 }, { "epoch": 0.5710594315245479, "grad_norm": 0.9490551352500916, "learning_rate": 8.196394114727586e-06, "loss": 1.1655, "step": 663 }, { "epoch": 0.5719207579672696, "grad_norm": 0.8966484069824219, "learning_rate": 8.168958254765275e-06, "loss": 1.1916, "step": 664 }, { "epoch": 0.5727820844099913, "grad_norm": 0.8795093894004822, "learning_rate": 8.14153664828355e-06, "loss": 1.2319, "step": 665 }, { "epoch": 0.5736434108527132, "grad_norm": 0.8979016542434692, "learning_rate": 8.114129508741947e-06, "loss": 1.1894, "step": 666 }, { "epoch": 0.5745047372954349, "grad_norm": 0.9347522258758545, "learning_rate": 8.086737049487398e-06, "loss": 1.1941, "step": 667 }, { "epoch": 0.5753660637381568, "grad_norm": 1.0714643001556396, "learning_rate": 8.059359483752551e-06, "loss": 1.2061, "step": 668 }, { "epoch": 0.5762273901808785, "grad_norm": 0.8999515175819397, "learning_rate": 8.031997024654123e-06, "loss": 1.1827, "step": 669 }, { "epoch": 0.5770887166236004, "grad_norm": 0.8698625564575195, "learning_rate": 8.00464988519123e-06, "loss": 1.1879, "step": 670 }, { "epoch": 0.5779500430663221, "grad_norm": 0.8773601651191711, "learning_rate": 7.977318278243742e-06, "loss": 1.1733, "step": 671 }, { "epoch": 0.5788113695090439, "grad_norm": 0.8871725797653198, "learning_rate": 7.950002416570614e-06, "loss": 1.1988, "step": 672 }, { "epoch": 0.5796726959517657, "grad_norm": 0.8334397673606873, "learning_rate": 7.92270251280823e-06, "loss": 1.1846, "step": 673 }, { "epoch": 0.5805340223944875, "grad_norm": 0.9135998487472534, "learning_rate": 7.895418779468766e-06, "loss": 1.1776, "step": 674 }, { "epoch": 0.5813953488372093, "grad_norm": 1.04806649684906, "learning_rate": 7.868151428938502e-06, "loss": 1.1788, "step": 675 }, { "epoch": 0.5822566752799311, "grad_norm": 0.8940375447273254, "learning_rate": 7.840900673476204e-06, "loss": 1.1911, "step": 676 }, { "epoch": 0.5831180017226529, "grad_norm": 0.8850206136703491, "learning_rate": 7.813666725211445e-06, "loss": 1.1773, "step": 677 }, { "epoch": 0.5839793281653747, "grad_norm": 0.9374780654907227, "learning_rate": 7.786449796142979e-06, "loss": 1.2115, "step": 678 }, { "epoch": 0.5848406546080964, "grad_norm": 0.9776993989944458, "learning_rate": 7.759250098137061e-06, "loss": 1.1713, "step": 679 }, { "epoch": 0.5857019810508183, "grad_norm": 0.8467763662338257, "learning_rate": 7.732067842925823e-06, "loss": 1.1584, "step": 680 }, { "epoch": 0.58656330749354, "grad_norm": 0.8266571164131165, "learning_rate": 7.704903242105616e-06, "loss": 1.1749, "step": 681 }, { "epoch": 0.5874246339362619, "grad_norm": 0.8675100803375244, "learning_rate": 7.677756507135354e-06, "loss": 1.1706, "step": 682 }, { "epoch": 0.5882859603789836, "grad_norm": 0.8470680713653564, "learning_rate": 7.650627849334881e-06, "loss": 1.186, "step": 683 }, { "epoch": 0.5891472868217055, "grad_norm": 0.8331758975982666, "learning_rate": 7.623517479883335e-06, "loss": 1.1826, "step": 684 }, { "epoch": 0.5900086132644272, "grad_norm": 0.8689389228820801, "learning_rate": 7.596425609817474e-06, "loss": 1.1931, "step": 685 }, { "epoch": 0.590869939707149, "grad_norm": 0.9240257740020752, "learning_rate": 7.569352450030054e-06, "loss": 1.188, "step": 686 }, { "epoch": 0.5917312661498708, "grad_norm": 0.9663352370262146, "learning_rate": 7.542298211268189e-06, "loss": 1.1903, "step": 687 }, { "epoch": 0.5925925925925926, "grad_norm": 0.9322826266288757, "learning_rate": 7.515263104131699e-06, "loss": 1.1991, "step": 688 }, { "epoch": 0.5934539190353144, "grad_norm": 0.8683158159255981, "learning_rate": 7.488247339071478e-06, "loss": 1.2022, "step": 689 }, { "epoch": 0.5943152454780362, "grad_norm": 0.9323744177818298, "learning_rate": 7.461251126387863e-06, "loss": 1.1404, "step": 690 }, { "epoch": 0.595176571920758, "grad_norm": 0.9025394320487976, "learning_rate": 7.434274676228973e-06, "loss": 1.2, "step": 691 }, { "epoch": 0.5960378983634798, "grad_norm": 0.858193039894104, "learning_rate": 7.407318198589102e-06, "loss": 1.1964, "step": 692 }, { "epoch": 0.5968992248062015, "grad_norm": 0.8841907382011414, "learning_rate": 7.380381903307061e-06, "loss": 1.1968, "step": 693 }, { "epoch": 0.5977605512489234, "grad_norm": 0.8853846788406372, "learning_rate": 7.353466000064563e-06, "loss": 1.222, "step": 694 }, { "epoch": 0.5986218776916451, "grad_norm": 0.9315546154975891, "learning_rate": 7.326570698384569e-06, "loss": 1.1806, "step": 695 }, { "epoch": 0.599483204134367, "grad_norm": 0.8450863361358643, "learning_rate": 7.299696207629692e-06, "loss": 1.1658, "step": 696 }, { "epoch": 0.6003445305770887, "grad_norm": 0.8357823491096497, "learning_rate": 7.2728427370005205e-06, "loss": 1.21, "step": 697 }, { "epoch": 0.6012058570198106, "grad_norm": 0.8645495772361755, "learning_rate": 7.246010495534036e-06, "loss": 1.2054, "step": 698 }, { "epoch": 0.6020671834625323, "grad_norm": 0.9765591621398926, "learning_rate": 7.2191996921019485e-06, "loss": 1.2005, "step": 699 }, { "epoch": 0.602928509905254, "grad_norm": 0.8903208374977112, "learning_rate": 7.1924105354090955e-06, "loss": 1.1458, "step": 700 }, { "epoch": 0.6037898363479759, "grad_norm": 0.8536180853843689, "learning_rate": 7.165643233991806e-06, "loss": 1.1817, "step": 701 }, { "epoch": 0.6046511627906976, "grad_norm": 0.9730960726737976, "learning_rate": 7.138897996216278e-06, "loss": 1.1883, "step": 702 }, { "epoch": 0.6055124892334195, "grad_norm": 0.9186335802078247, "learning_rate": 7.1121750302769685e-06, "loss": 1.1818, "step": 703 }, { "epoch": 0.6063738156761412, "grad_norm": 0.8939233422279358, "learning_rate": 7.085474544194946e-06, "loss": 1.1446, "step": 704 }, { "epoch": 0.6072351421188631, "grad_norm": 0.9298137426376343, "learning_rate": 7.058796745816303e-06, "loss": 1.1787, "step": 705 }, { "epoch": 0.6080964685615848, "grad_norm": 0.8468939065933228, "learning_rate": 7.03214184281051e-06, "loss": 1.2003, "step": 706 }, { "epoch": 0.6089577950043066, "grad_norm": 0.8879836201667786, "learning_rate": 7.0055100426688205e-06, "loss": 1.1677, "step": 707 }, { "epoch": 0.6098191214470284, "grad_norm": 0.8849911093711853, "learning_rate": 6.978901552702643e-06, "loss": 1.2002, "step": 708 }, { "epoch": 0.6106804478897502, "grad_norm": 0.9499090313911438, "learning_rate": 6.95231658004194e-06, "loss": 1.2391, "step": 709 }, { "epoch": 0.611541774332472, "grad_norm": 0.9845093488693237, "learning_rate": 6.925755331633592e-06, "loss": 1.1686, "step": 710 }, { "epoch": 0.6124031007751938, "grad_norm": 0.8965116739273071, "learning_rate": 6.899218014239815e-06, "loss": 1.1862, "step": 711 }, { "epoch": 0.6132644272179156, "grad_norm": 0.8552091121673584, "learning_rate": 6.872704834436526e-06, "loss": 1.2219, "step": 712 }, { "epoch": 0.6141257536606374, "grad_norm": 0.853168785572052, "learning_rate": 6.846215998611757e-06, "loss": 1.1754, "step": 713 }, { "epoch": 0.6149870801033591, "grad_norm": 0.8855804204940796, "learning_rate": 6.8197517129640265e-06, "loss": 1.222, "step": 714 }, { "epoch": 0.615848406546081, "grad_norm": 1.0299283266067505, "learning_rate": 6.79331218350076e-06, "loss": 1.1431, "step": 715 }, { "epoch": 0.6167097329888027, "grad_norm": 0.9015288352966309, "learning_rate": 6.766897616036661e-06, "loss": 1.1595, "step": 716 }, { "epoch": 0.6175710594315246, "grad_norm": 0.9094573855400085, "learning_rate": 6.740508216192121e-06, "loss": 1.202, "step": 717 }, { "epoch": 0.6184323858742463, "grad_norm": 0.8764570355415344, "learning_rate": 6.714144189391625e-06, "loss": 1.1975, "step": 718 }, { "epoch": 0.6192937123169682, "grad_norm": 0.8982758522033691, "learning_rate": 6.6878057408621345e-06, "loss": 1.1991, "step": 719 }, { "epoch": 0.6201550387596899, "grad_norm": 0.8324359059333801, "learning_rate": 6.661493075631506e-06, "loss": 1.1908, "step": 720 }, { "epoch": 0.6210163652024118, "grad_norm": 0.8555594682693481, "learning_rate": 6.635206398526895e-06, "loss": 1.1635, "step": 721 }, { "epoch": 0.6218776916451335, "grad_norm": 0.9507735371589661, "learning_rate": 6.60894591417315e-06, "loss": 1.2447, "step": 722 }, { "epoch": 0.6227390180878553, "grad_norm": 0.8586412072181702, "learning_rate": 6.582711826991226e-06, "loss": 1.1975, "step": 723 }, { "epoch": 0.6236003445305771, "grad_norm": 0.8518699407577515, "learning_rate": 6.556504341196592e-06, "loss": 1.2073, "step": 724 }, { "epoch": 0.6244616709732989, "grad_norm": 0.8891175389289856, "learning_rate": 6.5303236607976465e-06, "loss": 1.1805, "step": 725 }, { "epoch": 0.6253229974160207, "grad_norm": 1.035131812095642, "learning_rate": 6.504169989594121e-06, "loss": 1.1953, "step": 726 }, { "epoch": 0.6261843238587425, "grad_norm": 1.0852066278457642, "learning_rate": 6.4780435311754986e-06, "loss": 1.1504, "step": 727 }, { "epoch": 0.6270456503014643, "grad_norm": 0.848870575428009, "learning_rate": 6.451944488919433e-06, "loss": 1.1743, "step": 728 }, { "epoch": 0.627906976744186, "grad_norm": 0.8880192041397095, "learning_rate": 6.425873065990158e-06, "loss": 1.156, "step": 729 }, { "epoch": 0.6287683031869078, "grad_norm": 0.9287259578704834, "learning_rate": 6.3998294653369046e-06, "loss": 1.2049, "step": 730 }, { "epoch": 0.6296296296296297, "grad_norm": 0.9466278553009033, "learning_rate": 6.373813889692331e-06, "loss": 1.209, "step": 731 }, { "epoch": 0.6304909560723514, "grad_norm": 0.8213219046592712, "learning_rate": 6.347826541570936e-06, "loss": 1.1807, "step": 732 }, { "epoch": 0.6313522825150732, "grad_norm": 0.8916311860084534, "learning_rate": 6.3218676232674815e-06, "loss": 1.1974, "step": 733 }, { "epoch": 0.632213608957795, "grad_norm": 0.8614788055419922, "learning_rate": 6.295937336855433e-06, "loss": 1.1858, "step": 734 }, { "epoch": 0.6330749354005168, "grad_norm": 0.8437885642051697, "learning_rate": 6.270035884185367e-06, "loss": 1.1858, "step": 735 }, { "epoch": 0.6339362618432386, "grad_norm": 0.8793601989746094, "learning_rate": 6.244163466883405e-06, "loss": 1.1722, "step": 736 }, { "epoch": 0.6347975882859603, "grad_norm": 0.9624020457267761, "learning_rate": 6.218320286349655e-06, "loss": 1.1796, "step": 737 }, { "epoch": 0.6356589147286822, "grad_norm": 0.8338882923126221, "learning_rate": 6.192506543756626e-06, "loss": 1.2217, "step": 738 }, { "epoch": 0.6365202411714039, "grad_norm": 0.8388656377792358, "learning_rate": 6.1667224400476785e-06, "loss": 1.1981, "step": 739 }, { "epoch": 0.6373815676141258, "grad_norm": 0.8913929462432861, "learning_rate": 6.140968175935458e-06, "loss": 1.1906, "step": 740 }, { "epoch": 0.6382428940568475, "grad_norm": 0.8406807780265808, "learning_rate": 6.115243951900316e-06, "loss": 1.2023, "step": 741 }, { "epoch": 0.6391042204995694, "grad_norm": 0.9768598079681396, "learning_rate": 6.089549968188767e-06, "loss": 1.1909, "step": 742 }, { "epoch": 0.6399655469422911, "grad_norm": 0.9822099804878235, "learning_rate": 6.063886424811929e-06, "loss": 1.1945, "step": 743 }, { "epoch": 0.6408268733850129, "grad_norm": 0.8699774742126465, "learning_rate": 6.038253521543951e-06, "loss": 1.1947, "step": 744 }, { "epoch": 0.6416881998277347, "grad_norm": 0.875484049320221, "learning_rate": 6.01265145792048e-06, "loss": 1.1623, "step": 745 }, { "epoch": 0.6425495262704565, "grad_norm": 0.8913766145706177, "learning_rate": 5.987080433237082e-06, "loss": 1.1771, "step": 746 }, { "epoch": 0.6434108527131783, "grad_norm": 0.9033095240592957, "learning_rate": 5.961540646547722e-06, "loss": 1.1619, "step": 747 }, { "epoch": 0.6442721791559001, "grad_norm": 0.9094845056533813, "learning_rate": 5.936032296663188e-06, "loss": 1.1788, "step": 748 }, { "epoch": 0.6451335055986219, "grad_norm": 0.8750199675559998, "learning_rate": 5.9105555821495486e-06, "loss": 1.1971, "step": 749 }, { "epoch": 0.6459948320413437, "grad_norm": 0.8594929575920105, "learning_rate": 5.885110701326621e-06, "loss": 1.1961, "step": 750 }, { "epoch": 0.6468561584840654, "grad_norm": 0.8807885050773621, "learning_rate": 5.859697852266409e-06, "loss": 1.1702, "step": 751 }, { "epoch": 0.6477174849267873, "grad_norm": 0.8879327178001404, "learning_rate": 5.834317232791575e-06, "loss": 1.2105, "step": 752 }, { "epoch": 0.648578811369509, "grad_norm": 0.8946269154548645, "learning_rate": 5.808969040473893e-06, "loss": 1.1466, "step": 753 }, { "epoch": 0.6494401378122309, "grad_norm": 0.8713740706443787, "learning_rate": 5.7836534726327175e-06, "loss": 1.1555, "step": 754 }, { "epoch": 0.6503014642549526, "grad_norm": 0.9063122272491455, "learning_rate": 5.758370726333434e-06, "loss": 1.2304, "step": 755 }, { "epoch": 0.6511627906976745, "grad_norm": 0.9117863774299622, "learning_rate": 5.733120998385935e-06, "loss": 1.1567, "step": 756 }, { "epoch": 0.6520241171403962, "grad_norm": 0.8670936822891235, "learning_rate": 5.707904485343094e-06, "loss": 1.1848, "step": 757 }, { "epoch": 0.652885443583118, "grad_norm": 0.8828465342521667, "learning_rate": 5.682721383499217e-06, "loss": 1.1945, "step": 758 }, { "epoch": 0.6537467700258398, "grad_norm": 0.8751693964004517, "learning_rate": 5.657571888888538e-06, "loss": 1.1815, "step": 759 }, { "epoch": 0.6546080964685616, "grad_norm": 0.9212945699691772, "learning_rate": 5.63245619728367e-06, "loss": 1.1823, "step": 760 }, { "epoch": 0.6554694229112834, "grad_norm": 0.9038519263267517, "learning_rate": 5.60737450419409e-06, "loss": 1.1918, "step": 761 }, { "epoch": 0.6563307493540051, "grad_norm": 0.9209284782409668, "learning_rate": 5.582327004864631e-06, "loss": 1.1645, "step": 762 }, { "epoch": 0.657192075796727, "grad_norm": 0.8759058713912964, "learning_rate": 5.557313894273937e-06, "loss": 1.1895, "step": 763 }, { "epoch": 0.6580534022394487, "grad_norm": 1.0104118585586548, "learning_rate": 5.532335367132962e-06, "loss": 1.1919, "step": 764 }, { "epoch": 0.6589147286821705, "grad_norm": 0.8833430409431458, "learning_rate": 5.507391617883454e-06, "loss": 1.2118, "step": 765 }, { "epoch": 0.6597760551248923, "grad_norm": 0.9313235878944397, "learning_rate": 5.4824828406964305e-06, "loss": 1.1817, "step": 766 }, { "epoch": 0.6606373815676141, "grad_norm": 0.8975321054458618, "learning_rate": 5.457609229470681e-06, "loss": 1.182, "step": 767 }, { "epoch": 0.661498708010336, "grad_norm": 0.8778689503669739, "learning_rate": 5.4327709778312484e-06, "loss": 1.1567, "step": 768 }, { "epoch": 0.6623600344530577, "grad_norm": 0.9323556423187256, "learning_rate": 5.407968279127915e-06, "loss": 1.1592, "step": 769 }, { "epoch": 0.6632213608957795, "grad_norm": 0.8266853094100952, "learning_rate": 5.383201326433727e-06, "loss": 1.1793, "step": 770 }, { "epoch": 0.6640826873385013, "grad_norm": 0.8756425976753235, "learning_rate": 5.358470312543445e-06, "loss": 1.1877, "step": 771 }, { "epoch": 0.664944013781223, "grad_norm": 0.8671419024467468, "learning_rate": 5.3337754299720925e-06, "loss": 1.2267, "step": 772 }, { "epoch": 0.6658053402239449, "grad_norm": 0.9318386912345886, "learning_rate": 5.30911687095342e-06, "loss": 1.1453, "step": 773 }, { "epoch": 0.6666666666666666, "grad_norm": 0.8705496191978455, "learning_rate": 5.284494827438423e-06, "loss": 1.1715, "step": 774 }, { "epoch": 0.6675279931093885, "grad_norm": 0.8844230771064758, "learning_rate": 5.25990949109386e-06, "loss": 1.1755, "step": 775 }, { "epoch": 0.6683893195521102, "grad_norm": 0.8812039494514465, "learning_rate": 5.2353610533007305e-06, "loss": 1.1888, "step": 776 }, { "epoch": 0.6692506459948321, "grad_norm": 0.9645406603813171, "learning_rate": 5.210849705152809e-06, "loss": 1.1828, "step": 777 }, { "epoch": 0.6701119724375538, "grad_norm": 0.9021162390708923, "learning_rate": 5.186375637455159e-06, "loss": 1.158, "step": 778 }, { "epoch": 0.6709732988802756, "grad_norm": 0.9339265823364258, "learning_rate": 5.161939040722634e-06, "loss": 1.1629, "step": 779 }, { "epoch": 0.6718346253229974, "grad_norm": 0.8927223682403564, "learning_rate": 5.137540105178396e-06, "loss": 1.1745, "step": 780 }, { "epoch": 0.6726959517657192, "grad_norm": 0.9335219264030457, "learning_rate": 5.113179020752443e-06, "loss": 1.1858, "step": 781 }, { "epoch": 0.673557278208441, "grad_norm": 0.8412618041038513, "learning_rate": 5.088855977080123e-06, "loss": 1.1759, "step": 782 }, { "epoch": 0.6744186046511628, "grad_norm": 0.9003159999847412, "learning_rate": 5.064571163500667e-06, "loss": 1.1958, "step": 783 }, { "epoch": 0.6752799310938846, "grad_norm": 0.9444640278816223, "learning_rate": 5.040324769055709e-06, "loss": 1.1988, "step": 784 }, { "epoch": 0.6761412575366064, "grad_norm": 0.833273708820343, "learning_rate": 5.016116982487811e-06, "loss": 1.1886, "step": 785 }, { "epoch": 0.6770025839793282, "grad_norm": 0.9385854005813599, "learning_rate": 4.991947992238997e-06, "loss": 1.1769, "step": 786 }, { "epoch": 0.67786391042205, "grad_norm": 0.8889097571372986, "learning_rate": 4.967817986449284e-06, "loss": 1.1898, "step": 787 }, { "epoch": 0.6787252368647717, "grad_norm": 0.8416062593460083, "learning_rate": 4.943727152955235e-06, "loss": 1.1899, "step": 788 }, { "epoch": 0.6795865633074936, "grad_norm": 0.9640506505966187, "learning_rate": 4.9196756792884605e-06, "loss": 1.1683, "step": 789 }, { "epoch": 0.6804478897502153, "grad_norm": 0.8979779481887817, "learning_rate": 4.8956637526742e-06, "loss": 1.1926, "step": 790 }, { "epoch": 0.6813092161929372, "grad_norm": 0.8395998477935791, "learning_rate": 4.871691560029828e-06, "loss": 1.1593, "step": 791 }, { "epoch": 0.6821705426356589, "grad_norm": 0.8391218185424805, "learning_rate": 4.847759287963432e-06, "loss": 1.1615, "step": 792 }, { "epoch": 0.6830318690783808, "grad_norm": 0.8740763068199158, "learning_rate": 4.8238671227723285e-06, "loss": 1.1717, "step": 793 }, { "epoch": 0.6838931955211025, "grad_norm": 0.8377690315246582, "learning_rate": 4.800015250441638e-06, "loss": 1.1757, "step": 794 }, { "epoch": 0.6847545219638242, "grad_norm": 0.8415879011154175, "learning_rate": 4.7762038566428155e-06, "loss": 1.2087, "step": 795 }, { "epoch": 0.6856158484065461, "grad_norm": 0.8535274863243103, "learning_rate": 4.752433126732231e-06, "loss": 1.1714, "step": 796 }, { "epoch": 0.6864771748492678, "grad_norm": 0.8958263397216797, "learning_rate": 4.728703245749707e-06, "loss": 1.1477, "step": 797 }, { "epoch": 0.6873385012919897, "grad_norm": 0.9385268092155457, "learning_rate": 4.7050143984170805e-06, "loss": 1.2119, "step": 798 }, { "epoch": 0.6881998277347114, "grad_norm": 0.9213505983352661, "learning_rate": 4.681366769136769e-06, "loss": 1.1759, "step": 799 }, { "epoch": 0.6890611541774333, "grad_norm": 0.9221978187561035, "learning_rate": 4.657760541990329e-06, "loss": 1.1704, "step": 800 }, { "epoch": 0.689922480620155, "grad_norm": 0.8620073795318604, "learning_rate": 4.634195900737045e-06, "loss": 1.197, "step": 801 }, { "epoch": 0.6907838070628768, "grad_norm": 0.8191630244255066, "learning_rate": 4.610673028812459e-06, "loss": 1.1385, "step": 802 }, { "epoch": 0.6916451335055986, "grad_norm": 0.8765779137611389, "learning_rate": 4.587192109326988e-06, "loss": 1.1687, "step": 803 }, { "epoch": 0.6925064599483204, "grad_norm": 0.7997327446937561, "learning_rate": 4.563753325064457e-06, "loss": 1.1897, "step": 804 }, { "epoch": 0.6933677863910422, "grad_norm": 0.8961016535758972, "learning_rate": 4.540356858480711e-06, "loss": 1.1849, "step": 805 }, { "epoch": 0.694229112833764, "grad_norm": 0.8605460524559021, "learning_rate": 4.5170028917021705e-06, "loss": 1.1793, "step": 806 }, { "epoch": 0.6950904392764858, "grad_norm": 0.9754102826118469, "learning_rate": 4.493691606524423e-06, "loss": 1.2003, "step": 807 }, { "epoch": 0.6959517657192076, "grad_norm": 0.7908036708831787, "learning_rate": 4.470423184410804e-06, "loss": 1.1333, "step": 808 }, { "epoch": 0.6968130921619293, "grad_norm": 0.8532270789146423, "learning_rate": 4.447197806490996e-06, "loss": 1.1468, "step": 809 }, { "epoch": 0.6976744186046512, "grad_norm": 0.8650361895561218, "learning_rate": 4.424015653559611e-06, "loss": 1.1958, "step": 810 }, { "epoch": 0.6985357450473729, "grad_norm": 1.0336636304855347, "learning_rate": 4.400876906074772e-06, "loss": 1.2279, "step": 811 }, { "epoch": 0.6993970714900948, "grad_norm": 0.8429791331291199, "learning_rate": 4.377781744156727e-06, "loss": 1.2041, "step": 812 }, { "epoch": 0.7002583979328165, "grad_norm": 0.8677304983139038, "learning_rate": 4.35473034758643e-06, "loss": 1.1873, "step": 813 }, { "epoch": 0.7011197243755384, "grad_norm": 0.9752134680747986, "learning_rate": 4.331722895804158e-06, "loss": 1.1734, "step": 814 }, { "epoch": 0.7019810508182601, "grad_norm": 0.8603991270065308, "learning_rate": 4.30875956790811e-06, "loss": 1.1984, "step": 815 }, { "epoch": 0.7028423772609819, "grad_norm": 0.8431304693222046, "learning_rate": 4.285840542652997e-06, "loss": 1.1635, "step": 816 }, { "epoch": 0.7037037037037037, "grad_norm": 0.8751398324966431, "learning_rate": 4.262965998448665e-06, "loss": 1.1715, "step": 817 }, { "epoch": 0.7045650301464255, "grad_norm": 0.8455151915550232, "learning_rate": 4.240136113358714e-06, "loss": 1.212, "step": 818 }, { "epoch": 0.7054263565891473, "grad_norm": 0.9403558373451233, "learning_rate": 4.2173510650990905e-06, "loss": 1.1627, "step": 819 }, { "epoch": 0.7062876830318691, "grad_norm": 0.941561222076416, "learning_rate": 4.194611031036718e-06, "loss": 1.1861, "step": 820 }, { "epoch": 0.7071490094745909, "grad_norm": 0.9220948815345764, "learning_rate": 4.171916188188113e-06, "loss": 1.1696, "step": 821 }, { "epoch": 0.7080103359173127, "grad_norm": 0.8759995698928833, "learning_rate": 4.149266713218011e-06, "loss": 1.1863, "step": 822 }, { "epoch": 0.7088716623600344, "grad_norm": 0.8697245717048645, "learning_rate": 4.12666278243799e-06, "loss": 1.1579, "step": 823 }, { "epoch": 0.7097329888027563, "grad_norm": 0.8867110013961792, "learning_rate": 4.104104571805088e-06, "loss": 1.1722, "step": 824 }, { "epoch": 0.710594315245478, "grad_norm": 0.8431983590126038, "learning_rate": 4.0815922569204435e-06, "loss": 1.1983, "step": 825 }, { "epoch": 0.7114556416881999, "grad_norm": 0.891090452671051, "learning_rate": 4.059126013027924e-06, "loss": 1.1585, "step": 826 }, { "epoch": 0.7123169681309216, "grad_norm": 0.8412421345710754, "learning_rate": 4.0367060150127726e-06, "loss": 1.1664, "step": 827 }, { "epoch": 0.7131782945736435, "grad_norm": 0.8604631423950195, "learning_rate": 4.014332437400235e-06, "loss": 1.163, "step": 828 }, { "epoch": 0.7140396210163652, "grad_norm": 0.9067566394805908, "learning_rate": 3.9920054543541976e-06, "loss": 1.1838, "step": 829 }, { "epoch": 0.7149009474590869, "grad_norm": 0.8477993607521057, "learning_rate": 3.96972523967584e-06, "loss": 1.1958, "step": 830 }, { "epoch": 0.7157622739018088, "grad_norm": 0.8343601226806641, "learning_rate": 3.947491966802288e-06, "loss": 1.1784, "step": 831 }, { "epoch": 0.7166236003445305, "grad_norm": 0.8563992381095886, "learning_rate": 3.925305808805247e-06, "loss": 1.1488, "step": 832 }, { "epoch": 0.7174849267872524, "grad_norm": 0.9131225943565369, "learning_rate": 3.903166938389664e-06, "loss": 1.1762, "step": 833 }, { "epoch": 0.7183462532299741, "grad_norm": 0.872132420539856, "learning_rate": 3.881075527892391e-06, "loss": 1.1717, "step": 834 }, { "epoch": 0.719207579672696, "grad_norm": 0.8504886627197266, "learning_rate": 3.859031749280824e-06, "loss": 1.1479, "step": 835 }, { "epoch": 0.7200689061154177, "grad_norm": 0.8588838577270508, "learning_rate": 3.837035774151585e-06, "loss": 1.19, "step": 836 }, { "epoch": 0.7209302325581395, "grad_norm": 0.8740001916885376, "learning_rate": 3.815087773729171e-06, "loss": 1.186, "step": 837 }, { "epoch": 0.7217915590008613, "grad_norm": 0.8781384825706482, "learning_rate": 3.793187918864627e-06, "loss": 1.1913, "step": 838 }, { "epoch": 0.7226528854435831, "grad_norm": 0.8898410201072693, "learning_rate": 3.771336380034211e-06, "loss": 1.1711, "step": 839 }, { "epoch": 0.7235142118863049, "grad_norm": 0.8932210206985474, "learning_rate": 3.749533327338091e-06, "loss": 1.1818, "step": 840 }, { "epoch": 0.7243755383290267, "grad_norm": 0.8856852650642395, "learning_rate": 3.727778930498982e-06, "loss": 1.1913, "step": 841 }, { "epoch": 0.7252368647717485, "grad_norm": 0.8199148774147034, "learning_rate": 3.706073358860851e-06, "loss": 1.1862, "step": 842 }, { "epoch": 0.7260981912144703, "grad_norm": 0.8369928002357483, "learning_rate": 3.6844167813875888e-06, "loss": 1.2055, "step": 843 }, { "epoch": 0.726959517657192, "grad_norm": 0.8797032237052917, "learning_rate": 3.6628093666617083e-06, "loss": 1.1767, "step": 844 }, { "epoch": 0.7278208440999139, "grad_norm": 0.8433752059936523, "learning_rate": 3.6412512828830114e-06, "loss": 1.1601, "step": 845 }, { "epoch": 0.7286821705426356, "grad_norm": 0.8134076595306396, "learning_rate": 3.6197426978672923e-06, "loss": 1.1797, "step": 846 }, { "epoch": 0.7295434969853575, "grad_norm": 0.8705214858055115, "learning_rate": 3.5982837790450376e-06, "loss": 1.1506, "step": 847 }, { "epoch": 0.7304048234280792, "grad_norm": 0.9077462553977966, "learning_rate": 3.5768746934601007e-06, "loss": 1.1796, "step": 848 }, { "epoch": 0.7312661498708011, "grad_norm": 0.8746139407157898, "learning_rate": 3.5555156077684317e-06, "loss": 1.1535, "step": 849 }, { "epoch": 0.7321274763135228, "grad_norm": 0.9012914896011353, "learning_rate": 3.5342066882367488e-06, "loss": 1.2069, "step": 850 }, { "epoch": 0.7329888027562446, "grad_norm": 0.8886313438415527, "learning_rate": 3.5129481007412634e-06, "loss": 1.1654, "step": 851 }, { "epoch": 0.7338501291989664, "grad_norm": 0.8336722254753113, "learning_rate": 3.491740010766389e-06, "loss": 1.1773, "step": 852 }, { "epoch": 0.7347114556416882, "grad_norm": 0.8749405741691589, "learning_rate": 3.47058258340345e-06, "loss": 1.18, "step": 853 }, { "epoch": 0.73557278208441, "grad_norm": 0.8950448632240295, "learning_rate": 3.449475983349385e-06, "loss": 1.1257, "step": 854 }, { "epoch": 0.7364341085271318, "grad_norm": 0.8262083530426025, "learning_rate": 3.4284203749054834e-06, "loss": 1.1863, "step": 855 }, { "epoch": 0.7372954349698536, "grad_norm": 0.8869133591651917, "learning_rate": 3.4074159219760884e-06, "loss": 1.1659, "step": 856 }, { "epoch": 0.7381567614125754, "grad_norm": 0.8032558560371399, "learning_rate": 3.386462788067344e-06, "loss": 1.1862, "step": 857 }, { "epoch": 0.7390180878552972, "grad_norm": 0.8882765173912048, "learning_rate": 3.3655611362858966e-06, "loss": 1.1694, "step": 858 }, { "epoch": 0.739879414298019, "grad_norm": 0.9683348536491394, "learning_rate": 3.3447111293376468e-06, "loss": 1.1428, "step": 859 }, { "epoch": 0.7407407407407407, "grad_norm": 0.823493480682373, "learning_rate": 3.323912929526465e-06, "loss": 1.1393, "step": 860 }, { "epoch": 0.7416020671834626, "grad_norm": 0.8513652682304382, "learning_rate": 3.30316669875294e-06, "loss": 1.1785, "step": 861 }, { "epoch": 0.7424633936261843, "grad_norm": 0.8614821434020996, "learning_rate": 3.282472598513119e-06, "loss": 1.1626, "step": 862 }, { "epoch": 0.7433247200689062, "grad_norm": 0.8872314095497131, "learning_rate": 3.2618307898972413e-06, "loss": 1.1466, "step": 863 }, { "epoch": 0.7441860465116279, "grad_norm": 0.8613788485527039, "learning_rate": 3.2412414335884866e-06, "loss": 1.1586, "step": 864 }, { "epoch": 0.7450473729543498, "grad_norm": 0.8109983205795288, "learning_rate": 3.2207046898617365e-06, "loss": 1.1701, "step": 865 }, { "epoch": 0.7459086993970715, "grad_norm": 0.9684152603149414, "learning_rate": 3.2002207185823155e-06, "loss": 1.1888, "step": 866 }, { "epoch": 0.7467700258397932, "grad_norm": 1.0356762409210205, "learning_rate": 3.1797896792047422e-06, "loss": 1.2229, "step": 867 }, { "epoch": 0.7476313522825151, "grad_norm": 0.9049926996231079, "learning_rate": 3.1594117307714977e-06, "loss": 1.2125, "step": 868 }, { "epoch": 0.7484926787252368, "grad_norm": 0.9765392541885376, "learning_rate": 3.1390870319117838e-06, "loss": 1.1854, "step": 869 }, { "epoch": 0.7493540051679587, "grad_norm": 0.8243566155433655, "learning_rate": 3.118815740840294e-06, "loss": 1.1737, "step": 870 }, { "epoch": 0.7502153316106804, "grad_norm": 0.8770548105239868, "learning_rate": 3.098598015355967e-06, "loss": 1.1927, "step": 871 }, { "epoch": 0.7510766580534023, "grad_norm": 0.8610149025917053, "learning_rate": 3.0784340128407786e-06, "loss": 1.1686, "step": 872 }, { "epoch": 0.751937984496124, "grad_norm": 1.0247671604156494, "learning_rate": 3.0583238902584976e-06, "loss": 1.1861, "step": 873 }, { "epoch": 0.7527993109388458, "grad_norm": 0.8867828845977783, "learning_rate": 3.038267804153472e-06, "loss": 1.1948, "step": 874 }, { "epoch": 0.7536606373815676, "grad_norm": 0.902263879776001, "learning_rate": 3.0182659106494195e-06, "loss": 1.1778, "step": 875 }, { "epoch": 0.7545219638242894, "grad_norm": 0.9579049348831177, "learning_rate": 2.998318365448194e-06, "loss": 1.188, "step": 876 }, { "epoch": 0.7553832902670112, "grad_norm": 0.8788950443267822, "learning_rate": 2.9784253238285844e-06, "loss": 1.1451, "step": 877 }, { "epoch": 0.756244616709733, "grad_norm": 0.9002311825752258, "learning_rate": 2.9585869406451083e-06, "loss": 1.1493, "step": 878 }, { "epoch": 0.7571059431524548, "grad_norm": 0.9176036715507507, "learning_rate": 2.938803370326804e-06, "loss": 1.1789, "step": 879 }, { "epoch": 0.7579672695951766, "grad_norm": 0.87285977602005, "learning_rate": 2.9190747668760213e-06, "loss": 1.1633, "step": 880 }, { "epoch": 0.7588285960378983, "grad_norm": 0.8504787087440491, "learning_rate": 2.899401283867229e-06, "loss": 1.1753, "step": 881 }, { "epoch": 0.7596899224806202, "grad_norm": 0.8374565243721008, "learning_rate": 2.8797830744458177e-06, "loss": 1.168, "step": 882 }, { "epoch": 0.7605512489233419, "grad_norm": 0.8359323143959045, "learning_rate": 2.860220291326915e-06, "loss": 1.1788, "step": 883 }, { "epoch": 0.7614125753660638, "grad_norm": 0.8868809342384338, "learning_rate": 2.840713086794189e-06, "loss": 1.1907, "step": 884 }, { "epoch": 0.7622739018087855, "grad_norm": 0.874671220779419, "learning_rate": 2.8212616126986604e-06, "loss": 1.1735, "step": 885 }, { "epoch": 0.7631352282515074, "grad_norm": 0.9140809774398804, "learning_rate": 2.801866020457521e-06, "loss": 1.1776, "step": 886 }, { "epoch": 0.7639965546942291, "grad_norm": 0.8395150303840637, "learning_rate": 2.7825264610529703e-06, "loss": 1.1691, "step": 887 }, { "epoch": 0.7648578811369509, "grad_norm": 0.9220564961433411, "learning_rate": 2.7632430850310175e-06, "loss": 1.1921, "step": 888 }, { "epoch": 0.7657192075796727, "grad_norm": 0.8539912104606628, "learning_rate": 2.7440160425003236e-06, "loss": 1.1454, "step": 889 }, { "epoch": 0.7665805340223945, "grad_norm": 0.8318409323692322, "learning_rate": 2.7248454831310335e-06, "loss": 1.1648, "step": 890 }, { "epoch": 0.7674418604651163, "grad_norm": 0.8465113639831543, "learning_rate": 2.7057315561536e-06, "loss": 1.1529, "step": 891 }, { "epoch": 0.768303186907838, "grad_norm": 0.8205885887145996, "learning_rate": 2.686674410357637e-06, "loss": 1.2069, "step": 892 }, { "epoch": 0.7691645133505599, "grad_norm": 0.9047953486442566, "learning_rate": 2.667674194090748e-06, "loss": 1.1727, "step": 893 }, { "epoch": 0.7700258397932817, "grad_norm": 0.928527295589447, "learning_rate": 2.6487310552573776e-06, "loss": 1.1687, "step": 894 }, { "epoch": 0.7708871662360034, "grad_norm": 0.9136389493942261, "learning_rate": 2.6298451413176564e-06, "loss": 1.1628, "step": 895 }, { "epoch": 0.7717484926787253, "grad_norm": 0.8879351615905762, "learning_rate": 2.6110165992862635e-06, "loss": 1.1576, "step": 896 }, { "epoch": 0.772609819121447, "grad_norm": 0.8699619770050049, "learning_rate": 2.592245575731274e-06, "loss": 1.2179, "step": 897 }, { "epoch": 0.7734711455641688, "grad_norm": 0.9080535173416138, "learning_rate": 2.5735322167730116e-06, "loss": 1.1618, "step": 898 }, { "epoch": 0.7743324720068906, "grad_norm": 0.8573684096336365, "learning_rate": 2.5548766680829207e-06, "loss": 1.141, "step": 899 }, { "epoch": 0.7751937984496124, "grad_norm": 0.9463301301002502, "learning_rate": 2.5362790748824363e-06, "loss": 1.1608, "step": 900 }, { "epoch": 0.7760551248923342, "grad_norm": 0.8480750918388367, "learning_rate": 2.517739581941839e-06, "loss": 1.1811, "step": 901 }, { "epoch": 0.7769164513350559, "grad_norm": 0.9230459928512573, "learning_rate": 2.4992583335791375e-06, "loss": 1.1404, "step": 902 }, { "epoch": 0.7777777777777778, "grad_norm": 0.8646366000175476, "learning_rate": 2.4808354736589525e-06, "loss": 1.1808, "step": 903 }, { "epoch": 0.7786391042204995, "grad_norm": 0.8544058203697205, "learning_rate": 2.4624711455913764e-06, "loss": 1.1333, "step": 904 }, { "epoch": 0.7795004306632214, "grad_norm": 0.8751500248908997, "learning_rate": 2.444165492330879e-06, "loss": 1.1674, "step": 905 }, { "epoch": 0.7803617571059431, "grad_norm": 0.8781688213348389, "learning_rate": 2.425918656375177e-06, "loss": 1.1802, "step": 906 }, { "epoch": 0.781223083548665, "grad_norm": 1.0020009279251099, "learning_rate": 2.4077307797641357e-06, "loss": 1.1823, "step": 907 }, { "epoch": 0.7820844099913867, "grad_norm": 0.7854005098342896, "learning_rate": 2.389602004078657e-06, "loss": 1.1864, "step": 908 }, { "epoch": 0.7829457364341085, "grad_norm": 0.8276894688606262, "learning_rate": 2.3715324704395846e-06, "loss": 1.1667, "step": 909 }, { "epoch": 0.7838070628768303, "grad_norm": 0.8462375402450562, "learning_rate": 2.3535223195066025e-06, "loss": 1.1318, "step": 910 }, { "epoch": 0.7846683893195521, "grad_norm": 0.8951948881149292, "learning_rate": 2.335571691477132e-06, "loss": 1.189, "step": 911 }, { "epoch": 0.7855297157622739, "grad_norm": 0.8499693870544434, "learning_rate": 2.3176807260852475e-06, "loss": 1.1322, "step": 912 }, { "epoch": 0.7863910422049957, "grad_norm": 0.8753166198730469, "learning_rate": 2.2998495626005955e-06, "loss": 1.186, "step": 913 }, { "epoch": 0.7872523686477175, "grad_norm": 0.8034504652023315, "learning_rate": 2.282078339827293e-06, "loss": 1.174, "step": 914 }, { "epoch": 0.7881136950904393, "grad_norm": 0.8719635605812073, "learning_rate": 2.264367196102869e-06, "loss": 1.2179, "step": 915 }, { "epoch": 0.788975021533161, "grad_norm": 0.8509535193443298, "learning_rate": 2.2467162692971655e-06, "loss": 1.1951, "step": 916 }, { "epoch": 0.7898363479758829, "grad_norm": 0.9040639996528625, "learning_rate": 2.229125696811275e-06, "loss": 1.2126, "step": 917 }, { "epoch": 0.7906976744186046, "grad_norm": 0.9148096442222595, "learning_rate": 2.2115956155764817e-06, "loss": 1.1558, "step": 918 }, { "epoch": 0.7915590008613265, "grad_norm": 0.8161072134971619, "learning_rate": 2.1941261620531718e-06, "loss": 1.1379, "step": 919 }, { "epoch": 0.7924203273040482, "grad_norm": 0.8694244027137756, "learning_rate": 2.17671747222979e-06, "loss": 1.1576, "step": 920 }, { "epoch": 0.7932816537467701, "grad_norm": 0.830352246761322, "learning_rate": 2.1593696816217667e-06, "loss": 1.1848, "step": 921 }, { "epoch": 0.7941429801894918, "grad_norm": 0.8742688298225403, "learning_rate": 2.142082925270489e-06, "loss": 1.1939, "step": 922 }, { "epoch": 0.7950043066322137, "grad_norm": 0.8883207440376282, "learning_rate": 2.1248573377422155e-06, "loss": 1.1894, "step": 923 }, { "epoch": 0.7958656330749354, "grad_norm": 0.9029132127761841, "learning_rate": 2.107693053127049e-06, "loss": 1.1283, "step": 924 }, { "epoch": 0.7967269595176572, "grad_norm": 0.9062528014183044, "learning_rate": 2.090590205037888e-06, "loss": 1.2179, "step": 925 }, { "epoch": 0.797588285960379, "grad_norm": 0.8248071670532227, "learning_rate": 2.0735489266093923e-06, "loss": 1.1756, "step": 926 }, { "epoch": 0.7984496124031008, "grad_norm": 0.8426958918571472, "learning_rate": 2.056569350496933e-06, "loss": 1.1924, "step": 927 }, { "epoch": 0.7993109388458226, "grad_norm": 0.8882315158843994, "learning_rate": 2.0396516088755804e-06, "loss": 1.1843, "step": 928 }, { "epoch": 0.8001722652885443, "grad_norm": 0.9714380502700806, "learning_rate": 2.0227958334390506e-06, "loss": 1.1617, "step": 929 }, { "epoch": 0.8010335917312662, "grad_norm": 0.9388904571533203, "learning_rate": 2.0060021553986974e-06, "loss": 1.1715, "step": 930 }, { "epoch": 0.801894918173988, "grad_norm": 0.9136961698532104, "learning_rate": 1.989270705482492e-06, "loss": 1.1402, "step": 931 }, { "epoch": 0.8027562446167097, "grad_norm": 0.9584447145462036, "learning_rate": 1.9726016139339934e-06, "loss": 1.1393, "step": 932 }, { "epoch": 0.8036175710594315, "grad_norm": 0.8596916198730469, "learning_rate": 1.955995010511338e-06, "loss": 1.1759, "step": 933 }, { "epoch": 0.8044788975021533, "grad_norm": 0.8455019593238831, "learning_rate": 1.9394510244862397e-06, "loss": 1.1616, "step": 934 }, { "epoch": 0.8053402239448751, "grad_norm": 0.9066017270088196, "learning_rate": 1.9229697846429773e-06, "loss": 1.1546, "step": 935 }, { "epoch": 0.8062015503875969, "grad_norm": 0.8392831683158875, "learning_rate": 1.9065514192773848e-06, "loss": 1.1998, "step": 936 }, { "epoch": 0.8070628768303187, "grad_norm": 0.9439470767974854, "learning_rate": 1.8901960561958588e-06, "loss": 1.1673, "step": 937 }, { "epoch": 0.8079242032730405, "grad_norm": 0.8691911697387695, "learning_rate": 1.8739038227143658e-06, "loss": 1.1632, "step": 938 }, { "epoch": 0.8087855297157622, "grad_norm": 0.8843985199928284, "learning_rate": 1.8576748456574512e-06, "loss": 1.1743, "step": 939 }, { "epoch": 0.8096468561584841, "grad_norm": 0.8459395170211792, "learning_rate": 1.8415092513572498e-06, "loss": 1.1606, "step": 940 }, { "epoch": 0.8105081826012058, "grad_norm": 0.8578802943229675, "learning_rate": 1.8254071656524997e-06, "loss": 1.1418, "step": 941 }, { "epoch": 0.8113695090439277, "grad_norm": 0.8318899869918823, "learning_rate": 1.8093687138875648e-06, "loss": 1.1623, "step": 942 }, { "epoch": 0.8122308354866494, "grad_norm": 0.9500533938407898, "learning_rate": 1.7933940209114597e-06, "loss": 1.1926, "step": 943 }, { "epoch": 0.8130921619293713, "grad_norm": 0.8835731148719788, "learning_rate": 1.7774832110768847e-06, "loss": 1.1874, "step": 944 }, { "epoch": 0.813953488372093, "grad_norm": 0.8308874368667603, "learning_rate": 1.7616364082392446e-06, "loss": 1.1762, "step": 945 }, { "epoch": 0.8148148148148148, "grad_norm": 0.9012060761451721, "learning_rate": 1.745853735755687e-06, "loss": 1.1723, "step": 946 }, { "epoch": 0.8156761412575366, "grad_norm": 0.8090865015983582, "learning_rate": 1.7301353164841562e-06, "loss": 1.1905, "step": 947 }, { "epoch": 0.8165374677002584, "grad_norm": 0.9133245348930359, "learning_rate": 1.7144812727824233e-06, "loss": 1.1314, "step": 948 }, { "epoch": 0.8173987941429802, "grad_norm": 0.817896842956543, "learning_rate": 1.6988917265071337e-06, "loss": 1.2059, "step": 949 }, { "epoch": 0.818260120585702, "grad_norm": 0.8479551076889038, "learning_rate": 1.6833667990128622e-06, "loss": 1.1256, "step": 950 }, { "epoch": 0.8191214470284238, "grad_norm": 0.9293598532676697, "learning_rate": 1.6679066111511677e-06, "loss": 1.1371, "step": 951 }, { "epoch": 0.8199827734711456, "grad_norm": 0.8253700733184814, "learning_rate": 1.6525112832696576e-06, "loss": 1.1672, "step": 952 }, { "epoch": 0.8208440999138673, "grad_norm": 0.8608076572418213, "learning_rate": 1.6371809352110446e-06, "loss": 1.1383, "step": 953 }, { "epoch": 0.8217054263565892, "grad_norm": 0.9392169117927551, "learning_rate": 1.6219156863122121e-06, "loss": 1.1867, "step": 954 }, { "epoch": 0.8225667527993109, "grad_norm": 0.8883967399597168, "learning_rate": 1.6067156554032893e-06, "loss": 1.2018, "step": 955 }, { "epoch": 0.8234280792420328, "grad_norm": 0.8919565677642822, "learning_rate": 1.5915809608067245e-06, "loss": 1.1873, "step": 956 }, { "epoch": 0.8242894056847545, "grad_norm": 0.8182902932167053, "learning_rate": 1.5765117203363723e-06, "loss": 1.1613, "step": 957 }, { "epoch": 0.8251507321274764, "grad_norm": 0.8579580783843994, "learning_rate": 1.5615080512965563e-06, "loss": 1.1883, "step": 958 }, { "epoch": 0.8260120585701981, "grad_norm": 0.8626570105552673, "learning_rate": 1.5465700704811825e-06, "loss": 1.171, "step": 959 }, { "epoch": 0.8268733850129198, "grad_norm": 0.8572131395339966, "learning_rate": 1.5316978941728045e-06, "loss": 1.1429, "step": 960 }, { "epoch": 0.8277347114556417, "grad_norm": 0.8980193734169006, "learning_rate": 1.5168916381417387e-06, "loss": 1.1485, "step": 961 }, { "epoch": 0.8285960378983634, "grad_norm": 0.9097509384155273, "learning_rate": 1.5021514176451514e-06, "loss": 1.1834, "step": 962 }, { "epoch": 0.8294573643410853, "grad_norm": 1.002036213874817, "learning_rate": 1.487477347426164e-06, "loss": 1.1611, "step": 963 }, { "epoch": 0.830318690783807, "grad_norm": 0.967788815498352, "learning_rate": 1.4728695417129579e-06, "loss": 1.1471, "step": 964 }, { "epoch": 0.8311800172265289, "grad_norm": 0.9008656740188599, "learning_rate": 1.458328114217894e-06, "loss": 1.1956, "step": 965 }, { "epoch": 0.8320413436692506, "grad_norm": 0.927203893661499, "learning_rate": 1.4438531781366226e-06, "loss": 1.1646, "step": 966 }, { "epoch": 0.8329026701119724, "grad_norm": 0.877873957157135, "learning_rate": 1.4294448461471943e-06, "loss": 1.1799, "step": 967 }, { "epoch": 0.8337639965546942, "grad_norm": 0.8068222999572754, "learning_rate": 1.4151032304091928e-06, "loss": 1.1723, "step": 968 }, { "epoch": 0.834625322997416, "grad_norm": 0.8391594290733337, "learning_rate": 1.400828442562857e-06, "loss": 1.1462, "step": 969 }, { "epoch": 0.8354866494401378, "grad_norm": 0.8191964030265808, "learning_rate": 1.3866205937282195e-06, "loss": 1.1757, "step": 970 }, { "epoch": 0.8363479758828596, "grad_norm": 0.8616624474525452, "learning_rate": 1.372479794504229e-06, "loss": 1.1719, "step": 971 }, { "epoch": 0.8372093023255814, "grad_norm": 0.9123625755310059, "learning_rate": 1.3584061549679017e-06, "loss": 1.1638, "step": 972 }, { "epoch": 0.8380706287683032, "grad_norm": 0.8329593539237976, "learning_rate": 1.3443997846734535e-06, "loss": 1.1267, "step": 973 }, { "epoch": 0.8389319552110249, "grad_norm": 0.9073025584220886, "learning_rate": 1.330460792651459e-06, "loss": 1.1523, "step": 974 }, { "epoch": 0.8397932816537468, "grad_norm": 0.8376272916793823, "learning_rate": 1.31658928740799e-06, "loss": 1.1472, "step": 975 }, { "epoch": 0.8406546080964685, "grad_norm": 0.8478926420211792, "learning_rate": 1.3027853769237808e-06, "loss": 1.1705, "step": 976 }, { "epoch": 0.8415159345391904, "grad_norm": 0.9134843945503235, "learning_rate": 1.2890491686533812e-06, "loss": 1.1835, "step": 977 }, { "epoch": 0.8423772609819121, "grad_norm": 0.8817895650863647, "learning_rate": 1.2753807695243258e-06, "loss": 1.1549, "step": 978 }, { "epoch": 0.843238587424634, "grad_norm": 0.9673689603805542, "learning_rate": 1.2617802859363016e-06, "loss": 1.1211, "step": 979 }, { "epoch": 0.8440999138673557, "grad_norm": 0.8391812443733215, "learning_rate": 1.2482478237603102e-06, "loss": 1.1272, "step": 980 }, { "epoch": 0.8449612403100775, "grad_norm": 0.9365954399108887, "learning_rate": 1.2347834883378518e-06, "loss": 1.1997, "step": 981 }, { "epoch": 0.8458225667527993, "grad_norm": 0.8548864722251892, "learning_rate": 1.2213873844801049e-06, "loss": 1.1804, "step": 982 }, { "epoch": 0.8466838931955211, "grad_norm": 0.7988507151603699, "learning_rate": 1.2080596164671098e-06, "loss": 1.1528, "step": 983 }, { "epoch": 0.8475452196382429, "grad_norm": 0.8993740677833557, "learning_rate": 1.1948002880469601e-06, "loss": 1.1818, "step": 984 }, { "epoch": 0.8484065460809647, "grad_norm": 1.0622564554214478, "learning_rate": 1.1816095024349828e-06, "loss": 1.2045, "step": 985 }, { "epoch": 0.8492678725236865, "grad_norm": 0.9041065573692322, "learning_rate": 1.1684873623129457e-06, "loss": 1.1534, "step": 986 }, { "epoch": 0.8501291989664083, "grad_norm": 0.8370366096496582, "learning_rate": 1.1554339698282623e-06, "loss": 1.1714, "step": 987 }, { "epoch": 0.85099052540913, "grad_norm": 0.8588913679122925, "learning_rate": 1.1424494265931829e-06, "loss": 1.1604, "step": 988 }, { "epoch": 0.8518518518518519, "grad_norm": 0.9080969095230103, "learning_rate": 1.1295338336840113e-06, "loss": 1.1827, "step": 989 }, { "epoch": 0.8527131782945736, "grad_norm": 0.8849515318870544, "learning_rate": 1.1166872916403226e-06, "loss": 1.2205, "step": 990 }, { "epoch": 0.8535745047372955, "grad_norm": 0.8305906653404236, "learning_rate": 1.1039099004641684e-06, "loss": 1.165, "step": 991 }, { "epoch": 0.8544358311800172, "grad_norm": 0.8653746843338013, "learning_rate": 1.0912017596193115e-06, "loss": 1.1975, "step": 992 }, { "epoch": 0.8552971576227391, "grad_norm": 0.9006616473197937, "learning_rate": 1.0785629680304433e-06, "loss": 1.1512, "step": 993 }, { "epoch": 0.8561584840654608, "grad_norm": 0.8347445726394653, "learning_rate": 1.065993624082411e-06, "loss": 1.162, "step": 994 }, { "epoch": 0.8570198105081827, "grad_norm": 0.8577432632446289, "learning_rate": 1.053493825619467e-06, "loss": 1.1691, "step": 995 }, { "epoch": 0.8578811369509044, "grad_norm": 0.8580631613731384, "learning_rate": 1.0410636699444855e-06, "loss": 1.1569, "step": 996 }, { "epoch": 0.8587424633936261, "grad_norm": 0.8457480669021606, "learning_rate": 1.0287032538182262e-06, "loss": 1.1585, "step": 997 }, { "epoch": 0.859603789836348, "grad_norm": 0.8660425543785095, "learning_rate": 1.0164126734585667e-06, "loss": 1.1755, "step": 998 }, { "epoch": 0.8604651162790697, "grad_norm": 0.8238218426704407, "learning_rate": 1.0041920245397552e-06, "loss": 1.1735, "step": 999 }, { "epoch": 0.8613264427217916, "grad_norm": 0.8318968415260315, "learning_rate": 9.920414021916747e-07, "loss": 1.1299, "step": 1000 }, { "epoch": 0.8621877691645133, "grad_norm": 0.8238946199417114, "learning_rate": 9.79960900999094e-07, "loss": 1.1459, "step": 1001 }, { "epoch": 0.8630490956072352, "grad_norm": 0.9450215101242065, "learning_rate": 9.67950615000931e-07, "loss": 1.1746, "step": 1002 }, { "epoch": 0.8639104220499569, "grad_norm": 0.8457410335540771, "learning_rate": 9.560106376895305e-07, "loss": 1.1797, "step": 1003 }, { "epoch": 0.8647717484926787, "grad_norm": 0.8758255243301392, "learning_rate": 9.441410620099201e-07, "loss": 1.1639, "step": 1004 }, { "epoch": 0.8656330749354005, "grad_norm": 0.9426845908164978, "learning_rate": 9.32341980359105e-07, "loss": 1.1763, "step": 1005 }, { "epoch": 0.8664944013781223, "grad_norm": 0.7973933219909668, "learning_rate": 9.206134845853343e-07, "loss": 1.172, "step": 1006 }, { "epoch": 0.8673557278208441, "grad_norm": 0.8713229298591614, "learning_rate": 9.089556659873921e-07, "loss": 1.1684, "step": 1007 }, { "epoch": 0.8682170542635659, "grad_norm": 0.8608222603797913, "learning_rate": 8.973686153138872e-07, "loss": 1.1564, "step": 1008 }, { "epoch": 0.8690783807062877, "grad_norm": 0.8522865176200867, "learning_rate": 8.858524227625498e-07, "loss": 1.1644, "step": 1009 }, { "epoch": 0.8699397071490095, "grad_norm": 0.8249273896217346, "learning_rate": 8.744071779795171e-07, "loss": 1.1637, "step": 1010 }, { "epoch": 0.8708010335917312, "grad_norm": 0.9221320152282715, "learning_rate": 8.630329700586481e-07, "loss": 1.1803, "step": 1011 }, { "epoch": 0.8716623600344531, "grad_norm": 0.8558515906333923, "learning_rate": 8.517298875408253e-07, "loss": 1.1779, "step": 1012 }, { "epoch": 0.8725236864771748, "grad_norm": 0.8073273301124573, "learning_rate": 8.40498018413266e-07, "loss": 1.1561, "step": 1013 }, { "epoch": 0.8733850129198967, "grad_norm": 0.7956110835075378, "learning_rate": 8.293374501088358e-07, "loss": 1.1593, "step": 1014 }, { "epoch": 0.8742463393626184, "grad_norm": 0.8590007424354553, "learning_rate": 8.182482695053728e-07, "loss": 1.1656, "step": 1015 }, { "epoch": 0.8751076658053403, "grad_norm": 0.867030143737793, "learning_rate": 8.072305629250033e-07, "loss": 1.1469, "step": 1016 }, { "epoch": 0.875968992248062, "grad_norm": 0.8188326358795166, "learning_rate": 7.962844161334748e-07, "loss": 1.1654, "step": 1017 }, { "epoch": 0.8768303186907838, "grad_norm": 0.8768230676651001, "learning_rate": 7.854099143394933e-07, "loss": 1.1844, "step": 1018 }, { "epoch": 0.8776916451335056, "grad_norm": 0.840861976146698, "learning_rate": 7.746071421940482e-07, "loss": 1.1787, "step": 1019 }, { "epoch": 0.8785529715762274, "grad_norm": 0.8956863880157471, "learning_rate": 7.638761837897612e-07, "loss": 1.1662, "step": 1020 }, { "epoch": 0.8794142980189492, "grad_norm": 0.867888867855072, "learning_rate": 7.532171226602336e-07, "loss": 1.1606, "step": 1021 }, { "epoch": 0.880275624461671, "grad_norm": 0.8646805286407471, "learning_rate": 7.426300417793919e-07, "loss": 1.1673, "step": 1022 }, { "epoch": 0.8811369509043928, "grad_norm": 0.8361663818359375, "learning_rate": 7.321150235608398e-07, "loss": 1.1581, "step": 1023 }, { "epoch": 0.8819982773471146, "grad_norm": 0.8646005988121033, "learning_rate": 7.21672149857221e-07, "loss": 1.2194, "step": 1024 }, { "epoch": 0.8828596037898363, "grad_norm": 0.829197883605957, "learning_rate": 7.113015019595793e-07, "loss": 1.1772, "step": 1025 }, { "epoch": 0.8837209302325582, "grad_norm": 0.8020138144493103, "learning_rate": 7.010031605967316e-07, "loss": 1.1722, "step": 1026 }, { "epoch": 0.8845822566752799, "grad_norm": 0.9305247664451599, "learning_rate": 6.907772059346285e-07, "loss": 1.1753, "step": 1027 }, { "epoch": 0.8854435831180018, "grad_norm": 0.8150957226753235, "learning_rate": 6.806237175757457e-07, "loss": 1.1573, "step": 1028 }, { "epoch": 0.8863049095607235, "grad_norm": 0.8378347158432007, "learning_rate": 6.705427745584469e-07, "loss": 1.1683, "step": 1029 }, { "epoch": 0.8871662360034454, "grad_norm": 0.8095634579658508, "learning_rate": 6.605344553563775e-07, "loss": 1.1681, "step": 1030 }, { "epoch": 0.8880275624461671, "grad_norm": 0.8543710708618164, "learning_rate": 6.505988378778616e-07, "loss": 1.1875, "step": 1031 }, { "epoch": 0.8888888888888888, "grad_norm": 0.9036997556686401, "learning_rate": 6.407359994652773e-07, "loss": 1.218, "step": 1032 }, { "epoch": 0.8897502153316107, "grad_norm": 0.8567870855331421, "learning_rate": 6.309460168944692e-07, "loss": 1.1726, "step": 1033 }, { "epoch": 0.8906115417743324, "grad_norm": 0.827688455581665, "learning_rate": 6.212289663741477e-07, "loss": 1.181, "step": 1034 }, { "epoch": 0.8914728682170543, "grad_norm": 0.8630082011222839, "learning_rate": 6.11584923545292e-07, "loss": 1.1624, "step": 1035 }, { "epoch": 0.892334194659776, "grad_norm": 0.8756182789802551, "learning_rate": 6.020139634805622e-07, "loss": 1.164, "step": 1036 }, { "epoch": 0.8931955211024979, "grad_norm": 0.8342011570930481, "learning_rate": 5.925161606837182e-07, "loss": 1.1552, "step": 1037 }, { "epoch": 0.8940568475452196, "grad_norm": 0.8205728530883789, "learning_rate": 5.830915890890366e-07, "loss": 1.1652, "step": 1038 }, { "epoch": 0.8949181739879414, "grad_norm": 0.8547372221946716, "learning_rate": 5.737403220607374e-07, "loss": 1.1837, "step": 1039 }, { "epoch": 0.8957795004306632, "grad_norm": 0.8426522612571716, "learning_rate": 5.644624323924108e-07, "loss": 1.2164, "step": 1040 }, { "epoch": 0.896640826873385, "grad_norm": 0.8302751183509827, "learning_rate": 5.552579923064516e-07, "loss": 1.1782, "step": 1041 }, { "epoch": 0.8975021533161068, "grad_norm": 0.8574025630950928, "learning_rate": 5.461270734534973e-07, "loss": 1.1599, "step": 1042 }, { "epoch": 0.8983634797588286, "grad_norm": 0.8473777770996094, "learning_rate": 5.370697469118713e-07, "loss": 1.1562, "step": 1043 }, { "epoch": 0.8992248062015504, "grad_norm": 0.8828932046890259, "learning_rate": 5.28086083187025e-07, "loss": 1.1839, "step": 1044 }, { "epoch": 0.9000861326442722, "grad_norm": 0.9018280506134033, "learning_rate": 5.191761522109939e-07, "loss": 1.1884, "step": 1045 }, { "epoch": 0.9009474590869939, "grad_norm": 0.8722638487815857, "learning_rate": 5.10340023341851e-07, "loss": 1.1788, "step": 1046 }, { "epoch": 0.9018087855297158, "grad_norm": 0.8644421696662903, "learning_rate": 5.015777653631693e-07, "loss": 1.1753, "step": 1047 }, { "epoch": 0.9026701119724375, "grad_norm": 0.892296552658081, "learning_rate": 4.928894464834843e-07, "loss": 1.1831, "step": 1048 }, { "epoch": 0.9035314384151594, "grad_norm": 0.8241122364997864, "learning_rate": 4.84275134335761e-07, "loss": 1.1732, "step": 1049 }, { "epoch": 0.9043927648578811, "grad_norm": 0.8443476557731628, "learning_rate": 4.757348959768704e-07, "loss": 1.2135, "step": 1050 }, { "epoch": 0.905254091300603, "grad_norm": 0.8274143934249878, "learning_rate": 4.67268797887066e-07, "loss": 1.1765, "step": 1051 }, { "epoch": 0.9061154177433247, "grad_norm": 0.8572297096252441, "learning_rate": 4.5887690596946975e-07, "loss": 1.1602, "step": 1052 }, { "epoch": 0.9069767441860465, "grad_norm": 0.7972581386566162, "learning_rate": 4.5055928554955665e-07, "loss": 1.1749, "step": 1053 }, { "epoch": 0.9078380706287683, "grad_norm": 0.8930532932281494, "learning_rate": 4.4231600137464305e-07, "loss": 1.16, "step": 1054 }, { "epoch": 0.9086993970714901, "grad_norm": 0.866405189037323, "learning_rate": 4.341471176133838e-07, "loss": 1.1627, "step": 1055 }, { "epoch": 0.9095607235142119, "grad_norm": 0.8310356736183167, "learning_rate": 4.2605269785528037e-07, "loss": 1.1672, "step": 1056 }, { "epoch": 0.9104220499569337, "grad_norm": 0.8165515661239624, "learning_rate": 4.1803280511017564e-07, "loss": 1.2008, "step": 1057 }, { "epoch": 0.9112833763996555, "grad_norm": 0.9229612946510315, "learning_rate": 4.100875018077688e-07, "loss": 1.1724, "step": 1058 }, { "epoch": 0.9121447028423773, "grad_norm": 0.8320847749710083, "learning_rate": 4.0221684979712906e-07, "loss": 1.167, "step": 1059 }, { "epoch": 0.9130060292850991, "grad_norm": 0.8543427586555481, "learning_rate": 3.9442091034621156e-07, "loss": 1.1767, "step": 1060 }, { "epoch": 0.9138673557278209, "grad_norm": 0.9095268249511719, "learning_rate": 3.8669974414138553e-07, "loss": 1.1781, "step": 1061 }, { "epoch": 0.9147286821705426, "grad_norm": 0.8602164387702942, "learning_rate": 3.7905341128695484e-07, "loss": 1.1204, "step": 1062 }, { "epoch": 0.9155900086132644, "grad_norm": 0.8160679936408997, "learning_rate": 3.7148197130469574e-07, "loss": 1.1729, "step": 1063 }, { "epoch": 0.9164513350559862, "grad_norm": 0.8283565640449524, "learning_rate": 3.639854831333911e-07, "loss": 1.1956, "step": 1064 }, { "epoch": 0.917312661498708, "grad_norm": 0.8678607940673828, "learning_rate": 3.5656400512837365e-07, "loss": 1.1625, "step": 1065 }, { "epoch": 0.9181739879414298, "grad_norm": 0.9065176844596863, "learning_rate": 3.4921759506106876e-07, "loss": 1.1178, "step": 1066 }, { "epoch": 0.9190353143841516, "grad_norm": 0.8343356251716614, "learning_rate": 3.4194631011854827e-07, "loss": 1.1498, "step": 1067 }, { "epoch": 0.9198966408268734, "grad_norm": 0.840149998664856, "learning_rate": 3.347502069030795e-07, "loss": 1.148, "step": 1068 }, { "epoch": 0.9207579672695951, "grad_norm": 1.0363847017288208, "learning_rate": 3.2762934143169333e-07, "loss": 1.1507, "step": 1069 }, { "epoch": 0.921619293712317, "grad_norm": 0.8863590955734253, "learning_rate": 3.205837691357405e-07, "loss": 1.1516, "step": 1070 }, { "epoch": 0.9224806201550387, "grad_norm": 0.8128695487976074, "learning_rate": 3.136135448604594e-07, "loss": 1.1545, "step": 1071 }, { "epoch": 0.9233419465977606, "grad_norm": 0.793180525302887, "learning_rate": 3.067187228645618e-07, "loss": 1.1439, "step": 1072 }, { "epoch": 0.9242032730404823, "grad_norm": 0.890590488910675, "learning_rate": 2.9989935681979165e-07, "loss": 1.1357, "step": 1073 }, { "epoch": 0.9250645994832042, "grad_norm": 0.840506374835968, "learning_rate": 2.931554998105235e-07, "loss": 1.1369, "step": 1074 }, { "epoch": 0.9259259259259259, "grad_norm": 0.9699150919914246, "learning_rate": 2.8648720433334e-07, "loss": 1.1735, "step": 1075 }, { "epoch": 0.9267872523686477, "grad_norm": 0.9378847479820251, "learning_rate": 2.798945222966265e-07, "loss": 1.1793, "step": 1076 }, { "epoch": 0.9276485788113695, "grad_norm": 0.9439303278923035, "learning_rate": 2.733775050201626e-07, "loss": 1.145, "step": 1077 }, { "epoch": 0.9285099052540913, "grad_norm": 0.9288724660873413, "learning_rate": 2.6693620323473556e-07, "loss": 1.1739, "step": 1078 }, { "epoch": 0.9293712316968131, "grad_norm": 0.8163039088249207, "learning_rate": 2.605706670817276e-07, "loss": 1.1643, "step": 1079 }, { "epoch": 0.9302325581395349, "grad_norm": 0.824459433555603, "learning_rate": 2.5428094611273825e-07, "loss": 1.1689, "step": 1080 }, { "epoch": 0.9310938845822567, "grad_norm": 0.8641778826713562, "learning_rate": 2.480670892891934e-07, "loss": 1.1534, "step": 1081 }, { "epoch": 0.9319552110249785, "grad_norm": 0.8537282943725586, "learning_rate": 2.419291449819683e-07, "loss": 1.195, "step": 1082 }, { "epoch": 0.9328165374677002, "grad_norm": 0.9170420169830322, "learning_rate": 2.358671609710017e-07, "loss": 1.1937, "step": 1083 }, { "epoch": 0.9336778639104221, "grad_norm": 0.973427951335907, "learning_rate": 2.298811844449389e-07, "loss": 1.1538, "step": 1084 }, { "epoch": 0.9345391903531438, "grad_norm": 0.9759657382965088, "learning_rate": 2.2397126200074837e-07, "loss": 1.171, "step": 1085 }, { "epoch": 0.9354005167958657, "grad_norm": 0.8713492751121521, "learning_rate": 2.1813743964336998e-07, "loss": 1.1599, "step": 1086 }, { "epoch": 0.9362618432385874, "grad_norm": 0.8439943790435791, "learning_rate": 2.1237976278535522e-07, "loss": 1.1923, "step": 1087 }, { "epoch": 0.9371231696813093, "grad_norm": 0.8235443234443665, "learning_rate": 2.066982762465075e-07, "loss": 1.207, "step": 1088 }, { "epoch": 0.937984496124031, "grad_norm": 0.8441301584243774, "learning_rate": 2.0109302425354139e-07, "loss": 1.1855, "step": 1089 }, { "epoch": 0.9388458225667528, "grad_norm": 0.8680086731910706, "learning_rate": 1.9556405043973158e-07, "loss": 1.1622, "step": 1090 }, { "epoch": 0.9397071490094746, "grad_norm": 1.0325438976287842, "learning_rate": 1.901113978445801e-07, "loss": 1.1902, "step": 1091 }, { "epoch": 0.9405684754521964, "grad_norm": 0.908032238483429, "learning_rate": 1.8473510891347412e-07, "loss": 1.1998, "step": 1092 }, { "epoch": 0.9414298018949182, "grad_norm": 0.8270619511604309, "learning_rate": 1.794352254973597e-07, "loss": 1.1663, "step": 1093 }, { "epoch": 0.94229112833764, "grad_norm": 0.8700262308120728, "learning_rate": 1.742117888524153e-07, "loss": 1.1732, "step": 1094 }, { "epoch": 0.9431524547803618, "grad_norm": 0.8367083668708801, "learning_rate": 1.6906483963973207e-07, "loss": 1.1595, "step": 1095 }, { "epoch": 0.9440137812230835, "grad_norm": 0.9255692958831787, "learning_rate": 1.6399441792499305e-07, "loss": 1.1828, "step": 1096 }, { "epoch": 0.9448751076658053, "grad_norm": 0.8711833357810974, "learning_rate": 1.590005631781666e-07, "loss": 1.189, "step": 1097 }, { "epoch": 0.9457364341085271, "grad_norm": 0.8770515322685242, "learning_rate": 1.5408331427319345e-07, "loss": 1.1903, "step": 1098 }, { "epoch": 0.9465977605512489, "grad_norm": 0.8291750550270081, "learning_rate": 1.4924270948769027e-07, "loss": 1.1713, "step": 1099 }, { "epoch": 0.9474590869939707, "grad_norm": 0.9927818179130554, "learning_rate": 1.4447878650264867e-07, "loss": 1.2139, "step": 1100 }, { "epoch": 0.9483204134366925, "grad_norm": 0.8271976709365845, "learning_rate": 1.3979158240213787e-07, "loss": 1.1788, "step": 1101 }, { "epoch": 0.9491817398794143, "grad_norm": 0.8746073842048645, "learning_rate": 1.3518113367302356e-07, "loss": 1.149, "step": 1102 }, { "epoch": 0.9500430663221361, "grad_norm": 0.8062018752098083, "learning_rate": 1.3064747620468053e-07, "loss": 1.1709, "step": 1103 }, { "epoch": 0.9509043927648578, "grad_norm": 0.8557183146476746, "learning_rate": 1.2619064528871185e-07, "loss": 1.1872, "step": 1104 }, { "epoch": 0.9517657192075797, "grad_norm": 0.8311978578567505, "learning_rate": 1.218106756186743e-07, "loss": 1.1717, "step": 1105 }, { "epoch": 0.9526270456503014, "grad_norm": 0.8271847367286682, "learning_rate": 1.1750760128981131e-07, "loss": 1.1682, "step": 1106 }, { "epoch": 0.9534883720930233, "grad_norm": 0.8393845558166504, "learning_rate": 1.132814557987838e-07, "loss": 1.1971, "step": 1107 }, { "epoch": 0.954349698535745, "grad_norm": 0.8846034407615662, "learning_rate": 1.0913227204341292e-07, "loss": 1.1751, "step": 1108 }, { "epoch": 0.9552110249784669, "grad_norm": 0.7971451282501221, "learning_rate": 1.0506008232242348e-07, "loss": 1.1872, "step": 1109 }, { "epoch": 0.9560723514211886, "grad_norm": 0.8897345066070557, "learning_rate": 1.010649183351875e-07, "loss": 1.1562, "step": 1110 }, { "epoch": 0.9569336778639104, "grad_norm": 0.8979521989822388, "learning_rate": 9.714681118148329e-08, "loss": 1.1412, "step": 1111 }, { "epoch": 0.9577950043066322, "grad_norm": 0.836390495300293, "learning_rate": 9.330579136125117e-08, "loss": 1.1585, "step": 1112 }, { "epoch": 0.958656330749354, "grad_norm": 0.8353316187858582, "learning_rate": 8.95418887743571e-08, "loss": 1.1722, "step": 1113 }, { "epoch": 0.9595176571920758, "grad_norm": 0.9047684073448181, "learning_rate": 8.58551327203594e-08, "loss": 1.1509, "step": 1114 }, { "epoch": 0.9603789836347976, "grad_norm": 0.8520369529724121, "learning_rate": 8.224555189827565e-08, "loss": 1.1899, "step": 1115 }, { "epoch": 0.9612403100775194, "grad_norm": 0.8386793732643127, "learning_rate": 7.871317440637072e-08, "loss": 1.1716, "step": 1116 }, { "epoch": 0.9621016365202412, "grad_norm": 0.819486677646637, "learning_rate": 7.525802774192791e-08, "loss": 1.1616, "step": 1117 }, { "epoch": 0.9629629629629629, "grad_norm": 0.8136334419250488, "learning_rate": 7.188013880103817e-08, "loss": 1.1653, "step": 1118 }, { "epoch": 0.9638242894056848, "grad_norm": 0.8472459316253662, "learning_rate": 6.857953387839012e-08, "loss": 1.1666, "step": 1119 }, { "epoch": 0.9646856158484065, "grad_norm": 0.924691915512085, "learning_rate": 6.535623866706698e-08, "loss": 1.1448, "step": 1120 }, { "epoch": 0.9655469422911284, "grad_norm": 0.7984655499458313, "learning_rate": 6.22102782583478e-08, "loss": 1.1922, "step": 1121 }, { "epoch": 0.9664082687338501, "grad_norm": 0.9025313854217529, "learning_rate": 5.9141677141506536e-08, "loss": 1.1607, "step": 1122 }, { "epoch": 0.967269595176572, "grad_norm": 0.8492075204849243, "learning_rate": 5.61504592036255e-08, "loss": 1.1702, "step": 1123 }, { "epoch": 0.9681309216192937, "grad_norm": 0.8583032488822937, "learning_rate": 5.323664772941217e-08, "loss": 1.1547, "step": 1124 }, { "epoch": 0.9689922480620154, "grad_norm": 0.8614605665206909, "learning_rate": 5.0400265401009395e-08, "loss": 1.2079, "step": 1125 }, { "epoch": 0.9698535745047373, "grad_norm": 0.7994558215141296, "learning_rate": 4.76413342978288e-08, "loss": 1.1677, "step": 1126 }, { "epoch": 0.970714900947459, "grad_norm": 0.8343400955200195, "learning_rate": 4.4959875896370965e-08, "loss": 1.1701, "step": 1127 }, { "epoch": 0.9715762273901809, "grad_norm": 0.8843151330947876, "learning_rate": 4.2355911070062205e-08, "loss": 1.1443, "step": 1128 }, { "epoch": 0.9724375538329026, "grad_norm": 0.8479021191596985, "learning_rate": 3.982946008909139e-08, "loss": 1.1963, "step": 1129 }, { "epoch": 0.9732988802756245, "grad_norm": 0.834740936756134, "learning_rate": 3.738054262025226e-08, "loss": 1.1788, "step": 1130 }, { "epoch": 0.9741602067183462, "grad_norm": 0.8662601709365845, "learning_rate": 3.500917772679025e-08, "loss": 1.1637, "step": 1131 }, { "epoch": 0.9750215331610681, "grad_norm": 1.2002208232879639, "learning_rate": 3.271538386825257e-08, "loss": 1.1841, "step": 1132 }, { "epoch": 0.9758828596037898, "grad_norm": 0.8810811638832092, "learning_rate": 3.049917890034837e-08, "loss": 1.1845, "step": 1133 }, { "epoch": 0.9767441860465116, "grad_norm": 0.8231048583984375, "learning_rate": 2.8360580074804355e-08, "loss": 1.1378, "step": 1134 }, { "epoch": 0.9776055124892334, "grad_norm": 0.8419777750968933, "learning_rate": 2.6299604039237147e-08, "loss": 1.1737, "step": 1135 }, { "epoch": 0.9784668389319552, "grad_norm": 0.8421429991722107, "learning_rate": 2.43162668370156e-08, "loss": 1.1403, "step": 1136 }, { "epoch": 0.979328165374677, "grad_norm": 0.8512549996376038, "learning_rate": 2.2410583907142012e-08, "loss": 1.1771, "step": 1137 }, { "epoch": 0.9801894918173988, "grad_norm": 0.8299239873886108, "learning_rate": 2.0582570084132224e-08, "loss": 1.1864, "step": 1138 }, { "epoch": 0.9810508182601206, "grad_norm": 0.8365635871887207, "learning_rate": 1.883223959789571e-08, "loss": 1.1988, "step": 1139 }, { "epoch": 0.9819121447028424, "grad_norm": 0.8790571689605713, "learning_rate": 1.7159606073627875e-08, "loss": 1.1613, "step": 1140 }, { "epoch": 0.9827734711455641, "grad_norm": 0.8469464778900146, "learning_rate": 1.5564682531702402e-08, "loss": 1.1488, "step": 1141 }, { "epoch": 0.983634797588286, "grad_norm": 0.9134259223937988, "learning_rate": 1.4047481387573503e-08, "loss": 1.1662, "step": 1142 }, { "epoch": 0.9844961240310077, "grad_norm": 0.8976234793663025, "learning_rate": 1.2608014451672701e-08, "loss": 1.2034, "step": 1143 }, { "epoch": 0.9853574504737296, "grad_norm": 1.0268605947494507, "learning_rate": 1.1246292929325552e-08, "loss": 1.1368, "step": 1144 }, { "epoch": 0.9862187769164513, "grad_norm": 0.8144389986991882, "learning_rate": 9.96232742065506e-09, "loss": 1.1789, "step": 1145 }, { "epoch": 0.9870801033591732, "grad_norm": 0.9007599353790283, "learning_rate": 8.756127920505065e-09, "loss": 1.1993, "step": 1146 }, { "epoch": 0.9879414298018949, "grad_norm": 0.8165833950042725, "learning_rate": 7.627703818363642e-09, "loss": 1.17, "step": 1147 }, { "epoch": 0.9888027562446167, "grad_norm": 0.9031330347061157, "learning_rate": 6.577063898285385e-09, "loss": 1.1961, "step": 1148 }, { "epoch": 0.9896640826873385, "grad_norm": 0.8834061026573181, "learning_rate": 5.604216338824797e-09, "loss": 1.2096, "step": 1149 }, { "epoch": 0.9905254091300603, "grad_norm": 0.8041283488273621, "learning_rate": 4.7091687129718896e-09, "loss": 1.1761, "step": 1150 }, { "epoch": 0.9913867355727821, "grad_norm": 0.9003199338912964, "learning_rate": 3.891927988098898e-09, "loss": 1.206, "step": 1151 }, { "epoch": 0.9922480620155039, "grad_norm": 0.82999587059021, "learning_rate": 3.1525005258969953e-09, "loss": 1.1214, "step": 1152 }, { "epoch": 0.9931093884582257, "grad_norm": 0.8493615984916687, "learning_rate": 2.490892082331886e-09, "loss": 1.2059, "step": 1153 }, { "epoch": 0.9939707149009475, "grad_norm": 0.8477335572242737, "learning_rate": 1.907107807600506e-09, "loss": 1.1827, "step": 1154 }, { "epoch": 0.9948320413436692, "grad_norm": 0.8850772380828857, "learning_rate": 1.4011522460866122e-09, "loss": 1.1666, "step": 1155 }, { "epoch": 0.9956933677863911, "grad_norm": 0.8809195756912231, "learning_rate": 9.730293363297006e-10, "loss": 1.1649, "step": 1156 }, { "epoch": 0.9965546942291128, "grad_norm": 0.891930103302002, "learning_rate": 6.227424109883639e-10, "loss": 1.1778, "step": 1157 }, { "epoch": 0.9974160206718347, "grad_norm": 0.8007426261901855, "learning_rate": 3.502941968225315e-10, "loss": 1.1696, "step": 1158 }, { "epoch": 0.9982773471145564, "grad_norm": 0.8914885520935059, "learning_rate": 1.5568681466682223e-10, "loss": 1.1262, "step": 1159 }, { "epoch": 0.9991386735572783, "grad_norm": 0.8544787168502808, "learning_rate": 3.8921779411671414e-11, "loss": 1.1751, "step": 1160 }, { "epoch": 1.0, "grad_norm": 0.9441701769828796, "learning_rate": 0.0, "loss": 1.1453, "step": 1161 }, { "epoch": 1.0, "step": 1161, "total_flos": 2.598284398506174e+19, "train_loss": 1.285096526043257, "train_runtime": 25375.9482, "train_samples_per_second": 35.131, "train_steps_per_second": 0.046 } ], "logging_steps": 1.0, "max_steps": 1161, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 3000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.598284398506174e+19, "train_batch_size": 8, "trial_name": null, "trial_params": null }