|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 1161, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0008613264427217916, |
|
"grad_norm": 599.9192504882812, |
|
"learning_rate": 5.714285714285715e-07, |
|
"loss": 6.0815, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0017226528854435831, |
|
"grad_norm": 575.8373413085938, |
|
"learning_rate": 1.142857142857143e-06, |
|
"loss": 5.5586, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.002583979328165375, |
|
"grad_norm": 564.647216796875, |
|
"learning_rate": 1.7142857142857145e-06, |
|
"loss": 5.7547, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0034453057708871662, |
|
"grad_norm": 407.28704833984375, |
|
"learning_rate": 2.285714285714286e-06, |
|
"loss": 5.438, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.004306632213608958, |
|
"grad_norm": 371.2016906738281, |
|
"learning_rate": 2.8571428571428573e-06, |
|
"loss": 4.8455, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.00516795865633075, |
|
"grad_norm": 371.3276672363281, |
|
"learning_rate": 3.428571428571429e-06, |
|
"loss": 3.7534, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.006029285099052541, |
|
"grad_norm": 1765.4417724609375, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 3.7887, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.0068906115417743325, |
|
"grad_norm": 293.5586242675781, |
|
"learning_rate": 4.571428571428572e-06, |
|
"loss": 3.0103, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.007751937984496124, |
|
"grad_norm": 490.7660827636719, |
|
"learning_rate": 5.142857142857142e-06, |
|
"loss": 2.7897, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.008613264427217916, |
|
"grad_norm": 211.81118774414062, |
|
"learning_rate": 5.7142857142857145e-06, |
|
"loss": 2.951, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.009474590869939707, |
|
"grad_norm": 83.9891128540039, |
|
"learning_rate": 6.285714285714286e-06, |
|
"loss": 2.6668, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.0103359173126615, |
|
"grad_norm": 296.01885986328125, |
|
"learning_rate": 6.857142857142858e-06, |
|
"loss": 2.4551, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.01119724375538329, |
|
"grad_norm": 44.15952682495117, |
|
"learning_rate": 7.428571428571429e-06, |
|
"loss": 2.3069, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.012058570198105082, |
|
"grad_norm": 28.404870986938477, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 2.1574, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.012919896640826873, |
|
"grad_norm": 18892.63671875, |
|
"learning_rate": 8.571428571428571e-06, |
|
"loss": 8.4092, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.013781223083548665, |
|
"grad_norm": 13.016464233398438, |
|
"learning_rate": 9.142857142857144e-06, |
|
"loss": 2.0747, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.014642549526270457, |
|
"grad_norm": 24.35378074645996, |
|
"learning_rate": 9.714285714285715e-06, |
|
"loss": 2.1555, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.015503875968992248, |
|
"grad_norm": 129.4594268798828, |
|
"learning_rate": 1.0285714285714285e-05, |
|
"loss": 2.2513, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.01636520241171404, |
|
"grad_norm": 312.46337890625, |
|
"learning_rate": 1.0857142857142858e-05, |
|
"loss": 2.4386, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.017226528854435832, |
|
"grad_norm": 1113.3402099609375, |
|
"learning_rate": 1.1428571428571429e-05, |
|
"loss": 2.6086, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.01808785529715762, |
|
"grad_norm": 758.8408203125, |
|
"learning_rate": 1.2e-05, |
|
"loss": 2.3477, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.018949181739879414, |
|
"grad_norm": 90.13806915283203, |
|
"learning_rate": 1.2571428571428572e-05, |
|
"loss": 2.0906, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.019810508182601206, |
|
"grad_norm": 44.2735710144043, |
|
"learning_rate": 1.3142857142857145e-05, |
|
"loss": 2.0562, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.020671834625323, |
|
"grad_norm": 42.16297149658203, |
|
"learning_rate": 1.3714285714285716e-05, |
|
"loss": 2.0478, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.02153316106804479, |
|
"grad_norm": 35.452392578125, |
|
"learning_rate": 1.4285714285714287e-05, |
|
"loss": 2.0479, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.02239448751076658, |
|
"grad_norm": 23.058853149414062, |
|
"learning_rate": 1.4857142857142858e-05, |
|
"loss": 1.9681, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.023255813953488372, |
|
"grad_norm": 17.171300888061523, |
|
"learning_rate": 1.542857142857143e-05, |
|
"loss": 1.917, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.024117140396210164, |
|
"grad_norm": 12.922857284545898, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 1.8992, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.024978466838931956, |
|
"grad_norm": 10.135607719421387, |
|
"learning_rate": 1.6571428571428574e-05, |
|
"loss": 1.8757, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.025839793281653745, |
|
"grad_norm": 8.323143005371094, |
|
"learning_rate": 1.7142857142857142e-05, |
|
"loss": 1.8451, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.026701119724375538, |
|
"grad_norm": 5.9443793296813965, |
|
"learning_rate": 1.7714285714285717e-05, |
|
"loss": 1.7739, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.02756244616709733, |
|
"grad_norm": 4.917455196380615, |
|
"learning_rate": 1.8285714285714288e-05, |
|
"loss": 1.7803, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.028423772609819122, |
|
"grad_norm": 5.2614240646362305, |
|
"learning_rate": 1.885714285714286e-05, |
|
"loss": 1.7771, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.029285099052540915, |
|
"grad_norm": 4.4380316734313965, |
|
"learning_rate": 1.942857142857143e-05, |
|
"loss": 1.7563, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.030146425495262703, |
|
"grad_norm": 4.11315393447876, |
|
"learning_rate": 2e-05, |
|
"loss": 1.7493, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.031007751937984496, |
|
"grad_norm": 3.8529300689697266, |
|
"learning_rate": 1.9999961078220587e-05, |
|
"loss": 1.7646, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.03186907838070629, |
|
"grad_norm": 3.429624080657959, |
|
"learning_rate": 1.9999844313185335e-05, |
|
"loss": 1.7406, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.03273040482342808, |
|
"grad_norm": 3.103177785873413, |
|
"learning_rate": 1.9999649705803178e-05, |
|
"loss": 1.7295, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.03359173126614987, |
|
"grad_norm": 2.6700687408447266, |
|
"learning_rate": 1.9999377257589012e-05, |
|
"loss": 1.6595, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.034453057708871665, |
|
"grad_norm": 2.35770583152771, |
|
"learning_rate": 1.999902697066367e-05, |
|
"loss": 1.6766, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.03531438415159346, |
|
"grad_norm": 2.01990008354187, |
|
"learning_rate": 1.9998598847753918e-05, |
|
"loss": 1.6472, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.03617571059431524, |
|
"grad_norm": 1.8652081489562988, |
|
"learning_rate": 1.9998092892192403e-05, |
|
"loss": 1.6448, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.037037037037037035, |
|
"grad_norm": 1.9675177335739136, |
|
"learning_rate": 1.999750910791767e-05, |
|
"loss": 1.6573, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.03789836347975883, |
|
"grad_norm": 2.176922082901001, |
|
"learning_rate": 1.9996847499474102e-05, |
|
"loss": 1.6364, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.03875968992248062, |
|
"grad_norm": 2.259882688522339, |
|
"learning_rate": 1.99961080720119e-05, |
|
"loss": 1.6164, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.03962101636520241, |
|
"grad_norm": 2.2850229740142822, |
|
"learning_rate": 1.9995290831287032e-05, |
|
"loss": 1.5988, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.040482342807924204, |
|
"grad_norm": 2.1443259716033936, |
|
"learning_rate": 1.9994395783661177e-05, |
|
"loss": 1.5978, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.041343669250646, |
|
"grad_norm": 2.1305596828460693, |
|
"learning_rate": 1.9993422936101715e-05, |
|
"loss": 1.6235, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.04220499569336779, |
|
"grad_norm": 2.253634214401245, |
|
"learning_rate": 1.9992372296181637e-05, |
|
"loss": 1.6344, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.04306632213608958, |
|
"grad_norm": 1.944365382194519, |
|
"learning_rate": 1.9991243872079495e-05, |
|
"loss": 1.5638, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.04392764857881137, |
|
"grad_norm": 1.8713358640670776, |
|
"learning_rate": 1.9990037672579347e-05, |
|
"loss": 1.59, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.04478897502153316, |
|
"grad_norm": 1.4943434000015259, |
|
"learning_rate": 1.9988753707070675e-05, |
|
"loss": 1.6001, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.04565030146425495, |
|
"grad_norm": 1.5794744491577148, |
|
"learning_rate": 1.9987391985548326e-05, |
|
"loss": 1.5618, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.046511627906976744, |
|
"grad_norm": 1.5509154796600342, |
|
"learning_rate": 1.998595251861243e-05, |
|
"loss": 1.5514, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.047372954349698536, |
|
"grad_norm": 1.7027164697647095, |
|
"learning_rate": 1.9984435317468298e-05, |
|
"loss": 1.5591, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.04823428079242033, |
|
"grad_norm": 1.851098656654358, |
|
"learning_rate": 1.9982840393926374e-05, |
|
"loss": 1.6171, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.04909560723514212, |
|
"grad_norm": 1.8336509466171265, |
|
"learning_rate": 1.9981167760402104e-05, |
|
"loss": 1.6013, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.04995693367786391, |
|
"grad_norm": 1.5762431621551514, |
|
"learning_rate": 1.997941742991587e-05, |
|
"loss": 1.5275, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.050818260120585705, |
|
"grad_norm": 1.5044102668762207, |
|
"learning_rate": 1.997758941609286e-05, |
|
"loss": 1.5214, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.05167958656330749, |
|
"grad_norm": 1.93209707736969, |
|
"learning_rate": 1.9975683733162987e-05, |
|
"loss": 1.5819, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.05254091300602928, |
|
"grad_norm": 1.545015811920166, |
|
"learning_rate": 1.9973700395960765e-05, |
|
"loss": 1.5227, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.053402239448751075, |
|
"grad_norm": 1.2270804643630981, |
|
"learning_rate": 1.9971639419925197e-05, |
|
"loss": 1.5119, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.05426356589147287, |
|
"grad_norm": 1.7906986474990845, |
|
"learning_rate": 1.9969500821099654e-05, |
|
"loss": 1.5528, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.05512489233419466, |
|
"grad_norm": 1.5916507244110107, |
|
"learning_rate": 1.996728461613175e-05, |
|
"loss": 1.5367, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.05598621877691645, |
|
"grad_norm": 1.413952350616455, |
|
"learning_rate": 1.996499082227321e-05, |
|
"loss": 1.5247, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.056847545219638244, |
|
"grad_norm": 1.3606282472610474, |
|
"learning_rate": 1.996261945737975e-05, |
|
"loss": 1.459, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.05770887166236004, |
|
"grad_norm": 1.327688455581665, |
|
"learning_rate": 1.996017053991091e-05, |
|
"loss": 1.5062, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.05857019810508183, |
|
"grad_norm": 1.2740212678909302, |
|
"learning_rate": 1.995764408892994e-05, |
|
"loss": 1.478, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.059431524547803614, |
|
"grad_norm": 1.1678736209869385, |
|
"learning_rate": 1.995504012410363e-05, |
|
"loss": 1.4755, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.06029285099052541, |
|
"grad_norm": 1.574155569076538, |
|
"learning_rate": 1.995235866570217e-05, |
|
"loss": 1.5191, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.0611541774332472, |
|
"grad_norm": 1.2211532592773438, |
|
"learning_rate": 1.9949599734598993e-05, |
|
"loss": 1.4561, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.06201550387596899, |
|
"grad_norm": 1.313081979751587, |
|
"learning_rate": 1.994676335227059e-05, |
|
"loss": 1.502, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.06287683031869079, |
|
"grad_norm": 1.2082189321517944, |
|
"learning_rate": 1.9943849540796375e-05, |
|
"loss": 1.4598, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.06373815676141258, |
|
"grad_norm": 1.2929805517196655, |
|
"learning_rate": 1.9940858322858493e-05, |
|
"loss": 1.4993, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.06459948320413436, |
|
"grad_norm": 1.199054479598999, |
|
"learning_rate": 1.9937789721741654e-05, |
|
"loss": 1.4904, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.06546080964685616, |
|
"grad_norm": 1.350379467010498, |
|
"learning_rate": 1.9934643761332933e-05, |
|
"loss": 1.4534, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.06632213608957795, |
|
"grad_norm": 1.1363669633865356, |
|
"learning_rate": 1.9931420466121613e-05, |
|
"loss": 1.453, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.06718346253229975, |
|
"grad_norm": 1.1514127254486084, |
|
"learning_rate": 1.9928119861198962e-05, |
|
"loss": 1.4607, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.06804478897502153, |
|
"grad_norm": 1.0154885053634644, |
|
"learning_rate": 1.9924741972258076e-05, |
|
"loss": 1.4401, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.06890611541774333, |
|
"grad_norm": 1.055841326713562, |
|
"learning_rate": 1.9921286825593632e-05, |
|
"loss": 1.48, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.06976744186046512, |
|
"grad_norm": 1.3121042251586914, |
|
"learning_rate": 1.9917754448101725e-05, |
|
"loss": 1.4263, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.07062876830318691, |
|
"grad_norm": 1.102817416191101, |
|
"learning_rate": 1.9914144867279644e-05, |
|
"loss": 1.4401, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.0714900947459087, |
|
"grad_norm": 1.0836132764816284, |
|
"learning_rate": 1.9910458111225645e-05, |
|
"loss": 1.4866, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.07235142118863049, |
|
"grad_norm": 1.0442920923233032, |
|
"learning_rate": 1.990669420863875e-05, |
|
"loss": 1.4162, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.07321274763135228, |
|
"grad_norm": 1.0844299793243408, |
|
"learning_rate": 1.9902853188818518e-05, |
|
"loss": 1.4707, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.07407407407407407, |
|
"grad_norm": 0.993617057800293, |
|
"learning_rate": 1.9898935081664814e-05, |
|
"loss": 1.4567, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.07493540051679587, |
|
"grad_norm": 1.103285312652588, |
|
"learning_rate": 1.9894939917677577e-05, |
|
"loss": 1.4249, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.07579672695951765, |
|
"grad_norm": 1.0927506685256958, |
|
"learning_rate": 1.9890867727956587e-05, |
|
"loss": 1.4294, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.07665805340223945, |
|
"grad_norm": 1.0682389736175537, |
|
"learning_rate": 1.988671854420122e-05, |
|
"loss": 1.3915, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.07751937984496124, |
|
"grad_norm": 1.0784050226211548, |
|
"learning_rate": 1.9882492398710192e-05, |
|
"loss": 1.4445, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.07838070628768304, |
|
"grad_norm": 1.0482838153839111, |
|
"learning_rate": 1.987818932438133e-05, |
|
"loss": 1.3939, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.07924203273040482, |
|
"grad_norm": 1.1594949960708618, |
|
"learning_rate": 1.987380935471129e-05, |
|
"loss": 1.4543, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.08010335917312661, |
|
"grad_norm": 1.0159746408462524, |
|
"learning_rate": 1.986935252379532e-05, |
|
"loss": 1.4048, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.08096468561584841, |
|
"grad_norm": 1.1461023092269897, |
|
"learning_rate": 1.9864818866326978e-05, |
|
"loss": 1.4314, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.0818260120585702, |
|
"grad_norm": 1.1117491722106934, |
|
"learning_rate": 1.9860208417597863e-05, |
|
"loss": 1.4383, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.082687338501292, |
|
"grad_norm": 0.9831107258796692, |
|
"learning_rate": 1.9855521213497355e-05, |
|
"loss": 1.4245, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.08354866494401378, |
|
"grad_norm": 1.0372180938720703, |
|
"learning_rate": 1.9850757290512313e-05, |
|
"loss": 1.3838, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.08440999138673558, |
|
"grad_norm": 0.9056967496871948, |
|
"learning_rate": 1.9845916685726808e-05, |
|
"loss": 1.4076, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.08527131782945736, |
|
"grad_norm": 1.002386450767517, |
|
"learning_rate": 1.9840999436821836e-05, |
|
"loss": 1.4301, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.08613264427217916, |
|
"grad_norm": 0.9989022612571716, |
|
"learning_rate": 1.983600558207501e-05, |
|
"loss": 1.3993, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.08699397071490095, |
|
"grad_norm": 1.0508233308792114, |
|
"learning_rate": 1.983093516036027e-05, |
|
"loss": 1.4241, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.08785529715762273, |
|
"grad_norm": 0.9549940824508667, |
|
"learning_rate": 1.9825788211147587e-05, |
|
"loss": 1.427, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.08871662360034453, |
|
"grad_norm": 1.0347970724105835, |
|
"learning_rate": 1.9820564774502644e-05, |
|
"loss": 1.4029, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.08957795004306632, |
|
"grad_norm": 1.0045099258422852, |
|
"learning_rate": 1.981526489108653e-05, |
|
"loss": 1.4146, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.09043927648578812, |
|
"grad_norm": 1.2777268886566162, |
|
"learning_rate": 1.980988860215542e-05, |
|
"loss": 1.3599, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.0913006029285099, |
|
"grad_norm": 0.9002528190612793, |
|
"learning_rate": 1.980443594956027e-05, |
|
"loss": 1.4008, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.0921619293712317, |
|
"grad_norm": 0.9888056516647339, |
|
"learning_rate": 1.9798906975746462e-05, |
|
"loss": 1.422, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.09302325581395349, |
|
"grad_norm": 0.9847419857978821, |
|
"learning_rate": 1.9793301723753494e-05, |
|
"loss": 1.3866, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.09388458225667529, |
|
"grad_norm": 0.9763849377632141, |
|
"learning_rate": 1.9787620237214648e-05, |
|
"loss": 1.3883, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.09474590869939707, |
|
"grad_norm": 1.1006907224655151, |
|
"learning_rate": 1.9781862560356632e-05, |
|
"loss": 1.3748, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.09560723514211886, |
|
"grad_norm": 1.0140228271484375, |
|
"learning_rate": 1.9776028737999256e-05, |
|
"loss": 1.3581, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.09646856158484066, |
|
"grad_norm": 1.0311439037322998, |
|
"learning_rate": 1.9770118815555063e-05, |
|
"loss": 1.4253, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.09732988802756244, |
|
"grad_norm": 0.9372150301933289, |
|
"learning_rate": 1.9764132839029e-05, |
|
"loss": 1.4078, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.09819121447028424, |
|
"grad_norm": 1.0055251121520996, |
|
"learning_rate": 1.9758070855018033e-05, |
|
"loss": 1.432, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.09905254091300603, |
|
"grad_norm": 0.9005001783370972, |
|
"learning_rate": 1.9751932910710808e-05, |
|
"loss": 1.3998, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.09991386735572783, |
|
"grad_norm": 0.958956778049469, |
|
"learning_rate": 1.9745719053887265e-05, |
|
"loss": 1.3867, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.10077519379844961, |
|
"grad_norm": 1.2123082876205444, |
|
"learning_rate": 1.9739429332918276e-05, |
|
"loss": 1.382, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.10163652024117141, |
|
"grad_norm": 0.858340859413147, |
|
"learning_rate": 1.9733063796765267e-05, |
|
"loss": 1.3794, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.1024978466838932, |
|
"grad_norm": 0.8608000874519348, |
|
"learning_rate": 1.972662249497984e-05, |
|
"loss": 1.3996, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.10335917312661498, |
|
"grad_norm": 0.8888645768165588, |
|
"learning_rate": 1.972010547770338e-05, |
|
"loss": 1.3874, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.10422049956933678, |
|
"grad_norm": 0.9219585061073303, |
|
"learning_rate": 1.9713512795666663e-05, |
|
"loss": 1.3881, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.10508182601205857, |
|
"grad_norm": 0.8192768096923828, |
|
"learning_rate": 1.970684450018948e-05, |
|
"loss": 1.3827, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.10594315245478036, |
|
"grad_norm": 1.0063087940216064, |
|
"learning_rate": 1.9700100643180213e-05, |
|
"loss": 1.3679, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.10680447889750215, |
|
"grad_norm": 0.9152660369873047, |
|
"learning_rate": 1.969328127713544e-05, |
|
"loss": 1.3637, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.10766580534022395, |
|
"grad_norm": 0.9176394939422607, |
|
"learning_rate": 1.9686386455139544e-05, |
|
"loss": 1.3972, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.10852713178294573, |
|
"grad_norm": 0.91953444480896, |
|
"learning_rate": 1.9679416230864265e-05, |
|
"loss": 1.3833, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.10938845822566753, |
|
"grad_norm": 0.967275083065033, |
|
"learning_rate": 1.9672370658568306e-05, |
|
"loss": 1.3628, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.11024978466838932, |
|
"grad_norm": 0.8972539305686951, |
|
"learning_rate": 1.966524979309692e-05, |
|
"loss": 1.3457, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.1111111111111111, |
|
"grad_norm": 0.926163911819458, |
|
"learning_rate": 1.9658053689881453e-05, |
|
"loss": 1.3278, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.1119724375538329, |
|
"grad_norm": 0.9265878200531006, |
|
"learning_rate": 1.9650782404938933e-05, |
|
"loss": 1.347, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.11283376399655469, |
|
"grad_norm": 1.088254451751709, |
|
"learning_rate": 1.9643435994871626e-05, |
|
"loss": 1.3498, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.11369509043927649, |
|
"grad_norm": 0.9928959608078003, |
|
"learning_rate": 1.963601451686661e-05, |
|
"loss": 1.372, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.11455641688199827, |
|
"grad_norm": 0.8130192756652832, |
|
"learning_rate": 1.9628518028695307e-05, |
|
"loss": 1.3185, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.11541774332472007, |
|
"grad_norm": 0.9813799262046814, |
|
"learning_rate": 1.9620946588713048e-05, |
|
"loss": 1.3723, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.11627906976744186, |
|
"grad_norm": 0.8440076112747192, |
|
"learning_rate": 1.9613300255858615e-05, |
|
"loss": 1.379, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.11714039621016366, |
|
"grad_norm": 0.9899678826332092, |
|
"learning_rate": 1.960557908965379e-05, |
|
"loss": 1.3396, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.11800172265288544, |
|
"grad_norm": 0.9470325708389282, |
|
"learning_rate": 1.9597783150202873e-05, |
|
"loss": 1.3631, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.11886304909560723, |
|
"grad_norm": 0.9099454879760742, |
|
"learning_rate": 1.9589912498192233e-05, |
|
"loss": 1.3623, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.11972437553832903, |
|
"grad_norm": 1.1269025802612305, |
|
"learning_rate": 1.9581967194889826e-05, |
|
"loss": 1.3615, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.12058570198105081, |
|
"grad_norm": 0.8688855767250061, |
|
"learning_rate": 1.957394730214472e-05, |
|
"loss": 1.36, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.12144702842377261, |
|
"grad_norm": 0.9251095652580261, |
|
"learning_rate": 1.956585288238662e-05, |
|
"loss": 1.3274, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.1223083548664944, |
|
"grad_norm": 1.2113760709762573, |
|
"learning_rate": 1.955768399862536e-05, |
|
"loss": 1.3687, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.1231696813092162, |
|
"grad_norm": 1.0069940090179443, |
|
"learning_rate": 1.9549440714450447e-05, |
|
"loss": 1.3602, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.12403100775193798, |
|
"grad_norm": 0.987554132938385, |
|
"learning_rate": 1.9541123094030528e-05, |
|
"loss": 1.3515, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.12489233419465978, |
|
"grad_norm": 0.8786775469779968, |
|
"learning_rate": 1.9532731202112935e-05, |
|
"loss": 1.3849, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.12575366063738158, |
|
"grad_norm": 1.0503038167953491, |
|
"learning_rate": 1.9524265104023133e-05, |
|
"loss": 1.3638, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.12661498708010335, |
|
"grad_norm": 0.9434555768966675, |
|
"learning_rate": 1.9515724865664242e-05, |
|
"loss": 1.3065, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.12747631352282515, |
|
"grad_norm": 1.0372234582901, |
|
"learning_rate": 1.9507110553516518e-05, |
|
"loss": 1.3398, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.12833763996554695, |
|
"grad_norm": 0.9711533188819885, |
|
"learning_rate": 1.949842223463683e-05, |
|
"loss": 1.3428, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.12919896640826872, |
|
"grad_norm": 0.8844906091690063, |
|
"learning_rate": 1.9489659976658152e-05, |
|
"loss": 1.3685, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.13006029285099052, |
|
"grad_norm": 0.9219164848327637, |
|
"learning_rate": 1.9480823847789007e-05, |
|
"loss": 1.3465, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.13092161929371232, |
|
"grad_norm": 0.9082310795783997, |
|
"learning_rate": 1.947191391681298e-05, |
|
"loss": 1.3799, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.13178294573643412, |
|
"grad_norm": 0.9203616380691528, |
|
"learning_rate": 1.946293025308813e-05, |
|
"loss": 1.3636, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.1326442721791559, |
|
"grad_norm": 0.8217676877975464, |
|
"learning_rate": 1.9453872926546505e-05, |
|
"loss": 1.3247, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.1335055986218777, |
|
"grad_norm": 0.8964298963546753, |
|
"learning_rate": 1.944474200769355e-05, |
|
"loss": 1.3357, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.1343669250645995, |
|
"grad_norm": 0.9523929357528687, |
|
"learning_rate": 1.943553756760759e-05, |
|
"loss": 1.3712, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.13522825150732126, |
|
"grad_norm": 0.8506253361701965, |
|
"learning_rate": 1.9426259677939264e-05, |
|
"loss": 1.3522, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.13608957795004306, |
|
"grad_norm": 0.9101868867874146, |
|
"learning_rate": 1.9416908410910965e-05, |
|
"loss": 1.3266, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.13695090439276486, |
|
"grad_norm": 0.8775342702865601, |
|
"learning_rate": 1.9407483839316284e-05, |
|
"loss": 1.3575, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.13781223083548666, |
|
"grad_norm": 0.9115650057792664, |
|
"learning_rate": 1.939798603651944e-05, |
|
"loss": 1.3578, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.13867355727820843, |
|
"grad_norm": 0.8985600471496582, |
|
"learning_rate": 1.938841507645471e-05, |
|
"loss": 1.3059, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.13953488372093023, |
|
"grad_norm": 1.053168773651123, |
|
"learning_rate": 1.9378771033625855e-05, |
|
"loss": 1.3736, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.14039621016365203, |
|
"grad_norm": 0.9247081279754639, |
|
"learning_rate": 1.9369053983105533e-05, |
|
"loss": 1.3268, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.14125753660637383, |
|
"grad_norm": 0.937142014503479, |
|
"learning_rate": 1.9359264000534726e-05, |
|
"loss": 1.3352, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.1421188630490956, |
|
"grad_norm": 0.8469811081886292, |
|
"learning_rate": 1.934940116212214e-05, |
|
"loss": 1.3277, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.1429801894918174, |
|
"grad_norm": 0.8988816142082214, |
|
"learning_rate": 1.9339465544643623e-05, |
|
"loss": 1.3447, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.1438415159345392, |
|
"grad_norm": 1.033051609992981, |
|
"learning_rate": 1.9329457225441554e-05, |
|
"loss": 1.3072, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.14470284237726097, |
|
"grad_norm": 0.9636563658714294, |
|
"learning_rate": 1.9319376282424255e-05, |
|
"loss": 1.3027, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.14556416881998277, |
|
"grad_norm": 0.9479746222496033, |
|
"learning_rate": 1.9309222794065373e-05, |
|
"loss": 1.3319, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.14642549526270457, |
|
"grad_norm": 0.9528256058692932, |
|
"learning_rate": 1.929899683940327e-05, |
|
"loss": 1.3737, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.14728682170542637, |
|
"grad_norm": 0.87147057056427, |
|
"learning_rate": 1.9288698498040423e-05, |
|
"loss": 1.3361, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.14814814814814814, |
|
"grad_norm": 0.9613904356956482, |
|
"learning_rate": 1.9278327850142783e-05, |
|
"loss": 1.3485, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.14900947459086994, |
|
"grad_norm": 1.0545318126678467, |
|
"learning_rate": 1.9267884976439163e-05, |
|
"loss": 1.333, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.14987080103359174, |
|
"grad_norm": 0.9187701940536499, |
|
"learning_rate": 1.9257369958220612e-05, |
|
"loss": 1.3241, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.1507321274763135, |
|
"grad_norm": 0.866005539894104, |
|
"learning_rate": 1.9246782877339767e-05, |
|
"loss": 1.3133, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.1515934539190353, |
|
"grad_norm": 0.9420531392097473, |
|
"learning_rate": 1.923612381621024e-05, |
|
"loss": 1.2784, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.1524547803617571, |
|
"grad_norm": 0.9032683968544006, |
|
"learning_rate": 1.9225392857805955e-05, |
|
"loss": 1.2992, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.1533161068044789, |
|
"grad_norm": 0.9758039116859436, |
|
"learning_rate": 1.921459008566051e-05, |
|
"loss": 1.3013, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.15417743324720068, |
|
"grad_norm": 0.9804133176803589, |
|
"learning_rate": 1.9203715583866527e-05, |
|
"loss": 1.2893, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.15503875968992248, |
|
"grad_norm": 0.9305117130279541, |
|
"learning_rate": 1.9192769437075e-05, |
|
"loss": 1.3307, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.15590008613264428, |
|
"grad_norm": 0.9009129405021667, |
|
"learning_rate": 1.918175173049463e-05, |
|
"loss": 1.3344, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.15676141257536608, |
|
"grad_norm": 0.9379671812057495, |
|
"learning_rate": 1.9170662549891162e-05, |
|
"loss": 1.3434, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.15762273901808785, |
|
"grad_norm": 0.8975892066955566, |
|
"learning_rate": 1.9159501981586738e-05, |
|
"loss": 1.305, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.15848406546080965, |
|
"grad_norm": 0.8844696283340454, |
|
"learning_rate": 1.9148270112459178e-05, |
|
"loss": 1.3373, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.15934539190353145, |
|
"grad_norm": 0.9148349761962891, |
|
"learning_rate": 1.9136967029941354e-05, |
|
"loss": 1.322, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.16020671834625322, |
|
"grad_norm": 0.9962027668952942, |
|
"learning_rate": 1.9125592822020485e-05, |
|
"loss": 1.3065, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.16106804478897502, |
|
"grad_norm": 0.9287891983985901, |
|
"learning_rate": 1.9114147577237452e-05, |
|
"loss": 1.3102, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.16192937123169682, |
|
"grad_norm": 0.9015977382659912, |
|
"learning_rate": 1.9102631384686116e-05, |
|
"loss": 1.3011, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.16279069767441862, |
|
"grad_norm": 1.0545498132705688, |
|
"learning_rate": 1.909104433401261e-05, |
|
"loss": 1.2916, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.1636520241171404, |
|
"grad_norm": 0.8966095447540283, |
|
"learning_rate": 1.9079386515414667e-05, |
|
"loss": 1.2899, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.1645133505598622, |
|
"grad_norm": 0.8880913853645325, |
|
"learning_rate": 1.9067658019640897e-05, |
|
"loss": 1.3187, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.165374677002584, |
|
"grad_norm": 1.007127285003662, |
|
"learning_rate": 1.9055858937990083e-05, |
|
"loss": 1.2909, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.16623600344530576, |
|
"grad_norm": 0.8095739483833313, |
|
"learning_rate": 1.9043989362310472e-05, |
|
"loss": 1.2956, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.16709732988802756, |
|
"grad_norm": 0.8505532145500183, |
|
"learning_rate": 1.903204938499907e-05, |
|
"loss": 1.344, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.16795865633074936, |
|
"grad_norm": 0.8656225204467773, |
|
"learning_rate": 1.902003909900091e-05, |
|
"loss": 1.3279, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.16881998277347116, |
|
"grad_norm": 0.8762969374656677, |
|
"learning_rate": 1.9007958597808326e-05, |
|
"loss": 1.313, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.16968130921619293, |
|
"grad_norm": 0.8867192268371582, |
|
"learning_rate": 1.8995807975460246e-05, |
|
"loss": 1.2943, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.17054263565891473, |
|
"grad_norm": 0.9131550192832947, |
|
"learning_rate": 1.8983587326541437e-05, |
|
"loss": 1.302, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.17140396210163653, |
|
"grad_norm": 0.9059438705444336, |
|
"learning_rate": 1.8971296746181774e-05, |
|
"loss": 1.2927, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.17226528854435832, |
|
"grad_norm": 0.912719190120697, |
|
"learning_rate": 1.8958936330055516e-05, |
|
"loss": 1.2986, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.1731266149870801, |
|
"grad_norm": 0.9960724711418152, |
|
"learning_rate": 1.894650617438054e-05, |
|
"loss": 1.2989, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.1739879414298019, |
|
"grad_norm": 0.9200662970542908, |
|
"learning_rate": 1.893400637591759e-05, |
|
"loss": 1.3312, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.1748492678725237, |
|
"grad_norm": 1.040659785270691, |
|
"learning_rate": 1.8921437031969557e-05, |
|
"loss": 1.3031, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.17571059431524547, |
|
"grad_norm": 0.8604618310928345, |
|
"learning_rate": 1.8908798240380692e-05, |
|
"loss": 1.341, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.17657192075796727, |
|
"grad_norm": 0.982661247253418, |
|
"learning_rate": 1.8896090099535834e-05, |
|
"loss": 1.305, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.17743324720068906, |
|
"grad_norm": 0.8687034249305725, |
|
"learning_rate": 1.888331270835968e-05, |
|
"loss": 1.3114, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.17829457364341086, |
|
"grad_norm": 0.9046248197555542, |
|
"learning_rate": 1.8870466166315992e-05, |
|
"loss": 1.3212, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.17915590008613264, |
|
"grad_norm": 0.8828570246696472, |
|
"learning_rate": 1.885755057340682e-05, |
|
"loss": 1.2929, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.18001722652885443, |
|
"grad_norm": 0.8848705291748047, |
|
"learning_rate": 1.8844566030171737e-05, |
|
"loss": 1.3057, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.18087855297157623, |
|
"grad_norm": 1.004079818725586, |
|
"learning_rate": 1.8831512637687054e-05, |
|
"loss": 1.2909, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.181739879414298, |
|
"grad_norm": 0.9026859998703003, |
|
"learning_rate": 1.881839049756502e-05, |
|
"loss": 1.3081, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.1826012058570198, |
|
"grad_norm": 1.0891921520233154, |
|
"learning_rate": 1.880519971195304e-05, |
|
"loss": 1.3273, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.1834625322997416, |
|
"grad_norm": 0.9379571676254272, |
|
"learning_rate": 1.879194038353289e-05, |
|
"loss": 1.3223, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.1843238587424634, |
|
"grad_norm": 0.9415060877799988, |
|
"learning_rate": 1.87786126155199e-05, |
|
"loss": 1.3464, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.18518518518518517, |
|
"grad_norm": 0.9501126408576965, |
|
"learning_rate": 1.8765216511662153e-05, |
|
"loss": 1.3008, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.18604651162790697, |
|
"grad_norm": 0.9133473634719849, |
|
"learning_rate": 1.8751752176239693e-05, |
|
"loss": 1.2966, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.18690783807062877, |
|
"grad_norm": 0.9399188160896301, |
|
"learning_rate": 1.87382197140637e-05, |
|
"loss": 1.3193, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.18776916451335057, |
|
"grad_norm": 0.9107801914215088, |
|
"learning_rate": 1.8724619230475675e-05, |
|
"loss": 1.3024, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.18863049095607234, |
|
"grad_norm": 0.894831120967865, |
|
"learning_rate": 1.8710950831346623e-05, |
|
"loss": 1.3163, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.18949181739879414, |
|
"grad_norm": 1.0155446529388428, |
|
"learning_rate": 1.8697214623076222e-05, |
|
"loss": 1.2575, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.19035314384151594, |
|
"grad_norm": 0.9866533279418945, |
|
"learning_rate": 1.8683410712592015e-05, |
|
"loss": 1.3209, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.19121447028423771, |
|
"grad_norm": 0.8821126222610474, |
|
"learning_rate": 1.8669539207348544e-05, |
|
"loss": 1.2828, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.1920757967269595, |
|
"grad_norm": 0.8987988233566284, |
|
"learning_rate": 1.8655600215326547e-05, |
|
"loss": 1.2898, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.1929371231696813, |
|
"grad_norm": 0.9085987210273743, |
|
"learning_rate": 1.8641593845032098e-05, |
|
"loss": 1.3065, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.1937984496124031, |
|
"grad_norm": 0.9193576574325562, |
|
"learning_rate": 1.8627520205495772e-05, |
|
"loss": 1.2837, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.19465977605512488, |
|
"grad_norm": 0.9095317721366882, |
|
"learning_rate": 1.8613379406271784e-05, |
|
"loss": 1.3171, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.19552110249784668, |
|
"grad_norm": 0.9969485402107239, |
|
"learning_rate": 1.8599171557437147e-05, |
|
"loss": 1.2819, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.19638242894056848, |
|
"grad_norm": 0.9653918743133545, |
|
"learning_rate": 1.858489676959081e-05, |
|
"loss": 1.2601, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.19724375538329025, |
|
"grad_norm": 0.977749764919281, |
|
"learning_rate": 1.8570555153852806e-05, |
|
"loss": 1.2861, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.19810508182601205, |
|
"grad_norm": 0.9468740820884705, |
|
"learning_rate": 1.855614682186338e-05, |
|
"loss": 1.2869, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.19896640826873385, |
|
"grad_norm": 1.1150413751602173, |
|
"learning_rate": 1.8541671885782106e-05, |
|
"loss": 1.2798, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.19982773471145565, |
|
"grad_norm": 0.9200144410133362, |
|
"learning_rate": 1.8527130458287047e-05, |
|
"loss": 1.2855, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.20068906115417742, |
|
"grad_norm": 0.937813401222229, |
|
"learning_rate": 1.851252265257384e-05, |
|
"loss": 1.2598, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.20155038759689922, |
|
"grad_norm": 1.02151620388031, |
|
"learning_rate": 1.8497848582354852e-05, |
|
"loss": 1.2828, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.20241171403962102, |
|
"grad_norm": 0.9327558279037476, |
|
"learning_rate": 1.8483108361858263e-05, |
|
"loss": 1.2898, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.20327304048234282, |
|
"grad_norm": 1.0296434164047241, |
|
"learning_rate": 1.8468302105827195e-05, |
|
"loss": 1.2869, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.2041343669250646, |
|
"grad_norm": 0.9533336758613586, |
|
"learning_rate": 1.845342992951882e-05, |
|
"loss": 1.2768, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.2049956933677864, |
|
"grad_norm": 0.9365238547325134, |
|
"learning_rate": 1.8438491948703445e-05, |
|
"loss": 1.2783, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.2058570198105082, |
|
"grad_norm": 1.104421854019165, |
|
"learning_rate": 1.842348827966363e-05, |
|
"loss": 1.293, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.20671834625322996, |
|
"grad_norm": 0.8812316060066223, |
|
"learning_rate": 1.840841903919328e-05, |
|
"loss": 1.2856, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.20757967269595176, |
|
"grad_norm": 0.9076652526855469, |
|
"learning_rate": 1.8393284344596715e-05, |
|
"loss": 1.3453, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.20844099913867356, |
|
"grad_norm": 0.9731348156929016, |
|
"learning_rate": 1.837808431368779e-05, |
|
"loss": 1.3084, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.20930232558139536, |
|
"grad_norm": 0.8820015788078308, |
|
"learning_rate": 1.8362819064788956e-05, |
|
"loss": 1.2753, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.21016365202411713, |
|
"grad_norm": 1.0030732154846191, |
|
"learning_rate": 1.8347488716730343e-05, |
|
"loss": 1.302, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.21102497846683893, |
|
"grad_norm": 0.9901053309440613, |
|
"learning_rate": 1.8332093388848836e-05, |
|
"loss": 1.31, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.21188630490956073, |
|
"grad_norm": 0.81141597032547, |
|
"learning_rate": 1.8316633200987143e-05, |
|
"loss": 1.2973, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.2127476313522825, |
|
"grad_norm": 0.9569693803787231, |
|
"learning_rate": 1.830110827349287e-05, |
|
"loss": 1.2805, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.2136089577950043, |
|
"grad_norm": 0.9289495348930359, |
|
"learning_rate": 1.8285518727217578e-05, |
|
"loss": 1.2735, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.2144702842377261, |
|
"grad_norm": 0.8315669894218445, |
|
"learning_rate": 1.8269864683515847e-05, |
|
"loss": 1.2775, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.2153316106804479, |
|
"grad_norm": 0.9181628823280334, |
|
"learning_rate": 1.8254146264244316e-05, |
|
"loss": 1.2328, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.21619293712316967, |
|
"grad_norm": 0.8977162837982178, |
|
"learning_rate": 1.8238363591760758e-05, |
|
"loss": 1.2346, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.21705426356589147, |
|
"grad_norm": 0.9173194169998169, |
|
"learning_rate": 1.822251678892312e-05, |
|
"loss": 1.2653, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.21791559000861327, |
|
"grad_norm": 0.9074868559837341, |
|
"learning_rate": 1.8206605979088545e-05, |
|
"loss": 1.2514, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.21877691645133507, |
|
"grad_norm": 0.8458305597305298, |
|
"learning_rate": 1.819063128611244e-05, |
|
"loss": 1.2885, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.21963824289405684, |
|
"grad_norm": 0.914714515209198, |
|
"learning_rate": 1.8174592834347503e-05, |
|
"loss": 1.2431, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.22049956933677864, |
|
"grad_norm": 0.8686874508857727, |
|
"learning_rate": 1.815849074864275e-05, |
|
"loss": 1.2554, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.22136089577950044, |
|
"grad_norm": 0.8784075379371643, |
|
"learning_rate": 1.814232515434255e-05, |
|
"loss": 1.3005, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.2222222222222222, |
|
"grad_norm": 0.919947624206543, |
|
"learning_rate": 1.8126096177285637e-05, |
|
"loss": 1.2968, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.223083548664944, |
|
"grad_norm": 0.9774091839790344, |
|
"learning_rate": 1.8109803943804146e-05, |
|
"loss": 1.2763, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.2239448751076658, |
|
"grad_norm": 0.8449957370758057, |
|
"learning_rate": 1.8093448580722617e-05, |
|
"loss": 1.2653, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.2248062015503876, |
|
"grad_norm": 0.9756072163581848, |
|
"learning_rate": 1.8077030215357024e-05, |
|
"loss": 1.272, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.22566752799310938, |
|
"grad_norm": 0.8839893341064453, |
|
"learning_rate": 1.806054897551376e-05, |
|
"loss": 1.2743, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.22652885443583118, |
|
"grad_norm": 0.8637559413909912, |
|
"learning_rate": 1.8044004989488662e-05, |
|
"loss": 1.2862, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.22739018087855298, |
|
"grad_norm": 0.9832426309585571, |
|
"learning_rate": 1.802739838606601e-05, |
|
"loss": 1.2886, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.22825150732127478, |
|
"grad_norm": 0.9957407116889954, |
|
"learning_rate": 1.801072929451751e-05, |
|
"loss": 1.2629, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.22911283376399655, |
|
"grad_norm": 0.9450539946556091, |
|
"learning_rate": 1.7993997844601305e-05, |
|
"loss": 1.2606, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.22997416020671835, |
|
"grad_norm": 0.9669734239578247, |
|
"learning_rate": 1.7977204166560954e-05, |
|
"loss": 1.2614, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.23083548664944015, |
|
"grad_norm": 0.9341006278991699, |
|
"learning_rate": 1.7960348391124422e-05, |
|
"loss": 1.2826, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.23169681309216192, |
|
"grad_norm": 0.8190209865570068, |
|
"learning_rate": 1.7943430649503065e-05, |
|
"loss": 1.2575, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.23255813953488372, |
|
"grad_norm": 0.9061177968978882, |
|
"learning_rate": 1.7926451073390612e-05, |
|
"loss": 1.2458, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.23341946597760552, |
|
"grad_norm": 0.8741104602813721, |
|
"learning_rate": 1.7909409794962115e-05, |
|
"loss": 1.2805, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.23428079242032732, |
|
"grad_norm": 0.8899810910224915, |
|
"learning_rate": 1.7892306946872952e-05, |
|
"loss": 1.2581, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.2351421188630491, |
|
"grad_norm": 0.9245790243148804, |
|
"learning_rate": 1.7875142662257788e-05, |
|
"loss": 1.2877, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.2360034453057709, |
|
"grad_norm": 0.8939240574836731, |
|
"learning_rate": 1.7857917074729513e-05, |
|
"loss": 1.2461, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.2368647717484927, |
|
"grad_norm": 0.8694809079170227, |
|
"learning_rate": 1.7840630318378233e-05, |
|
"loss": 1.2458, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.23772609819121446, |
|
"grad_norm": 0.8689870238304138, |
|
"learning_rate": 1.7823282527770214e-05, |
|
"loss": 1.2821, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.23858742463393626, |
|
"grad_norm": 0.8938902616500854, |
|
"learning_rate": 1.7805873837946833e-05, |
|
"loss": 1.2691, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.23944875107665806, |
|
"grad_norm": 1.0327783823013306, |
|
"learning_rate": 1.778840438442352e-05, |
|
"loss": 1.2732, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.24031007751937986, |
|
"grad_norm": 0.9891493916511536, |
|
"learning_rate": 1.7770874303188727e-05, |
|
"loss": 1.253, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.24117140396210163, |
|
"grad_norm": 0.9043530821800232, |
|
"learning_rate": 1.7753283730702837e-05, |
|
"loss": 1.2731, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.24203273040482343, |
|
"grad_norm": 0.9076008796691895, |
|
"learning_rate": 1.7735632803897135e-05, |
|
"loss": 1.2684, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.24289405684754523, |
|
"grad_norm": 0.8306509852409363, |
|
"learning_rate": 1.7717921660172708e-05, |
|
"loss": 1.2577, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.24375538329026702, |
|
"grad_norm": 0.8941156268119812, |
|
"learning_rate": 1.7700150437399405e-05, |
|
"loss": 1.2682, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.2446167097329888, |
|
"grad_norm": 0.8855567574501038, |
|
"learning_rate": 1.7682319273914755e-05, |
|
"loss": 1.2904, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.2454780361757106, |
|
"grad_norm": 1.0233161449432373, |
|
"learning_rate": 1.766442830852287e-05, |
|
"loss": 1.2907, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.2463393626184324, |
|
"grad_norm": 1.106757402420044, |
|
"learning_rate": 1.76464776804934e-05, |
|
"loss": 1.245, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.24720068906115417, |
|
"grad_norm": 0.92214435338974, |
|
"learning_rate": 1.7628467529560417e-05, |
|
"loss": 1.254, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.24806201550387597, |
|
"grad_norm": 0.8845908045768738, |
|
"learning_rate": 1.7610397995921348e-05, |
|
"loss": 1.314, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.24892334194659776, |
|
"grad_norm": 0.837216317653656, |
|
"learning_rate": 1.759226922023587e-05, |
|
"loss": 1.2943, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.24978466838931956, |
|
"grad_norm": 0.9128316044807434, |
|
"learning_rate": 1.7574081343624827e-05, |
|
"loss": 1.2881, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.25064599483204136, |
|
"grad_norm": 0.9482645392417908, |
|
"learning_rate": 1.7555834507669124e-05, |
|
"loss": 1.2846, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.25150732127476316, |
|
"grad_norm": 0.9201834797859192, |
|
"learning_rate": 1.7537528854408625e-05, |
|
"loss": 1.2729, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.2523686477174849, |
|
"grad_norm": 0.8601769804954529, |
|
"learning_rate": 1.751916452634105e-05, |
|
"loss": 1.2601, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.2532299741602067, |
|
"grad_norm": 0.8500188589096069, |
|
"learning_rate": 1.7500741666420863e-05, |
|
"loss": 1.2627, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.2540913006029285, |
|
"grad_norm": 0.9467564821243286, |
|
"learning_rate": 1.7482260418058167e-05, |
|
"loss": 1.2572, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.2549526270456503, |
|
"grad_norm": 0.9599546790122986, |
|
"learning_rate": 1.7463720925117565e-05, |
|
"loss": 1.2445, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.2558139534883721, |
|
"grad_norm": 0.8565171957015991, |
|
"learning_rate": 1.744512333191708e-05, |
|
"loss": 1.2222, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.2566752799310939, |
|
"grad_norm": 1.068682074546814, |
|
"learning_rate": 1.7426467783226992e-05, |
|
"loss": 1.2574, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.2575366063738157, |
|
"grad_norm": 0.9868722558021545, |
|
"learning_rate": 1.7407754424268727e-05, |
|
"loss": 1.2234, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.25839793281653745, |
|
"grad_norm": 0.8837171196937561, |
|
"learning_rate": 1.7388983400713736e-05, |
|
"loss": 1.2489, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.25925925925925924, |
|
"grad_norm": 1.0964189767837524, |
|
"learning_rate": 1.7370154858682347e-05, |
|
"loss": 1.2658, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.26012058570198104, |
|
"grad_norm": 0.9035983085632324, |
|
"learning_rate": 1.7351268944742626e-05, |
|
"loss": 1.261, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.26098191214470284, |
|
"grad_norm": 1.021364688873291, |
|
"learning_rate": 1.7332325805909256e-05, |
|
"loss": 1.2542, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.26184323858742464, |
|
"grad_norm": 0.8748368620872498, |
|
"learning_rate": 1.7313325589642363e-05, |
|
"loss": 1.2155, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.26270456503014644, |
|
"grad_norm": 1.0000536441802979, |
|
"learning_rate": 1.7294268443846403e-05, |
|
"loss": 1.2386, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.26356589147286824, |
|
"grad_norm": 0.8522447943687439, |
|
"learning_rate": 1.727515451686897e-05, |
|
"loss": 1.262, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.26442721791559, |
|
"grad_norm": 1.03394615650177, |
|
"learning_rate": 1.7255983957499676e-05, |
|
"loss": 1.2607, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.2652885443583118, |
|
"grad_norm": 0.9327899217605591, |
|
"learning_rate": 1.7236756914968985e-05, |
|
"loss": 1.239, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.2661498708010336, |
|
"grad_norm": 0.8953905701637268, |
|
"learning_rate": 1.7217473538947032e-05, |
|
"loss": 1.2584, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.2670111972437554, |
|
"grad_norm": 0.877118706703186, |
|
"learning_rate": 1.719813397954248e-05, |
|
"loss": 1.2692, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.2678725236864772, |
|
"grad_norm": 0.8814844489097595, |
|
"learning_rate": 1.7178738387301342e-05, |
|
"loss": 1.258, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.268733850129199, |
|
"grad_norm": 0.882537305355072, |
|
"learning_rate": 1.7159286913205813e-05, |
|
"loss": 1.2923, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.2695951765719208, |
|
"grad_norm": 0.9672994613647461, |
|
"learning_rate": 1.7139779708673084e-05, |
|
"loss": 1.2725, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.2704565030146425, |
|
"grad_norm": 0.8873171806335449, |
|
"learning_rate": 1.7120216925554185e-05, |
|
"loss": 1.2268, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.2713178294573643, |
|
"grad_norm": 0.9454041123390198, |
|
"learning_rate": 1.7100598716132775e-05, |
|
"loss": 1.2358, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.2721791559000861, |
|
"grad_norm": 0.943789005279541, |
|
"learning_rate": 1.708092523312398e-05, |
|
"loss": 1.2138, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.2730404823428079, |
|
"grad_norm": 0.8467637300491333, |
|
"learning_rate": 1.7061196629673198e-05, |
|
"loss": 1.2495, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.2739018087855297, |
|
"grad_norm": 0.8850976228713989, |
|
"learning_rate": 1.7041413059354893e-05, |
|
"loss": 1.2586, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.2747631352282515, |
|
"grad_norm": 0.8928930759429932, |
|
"learning_rate": 1.7021574676171418e-05, |
|
"loss": 1.2279, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.2756244616709733, |
|
"grad_norm": 0.9026039242744446, |
|
"learning_rate": 1.7001681634551813e-05, |
|
"loss": 1.2361, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.27648578811369506, |
|
"grad_norm": 0.90481036901474, |
|
"learning_rate": 1.6981734089350585e-05, |
|
"loss": 1.2308, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.27734711455641686, |
|
"grad_norm": 0.8753896951675415, |
|
"learning_rate": 1.696173219584653e-05, |
|
"loss": 1.2383, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.27820844099913866, |
|
"grad_norm": 1.0231022834777832, |
|
"learning_rate": 1.6941676109741506e-05, |
|
"loss": 1.2472, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.27906976744186046, |
|
"grad_norm": 0.8703423142433167, |
|
"learning_rate": 1.6921565987159226e-05, |
|
"loss": 1.2712, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.27993109388458226, |
|
"grad_norm": 0.9238699078559875, |
|
"learning_rate": 1.6901401984644034e-05, |
|
"loss": 1.2279, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.28079242032730406, |
|
"grad_norm": 0.9144941568374634, |
|
"learning_rate": 1.6881184259159708e-05, |
|
"loss": 1.2416, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.28165374677002586, |
|
"grad_norm": 0.8758153915405273, |
|
"learning_rate": 1.686091296808822e-05, |
|
"loss": 1.2736, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.28251507321274766, |
|
"grad_norm": 0.8742189407348633, |
|
"learning_rate": 1.6840588269228507e-05, |
|
"loss": 1.2876, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.2833763996554694, |
|
"grad_norm": 0.9139267206192017, |
|
"learning_rate": 1.682021032079526e-05, |
|
"loss": 1.2919, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.2842377260981912, |
|
"grad_norm": 0.8497806191444397, |
|
"learning_rate": 1.6799779281417685e-05, |
|
"loss": 1.2275, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.285099052540913, |
|
"grad_norm": 0.9016520380973816, |
|
"learning_rate": 1.6779295310138264e-05, |
|
"loss": 1.2126, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.2859603789836348, |
|
"grad_norm": 0.8737862706184387, |
|
"learning_rate": 1.6758758566411516e-05, |
|
"loss": 1.2641, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.2868217054263566, |
|
"grad_norm": 0.9002192616462708, |
|
"learning_rate": 1.6738169210102765e-05, |
|
"loss": 1.2872, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.2876830318690784, |
|
"grad_norm": 0.8395231366157532, |
|
"learning_rate": 1.6717527401486882e-05, |
|
"loss": 1.2808, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.2885443583118002, |
|
"grad_norm": 0.9024813771247864, |
|
"learning_rate": 1.669683330124706e-05, |
|
"loss": 1.2415, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.28940568475452194, |
|
"grad_norm": 0.9167109131813049, |
|
"learning_rate": 1.667608707047354e-05, |
|
"loss": 1.2375, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.29026701119724374, |
|
"grad_norm": 0.8365147113800049, |
|
"learning_rate": 1.6655288870662354e-05, |
|
"loss": 1.2451, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.29112833763996554, |
|
"grad_norm": 0.9034088253974915, |
|
"learning_rate": 1.6634438863714108e-05, |
|
"loss": 1.2237, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.29198966408268734, |
|
"grad_norm": 0.8997372984886169, |
|
"learning_rate": 1.661353721193266e-05, |
|
"loss": 1.2358, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.29285099052540914, |
|
"grad_norm": 0.9689245820045471, |
|
"learning_rate": 1.6592584078023915e-05, |
|
"loss": 1.2523, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.29371231696813094, |
|
"grad_norm": 0.9412267208099365, |
|
"learning_rate": 1.657157962509452e-05, |
|
"loss": 1.2445, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.29457364341085274, |
|
"grad_norm": 0.8464860916137695, |
|
"learning_rate": 1.6550524016650617e-05, |
|
"loss": 1.2102, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.2954349698535745, |
|
"grad_norm": 0.8929284811019897, |
|
"learning_rate": 1.652941741659655e-05, |
|
"loss": 1.2433, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.2962962962962963, |
|
"grad_norm": 1.0461153984069824, |
|
"learning_rate": 1.650825998923361e-05, |
|
"loss": 1.2289, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.2971576227390181, |
|
"grad_norm": 0.8382362127304077, |
|
"learning_rate": 1.6487051899258738e-05, |
|
"loss": 1.2455, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.2980189491817399, |
|
"grad_norm": 0.9788974523544312, |
|
"learning_rate": 1.6465793311763255e-05, |
|
"loss": 1.2314, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.2988802756244617, |
|
"grad_norm": 0.9103128910064697, |
|
"learning_rate": 1.6444484392231574e-05, |
|
"loss": 1.2331, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.2997416020671835, |
|
"grad_norm": 0.9638088345527649, |
|
"learning_rate": 1.6423125306539903e-05, |
|
"loss": 1.2228, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.3006029285099053, |
|
"grad_norm": 0.9347442388534546, |
|
"learning_rate": 1.6401716220954968e-05, |
|
"loss": 1.227, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.301464254952627, |
|
"grad_norm": 0.9683001637458801, |
|
"learning_rate": 1.638025730213271e-05, |
|
"loss": 1.2253, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.3023255813953488, |
|
"grad_norm": 0.9251329898834229, |
|
"learning_rate": 1.6358748717116993e-05, |
|
"loss": 1.2698, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.3031869078380706, |
|
"grad_norm": 0.9530742168426514, |
|
"learning_rate": 1.6337190633338294e-05, |
|
"loss": 1.2407, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.3040482342807924, |
|
"grad_norm": 0.9009522199630737, |
|
"learning_rate": 1.631558321861241e-05, |
|
"loss": 1.2478, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.3049095607235142, |
|
"grad_norm": 1.0106037855148315, |
|
"learning_rate": 1.6293926641139154e-05, |
|
"loss": 1.2298, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.305770887166236, |
|
"grad_norm": 0.8804092407226562, |
|
"learning_rate": 1.627222106950102e-05, |
|
"loss": 1.2324, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.3066322136089578, |
|
"grad_norm": 0.8187827467918396, |
|
"learning_rate": 1.625046667266191e-05, |
|
"loss": 1.2714, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.30749354005167956, |
|
"grad_norm": 1.0687847137451172, |
|
"learning_rate": 1.6228663619965787e-05, |
|
"loss": 1.2423, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.30835486649440136, |
|
"grad_norm": 0.8696728348731995, |
|
"learning_rate": 1.620681208113538e-05, |
|
"loss": 1.241, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.30921619293712316, |
|
"grad_norm": 0.8725491762161255, |
|
"learning_rate": 1.6184912226270833e-05, |
|
"loss": 1.2741, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.31007751937984496, |
|
"grad_norm": 1.0487327575683594, |
|
"learning_rate": 1.6162964225848416e-05, |
|
"loss": 1.2382, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.31093884582256676, |
|
"grad_norm": 0.8088880181312561, |
|
"learning_rate": 1.6140968250719177e-05, |
|
"loss": 1.2441, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.31180017226528856, |
|
"grad_norm": 0.8771573305130005, |
|
"learning_rate": 1.611892447210761e-05, |
|
"loss": 1.2393, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.31266149870801035, |
|
"grad_norm": 0.9969056248664856, |
|
"learning_rate": 1.609683306161034e-05, |
|
"loss": 1.2319, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.31352282515073215, |
|
"grad_norm": 0.9056423902511597, |
|
"learning_rate": 1.6074694191194758e-05, |
|
"loss": 1.2191, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.3143841515934539, |
|
"grad_norm": 0.8870687484741211, |
|
"learning_rate": 1.6052508033197713e-05, |
|
"loss": 1.249, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.3152454780361757, |
|
"grad_norm": 0.9680421352386475, |
|
"learning_rate": 1.6030274760324163e-05, |
|
"loss": 1.2299, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.3161068044788975, |
|
"grad_norm": 0.9495812058448792, |
|
"learning_rate": 1.6007994545645807e-05, |
|
"loss": 1.2574, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.3169681309216193, |
|
"grad_norm": 0.8676847219467163, |
|
"learning_rate": 1.598566756259977e-05, |
|
"loss": 1.2284, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.3178294573643411, |
|
"grad_norm": 0.841946005821228, |
|
"learning_rate": 1.596329398498723e-05, |
|
"loss": 1.2304, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.3186907838070629, |
|
"grad_norm": 1.0011506080627441, |
|
"learning_rate": 1.5940873986972078e-05, |
|
"loss": 1.2454, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.3195521102497847, |
|
"grad_norm": 0.9674389958381653, |
|
"learning_rate": 1.5918407743079564e-05, |
|
"loss": 1.241, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.32041343669250644, |
|
"grad_norm": 1.0047541856765747, |
|
"learning_rate": 1.5895895428194915e-05, |
|
"loss": 1.2541, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.32127476313522824, |
|
"grad_norm": 0.9397938847541809, |
|
"learning_rate": 1.5873337217562012e-05, |
|
"loss": 1.2182, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.32213608957795004, |
|
"grad_norm": 0.8715697526931763, |
|
"learning_rate": 1.585073328678199e-05, |
|
"loss": 1.243, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.32299741602067183, |
|
"grad_norm": 0.9596509337425232, |
|
"learning_rate": 1.582808381181189e-05, |
|
"loss": 1.2417, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.32385874246339363, |
|
"grad_norm": 0.8386455178260803, |
|
"learning_rate": 1.5805388968963286e-05, |
|
"loss": 1.2625, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.32472006890611543, |
|
"grad_norm": 0.897318422794342, |
|
"learning_rate": 1.5782648934900915e-05, |
|
"loss": 1.2233, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.32558139534883723, |
|
"grad_norm": 0.8364303112030029, |
|
"learning_rate": 1.575986388664129e-05, |
|
"loss": 1.2323, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.326442721791559, |
|
"grad_norm": 0.8544413447380066, |
|
"learning_rate": 1.5737034001551336e-05, |
|
"loss": 1.2373, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.3273040482342808, |
|
"grad_norm": 0.9020499587059021, |
|
"learning_rate": 1.5714159457347007e-05, |
|
"loss": 1.2432, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.3281653746770026, |
|
"grad_norm": 0.8884045481681824, |
|
"learning_rate": 1.5691240432091892e-05, |
|
"loss": 1.2144, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.3290267011197244, |
|
"grad_norm": 0.9886582493782043, |
|
"learning_rate": 1.566827710419584e-05, |
|
"loss": 1.2344, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.3298880275624462, |
|
"grad_norm": 0.8105961680412292, |
|
"learning_rate": 1.5645269652413574e-05, |
|
"loss": 1.2441, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.330749354005168, |
|
"grad_norm": 0.9423347115516663, |
|
"learning_rate": 1.5622218255843276e-05, |
|
"loss": 1.2172, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.33161068044788977, |
|
"grad_norm": 0.970775842666626, |
|
"learning_rate": 1.559912309392523e-05, |
|
"loss": 1.2587, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.3324720068906115, |
|
"grad_norm": 0.9015215039253235, |
|
"learning_rate": 1.5575984346440393e-05, |
|
"loss": 1.2388, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.3333333333333333, |
|
"grad_norm": 0.9856666326522827, |
|
"learning_rate": 1.5552802193509003e-05, |
|
"loss": 1.2181, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.3341946597760551, |
|
"grad_norm": 0.9138821959495544, |
|
"learning_rate": 1.55295768155892e-05, |
|
"loss": 1.2418, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.3350559862187769, |
|
"grad_norm": 0.8847649693489075, |
|
"learning_rate": 1.5506308393475582e-05, |
|
"loss": 1.1956, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.3359173126614987, |
|
"grad_norm": 0.847986102104187, |
|
"learning_rate": 1.5482997108297834e-05, |
|
"loss": 1.2041, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.3367786391042205, |
|
"grad_norm": 1.1017706394195557, |
|
"learning_rate": 1.545964314151929e-05, |
|
"loss": 1.2646, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.3376399655469423, |
|
"grad_norm": 0.8774247169494629, |
|
"learning_rate": 1.5436246674935543e-05, |
|
"loss": 1.2236, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.3385012919896641, |
|
"grad_norm": 0.8937813639640808, |
|
"learning_rate": 1.5412807890673015e-05, |
|
"loss": 1.2192, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.33936261843238585, |
|
"grad_norm": 0.8541170358657837, |
|
"learning_rate": 1.5389326971187543e-05, |
|
"loss": 1.2268, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.34022394487510765, |
|
"grad_norm": 0.9567473530769348, |
|
"learning_rate": 1.536580409926296e-05, |
|
"loss": 1.2489, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.34108527131782945, |
|
"grad_norm": 0.8888243436813354, |
|
"learning_rate": 1.5342239458009675e-05, |
|
"loss": 1.2308, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.34194659776055125, |
|
"grad_norm": 0.893013060092926, |
|
"learning_rate": 1.5318633230863237e-05, |
|
"loss": 1.2237, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.34280792420327305, |
|
"grad_norm": 0.874277651309967, |
|
"learning_rate": 1.5294985601582922e-05, |
|
"loss": 1.186, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.34366925064599485, |
|
"grad_norm": 0.8381382822990417, |
|
"learning_rate": 1.5271296754250296e-05, |
|
"loss": 1.2454, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.34453057708871665, |
|
"grad_norm": 0.9372345209121704, |
|
"learning_rate": 1.524756687326777e-05, |
|
"loss": 1.2233, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.3453919035314384, |
|
"grad_norm": 0.869147002696991, |
|
"learning_rate": 1.5223796143357188e-05, |
|
"loss": 1.2084, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.3462532299741602, |
|
"grad_norm": 0.8200699687004089, |
|
"learning_rate": 1.5199984749558367e-05, |
|
"loss": 1.2355, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.347114556416882, |
|
"grad_norm": 0.901074230670929, |
|
"learning_rate": 1.5176132877227674e-05, |
|
"loss": 1.2653, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.3479758828596038, |
|
"grad_norm": 0.871003270149231, |
|
"learning_rate": 1.5152240712036573e-05, |
|
"loss": 1.2289, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.3488372093023256, |
|
"grad_norm": 0.9896201491355896, |
|
"learning_rate": 1.5128308439970174e-05, |
|
"loss": 1.2388, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.3496985357450474, |
|
"grad_norm": 0.9683994054794312, |
|
"learning_rate": 1.5104336247325803e-05, |
|
"loss": 1.2178, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.3505598621877692, |
|
"grad_norm": 0.8784785270690918, |
|
"learning_rate": 1.5080324320711542e-05, |
|
"loss": 1.2929, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.35142118863049093, |
|
"grad_norm": 0.9590929746627808, |
|
"learning_rate": 1.505627284704477e-05, |
|
"loss": 1.2284, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.35228251507321273, |
|
"grad_norm": 0.8345155715942383, |
|
"learning_rate": 1.5032182013550719e-05, |
|
"loss": 1.2614, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.35314384151593453, |
|
"grad_norm": 0.8421818017959595, |
|
"learning_rate": 1.5008052007761009e-05, |
|
"loss": 1.2258, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.35400516795865633, |
|
"grad_norm": 1.0068047046661377, |
|
"learning_rate": 1.498388301751219e-05, |
|
"loss": 1.2686, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.35486649440137813, |
|
"grad_norm": 0.8941269516944885, |
|
"learning_rate": 1.495967523094429e-05, |
|
"loss": 1.2268, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.35572782084409993, |
|
"grad_norm": 0.9728747606277466, |
|
"learning_rate": 1.4935428836499333e-05, |
|
"loss": 1.2359, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.35658914728682173, |
|
"grad_norm": 0.946317732334137, |
|
"learning_rate": 1.4911144022919879e-05, |
|
"loss": 1.2062, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.35745047372954347, |
|
"grad_norm": 0.8988799452781677, |
|
"learning_rate": 1.4886820979247561e-05, |
|
"loss": 1.2389, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.35831180017226527, |
|
"grad_norm": 1.0396331548690796, |
|
"learning_rate": 1.4862459894821606e-05, |
|
"loss": 1.2247, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.35917312661498707, |
|
"grad_norm": 0.8363151550292969, |
|
"learning_rate": 1.483806095927737e-05, |
|
"loss": 1.1869, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.36003445305770887, |
|
"grad_norm": 0.9274395704269409, |
|
"learning_rate": 1.481362436254484e-05, |
|
"loss": 1.1593, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.36089577950043067, |
|
"grad_norm": 1.005424976348877, |
|
"learning_rate": 1.4789150294847192e-05, |
|
"loss": 1.2132, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.36175710594315247, |
|
"grad_norm": 0.9512991309165955, |
|
"learning_rate": 1.4764638946699275e-05, |
|
"loss": 1.2082, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.36261843238587427, |
|
"grad_norm": 0.8919705152511597, |
|
"learning_rate": 1.4740090508906147e-05, |
|
"loss": 1.1693, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.363479758828596, |
|
"grad_norm": 0.8347681760787964, |
|
"learning_rate": 1.4715505172561577e-05, |
|
"loss": 1.2604, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.3643410852713178, |
|
"grad_norm": 0.8956724405288696, |
|
"learning_rate": 1.4690883129046585e-05, |
|
"loss": 1.2241, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.3652024117140396, |
|
"grad_norm": 0.9047289490699768, |
|
"learning_rate": 1.466622457002791e-05, |
|
"loss": 1.2257, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.3660637381567614, |
|
"grad_norm": 1.0073318481445312, |
|
"learning_rate": 1.4641529687456558e-05, |
|
"loss": 1.2385, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.3669250645994832, |
|
"grad_norm": 0.9420737028121948, |
|
"learning_rate": 1.4616798673566276e-05, |
|
"loss": 1.2037, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.367786391042205, |
|
"grad_norm": 0.9980217218399048, |
|
"learning_rate": 1.4592031720872086e-05, |
|
"loss": 1.2163, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.3686477174849268, |
|
"grad_norm": 0.8870885968208313, |
|
"learning_rate": 1.4567229022168756e-05, |
|
"loss": 1.2356, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.3695090439276486, |
|
"grad_norm": 0.9896050095558167, |
|
"learning_rate": 1.454239077052932e-05, |
|
"loss": 1.2334, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.37037037037037035, |
|
"grad_norm": 1.0749104022979736, |
|
"learning_rate": 1.4517517159303573e-05, |
|
"loss": 1.2024, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.37123169681309215, |
|
"grad_norm": 0.8981916904449463, |
|
"learning_rate": 1.4492608382116548e-05, |
|
"loss": 1.1876, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.37209302325581395, |
|
"grad_norm": 0.9399816393852234, |
|
"learning_rate": 1.4467664632867042e-05, |
|
"loss": 1.2465, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.37295434969853575, |
|
"grad_norm": 0.9650474786758423, |
|
"learning_rate": 1.4442686105726066e-05, |
|
"loss": 1.2525, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.37381567614125755, |
|
"grad_norm": 0.9051154255867004, |
|
"learning_rate": 1.4417672995135372e-05, |
|
"loss": 1.2653, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.37467700258397935, |
|
"grad_norm": 0.8648033142089844, |
|
"learning_rate": 1.4392625495805913e-05, |
|
"loss": 1.2221, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.37553832902670115, |
|
"grad_norm": 0.942790687084198, |
|
"learning_rate": 1.4367543802716334e-05, |
|
"loss": 1.2258, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.3763996554694229, |
|
"grad_norm": 0.849907398223877, |
|
"learning_rate": 1.4342428111111461e-05, |
|
"loss": 1.2307, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.3772609819121447, |
|
"grad_norm": 0.9022756218910217, |
|
"learning_rate": 1.4317278616500785e-05, |
|
"loss": 1.2356, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.3781223083548665, |
|
"grad_norm": 1.0210813283920288, |
|
"learning_rate": 1.4292095514656907e-05, |
|
"loss": 1.2099, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.3789836347975883, |
|
"grad_norm": 0.8442511558532715, |
|
"learning_rate": 1.4266879001614067e-05, |
|
"loss": 1.2509, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.3798449612403101, |
|
"grad_norm": 0.8751745223999023, |
|
"learning_rate": 1.424162927366657e-05, |
|
"loss": 1.2039, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.3807062876830319, |
|
"grad_norm": 0.8547453880310059, |
|
"learning_rate": 1.4216346527367284e-05, |
|
"loss": 1.239, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.3815676141257537, |
|
"grad_norm": 0.8485442399978638, |
|
"learning_rate": 1.4191030959526106e-05, |
|
"loss": 1.2208, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.38242894056847543, |
|
"grad_norm": 0.9222948551177979, |
|
"learning_rate": 1.4165682767208426e-05, |
|
"loss": 1.2407, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.3832902670111972, |
|
"grad_norm": 0.8452631235122681, |
|
"learning_rate": 1.4140302147733596e-05, |
|
"loss": 1.2388, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.384151593453919, |
|
"grad_norm": 0.8585206270217896, |
|
"learning_rate": 1.4114889298673383e-05, |
|
"loss": 1.2138, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.3850129198966408, |
|
"grad_norm": 0.8404960036277771, |
|
"learning_rate": 1.4089444417850455e-05, |
|
"loss": 1.2613, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.3858742463393626, |
|
"grad_norm": 0.8335172533988953, |
|
"learning_rate": 1.4063967703336814e-05, |
|
"loss": 1.2188, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.3867355727820844, |
|
"grad_norm": 0.8247851729393005, |
|
"learning_rate": 1.403845935345228e-05, |
|
"loss": 1.2334, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.3875968992248062, |
|
"grad_norm": 0.9984527230262756, |
|
"learning_rate": 1.401291956676292e-05, |
|
"loss": 1.2272, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.38845822566752797, |
|
"grad_norm": 0.8594214916229248, |
|
"learning_rate": 1.3987348542079526e-05, |
|
"loss": 1.1752, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.38931955211024977, |
|
"grad_norm": 0.95940101146698, |
|
"learning_rate": 1.396174647845605e-05, |
|
"loss": 1.2172, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.39018087855297157, |
|
"grad_norm": 0.8622090816497803, |
|
"learning_rate": 1.3936113575188074e-05, |
|
"loss": 1.206, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.39104220499569337, |
|
"grad_norm": 0.8805087804794312, |
|
"learning_rate": 1.3910450031811235e-05, |
|
"loss": 1.2339, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.39190353143841516, |
|
"grad_norm": 0.8449251651763916, |
|
"learning_rate": 1.3884756048099688e-05, |
|
"loss": 1.2253, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.39276485788113696, |
|
"grad_norm": 0.9744880199432373, |
|
"learning_rate": 1.3859031824064543e-05, |
|
"loss": 1.237, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.39362618432385876, |
|
"grad_norm": 0.8672634959220886, |
|
"learning_rate": 1.3833277559952323e-05, |
|
"loss": 1.2199, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.3944875107665805, |
|
"grad_norm": 0.8973323702812195, |
|
"learning_rate": 1.380749345624338e-05, |
|
"loss": 1.1859, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.3953488372093023, |
|
"grad_norm": 0.9631936550140381, |
|
"learning_rate": 1.3781679713650349e-05, |
|
"loss": 1.2133, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.3962101636520241, |
|
"grad_norm": 0.8389044404029846, |
|
"learning_rate": 1.3755836533116597e-05, |
|
"loss": 1.2095, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.3970714900947459, |
|
"grad_norm": 0.8947293162345886, |
|
"learning_rate": 1.3729964115814636e-05, |
|
"loss": 1.1841, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.3979328165374677, |
|
"grad_norm": 0.9099428057670593, |
|
"learning_rate": 1.3704062663144569e-05, |
|
"loss": 1.2188, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.3987941429801895, |
|
"grad_norm": 0.9000226855278015, |
|
"learning_rate": 1.3678132376732518e-05, |
|
"loss": 1.2673, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.3996554694229113, |
|
"grad_norm": 0.9490253925323486, |
|
"learning_rate": 1.3652173458429068e-05, |
|
"loss": 1.2233, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.4005167958656331, |
|
"grad_norm": 0.9906533360481262, |
|
"learning_rate": 1.3626186110307673e-05, |
|
"loss": 1.2125, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.40137812230835485, |
|
"grad_norm": 0.9203736186027527, |
|
"learning_rate": 1.3600170534663097e-05, |
|
"loss": 1.1732, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.40223944875107664, |
|
"grad_norm": 0.8627713918685913, |
|
"learning_rate": 1.3574126934009843e-05, |
|
"loss": 1.2441, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.40310077519379844, |
|
"grad_norm": 0.8728564977645874, |
|
"learning_rate": 1.3548055511080568e-05, |
|
"loss": 1.1814, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.40396210163652024, |
|
"grad_norm": 0.9065693020820618, |
|
"learning_rate": 1.3521956468824505e-05, |
|
"loss": 1.2281, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.40482342807924204, |
|
"grad_norm": 0.860071063041687, |
|
"learning_rate": 1.3495830010405884e-05, |
|
"loss": 1.2438, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.40568475452196384, |
|
"grad_norm": 0.8716009855270386, |
|
"learning_rate": 1.346967633920236e-05, |
|
"loss": 1.2185, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.40654608096468564, |
|
"grad_norm": 1.0129085779190063, |
|
"learning_rate": 1.344349565880341e-05, |
|
"loss": 1.2253, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.4074074074074074, |
|
"grad_norm": 0.8441084623336792, |
|
"learning_rate": 1.3417288173008778e-05, |
|
"loss": 1.2128, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.4082687338501292, |
|
"grad_norm": 0.9617881178855896, |
|
"learning_rate": 1.339105408582685e-05, |
|
"loss": 1.2085, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.409130060292851, |
|
"grad_norm": 0.8780672550201416, |
|
"learning_rate": 1.3364793601473105e-05, |
|
"loss": 1.1989, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.4099913867355728, |
|
"grad_norm": 0.9372087717056274, |
|
"learning_rate": 1.3338506924368494e-05, |
|
"loss": 1.2343, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.4108527131782946, |
|
"grad_norm": 0.8718287348747253, |
|
"learning_rate": 1.331219425913787e-05, |
|
"loss": 1.2009, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.4117140396210164, |
|
"grad_norm": 0.8041454553604126, |
|
"learning_rate": 1.3285855810608377e-05, |
|
"loss": 1.2006, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.4125753660637382, |
|
"grad_norm": 0.8639675378799438, |
|
"learning_rate": 1.325949178380788e-05, |
|
"loss": 1.2244, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.4134366925064599, |
|
"grad_norm": 0.9752419590950012, |
|
"learning_rate": 1.3233102383963341e-05, |
|
"loss": 1.2374, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.4142980189491817, |
|
"grad_norm": 0.9429941177368164, |
|
"learning_rate": 1.3206687816499242e-05, |
|
"loss": 1.2437, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.4151593453919035, |
|
"grad_norm": 0.830390214920044, |
|
"learning_rate": 1.3180248287035977e-05, |
|
"loss": 1.2173, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.4160206718346253, |
|
"grad_norm": 0.8312661051750183, |
|
"learning_rate": 1.3153784001388249e-05, |
|
"loss": 1.2432, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.4168819982773471, |
|
"grad_norm": 0.8787814378738403, |
|
"learning_rate": 1.3127295165563476e-05, |
|
"loss": 1.2465, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.4177433247200689, |
|
"grad_norm": 0.8632099628448486, |
|
"learning_rate": 1.3100781985760188e-05, |
|
"loss": 1.2347, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.4186046511627907, |
|
"grad_norm": 1.0281778573989868, |
|
"learning_rate": 1.3074244668366412e-05, |
|
"loss": 1.2193, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.41946597760551246, |
|
"grad_norm": 0.9415022730827332, |
|
"learning_rate": 1.3047683419958062e-05, |
|
"loss": 1.2021, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.42032730404823426, |
|
"grad_norm": 0.9123542904853821, |
|
"learning_rate": 1.3021098447297358e-05, |
|
"loss": 1.216, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.42118863049095606, |
|
"grad_norm": 0.9520506858825684, |
|
"learning_rate": 1.2994489957331183e-05, |
|
"loss": 1.2148, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.42204995693367786, |
|
"grad_norm": 0.8072666525840759, |
|
"learning_rate": 1.2967858157189495e-05, |
|
"loss": 1.2404, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.42291128337639966, |
|
"grad_norm": 0.9025439620018005, |
|
"learning_rate": 1.29412032541837e-05, |
|
"loss": 1.2183, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.42377260981912146, |
|
"grad_norm": 0.8652177453041077, |
|
"learning_rate": 1.2914525455805056e-05, |
|
"loss": 1.1991, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.42463393626184326, |
|
"grad_norm": 1.0470143556594849, |
|
"learning_rate": 1.2887824969723035e-05, |
|
"loss": 1.2286, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.425495262704565, |
|
"grad_norm": 0.9123655557632446, |
|
"learning_rate": 1.2861102003783722e-05, |
|
"loss": 1.2405, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.4263565891472868, |
|
"grad_norm": 0.8399021625518799, |
|
"learning_rate": 1.2834356766008198e-05, |
|
"loss": 1.1718, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.4272179155900086, |
|
"grad_norm": 1.0004807710647583, |
|
"learning_rate": 1.2807589464590908e-05, |
|
"loss": 1.1792, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.4280792420327304, |
|
"grad_norm": 0.8504483103752136, |
|
"learning_rate": 1.2780800307898057e-05, |
|
"loss": 1.1962, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.4289405684754522, |
|
"grad_norm": 0.9600708484649658, |
|
"learning_rate": 1.2753989504465967e-05, |
|
"loss": 1.2396, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.429801894918174, |
|
"grad_norm": 0.8892411589622498, |
|
"learning_rate": 1.2727157262999481e-05, |
|
"loss": 1.1949, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.4306632213608958, |
|
"grad_norm": 0.9092352986335754, |
|
"learning_rate": 1.270030379237031e-05, |
|
"loss": 1.2467, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.4315245478036176, |
|
"grad_norm": 0.8833507299423218, |
|
"learning_rate": 1.2673429301615431e-05, |
|
"loss": 1.1803, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.43238587424633934, |
|
"grad_norm": 0.8737507462501526, |
|
"learning_rate": 1.2646533999935442e-05, |
|
"loss": 1.2517, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.43324720068906114, |
|
"grad_norm": 0.8569416999816895, |
|
"learning_rate": 1.2619618096692942e-05, |
|
"loss": 1.2096, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.43410852713178294, |
|
"grad_norm": 0.9263505935668945, |
|
"learning_rate": 1.25926818014109e-05, |
|
"loss": 1.1993, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.43496985357450474, |
|
"grad_norm": 1.0186364650726318, |
|
"learning_rate": 1.256572532377103e-05, |
|
"loss": 1.2233, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.43583118001722654, |
|
"grad_norm": 0.83504319190979, |
|
"learning_rate": 1.253874887361214e-05, |
|
"loss": 1.2251, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.43669250645994834, |
|
"grad_norm": 0.9106087684631348, |
|
"learning_rate": 1.2511752660928523e-05, |
|
"loss": 1.2103, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.43755383290267014, |
|
"grad_norm": 0.8864356279373169, |
|
"learning_rate": 1.2484736895868306e-05, |
|
"loss": 1.2015, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.4384151593453919, |
|
"grad_norm": 0.9549910426139832, |
|
"learning_rate": 1.2457701788731812e-05, |
|
"loss": 1.2394, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.4392764857881137, |
|
"grad_norm": 0.9277191758155823, |
|
"learning_rate": 1.2430647549969949e-05, |
|
"loss": 1.2186, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.4401378122308355, |
|
"grad_norm": 0.9281412363052368, |
|
"learning_rate": 1.2403574390182529e-05, |
|
"loss": 1.2122, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.4409991386735573, |
|
"grad_norm": 0.8452510833740234, |
|
"learning_rate": 1.2376482520116666e-05, |
|
"loss": 1.2203, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.4418604651162791, |
|
"grad_norm": 0.9457308650016785, |
|
"learning_rate": 1.2349372150665117e-05, |
|
"loss": 1.2073, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.4427217915590009, |
|
"grad_norm": 0.8949629664421082, |
|
"learning_rate": 1.2322243492864651e-05, |
|
"loss": 1.2078, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.4435831180017227, |
|
"grad_norm": 0.9043914079666138, |
|
"learning_rate": 1.2295096757894389e-05, |
|
"loss": 1.2159, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.4444444444444444, |
|
"grad_norm": 0.8685716390609741, |
|
"learning_rate": 1.2267932157074178e-05, |
|
"loss": 1.2262, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.4453057708871662, |
|
"grad_norm": 0.8798421621322632, |
|
"learning_rate": 1.224074990186294e-05, |
|
"loss": 1.2551, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.446167097329888, |
|
"grad_norm": 0.9253307580947876, |
|
"learning_rate": 1.2213550203857025e-05, |
|
"loss": 1.2072, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.4470284237726098, |
|
"grad_norm": 0.9633350372314453, |
|
"learning_rate": 1.2186333274788558e-05, |
|
"loss": 1.2122, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.4478897502153316, |
|
"grad_norm": 0.8635453581809998, |
|
"learning_rate": 1.21590993265238e-05, |
|
"loss": 1.2107, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.4487510766580534, |
|
"grad_norm": 0.8736230134963989, |
|
"learning_rate": 1.2131848571061501e-05, |
|
"loss": 1.2315, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.4496124031007752, |
|
"grad_norm": 0.8366631269454956, |
|
"learning_rate": 1.2104581220531237e-05, |
|
"loss": 1.2226, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.45047372954349696, |
|
"grad_norm": 0.9264794588088989, |
|
"learning_rate": 1.2077297487191771e-05, |
|
"loss": 1.2, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.45133505598621876, |
|
"grad_norm": 0.8290119171142578, |
|
"learning_rate": 1.2049997583429389e-05, |
|
"loss": 1.2188, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.45219638242894056, |
|
"grad_norm": 0.9655864238739014, |
|
"learning_rate": 1.202268172175626e-05, |
|
"loss": 1.2326, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.45305770887166236, |
|
"grad_norm": 0.8329289555549622, |
|
"learning_rate": 1.1995350114808772e-05, |
|
"loss": 1.2218, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.45391903531438416, |
|
"grad_norm": 0.9534193873405457, |
|
"learning_rate": 1.1968002975345882e-05, |
|
"loss": 1.2072, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.45478036175710596, |
|
"grad_norm": 0.8719503283500671, |
|
"learning_rate": 1.194064051624745e-05, |
|
"loss": 1.2469, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.45564168819982775, |
|
"grad_norm": 0.9249786138534546, |
|
"learning_rate": 1.1913262950512605e-05, |
|
"loss": 1.2378, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.45650301464254955, |
|
"grad_norm": 0.8721954226493835, |
|
"learning_rate": 1.1885870491258054e-05, |
|
"loss": 1.1767, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.4573643410852713, |
|
"grad_norm": 0.9456660747528076, |
|
"learning_rate": 1.185846335171645e-05, |
|
"loss": 1.1892, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.4582256675279931, |
|
"grad_norm": 0.856704592704773, |
|
"learning_rate": 1.1831041745234728e-05, |
|
"loss": 1.2102, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.4590869939707149, |
|
"grad_norm": 0.8333457112312317, |
|
"learning_rate": 1.180360588527242e-05, |
|
"loss": 1.2255, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.4599483204134367, |
|
"grad_norm": 0.9460641741752625, |
|
"learning_rate": 1.177615598540003e-05, |
|
"loss": 1.1819, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.4608096468561585, |
|
"grad_norm": 0.830569326877594, |
|
"learning_rate": 1.1748692259297347e-05, |
|
"loss": 1.2227, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.4616709732988803, |
|
"grad_norm": 0.8055482506752014, |
|
"learning_rate": 1.172121492075179e-05, |
|
"loss": 1.2011, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.4625322997416021, |
|
"grad_norm": 0.9891955256462097, |
|
"learning_rate": 1.169372418365674e-05, |
|
"loss": 1.1889, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.46339362618432384, |
|
"grad_norm": 0.9944434762001038, |
|
"learning_rate": 1.1666220262009877e-05, |
|
"loss": 1.185, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.46425495262704564, |
|
"grad_norm": 0.8905592560768127, |
|
"learning_rate": 1.1638703369911517e-05, |
|
"loss": 1.2404, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.46511627906976744, |
|
"grad_norm": 0.8993557691574097, |
|
"learning_rate": 1.161117372156294e-05, |
|
"loss": 1.1786, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.46597760551248923, |
|
"grad_norm": 0.9197692275047302, |
|
"learning_rate": 1.1583631531264723e-05, |
|
"loss": 1.2178, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.46683893195521103, |
|
"grad_norm": 0.8710423111915588, |
|
"learning_rate": 1.1556077013415084e-05, |
|
"loss": 1.2422, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.46770025839793283, |
|
"grad_norm": 0.8500446081161499, |
|
"learning_rate": 1.152851038250819e-05, |
|
"loss": 1.2196, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.46856158484065463, |
|
"grad_norm": 0.88962721824646, |
|
"learning_rate": 1.150093185313251e-05, |
|
"loss": 1.1858, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.4694229112833764, |
|
"grad_norm": 0.907476544380188, |
|
"learning_rate": 1.147334163996913e-05, |
|
"loss": 1.1906, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.4702842377260982, |
|
"grad_norm": 0.9100911617279053, |
|
"learning_rate": 1.1445739957790087e-05, |
|
"loss": 1.2185, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.47114556416882, |
|
"grad_norm": 0.9100723266601562, |
|
"learning_rate": 1.14181270214567e-05, |
|
"loss": 1.2539, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.4720068906115418, |
|
"grad_norm": 0.9603404998779297, |
|
"learning_rate": 1.1390503045917892e-05, |
|
"loss": 1.1917, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.4728682170542636, |
|
"grad_norm": 0.8922184109687805, |
|
"learning_rate": 1.1362868246208519e-05, |
|
"loss": 1.1583, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.4737295434969854, |
|
"grad_norm": 0.9883030652999878, |
|
"learning_rate": 1.1335222837447692e-05, |
|
"loss": 1.2329, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.47459086993970717, |
|
"grad_norm": 0.9090227484703064, |
|
"learning_rate": 1.1307567034837123e-05, |
|
"loss": 1.2424, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.4754521963824289, |
|
"grad_norm": 0.8161377906799316, |
|
"learning_rate": 1.127990105365941e-05, |
|
"loss": 1.2166, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.4763135228251507, |
|
"grad_norm": 0.9323464632034302, |
|
"learning_rate": 1.1252225109276404e-05, |
|
"loss": 1.1941, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.4771748492678725, |
|
"grad_norm": 0.9303016662597656, |
|
"learning_rate": 1.12245394171275e-05, |
|
"loss": 1.1916, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.4780361757105943, |
|
"grad_norm": 0.9320595264434814, |
|
"learning_rate": 1.1196844192727984e-05, |
|
"loss": 1.2087, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.4788975021533161, |
|
"grad_norm": 0.848597526550293, |
|
"learning_rate": 1.1169139651667334e-05, |
|
"loss": 1.1857, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.4797588285960379, |
|
"grad_norm": 0.9183049201965332, |
|
"learning_rate": 1.1141426009607562e-05, |
|
"loss": 1.2185, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.4806201550387597, |
|
"grad_norm": 0.9825764298439026, |
|
"learning_rate": 1.1113703482281515e-05, |
|
"loss": 1.2045, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.48148148148148145, |
|
"grad_norm": 0.8843861818313599, |
|
"learning_rate": 1.1085972285491213e-05, |
|
"loss": 1.198, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.48234280792420325, |
|
"grad_norm": 0.8294288516044617, |
|
"learning_rate": 1.1058232635106167e-05, |
|
"loss": 1.1729, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.48320413436692505, |
|
"grad_norm": 0.810230553150177, |
|
"learning_rate": 1.103048474706168e-05, |
|
"loss": 1.2055, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.48406546080964685, |
|
"grad_norm": 0.8811335563659668, |
|
"learning_rate": 1.1002728837357192e-05, |
|
"loss": 1.2168, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.48492678725236865, |
|
"grad_norm": 0.9530275464057922, |
|
"learning_rate": 1.097496512205458e-05, |
|
"loss": 1.2179, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.48578811369509045, |
|
"grad_norm": 0.9637301564216614, |
|
"learning_rate": 1.0947193817276485e-05, |
|
"loss": 1.1692, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.48664944013781225, |
|
"grad_norm": 0.8479545712471008, |
|
"learning_rate": 1.0919415139204625e-05, |
|
"loss": 1.2172, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.48751076658053405, |
|
"grad_norm": 0.9883643388748169, |
|
"learning_rate": 1.089162930407812e-05, |
|
"loss": 1.2697, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.4883720930232558, |
|
"grad_norm": 0.8231867551803589, |
|
"learning_rate": 1.0863836528191795e-05, |
|
"loss": 1.215, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.4892334194659776, |
|
"grad_norm": 0.8607746362686157, |
|
"learning_rate": 1.0836037027894515e-05, |
|
"loss": 1.2322, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.4900947459086994, |
|
"grad_norm": 0.85414719581604, |
|
"learning_rate": 1.0808231019587472e-05, |
|
"loss": 1.2044, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.4909560723514212, |
|
"grad_norm": 0.835567831993103, |
|
"learning_rate": 1.0780418719722544e-05, |
|
"loss": 1.218, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.491817398794143, |
|
"grad_norm": 0.841223418712616, |
|
"learning_rate": 1.075260034480056e-05, |
|
"loss": 1.2253, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.4926787252368648, |
|
"grad_norm": 0.8959380984306335, |
|
"learning_rate": 1.0724776111369654e-05, |
|
"loss": 1.1758, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.4935400516795866, |
|
"grad_norm": 0.9385613799095154, |
|
"learning_rate": 1.0696946236023566e-05, |
|
"loss": 1.2174, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.49440137812230833, |
|
"grad_norm": 0.9247633814811707, |
|
"learning_rate": 1.0669110935399944e-05, |
|
"loss": 1.2104, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.49526270456503013, |
|
"grad_norm": 0.8738383650779724, |
|
"learning_rate": 1.0641270426178677e-05, |
|
"loss": 1.1825, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.49612403100775193, |
|
"grad_norm": 1.0910409688949585, |
|
"learning_rate": 1.0613424925080194e-05, |
|
"loss": 1.1731, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.49698535745047373, |
|
"grad_norm": 0.8407815098762512, |
|
"learning_rate": 1.058557464886379e-05, |
|
"loss": 1.2002, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 0.49784668389319553, |
|
"grad_norm": 0.9195674657821655, |
|
"learning_rate": 1.055771981432592e-05, |
|
"loss": 1.2013, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.49870801033591733, |
|
"grad_norm": 0.8779056072235107, |
|
"learning_rate": 1.0529860638298535e-05, |
|
"loss": 1.2098, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.49956933677863913, |
|
"grad_norm": 0.8528910875320435, |
|
"learning_rate": 1.0501997337647372e-05, |
|
"loss": 1.1676, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.5004306632213609, |
|
"grad_norm": 0.8757944703102112, |
|
"learning_rate": 1.0474130129270281e-05, |
|
"loss": 1.1858, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 0.5012919896640827, |
|
"grad_norm": 0.8935431241989136, |
|
"learning_rate": 1.0446259230095531e-05, |
|
"loss": 1.1776, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.5021533161068045, |
|
"grad_norm": 0.9777861833572388, |
|
"learning_rate": 1.0418384857080118e-05, |
|
"loss": 1.1756, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 0.5030146425495263, |
|
"grad_norm": 0.9448710680007935, |
|
"learning_rate": 1.039050722720808e-05, |
|
"loss": 1.2015, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.5038759689922481, |
|
"grad_norm": 0.8520198464393616, |
|
"learning_rate": 1.0362626557488811e-05, |
|
"loss": 1.1893, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.5047372954349698, |
|
"grad_norm": 0.9623661041259766, |
|
"learning_rate": 1.0334743064955367e-05, |
|
"loss": 1.1882, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.5055986218776917, |
|
"grad_norm": 0.8741092681884766, |
|
"learning_rate": 1.0306856966662776e-05, |
|
"loss": 1.17, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 0.5064599483204134, |
|
"grad_norm": 0.8488198518753052, |
|
"learning_rate": 1.027896847968635e-05, |
|
"loss": 1.2303, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.5073212747631353, |
|
"grad_norm": 0.9547082185745239, |
|
"learning_rate": 1.0251077821119998e-05, |
|
"loss": 1.2269, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.508182601205857, |
|
"grad_norm": 1.0101442337036133, |
|
"learning_rate": 1.0223185208074538e-05, |
|
"loss": 1.191, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.5090439276485789, |
|
"grad_norm": 0.8840098977088928, |
|
"learning_rate": 1.0195290857675982e-05, |
|
"loss": 1.1732, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.5099052540913006, |
|
"grad_norm": 0.7868524789810181, |
|
"learning_rate": 1.0167394987063894e-05, |
|
"loss": 1.1934, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.5107665805340224, |
|
"grad_norm": 0.8917614817619324, |
|
"learning_rate": 1.0139497813389654e-05, |
|
"loss": 1.2006, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 0.5116279069767442, |
|
"grad_norm": 0.8367542624473572, |
|
"learning_rate": 1.0111599553814788e-05, |
|
"loss": 1.1675, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.512489233419466, |
|
"grad_norm": 0.8534975647926331, |
|
"learning_rate": 1.008370042550928e-05, |
|
"loss": 1.1979, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.5133505598621878, |
|
"grad_norm": 0.8273769617080688, |
|
"learning_rate": 1.0055800645649874e-05, |
|
"loss": 1.1826, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.5142118863049095, |
|
"grad_norm": 0.9318427443504333, |
|
"learning_rate": 1.002790043141838e-05, |
|
"loss": 1.18, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.5150732127476314, |
|
"grad_norm": 0.9464846849441528, |
|
"learning_rate": 1e-05, |
|
"loss": 1.1945, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.5159345391903531, |
|
"grad_norm": 0.893747866153717, |
|
"learning_rate": 9.972099568581621e-06, |
|
"loss": 1.2057, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 0.5167958656330749, |
|
"grad_norm": 0.9831318259239197, |
|
"learning_rate": 9.94419935435013e-06, |
|
"loss": 1.2543, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.5176571920757967, |
|
"grad_norm": 0.8675307035446167, |
|
"learning_rate": 9.916299574490722e-06, |
|
"loss": 1.1924, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 0.5185185185185185, |
|
"grad_norm": 0.8235703706741333, |
|
"learning_rate": 9.888400446185212e-06, |
|
"loss": 1.205, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.5193798449612403, |
|
"grad_norm": 0.9362935423851013, |
|
"learning_rate": 9.860502186610349e-06, |
|
"loss": 1.2045, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.5202411714039621, |
|
"grad_norm": 1.0003681182861328, |
|
"learning_rate": 9.832605012936107e-06, |
|
"loss": 1.1725, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.5211024978466839, |
|
"grad_norm": 0.8889096975326538, |
|
"learning_rate": 9.80470914232402e-06, |
|
"loss": 1.2078, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.5219638242894057, |
|
"grad_norm": 0.9049688577651978, |
|
"learning_rate": 9.77681479192547e-06, |
|
"loss": 1.1603, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.5228251507321274, |
|
"grad_norm": 0.8464852571487427, |
|
"learning_rate": 9.748922178880005e-06, |
|
"loss": 1.2056, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 0.5236864771748493, |
|
"grad_norm": 0.9175320267677307, |
|
"learning_rate": 9.721031520313653e-06, |
|
"loss": 1.2323, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.524547803617571, |
|
"grad_norm": 0.8868247270584106, |
|
"learning_rate": 9.693143033337228e-06, |
|
"loss": 1.186, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.5254091300602929, |
|
"grad_norm": 0.8364291191101074, |
|
"learning_rate": 9.665256935044636e-06, |
|
"loss": 1.1761, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.5262704565030146, |
|
"grad_norm": 0.8297803401947021, |
|
"learning_rate": 9.637373442511192e-06, |
|
"loss": 1.1775, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 0.5271317829457365, |
|
"grad_norm": 0.9279587268829346, |
|
"learning_rate": 9.609492772791924e-06, |
|
"loss": 1.1789, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.5279931093884582, |
|
"grad_norm": 0.8592475056648254, |
|
"learning_rate": 9.581615142919887e-06, |
|
"loss": 1.2299, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 0.52885443583118, |
|
"grad_norm": 0.8591251373291016, |
|
"learning_rate": 9.55374076990447e-06, |
|
"loss": 1.1792, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.5297157622739018, |
|
"grad_norm": 0.9070160388946533, |
|
"learning_rate": 9.525869870729719e-06, |
|
"loss": 1.1989, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.5305770887166236, |
|
"grad_norm": 0.8992422223091125, |
|
"learning_rate": 9.49800266235263e-06, |
|
"loss": 1.1947, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.5314384151593454, |
|
"grad_norm": 0.8558951616287231, |
|
"learning_rate": 9.470139361701469e-06, |
|
"loss": 1.2101, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 0.5322997416020672, |
|
"grad_norm": 0.9297447800636292, |
|
"learning_rate": 9.442280185674084e-06, |
|
"loss": 1.1918, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.533161068044789, |
|
"grad_norm": 0.8094662427902222, |
|
"learning_rate": 9.414425351136215e-06, |
|
"loss": 1.2102, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 0.5340223944875108, |
|
"grad_norm": 0.8350943326950073, |
|
"learning_rate": 9.386575074919806e-06, |
|
"loss": 1.1809, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.5348837209302325, |
|
"grad_norm": 0.9019107818603516, |
|
"learning_rate": 9.358729573821325e-06, |
|
"loss": 1.1814, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.5357450473729544, |
|
"grad_norm": 0.8287387490272522, |
|
"learning_rate": 9.330889064600058e-06, |
|
"loss": 1.2297, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.5366063738156761, |
|
"grad_norm": 0.8505203723907471, |
|
"learning_rate": 9.303053763976436e-06, |
|
"loss": 1.2028, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 0.537467700258398, |
|
"grad_norm": 0.8712747097015381, |
|
"learning_rate": 9.275223888630348e-06, |
|
"loss": 1.1718, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.5383290267011197, |
|
"grad_norm": 0.907810389995575, |
|
"learning_rate": 9.247399655199444e-06, |
|
"loss": 1.2403, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.5391903531438416, |
|
"grad_norm": 0.8972229957580566, |
|
"learning_rate": 9.219581280277463e-06, |
|
"loss": 1.1864, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 0.5400516795865633, |
|
"grad_norm": 0.897445023059845, |
|
"learning_rate": 9.191768980412528e-06, |
|
"loss": 1.1663, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 0.540913006029285, |
|
"grad_norm": 0.813290536403656, |
|
"learning_rate": 9.163962972105488e-06, |
|
"loss": 1.2084, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 0.5417743324720069, |
|
"grad_norm": 0.8626779913902283, |
|
"learning_rate": 9.136163471808207e-06, |
|
"loss": 1.2222, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 0.5426356589147286, |
|
"grad_norm": 0.949863076210022, |
|
"learning_rate": 9.108370695921884e-06, |
|
"loss": 1.2068, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.5434969853574505, |
|
"grad_norm": 1.0227597951889038, |
|
"learning_rate": 9.080584860795378e-06, |
|
"loss": 1.2174, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 0.5443583118001722, |
|
"grad_norm": 0.8536685705184937, |
|
"learning_rate": 9.05280618272352e-06, |
|
"loss": 1.1978, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.5452196382428941, |
|
"grad_norm": 0.8163442015647888, |
|
"learning_rate": 9.025034877945422e-06, |
|
"loss": 1.1749, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 0.5460809646856158, |
|
"grad_norm": 0.8537796139717102, |
|
"learning_rate": 8.99727116264281e-06, |
|
"loss": 1.1979, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 0.5469422911283376, |
|
"grad_norm": 0.8340730667114258, |
|
"learning_rate": 8.969515252938323e-06, |
|
"loss": 1.1756, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.5478036175710594, |
|
"grad_norm": 0.8549203276634216, |
|
"learning_rate": 8.941767364893836e-06, |
|
"loss": 1.1931, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.5486649440137812, |
|
"grad_norm": 0.8984569311141968, |
|
"learning_rate": 8.914027714508788e-06, |
|
"loss": 1.167, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 0.549526270456503, |
|
"grad_norm": 0.8956097960472107, |
|
"learning_rate": 8.88629651771849e-06, |
|
"loss": 1.1972, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.5503875968992248, |
|
"grad_norm": 0.889928936958313, |
|
"learning_rate": 8.85857399039244e-06, |
|
"loss": 1.1852, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 0.5512489233419466, |
|
"grad_norm": 0.8526130318641663, |
|
"learning_rate": 8.830860348332666e-06, |
|
"loss": 1.1825, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.5521102497846684, |
|
"grad_norm": 0.9533372521400452, |
|
"learning_rate": 8.803155807272019e-06, |
|
"loss": 1.208, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 0.5529715762273901, |
|
"grad_norm": 1.0128767490386963, |
|
"learning_rate": 8.775460582872502e-06, |
|
"loss": 1.2077, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.553832902670112, |
|
"grad_norm": 0.9490034580230713, |
|
"learning_rate": 8.7477748907236e-06, |
|
"loss": 1.2078, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 0.5546942291128337, |
|
"grad_norm": 0.9446279406547546, |
|
"learning_rate": 8.720098946340594e-06, |
|
"loss": 1.1819, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.5555555555555556, |
|
"grad_norm": 0.9068703055381775, |
|
"learning_rate": 8.69243296516288e-06, |
|
"loss": 1.2096, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.5564168819982773, |
|
"grad_norm": 0.8627210259437561, |
|
"learning_rate": 8.664777162552308e-06, |
|
"loss": 1.1981, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 0.5572782084409992, |
|
"grad_norm": 0.8679218292236328, |
|
"learning_rate": 8.637131753791485e-06, |
|
"loss": 1.2136, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 0.5581395348837209, |
|
"grad_norm": 0.8559207916259766, |
|
"learning_rate": 8.609496954082113e-06, |
|
"loss": 1.184, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 0.5590008613264428, |
|
"grad_norm": 0.8170911073684692, |
|
"learning_rate": 8.581872978543305e-06, |
|
"loss": 1.1849, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 0.5598621877691645, |
|
"grad_norm": 0.9225674867630005, |
|
"learning_rate": 8.554260042209918e-06, |
|
"loss": 1.1612, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.5607235142118863, |
|
"grad_norm": 0.8442004919052124, |
|
"learning_rate": 8.526658360030876e-06, |
|
"loss": 1.195, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.5615848406546081, |
|
"grad_norm": 0.9129129648208618, |
|
"learning_rate": 8.499068146867492e-06, |
|
"loss": 1.2332, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 0.5624461670973299, |
|
"grad_norm": 0.8818443417549133, |
|
"learning_rate": 8.471489617491813e-06, |
|
"loss": 1.1633, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 0.5633074935400517, |
|
"grad_norm": 0.8677276968955994, |
|
"learning_rate": 8.44392298658492e-06, |
|
"loss": 1.2062, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 0.5641688199827735, |
|
"grad_norm": 0.9707540273666382, |
|
"learning_rate": 8.41636846873528e-06, |
|
"loss": 1.1585, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.5650301464254953, |
|
"grad_norm": 0.929904580116272, |
|
"learning_rate": 8.388826278437066e-06, |
|
"loss": 1.1807, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 0.5658914728682171, |
|
"grad_norm": 0.8509595394134521, |
|
"learning_rate": 8.361296630088488e-06, |
|
"loss": 1.1799, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 0.5667527993109388, |
|
"grad_norm": 0.9297439455986023, |
|
"learning_rate": 8.333779737990124e-06, |
|
"loss": 1.2112, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 0.5676141257536607, |
|
"grad_norm": 0.913957953453064, |
|
"learning_rate": 8.306275816343262e-06, |
|
"loss": 1.1909, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 0.5684754521963824, |
|
"grad_norm": 0.8199489116668701, |
|
"learning_rate": 8.278785079248211e-06, |
|
"loss": 1.1718, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.5693367786391043, |
|
"grad_norm": 0.9394577741622925, |
|
"learning_rate": 8.251307740702656e-06, |
|
"loss": 1.1927, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 0.570198105081826, |
|
"grad_norm": 1.024397611618042, |
|
"learning_rate": 8.223844014599973e-06, |
|
"loss": 1.1849, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 0.5710594315245479, |
|
"grad_norm": 0.9490551352500916, |
|
"learning_rate": 8.196394114727586e-06, |
|
"loss": 1.1655, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 0.5719207579672696, |
|
"grad_norm": 0.8966484069824219, |
|
"learning_rate": 8.168958254765275e-06, |
|
"loss": 1.1916, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 0.5727820844099913, |
|
"grad_norm": 0.8795093894004822, |
|
"learning_rate": 8.14153664828355e-06, |
|
"loss": 1.2319, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.5736434108527132, |
|
"grad_norm": 0.8979016542434692, |
|
"learning_rate": 8.114129508741947e-06, |
|
"loss": 1.1894, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 0.5745047372954349, |
|
"grad_norm": 0.9347522258758545, |
|
"learning_rate": 8.086737049487398e-06, |
|
"loss": 1.1941, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 0.5753660637381568, |
|
"grad_norm": 1.0714643001556396, |
|
"learning_rate": 8.059359483752551e-06, |
|
"loss": 1.2061, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 0.5762273901808785, |
|
"grad_norm": 0.8999515175819397, |
|
"learning_rate": 8.031997024654123e-06, |
|
"loss": 1.1827, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 0.5770887166236004, |
|
"grad_norm": 0.8698625564575195, |
|
"learning_rate": 8.00464988519123e-06, |
|
"loss": 1.1879, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.5779500430663221, |
|
"grad_norm": 0.8773601651191711, |
|
"learning_rate": 7.977318278243742e-06, |
|
"loss": 1.1733, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 0.5788113695090439, |
|
"grad_norm": 0.8871725797653198, |
|
"learning_rate": 7.950002416570614e-06, |
|
"loss": 1.1988, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.5796726959517657, |
|
"grad_norm": 0.8334397673606873, |
|
"learning_rate": 7.92270251280823e-06, |
|
"loss": 1.1846, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 0.5805340223944875, |
|
"grad_norm": 0.9135998487472534, |
|
"learning_rate": 7.895418779468766e-06, |
|
"loss": 1.1776, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 0.5813953488372093, |
|
"grad_norm": 1.04806649684906, |
|
"learning_rate": 7.868151428938502e-06, |
|
"loss": 1.1788, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.5822566752799311, |
|
"grad_norm": 0.8940375447273254, |
|
"learning_rate": 7.840900673476204e-06, |
|
"loss": 1.1911, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 0.5831180017226529, |
|
"grad_norm": 0.8850206136703491, |
|
"learning_rate": 7.813666725211445e-06, |
|
"loss": 1.1773, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 0.5839793281653747, |
|
"grad_norm": 0.9374780654907227, |
|
"learning_rate": 7.786449796142979e-06, |
|
"loss": 1.2115, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 0.5848406546080964, |
|
"grad_norm": 0.9776993989944458, |
|
"learning_rate": 7.759250098137061e-06, |
|
"loss": 1.1713, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 0.5857019810508183, |
|
"grad_norm": 0.8467763662338257, |
|
"learning_rate": 7.732067842925823e-06, |
|
"loss": 1.1584, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.58656330749354, |
|
"grad_norm": 0.8266571164131165, |
|
"learning_rate": 7.704903242105616e-06, |
|
"loss": 1.1749, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 0.5874246339362619, |
|
"grad_norm": 0.8675100803375244, |
|
"learning_rate": 7.677756507135354e-06, |
|
"loss": 1.1706, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 0.5882859603789836, |
|
"grad_norm": 0.8470680713653564, |
|
"learning_rate": 7.650627849334881e-06, |
|
"loss": 1.186, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 0.5891472868217055, |
|
"grad_norm": 0.8331758975982666, |
|
"learning_rate": 7.623517479883335e-06, |
|
"loss": 1.1826, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 0.5900086132644272, |
|
"grad_norm": 0.8689389228820801, |
|
"learning_rate": 7.596425609817474e-06, |
|
"loss": 1.1931, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.590869939707149, |
|
"grad_norm": 0.9240257740020752, |
|
"learning_rate": 7.569352450030054e-06, |
|
"loss": 1.188, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 0.5917312661498708, |
|
"grad_norm": 0.9663352370262146, |
|
"learning_rate": 7.542298211268189e-06, |
|
"loss": 1.1903, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 0.5925925925925926, |
|
"grad_norm": 0.9322826266288757, |
|
"learning_rate": 7.515263104131699e-06, |
|
"loss": 1.1991, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 0.5934539190353144, |
|
"grad_norm": 0.8683158159255981, |
|
"learning_rate": 7.488247339071478e-06, |
|
"loss": 1.2022, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 0.5943152454780362, |
|
"grad_norm": 0.9323744177818298, |
|
"learning_rate": 7.461251126387863e-06, |
|
"loss": 1.1404, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.595176571920758, |
|
"grad_norm": 0.9025394320487976, |
|
"learning_rate": 7.434274676228973e-06, |
|
"loss": 1.2, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 0.5960378983634798, |
|
"grad_norm": 0.858193039894104, |
|
"learning_rate": 7.407318198589102e-06, |
|
"loss": 1.1964, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 0.5968992248062015, |
|
"grad_norm": 0.8841907382011414, |
|
"learning_rate": 7.380381903307061e-06, |
|
"loss": 1.1968, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 0.5977605512489234, |
|
"grad_norm": 0.8853846788406372, |
|
"learning_rate": 7.353466000064563e-06, |
|
"loss": 1.222, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 0.5986218776916451, |
|
"grad_norm": 0.9315546154975891, |
|
"learning_rate": 7.326570698384569e-06, |
|
"loss": 1.1806, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.599483204134367, |
|
"grad_norm": 0.8450863361358643, |
|
"learning_rate": 7.299696207629692e-06, |
|
"loss": 1.1658, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.6003445305770887, |
|
"grad_norm": 0.8357823491096497, |
|
"learning_rate": 7.2728427370005205e-06, |
|
"loss": 1.21, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 0.6012058570198106, |
|
"grad_norm": 0.8645495772361755, |
|
"learning_rate": 7.246010495534036e-06, |
|
"loss": 1.2054, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 0.6020671834625323, |
|
"grad_norm": 0.9765591621398926, |
|
"learning_rate": 7.2191996921019485e-06, |
|
"loss": 1.2005, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 0.602928509905254, |
|
"grad_norm": 0.8903208374977112, |
|
"learning_rate": 7.1924105354090955e-06, |
|
"loss": 1.1458, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.6037898363479759, |
|
"grad_norm": 0.8536180853843689, |
|
"learning_rate": 7.165643233991806e-06, |
|
"loss": 1.1817, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 0.6046511627906976, |
|
"grad_norm": 0.9730960726737976, |
|
"learning_rate": 7.138897996216278e-06, |
|
"loss": 1.1883, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 0.6055124892334195, |
|
"grad_norm": 0.9186335802078247, |
|
"learning_rate": 7.1121750302769685e-06, |
|
"loss": 1.1818, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 0.6063738156761412, |
|
"grad_norm": 0.8939233422279358, |
|
"learning_rate": 7.085474544194946e-06, |
|
"loss": 1.1446, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 0.6072351421188631, |
|
"grad_norm": 0.9298137426376343, |
|
"learning_rate": 7.058796745816303e-06, |
|
"loss": 1.1787, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.6080964685615848, |
|
"grad_norm": 0.8468939065933228, |
|
"learning_rate": 7.03214184281051e-06, |
|
"loss": 1.2003, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 0.6089577950043066, |
|
"grad_norm": 0.8879836201667786, |
|
"learning_rate": 7.0055100426688205e-06, |
|
"loss": 1.1677, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 0.6098191214470284, |
|
"grad_norm": 0.8849911093711853, |
|
"learning_rate": 6.978901552702643e-06, |
|
"loss": 1.2002, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 0.6106804478897502, |
|
"grad_norm": 0.9499090313911438, |
|
"learning_rate": 6.95231658004194e-06, |
|
"loss": 1.2391, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 0.611541774332472, |
|
"grad_norm": 0.9845093488693237, |
|
"learning_rate": 6.925755331633592e-06, |
|
"loss": 1.1686, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.6124031007751938, |
|
"grad_norm": 0.8965116739273071, |
|
"learning_rate": 6.899218014239815e-06, |
|
"loss": 1.1862, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 0.6132644272179156, |
|
"grad_norm": 0.8552091121673584, |
|
"learning_rate": 6.872704834436526e-06, |
|
"loss": 1.2219, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 0.6141257536606374, |
|
"grad_norm": 0.853168785572052, |
|
"learning_rate": 6.846215998611757e-06, |
|
"loss": 1.1754, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 0.6149870801033591, |
|
"grad_norm": 0.8855804204940796, |
|
"learning_rate": 6.8197517129640265e-06, |
|
"loss": 1.222, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 0.615848406546081, |
|
"grad_norm": 1.0299283266067505, |
|
"learning_rate": 6.79331218350076e-06, |
|
"loss": 1.1431, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.6167097329888027, |
|
"grad_norm": 0.9015288352966309, |
|
"learning_rate": 6.766897616036661e-06, |
|
"loss": 1.1595, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 0.6175710594315246, |
|
"grad_norm": 0.9094573855400085, |
|
"learning_rate": 6.740508216192121e-06, |
|
"loss": 1.202, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 0.6184323858742463, |
|
"grad_norm": 0.8764570355415344, |
|
"learning_rate": 6.714144189391625e-06, |
|
"loss": 1.1975, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 0.6192937123169682, |
|
"grad_norm": 0.8982758522033691, |
|
"learning_rate": 6.6878057408621345e-06, |
|
"loss": 1.1991, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 0.6201550387596899, |
|
"grad_norm": 0.8324359059333801, |
|
"learning_rate": 6.661493075631506e-06, |
|
"loss": 1.1908, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.6210163652024118, |
|
"grad_norm": 0.8555594682693481, |
|
"learning_rate": 6.635206398526895e-06, |
|
"loss": 1.1635, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 0.6218776916451335, |
|
"grad_norm": 0.9507735371589661, |
|
"learning_rate": 6.60894591417315e-06, |
|
"loss": 1.2447, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 0.6227390180878553, |
|
"grad_norm": 0.8586412072181702, |
|
"learning_rate": 6.582711826991226e-06, |
|
"loss": 1.1975, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 0.6236003445305771, |
|
"grad_norm": 0.8518699407577515, |
|
"learning_rate": 6.556504341196592e-06, |
|
"loss": 1.2073, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 0.6244616709732989, |
|
"grad_norm": 0.8891175389289856, |
|
"learning_rate": 6.5303236607976465e-06, |
|
"loss": 1.1805, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.6253229974160207, |
|
"grad_norm": 1.035131812095642, |
|
"learning_rate": 6.504169989594121e-06, |
|
"loss": 1.1953, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 0.6261843238587425, |
|
"grad_norm": 1.0852066278457642, |
|
"learning_rate": 6.4780435311754986e-06, |
|
"loss": 1.1504, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 0.6270456503014643, |
|
"grad_norm": 0.848870575428009, |
|
"learning_rate": 6.451944488919433e-06, |
|
"loss": 1.1743, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 0.627906976744186, |
|
"grad_norm": 0.8880192041397095, |
|
"learning_rate": 6.425873065990158e-06, |
|
"loss": 1.156, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 0.6287683031869078, |
|
"grad_norm": 0.9287259578704834, |
|
"learning_rate": 6.3998294653369046e-06, |
|
"loss": 1.2049, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.6296296296296297, |
|
"grad_norm": 0.9466278553009033, |
|
"learning_rate": 6.373813889692331e-06, |
|
"loss": 1.209, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 0.6304909560723514, |
|
"grad_norm": 0.8213219046592712, |
|
"learning_rate": 6.347826541570936e-06, |
|
"loss": 1.1807, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 0.6313522825150732, |
|
"grad_norm": 0.8916311860084534, |
|
"learning_rate": 6.3218676232674815e-06, |
|
"loss": 1.1974, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 0.632213608957795, |
|
"grad_norm": 0.8614788055419922, |
|
"learning_rate": 6.295937336855433e-06, |
|
"loss": 1.1858, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 0.6330749354005168, |
|
"grad_norm": 0.8437885642051697, |
|
"learning_rate": 6.270035884185367e-06, |
|
"loss": 1.1858, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.6339362618432386, |
|
"grad_norm": 0.8793601989746094, |
|
"learning_rate": 6.244163466883405e-06, |
|
"loss": 1.1722, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 0.6347975882859603, |
|
"grad_norm": 0.9624020457267761, |
|
"learning_rate": 6.218320286349655e-06, |
|
"loss": 1.1796, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 0.6356589147286822, |
|
"grad_norm": 0.8338882923126221, |
|
"learning_rate": 6.192506543756626e-06, |
|
"loss": 1.2217, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 0.6365202411714039, |
|
"grad_norm": 0.8388656377792358, |
|
"learning_rate": 6.1667224400476785e-06, |
|
"loss": 1.1981, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 0.6373815676141258, |
|
"grad_norm": 0.8913929462432861, |
|
"learning_rate": 6.140968175935458e-06, |
|
"loss": 1.1906, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.6382428940568475, |
|
"grad_norm": 0.8406807780265808, |
|
"learning_rate": 6.115243951900316e-06, |
|
"loss": 1.2023, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 0.6391042204995694, |
|
"grad_norm": 0.9768598079681396, |
|
"learning_rate": 6.089549968188767e-06, |
|
"loss": 1.1909, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 0.6399655469422911, |
|
"grad_norm": 0.9822099804878235, |
|
"learning_rate": 6.063886424811929e-06, |
|
"loss": 1.1945, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 0.6408268733850129, |
|
"grad_norm": 0.8699774742126465, |
|
"learning_rate": 6.038253521543951e-06, |
|
"loss": 1.1947, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.6416881998277347, |
|
"grad_norm": 0.875484049320221, |
|
"learning_rate": 6.01265145792048e-06, |
|
"loss": 1.1623, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.6425495262704565, |
|
"grad_norm": 0.8913766145706177, |
|
"learning_rate": 5.987080433237082e-06, |
|
"loss": 1.1771, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 0.6434108527131783, |
|
"grad_norm": 0.9033095240592957, |
|
"learning_rate": 5.961540646547722e-06, |
|
"loss": 1.1619, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 0.6442721791559001, |
|
"grad_norm": 0.9094845056533813, |
|
"learning_rate": 5.936032296663188e-06, |
|
"loss": 1.1788, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 0.6451335055986219, |
|
"grad_norm": 0.8750199675559998, |
|
"learning_rate": 5.9105555821495486e-06, |
|
"loss": 1.1971, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 0.6459948320413437, |
|
"grad_norm": 0.8594929575920105, |
|
"learning_rate": 5.885110701326621e-06, |
|
"loss": 1.1961, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.6468561584840654, |
|
"grad_norm": 0.8807885050773621, |
|
"learning_rate": 5.859697852266409e-06, |
|
"loss": 1.1702, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 0.6477174849267873, |
|
"grad_norm": 0.8879327178001404, |
|
"learning_rate": 5.834317232791575e-06, |
|
"loss": 1.2105, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 0.648578811369509, |
|
"grad_norm": 0.8946269154548645, |
|
"learning_rate": 5.808969040473893e-06, |
|
"loss": 1.1466, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 0.6494401378122309, |
|
"grad_norm": 0.8713740706443787, |
|
"learning_rate": 5.7836534726327175e-06, |
|
"loss": 1.1555, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 0.6503014642549526, |
|
"grad_norm": 0.9063122272491455, |
|
"learning_rate": 5.758370726333434e-06, |
|
"loss": 1.2304, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.6511627906976745, |
|
"grad_norm": 0.9117863774299622, |
|
"learning_rate": 5.733120998385935e-06, |
|
"loss": 1.1567, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 0.6520241171403962, |
|
"grad_norm": 0.8670936822891235, |
|
"learning_rate": 5.707904485343094e-06, |
|
"loss": 1.1848, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 0.652885443583118, |
|
"grad_norm": 0.8828465342521667, |
|
"learning_rate": 5.682721383499217e-06, |
|
"loss": 1.1945, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 0.6537467700258398, |
|
"grad_norm": 0.8751693964004517, |
|
"learning_rate": 5.657571888888538e-06, |
|
"loss": 1.1815, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 0.6546080964685616, |
|
"grad_norm": 0.9212945699691772, |
|
"learning_rate": 5.63245619728367e-06, |
|
"loss": 1.1823, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.6554694229112834, |
|
"grad_norm": 0.9038519263267517, |
|
"learning_rate": 5.60737450419409e-06, |
|
"loss": 1.1918, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 0.6563307493540051, |
|
"grad_norm": 0.9209284782409668, |
|
"learning_rate": 5.582327004864631e-06, |
|
"loss": 1.1645, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 0.657192075796727, |
|
"grad_norm": 0.8759058713912964, |
|
"learning_rate": 5.557313894273937e-06, |
|
"loss": 1.1895, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 0.6580534022394487, |
|
"grad_norm": 1.0104118585586548, |
|
"learning_rate": 5.532335367132962e-06, |
|
"loss": 1.1919, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 0.6589147286821705, |
|
"grad_norm": 0.8833430409431458, |
|
"learning_rate": 5.507391617883454e-06, |
|
"loss": 1.2118, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.6597760551248923, |
|
"grad_norm": 0.9313235878944397, |
|
"learning_rate": 5.4824828406964305e-06, |
|
"loss": 1.1817, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 0.6606373815676141, |
|
"grad_norm": 0.8975321054458618, |
|
"learning_rate": 5.457609229470681e-06, |
|
"loss": 1.182, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 0.661498708010336, |
|
"grad_norm": 0.8778689503669739, |
|
"learning_rate": 5.4327709778312484e-06, |
|
"loss": 1.1567, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 0.6623600344530577, |
|
"grad_norm": 0.9323556423187256, |
|
"learning_rate": 5.407968279127915e-06, |
|
"loss": 1.1592, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 0.6632213608957795, |
|
"grad_norm": 0.8266853094100952, |
|
"learning_rate": 5.383201326433727e-06, |
|
"loss": 1.1793, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.6640826873385013, |
|
"grad_norm": 0.8756425976753235, |
|
"learning_rate": 5.358470312543445e-06, |
|
"loss": 1.1877, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 0.664944013781223, |
|
"grad_norm": 0.8671419024467468, |
|
"learning_rate": 5.3337754299720925e-06, |
|
"loss": 1.2267, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 0.6658053402239449, |
|
"grad_norm": 0.9318386912345886, |
|
"learning_rate": 5.30911687095342e-06, |
|
"loss": 1.1453, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 0.6666666666666666, |
|
"grad_norm": 0.8705496191978455, |
|
"learning_rate": 5.284494827438423e-06, |
|
"loss": 1.1715, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 0.6675279931093885, |
|
"grad_norm": 0.8844230771064758, |
|
"learning_rate": 5.25990949109386e-06, |
|
"loss": 1.1755, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.6683893195521102, |
|
"grad_norm": 0.8812039494514465, |
|
"learning_rate": 5.2353610533007305e-06, |
|
"loss": 1.1888, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 0.6692506459948321, |
|
"grad_norm": 0.9645406603813171, |
|
"learning_rate": 5.210849705152809e-06, |
|
"loss": 1.1828, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 0.6701119724375538, |
|
"grad_norm": 0.9021162390708923, |
|
"learning_rate": 5.186375637455159e-06, |
|
"loss": 1.158, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 0.6709732988802756, |
|
"grad_norm": 0.9339265823364258, |
|
"learning_rate": 5.161939040722634e-06, |
|
"loss": 1.1629, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 0.6718346253229974, |
|
"grad_norm": 0.8927223682403564, |
|
"learning_rate": 5.137540105178396e-06, |
|
"loss": 1.1745, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.6726959517657192, |
|
"grad_norm": 0.9335219264030457, |
|
"learning_rate": 5.113179020752443e-06, |
|
"loss": 1.1858, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 0.673557278208441, |
|
"grad_norm": 0.8412618041038513, |
|
"learning_rate": 5.088855977080123e-06, |
|
"loss": 1.1759, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 0.6744186046511628, |
|
"grad_norm": 0.9003159999847412, |
|
"learning_rate": 5.064571163500667e-06, |
|
"loss": 1.1958, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 0.6752799310938846, |
|
"grad_norm": 0.9444640278816223, |
|
"learning_rate": 5.040324769055709e-06, |
|
"loss": 1.1988, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 0.6761412575366064, |
|
"grad_norm": 0.833273708820343, |
|
"learning_rate": 5.016116982487811e-06, |
|
"loss": 1.1886, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.6770025839793282, |
|
"grad_norm": 0.9385854005813599, |
|
"learning_rate": 4.991947992238997e-06, |
|
"loss": 1.1769, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 0.67786391042205, |
|
"grad_norm": 0.8889097571372986, |
|
"learning_rate": 4.967817986449284e-06, |
|
"loss": 1.1898, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 0.6787252368647717, |
|
"grad_norm": 0.8416062593460083, |
|
"learning_rate": 4.943727152955235e-06, |
|
"loss": 1.1899, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 0.6795865633074936, |
|
"grad_norm": 0.9640506505966187, |
|
"learning_rate": 4.9196756792884605e-06, |
|
"loss": 1.1683, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 0.6804478897502153, |
|
"grad_norm": 0.8979779481887817, |
|
"learning_rate": 4.8956637526742e-06, |
|
"loss": 1.1926, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.6813092161929372, |
|
"grad_norm": 0.8395998477935791, |
|
"learning_rate": 4.871691560029828e-06, |
|
"loss": 1.1593, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 0.6821705426356589, |
|
"grad_norm": 0.8391218185424805, |
|
"learning_rate": 4.847759287963432e-06, |
|
"loss": 1.1615, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 0.6830318690783808, |
|
"grad_norm": 0.8740763068199158, |
|
"learning_rate": 4.8238671227723285e-06, |
|
"loss": 1.1717, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 0.6838931955211025, |
|
"grad_norm": 0.8377690315246582, |
|
"learning_rate": 4.800015250441638e-06, |
|
"loss": 1.1757, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 0.6847545219638242, |
|
"grad_norm": 0.8415879011154175, |
|
"learning_rate": 4.7762038566428155e-06, |
|
"loss": 1.2087, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.6856158484065461, |
|
"grad_norm": 0.8535274863243103, |
|
"learning_rate": 4.752433126732231e-06, |
|
"loss": 1.1714, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 0.6864771748492678, |
|
"grad_norm": 0.8958263397216797, |
|
"learning_rate": 4.728703245749707e-06, |
|
"loss": 1.1477, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 0.6873385012919897, |
|
"grad_norm": 0.9385268092155457, |
|
"learning_rate": 4.7050143984170805e-06, |
|
"loss": 1.2119, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 0.6881998277347114, |
|
"grad_norm": 0.9213505983352661, |
|
"learning_rate": 4.681366769136769e-06, |
|
"loss": 1.1759, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 0.6890611541774333, |
|
"grad_norm": 0.9221978187561035, |
|
"learning_rate": 4.657760541990329e-06, |
|
"loss": 1.1704, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.689922480620155, |
|
"grad_norm": 0.8620073795318604, |
|
"learning_rate": 4.634195900737045e-06, |
|
"loss": 1.197, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 0.6907838070628768, |
|
"grad_norm": 0.8191630244255066, |
|
"learning_rate": 4.610673028812459e-06, |
|
"loss": 1.1385, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 0.6916451335055986, |
|
"grad_norm": 0.8765779137611389, |
|
"learning_rate": 4.587192109326988e-06, |
|
"loss": 1.1687, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 0.6925064599483204, |
|
"grad_norm": 0.7997327446937561, |
|
"learning_rate": 4.563753325064457e-06, |
|
"loss": 1.1897, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 0.6933677863910422, |
|
"grad_norm": 0.8961016535758972, |
|
"learning_rate": 4.540356858480711e-06, |
|
"loss": 1.1849, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.694229112833764, |
|
"grad_norm": 0.8605460524559021, |
|
"learning_rate": 4.5170028917021705e-06, |
|
"loss": 1.1793, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 0.6950904392764858, |
|
"grad_norm": 0.9754102826118469, |
|
"learning_rate": 4.493691606524423e-06, |
|
"loss": 1.2003, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 0.6959517657192076, |
|
"grad_norm": 0.7908036708831787, |
|
"learning_rate": 4.470423184410804e-06, |
|
"loss": 1.1333, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 0.6968130921619293, |
|
"grad_norm": 0.8532270789146423, |
|
"learning_rate": 4.447197806490996e-06, |
|
"loss": 1.1468, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 0.6976744186046512, |
|
"grad_norm": 0.8650361895561218, |
|
"learning_rate": 4.424015653559611e-06, |
|
"loss": 1.1958, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.6985357450473729, |
|
"grad_norm": 1.0336636304855347, |
|
"learning_rate": 4.400876906074772e-06, |
|
"loss": 1.2279, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 0.6993970714900948, |
|
"grad_norm": 0.8429791331291199, |
|
"learning_rate": 4.377781744156727e-06, |
|
"loss": 1.2041, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 0.7002583979328165, |
|
"grad_norm": 0.8677304983139038, |
|
"learning_rate": 4.35473034758643e-06, |
|
"loss": 1.1873, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 0.7011197243755384, |
|
"grad_norm": 0.9752134680747986, |
|
"learning_rate": 4.331722895804158e-06, |
|
"loss": 1.1734, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 0.7019810508182601, |
|
"grad_norm": 0.8603991270065308, |
|
"learning_rate": 4.30875956790811e-06, |
|
"loss": 1.1984, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.7028423772609819, |
|
"grad_norm": 0.8431304693222046, |
|
"learning_rate": 4.285840542652997e-06, |
|
"loss": 1.1635, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 0.7037037037037037, |
|
"grad_norm": 0.8751398324966431, |
|
"learning_rate": 4.262965998448665e-06, |
|
"loss": 1.1715, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 0.7045650301464255, |
|
"grad_norm": 0.8455151915550232, |
|
"learning_rate": 4.240136113358714e-06, |
|
"loss": 1.212, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 0.7054263565891473, |
|
"grad_norm": 0.9403558373451233, |
|
"learning_rate": 4.2173510650990905e-06, |
|
"loss": 1.1627, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 0.7062876830318691, |
|
"grad_norm": 0.941561222076416, |
|
"learning_rate": 4.194611031036718e-06, |
|
"loss": 1.1861, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.7071490094745909, |
|
"grad_norm": 0.9220948815345764, |
|
"learning_rate": 4.171916188188113e-06, |
|
"loss": 1.1696, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 0.7080103359173127, |
|
"grad_norm": 0.8759995698928833, |
|
"learning_rate": 4.149266713218011e-06, |
|
"loss": 1.1863, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 0.7088716623600344, |
|
"grad_norm": 0.8697245717048645, |
|
"learning_rate": 4.12666278243799e-06, |
|
"loss": 1.1579, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 0.7097329888027563, |
|
"grad_norm": 0.8867110013961792, |
|
"learning_rate": 4.104104571805088e-06, |
|
"loss": 1.1722, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 0.710594315245478, |
|
"grad_norm": 0.8431983590126038, |
|
"learning_rate": 4.0815922569204435e-06, |
|
"loss": 1.1983, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.7114556416881999, |
|
"grad_norm": 0.891090452671051, |
|
"learning_rate": 4.059126013027924e-06, |
|
"loss": 1.1585, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 0.7123169681309216, |
|
"grad_norm": 0.8412421345710754, |
|
"learning_rate": 4.0367060150127726e-06, |
|
"loss": 1.1664, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 0.7131782945736435, |
|
"grad_norm": 0.8604631423950195, |
|
"learning_rate": 4.014332437400235e-06, |
|
"loss": 1.163, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 0.7140396210163652, |
|
"grad_norm": 0.9067566394805908, |
|
"learning_rate": 3.9920054543541976e-06, |
|
"loss": 1.1838, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 0.7149009474590869, |
|
"grad_norm": 0.8477993607521057, |
|
"learning_rate": 3.96972523967584e-06, |
|
"loss": 1.1958, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.7157622739018088, |
|
"grad_norm": 0.8343601226806641, |
|
"learning_rate": 3.947491966802288e-06, |
|
"loss": 1.1784, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 0.7166236003445305, |
|
"grad_norm": 0.8563992381095886, |
|
"learning_rate": 3.925305808805247e-06, |
|
"loss": 1.1488, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 0.7174849267872524, |
|
"grad_norm": 0.9131225943565369, |
|
"learning_rate": 3.903166938389664e-06, |
|
"loss": 1.1762, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 0.7183462532299741, |
|
"grad_norm": 0.872132420539856, |
|
"learning_rate": 3.881075527892391e-06, |
|
"loss": 1.1717, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 0.719207579672696, |
|
"grad_norm": 0.8504886627197266, |
|
"learning_rate": 3.859031749280824e-06, |
|
"loss": 1.1479, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.7200689061154177, |
|
"grad_norm": 0.8588838577270508, |
|
"learning_rate": 3.837035774151585e-06, |
|
"loss": 1.19, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 0.7209302325581395, |
|
"grad_norm": 0.8740001916885376, |
|
"learning_rate": 3.815087773729171e-06, |
|
"loss": 1.186, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 0.7217915590008613, |
|
"grad_norm": 0.8781384825706482, |
|
"learning_rate": 3.793187918864627e-06, |
|
"loss": 1.1913, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 0.7226528854435831, |
|
"grad_norm": 0.8898410201072693, |
|
"learning_rate": 3.771336380034211e-06, |
|
"loss": 1.1711, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 0.7235142118863049, |
|
"grad_norm": 0.8932210206985474, |
|
"learning_rate": 3.749533327338091e-06, |
|
"loss": 1.1818, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.7243755383290267, |
|
"grad_norm": 0.8856852650642395, |
|
"learning_rate": 3.727778930498982e-06, |
|
"loss": 1.1913, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 0.7252368647717485, |
|
"grad_norm": 0.8199148774147034, |
|
"learning_rate": 3.706073358860851e-06, |
|
"loss": 1.1862, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 0.7260981912144703, |
|
"grad_norm": 0.8369928002357483, |
|
"learning_rate": 3.6844167813875888e-06, |
|
"loss": 1.2055, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 0.726959517657192, |
|
"grad_norm": 0.8797032237052917, |
|
"learning_rate": 3.6628093666617083e-06, |
|
"loss": 1.1767, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 0.7278208440999139, |
|
"grad_norm": 0.8433752059936523, |
|
"learning_rate": 3.6412512828830114e-06, |
|
"loss": 1.1601, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.7286821705426356, |
|
"grad_norm": 0.8134076595306396, |
|
"learning_rate": 3.6197426978672923e-06, |
|
"loss": 1.1797, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 0.7295434969853575, |
|
"grad_norm": 0.8705214858055115, |
|
"learning_rate": 3.5982837790450376e-06, |
|
"loss": 1.1506, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 0.7304048234280792, |
|
"grad_norm": 0.9077462553977966, |
|
"learning_rate": 3.5768746934601007e-06, |
|
"loss": 1.1796, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 0.7312661498708011, |
|
"grad_norm": 0.8746139407157898, |
|
"learning_rate": 3.5555156077684317e-06, |
|
"loss": 1.1535, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 0.7321274763135228, |
|
"grad_norm": 0.9012914896011353, |
|
"learning_rate": 3.5342066882367488e-06, |
|
"loss": 1.2069, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.7329888027562446, |
|
"grad_norm": 0.8886313438415527, |
|
"learning_rate": 3.5129481007412634e-06, |
|
"loss": 1.1654, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 0.7338501291989664, |
|
"grad_norm": 0.8336722254753113, |
|
"learning_rate": 3.491740010766389e-06, |
|
"loss": 1.1773, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 0.7347114556416882, |
|
"grad_norm": 0.8749405741691589, |
|
"learning_rate": 3.47058258340345e-06, |
|
"loss": 1.18, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 0.73557278208441, |
|
"grad_norm": 0.8950448632240295, |
|
"learning_rate": 3.449475983349385e-06, |
|
"loss": 1.1257, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 0.7364341085271318, |
|
"grad_norm": 0.8262083530426025, |
|
"learning_rate": 3.4284203749054834e-06, |
|
"loss": 1.1863, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.7372954349698536, |
|
"grad_norm": 0.8869133591651917, |
|
"learning_rate": 3.4074159219760884e-06, |
|
"loss": 1.1659, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 0.7381567614125754, |
|
"grad_norm": 0.8032558560371399, |
|
"learning_rate": 3.386462788067344e-06, |
|
"loss": 1.1862, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 0.7390180878552972, |
|
"grad_norm": 0.8882765173912048, |
|
"learning_rate": 3.3655611362858966e-06, |
|
"loss": 1.1694, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 0.739879414298019, |
|
"grad_norm": 0.9683348536491394, |
|
"learning_rate": 3.3447111293376468e-06, |
|
"loss": 1.1428, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 0.7407407407407407, |
|
"grad_norm": 0.823493480682373, |
|
"learning_rate": 3.323912929526465e-06, |
|
"loss": 1.1393, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.7416020671834626, |
|
"grad_norm": 0.8513652682304382, |
|
"learning_rate": 3.30316669875294e-06, |
|
"loss": 1.1785, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 0.7424633936261843, |
|
"grad_norm": 0.8614821434020996, |
|
"learning_rate": 3.282472598513119e-06, |
|
"loss": 1.1626, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 0.7433247200689062, |
|
"grad_norm": 0.8872314095497131, |
|
"learning_rate": 3.2618307898972413e-06, |
|
"loss": 1.1466, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 0.7441860465116279, |
|
"grad_norm": 0.8613788485527039, |
|
"learning_rate": 3.2412414335884866e-06, |
|
"loss": 1.1586, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 0.7450473729543498, |
|
"grad_norm": 0.8109983205795288, |
|
"learning_rate": 3.2207046898617365e-06, |
|
"loss": 1.1701, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.7459086993970715, |
|
"grad_norm": 0.9684152603149414, |
|
"learning_rate": 3.2002207185823155e-06, |
|
"loss": 1.1888, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 0.7467700258397932, |
|
"grad_norm": 1.0356762409210205, |
|
"learning_rate": 3.1797896792047422e-06, |
|
"loss": 1.2229, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 0.7476313522825151, |
|
"grad_norm": 0.9049926996231079, |
|
"learning_rate": 3.1594117307714977e-06, |
|
"loss": 1.2125, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 0.7484926787252368, |
|
"grad_norm": 0.9765392541885376, |
|
"learning_rate": 3.1390870319117838e-06, |
|
"loss": 1.1854, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 0.7493540051679587, |
|
"grad_norm": 0.8243566155433655, |
|
"learning_rate": 3.118815740840294e-06, |
|
"loss": 1.1737, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.7502153316106804, |
|
"grad_norm": 0.8770548105239868, |
|
"learning_rate": 3.098598015355967e-06, |
|
"loss": 1.1927, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 0.7510766580534023, |
|
"grad_norm": 0.8610149025917053, |
|
"learning_rate": 3.0784340128407786e-06, |
|
"loss": 1.1686, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 0.751937984496124, |
|
"grad_norm": 1.0247671604156494, |
|
"learning_rate": 3.0583238902584976e-06, |
|
"loss": 1.1861, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 0.7527993109388458, |
|
"grad_norm": 0.8867828845977783, |
|
"learning_rate": 3.038267804153472e-06, |
|
"loss": 1.1948, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 0.7536606373815676, |
|
"grad_norm": 0.902263879776001, |
|
"learning_rate": 3.0182659106494195e-06, |
|
"loss": 1.1778, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.7545219638242894, |
|
"grad_norm": 0.9579049348831177, |
|
"learning_rate": 2.998318365448194e-06, |
|
"loss": 1.188, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 0.7553832902670112, |
|
"grad_norm": 0.8788950443267822, |
|
"learning_rate": 2.9784253238285844e-06, |
|
"loss": 1.1451, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 0.756244616709733, |
|
"grad_norm": 0.9002311825752258, |
|
"learning_rate": 2.9585869406451083e-06, |
|
"loss": 1.1493, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 0.7571059431524548, |
|
"grad_norm": 0.9176036715507507, |
|
"learning_rate": 2.938803370326804e-06, |
|
"loss": 1.1789, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 0.7579672695951766, |
|
"grad_norm": 0.87285977602005, |
|
"learning_rate": 2.9190747668760213e-06, |
|
"loss": 1.1633, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.7588285960378983, |
|
"grad_norm": 0.8504787087440491, |
|
"learning_rate": 2.899401283867229e-06, |
|
"loss": 1.1753, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 0.7596899224806202, |
|
"grad_norm": 0.8374565243721008, |
|
"learning_rate": 2.8797830744458177e-06, |
|
"loss": 1.168, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 0.7605512489233419, |
|
"grad_norm": 0.8359323143959045, |
|
"learning_rate": 2.860220291326915e-06, |
|
"loss": 1.1788, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 0.7614125753660638, |
|
"grad_norm": 0.8868809342384338, |
|
"learning_rate": 2.840713086794189e-06, |
|
"loss": 1.1907, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 0.7622739018087855, |
|
"grad_norm": 0.874671220779419, |
|
"learning_rate": 2.8212616126986604e-06, |
|
"loss": 1.1735, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.7631352282515074, |
|
"grad_norm": 0.9140809774398804, |
|
"learning_rate": 2.801866020457521e-06, |
|
"loss": 1.1776, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 0.7639965546942291, |
|
"grad_norm": 0.8395150303840637, |
|
"learning_rate": 2.7825264610529703e-06, |
|
"loss": 1.1691, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 0.7648578811369509, |
|
"grad_norm": 0.9220564961433411, |
|
"learning_rate": 2.7632430850310175e-06, |
|
"loss": 1.1921, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 0.7657192075796727, |
|
"grad_norm": 0.8539912104606628, |
|
"learning_rate": 2.7440160425003236e-06, |
|
"loss": 1.1454, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 0.7665805340223945, |
|
"grad_norm": 0.8318409323692322, |
|
"learning_rate": 2.7248454831310335e-06, |
|
"loss": 1.1648, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.7674418604651163, |
|
"grad_norm": 0.8465113639831543, |
|
"learning_rate": 2.7057315561536e-06, |
|
"loss": 1.1529, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 0.768303186907838, |
|
"grad_norm": 0.8205885887145996, |
|
"learning_rate": 2.686674410357637e-06, |
|
"loss": 1.2069, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 0.7691645133505599, |
|
"grad_norm": 0.9047953486442566, |
|
"learning_rate": 2.667674194090748e-06, |
|
"loss": 1.1727, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 0.7700258397932817, |
|
"grad_norm": 0.928527295589447, |
|
"learning_rate": 2.6487310552573776e-06, |
|
"loss": 1.1687, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 0.7708871662360034, |
|
"grad_norm": 0.9136389493942261, |
|
"learning_rate": 2.6298451413176564e-06, |
|
"loss": 1.1628, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.7717484926787253, |
|
"grad_norm": 0.8879351615905762, |
|
"learning_rate": 2.6110165992862635e-06, |
|
"loss": 1.1576, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 0.772609819121447, |
|
"grad_norm": 0.8699619770050049, |
|
"learning_rate": 2.592245575731274e-06, |
|
"loss": 1.2179, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 0.7734711455641688, |
|
"grad_norm": 0.9080535173416138, |
|
"learning_rate": 2.5735322167730116e-06, |
|
"loss": 1.1618, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 0.7743324720068906, |
|
"grad_norm": 0.8573684096336365, |
|
"learning_rate": 2.5548766680829207e-06, |
|
"loss": 1.141, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 0.7751937984496124, |
|
"grad_norm": 0.9463301301002502, |
|
"learning_rate": 2.5362790748824363e-06, |
|
"loss": 1.1608, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.7760551248923342, |
|
"grad_norm": 0.8480750918388367, |
|
"learning_rate": 2.517739581941839e-06, |
|
"loss": 1.1811, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 0.7769164513350559, |
|
"grad_norm": 0.9230459928512573, |
|
"learning_rate": 2.4992583335791375e-06, |
|
"loss": 1.1404, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 0.7777777777777778, |
|
"grad_norm": 0.8646366000175476, |
|
"learning_rate": 2.4808354736589525e-06, |
|
"loss": 1.1808, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 0.7786391042204995, |
|
"grad_norm": 0.8544058203697205, |
|
"learning_rate": 2.4624711455913764e-06, |
|
"loss": 1.1333, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 0.7795004306632214, |
|
"grad_norm": 0.8751500248908997, |
|
"learning_rate": 2.444165492330879e-06, |
|
"loss": 1.1674, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.7803617571059431, |
|
"grad_norm": 0.8781688213348389, |
|
"learning_rate": 2.425918656375177e-06, |
|
"loss": 1.1802, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 0.781223083548665, |
|
"grad_norm": 1.0020009279251099, |
|
"learning_rate": 2.4077307797641357e-06, |
|
"loss": 1.1823, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 0.7820844099913867, |
|
"grad_norm": 0.7854005098342896, |
|
"learning_rate": 2.389602004078657e-06, |
|
"loss": 1.1864, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 0.7829457364341085, |
|
"grad_norm": 0.8276894688606262, |
|
"learning_rate": 2.3715324704395846e-06, |
|
"loss": 1.1667, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 0.7838070628768303, |
|
"grad_norm": 0.8462375402450562, |
|
"learning_rate": 2.3535223195066025e-06, |
|
"loss": 1.1318, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.7846683893195521, |
|
"grad_norm": 0.8951948881149292, |
|
"learning_rate": 2.335571691477132e-06, |
|
"loss": 1.189, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 0.7855297157622739, |
|
"grad_norm": 0.8499693870544434, |
|
"learning_rate": 2.3176807260852475e-06, |
|
"loss": 1.1322, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 0.7863910422049957, |
|
"grad_norm": 0.8753166198730469, |
|
"learning_rate": 2.2998495626005955e-06, |
|
"loss": 1.186, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 0.7872523686477175, |
|
"grad_norm": 0.8034504652023315, |
|
"learning_rate": 2.282078339827293e-06, |
|
"loss": 1.174, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 0.7881136950904393, |
|
"grad_norm": 0.8719635605812073, |
|
"learning_rate": 2.264367196102869e-06, |
|
"loss": 1.2179, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.788975021533161, |
|
"grad_norm": 0.8509535193443298, |
|
"learning_rate": 2.2467162692971655e-06, |
|
"loss": 1.1951, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 0.7898363479758829, |
|
"grad_norm": 0.9040639996528625, |
|
"learning_rate": 2.229125696811275e-06, |
|
"loss": 1.2126, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 0.7906976744186046, |
|
"grad_norm": 0.9148096442222595, |
|
"learning_rate": 2.2115956155764817e-06, |
|
"loss": 1.1558, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 0.7915590008613265, |
|
"grad_norm": 0.8161072134971619, |
|
"learning_rate": 2.1941261620531718e-06, |
|
"loss": 1.1379, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 0.7924203273040482, |
|
"grad_norm": 0.8694244027137756, |
|
"learning_rate": 2.17671747222979e-06, |
|
"loss": 1.1576, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.7932816537467701, |
|
"grad_norm": 0.830352246761322, |
|
"learning_rate": 2.1593696816217667e-06, |
|
"loss": 1.1848, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 0.7941429801894918, |
|
"grad_norm": 0.8742688298225403, |
|
"learning_rate": 2.142082925270489e-06, |
|
"loss": 1.1939, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 0.7950043066322137, |
|
"grad_norm": 0.8883207440376282, |
|
"learning_rate": 2.1248573377422155e-06, |
|
"loss": 1.1894, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 0.7958656330749354, |
|
"grad_norm": 0.9029132127761841, |
|
"learning_rate": 2.107693053127049e-06, |
|
"loss": 1.1283, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 0.7967269595176572, |
|
"grad_norm": 0.9062528014183044, |
|
"learning_rate": 2.090590205037888e-06, |
|
"loss": 1.2179, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.797588285960379, |
|
"grad_norm": 0.8248071670532227, |
|
"learning_rate": 2.0735489266093923e-06, |
|
"loss": 1.1756, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 0.7984496124031008, |
|
"grad_norm": 0.8426958918571472, |
|
"learning_rate": 2.056569350496933e-06, |
|
"loss": 1.1924, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 0.7993109388458226, |
|
"grad_norm": 0.8882315158843994, |
|
"learning_rate": 2.0396516088755804e-06, |
|
"loss": 1.1843, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 0.8001722652885443, |
|
"grad_norm": 0.9714380502700806, |
|
"learning_rate": 2.0227958334390506e-06, |
|
"loss": 1.1617, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 0.8010335917312662, |
|
"grad_norm": 0.9388904571533203, |
|
"learning_rate": 2.0060021553986974e-06, |
|
"loss": 1.1715, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.801894918173988, |
|
"grad_norm": 0.9136961698532104, |
|
"learning_rate": 1.989270705482492e-06, |
|
"loss": 1.1402, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 0.8027562446167097, |
|
"grad_norm": 0.9584447145462036, |
|
"learning_rate": 1.9726016139339934e-06, |
|
"loss": 1.1393, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 0.8036175710594315, |
|
"grad_norm": 0.8596916198730469, |
|
"learning_rate": 1.955995010511338e-06, |
|
"loss": 1.1759, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 0.8044788975021533, |
|
"grad_norm": 0.8455019593238831, |
|
"learning_rate": 1.9394510244862397e-06, |
|
"loss": 1.1616, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 0.8053402239448751, |
|
"grad_norm": 0.9066017270088196, |
|
"learning_rate": 1.9229697846429773e-06, |
|
"loss": 1.1546, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.8062015503875969, |
|
"grad_norm": 0.8392831683158875, |
|
"learning_rate": 1.9065514192773848e-06, |
|
"loss": 1.1998, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 0.8070628768303187, |
|
"grad_norm": 0.9439470767974854, |
|
"learning_rate": 1.8901960561958588e-06, |
|
"loss": 1.1673, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 0.8079242032730405, |
|
"grad_norm": 0.8691911697387695, |
|
"learning_rate": 1.8739038227143658e-06, |
|
"loss": 1.1632, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 0.8087855297157622, |
|
"grad_norm": 0.8843985199928284, |
|
"learning_rate": 1.8576748456574512e-06, |
|
"loss": 1.1743, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 0.8096468561584841, |
|
"grad_norm": 0.8459395170211792, |
|
"learning_rate": 1.8415092513572498e-06, |
|
"loss": 1.1606, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.8105081826012058, |
|
"grad_norm": 0.8578802943229675, |
|
"learning_rate": 1.8254071656524997e-06, |
|
"loss": 1.1418, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 0.8113695090439277, |
|
"grad_norm": 0.8318899869918823, |
|
"learning_rate": 1.8093687138875648e-06, |
|
"loss": 1.1623, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 0.8122308354866494, |
|
"grad_norm": 0.9500533938407898, |
|
"learning_rate": 1.7933940209114597e-06, |
|
"loss": 1.1926, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 0.8130921619293713, |
|
"grad_norm": 0.8835731148719788, |
|
"learning_rate": 1.7774832110768847e-06, |
|
"loss": 1.1874, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 0.813953488372093, |
|
"grad_norm": 0.8308874368667603, |
|
"learning_rate": 1.7616364082392446e-06, |
|
"loss": 1.1762, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.8148148148148148, |
|
"grad_norm": 0.9012060761451721, |
|
"learning_rate": 1.745853735755687e-06, |
|
"loss": 1.1723, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 0.8156761412575366, |
|
"grad_norm": 0.8090865015983582, |
|
"learning_rate": 1.7301353164841562e-06, |
|
"loss": 1.1905, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 0.8165374677002584, |
|
"grad_norm": 0.9133245348930359, |
|
"learning_rate": 1.7144812727824233e-06, |
|
"loss": 1.1314, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 0.8173987941429802, |
|
"grad_norm": 0.817896842956543, |
|
"learning_rate": 1.6988917265071337e-06, |
|
"loss": 1.2059, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 0.818260120585702, |
|
"grad_norm": 0.8479551076889038, |
|
"learning_rate": 1.6833667990128622e-06, |
|
"loss": 1.1256, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.8191214470284238, |
|
"grad_norm": 0.9293598532676697, |
|
"learning_rate": 1.6679066111511677e-06, |
|
"loss": 1.1371, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 0.8199827734711456, |
|
"grad_norm": 0.8253700733184814, |
|
"learning_rate": 1.6525112832696576e-06, |
|
"loss": 1.1672, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 0.8208440999138673, |
|
"grad_norm": 0.8608076572418213, |
|
"learning_rate": 1.6371809352110446e-06, |
|
"loss": 1.1383, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 0.8217054263565892, |
|
"grad_norm": 0.9392169117927551, |
|
"learning_rate": 1.6219156863122121e-06, |
|
"loss": 1.1867, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 0.8225667527993109, |
|
"grad_norm": 0.8883967399597168, |
|
"learning_rate": 1.6067156554032893e-06, |
|
"loss": 1.2018, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.8234280792420328, |
|
"grad_norm": 0.8919565677642822, |
|
"learning_rate": 1.5915809608067245e-06, |
|
"loss": 1.1873, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 0.8242894056847545, |
|
"grad_norm": 0.8182902932167053, |
|
"learning_rate": 1.5765117203363723e-06, |
|
"loss": 1.1613, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 0.8251507321274764, |
|
"grad_norm": 0.8579580783843994, |
|
"learning_rate": 1.5615080512965563e-06, |
|
"loss": 1.1883, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.8260120585701981, |
|
"grad_norm": 0.8626570105552673, |
|
"learning_rate": 1.5465700704811825e-06, |
|
"loss": 1.171, |
|
"step": 959 |
|
}, |
|
{ |
|
"epoch": 0.8268733850129198, |
|
"grad_norm": 0.8572131395339966, |
|
"learning_rate": 1.5316978941728045e-06, |
|
"loss": 1.1429, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.8277347114556417, |
|
"grad_norm": 0.8980193734169006, |
|
"learning_rate": 1.5168916381417387e-06, |
|
"loss": 1.1485, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 0.8285960378983634, |
|
"grad_norm": 0.9097509384155273, |
|
"learning_rate": 1.5021514176451514e-06, |
|
"loss": 1.1834, |
|
"step": 962 |
|
}, |
|
{ |
|
"epoch": 0.8294573643410853, |
|
"grad_norm": 1.002036213874817, |
|
"learning_rate": 1.487477347426164e-06, |
|
"loss": 1.1611, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 0.830318690783807, |
|
"grad_norm": 0.967788815498352, |
|
"learning_rate": 1.4728695417129579e-06, |
|
"loss": 1.1471, |
|
"step": 964 |
|
}, |
|
{ |
|
"epoch": 0.8311800172265289, |
|
"grad_norm": 0.9008656740188599, |
|
"learning_rate": 1.458328114217894e-06, |
|
"loss": 1.1956, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.8320413436692506, |
|
"grad_norm": 0.927203893661499, |
|
"learning_rate": 1.4438531781366226e-06, |
|
"loss": 1.1646, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 0.8329026701119724, |
|
"grad_norm": 0.877873957157135, |
|
"learning_rate": 1.4294448461471943e-06, |
|
"loss": 1.1799, |
|
"step": 967 |
|
}, |
|
{ |
|
"epoch": 0.8337639965546942, |
|
"grad_norm": 0.8068222999572754, |
|
"learning_rate": 1.4151032304091928e-06, |
|
"loss": 1.1723, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 0.834625322997416, |
|
"grad_norm": 0.8391594290733337, |
|
"learning_rate": 1.400828442562857e-06, |
|
"loss": 1.1462, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 0.8354866494401378, |
|
"grad_norm": 0.8191964030265808, |
|
"learning_rate": 1.3866205937282195e-06, |
|
"loss": 1.1757, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.8363479758828596, |
|
"grad_norm": 0.8616624474525452, |
|
"learning_rate": 1.372479794504229e-06, |
|
"loss": 1.1719, |
|
"step": 971 |
|
}, |
|
{ |
|
"epoch": 0.8372093023255814, |
|
"grad_norm": 0.9123625755310059, |
|
"learning_rate": 1.3584061549679017e-06, |
|
"loss": 1.1638, |
|
"step": 972 |
|
}, |
|
{ |
|
"epoch": 0.8380706287683032, |
|
"grad_norm": 0.8329593539237976, |
|
"learning_rate": 1.3443997846734535e-06, |
|
"loss": 1.1267, |
|
"step": 973 |
|
}, |
|
{ |
|
"epoch": 0.8389319552110249, |
|
"grad_norm": 0.9073025584220886, |
|
"learning_rate": 1.330460792651459e-06, |
|
"loss": 1.1523, |
|
"step": 974 |
|
}, |
|
{ |
|
"epoch": 0.8397932816537468, |
|
"grad_norm": 0.8376272916793823, |
|
"learning_rate": 1.31658928740799e-06, |
|
"loss": 1.1472, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.8406546080964685, |
|
"grad_norm": 0.8478926420211792, |
|
"learning_rate": 1.3027853769237808e-06, |
|
"loss": 1.1705, |
|
"step": 976 |
|
}, |
|
{ |
|
"epoch": 0.8415159345391904, |
|
"grad_norm": 0.9134843945503235, |
|
"learning_rate": 1.2890491686533812e-06, |
|
"loss": 1.1835, |
|
"step": 977 |
|
}, |
|
{ |
|
"epoch": 0.8423772609819121, |
|
"grad_norm": 0.8817895650863647, |
|
"learning_rate": 1.2753807695243258e-06, |
|
"loss": 1.1549, |
|
"step": 978 |
|
}, |
|
{ |
|
"epoch": 0.843238587424634, |
|
"grad_norm": 0.9673689603805542, |
|
"learning_rate": 1.2617802859363016e-06, |
|
"loss": 1.1211, |
|
"step": 979 |
|
}, |
|
{ |
|
"epoch": 0.8440999138673557, |
|
"grad_norm": 0.8391812443733215, |
|
"learning_rate": 1.2482478237603102e-06, |
|
"loss": 1.1272, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.8449612403100775, |
|
"grad_norm": 0.9365954399108887, |
|
"learning_rate": 1.2347834883378518e-06, |
|
"loss": 1.1997, |
|
"step": 981 |
|
}, |
|
{ |
|
"epoch": 0.8458225667527993, |
|
"grad_norm": 0.8548864722251892, |
|
"learning_rate": 1.2213873844801049e-06, |
|
"loss": 1.1804, |
|
"step": 982 |
|
}, |
|
{ |
|
"epoch": 0.8466838931955211, |
|
"grad_norm": 0.7988507151603699, |
|
"learning_rate": 1.2080596164671098e-06, |
|
"loss": 1.1528, |
|
"step": 983 |
|
}, |
|
{ |
|
"epoch": 0.8475452196382429, |
|
"grad_norm": 0.8993740677833557, |
|
"learning_rate": 1.1948002880469601e-06, |
|
"loss": 1.1818, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 0.8484065460809647, |
|
"grad_norm": 1.0622564554214478, |
|
"learning_rate": 1.1816095024349828e-06, |
|
"loss": 1.2045, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.8492678725236865, |
|
"grad_norm": 0.9041065573692322, |
|
"learning_rate": 1.1684873623129457e-06, |
|
"loss": 1.1534, |
|
"step": 986 |
|
}, |
|
{ |
|
"epoch": 0.8501291989664083, |
|
"grad_norm": 0.8370366096496582, |
|
"learning_rate": 1.1554339698282623e-06, |
|
"loss": 1.1714, |
|
"step": 987 |
|
}, |
|
{ |
|
"epoch": 0.85099052540913, |
|
"grad_norm": 0.8588913679122925, |
|
"learning_rate": 1.1424494265931829e-06, |
|
"loss": 1.1604, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 0.8518518518518519, |
|
"grad_norm": 0.9080969095230103, |
|
"learning_rate": 1.1295338336840113e-06, |
|
"loss": 1.1827, |
|
"step": 989 |
|
}, |
|
{ |
|
"epoch": 0.8527131782945736, |
|
"grad_norm": 0.8849515318870544, |
|
"learning_rate": 1.1166872916403226e-06, |
|
"loss": 1.2205, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.8535745047372955, |
|
"grad_norm": 0.8305906653404236, |
|
"learning_rate": 1.1039099004641684e-06, |
|
"loss": 1.165, |
|
"step": 991 |
|
}, |
|
{ |
|
"epoch": 0.8544358311800172, |
|
"grad_norm": 0.8653746843338013, |
|
"learning_rate": 1.0912017596193115e-06, |
|
"loss": 1.1975, |
|
"step": 992 |
|
}, |
|
{ |
|
"epoch": 0.8552971576227391, |
|
"grad_norm": 0.9006616473197937, |
|
"learning_rate": 1.0785629680304433e-06, |
|
"loss": 1.1512, |
|
"step": 993 |
|
}, |
|
{ |
|
"epoch": 0.8561584840654608, |
|
"grad_norm": 0.8347445726394653, |
|
"learning_rate": 1.065993624082411e-06, |
|
"loss": 1.162, |
|
"step": 994 |
|
}, |
|
{ |
|
"epoch": 0.8570198105081827, |
|
"grad_norm": 0.8577432632446289, |
|
"learning_rate": 1.053493825619467e-06, |
|
"loss": 1.1691, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.8578811369509044, |
|
"grad_norm": 0.8580631613731384, |
|
"learning_rate": 1.0410636699444855e-06, |
|
"loss": 1.1569, |
|
"step": 996 |
|
}, |
|
{ |
|
"epoch": 0.8587424633936261, |
|
"grad_norm": 0.8457480669021606, |
|
"learning_rate": 1.0287032538182262e-06, |
|
"loss": 1.1585, |
|
"step": 997 |
|
}, |
|
{ |
|
"epoch": 0.859603789836348, |
|
"grad_norm": 0.8660425543785095, |
|
"learning_rate": 1.0164126734585667e-06, |
|
"loss": 1.1755, |
|
"step": 998 |
|
}, |
|
{ |
|
"epoch": 0.8604651162790697, |
|
"grad_norm": 0.8238218426704407, |
|
"learning_rate": 1.0041920245397552e-06, |
|
"loss": 1.1735, |
|
"step": 999 |
|
}, |
|
{ |
|
"epoch": 0.8613264427217916, |
|
"grad_norm": 0.8318968415260315, |
|
"learning_rate": 9.920414021916747e-07, |
|
"loss": 1.1299, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.8621877691645133, |
|
"grad_norm": 0.8238946199417114, |
|
"learning_rate": 9.79960900999094e-07, |
|
"loss": 1.1459, |
|
"step": 1001 |
|
}, |
|
{ |
|
"epoch": 0.8630490956072352, |
|
"grad_norm": 0.9450215101242065, |
|
"learning_rate": 9.67950615000931e-07, |
|
"loss": 1.1746, |
|
"step": 1002 |
|
}, |
|
{ |
|
"epoch": 0.8639104220499569, |
|
"grad_norm": 0.8457410335540771, |
|
"learning_rate": 9.560106376895305e-07, |
|
"loss": 1.1797, |
|
"step": 1003 |
|
}, |
|
{ |
|
"epoch": 0.8647717484926787, |
|
"grad_norm": 0.8758255243301392, |
|
"learning_rate": 9.441410620099201e-07, |
|
"loss": 1.1639, |
|
"step": 1004 |
|
}, |
|
{ |
|
"epoch": 0.8656330749354005, |
|
"grad_norm": 0.9426845908164978, |
|
"learning_rate": 9.32341980359105e-07, |
|
"loss": 1.1763, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.8664944013781223, |
|
"grad_norm": 0.7973933219909668, |
|
"learning_rate": 9.206134845853343e-07, |
|
"loss": 1.172, |
|
"step": 1006 |
|
}, |
|
{ |
|
"epoch": 0.8673557278208441, |
|
"grad_norm": 0.8713229298591614, |
|
"learning_rate": 9.089556659873921e-07, |
|
"loss": 1.1684, |
|
"step": 1007 |
|
}, |
|
{ |
|
"epoch": 0.8682170542635659, |
|
"grad_norm": 0.8608222603797913, |
|
"learning_rate": 8.973686153138872e-07, |
|
"loss": 1.1564, |
|
"step": 1008 |
|
}, |
|
{ |
|
"epoch": 0.8690783807062877, |
|
"grad_norm": 0.8522865176200867, |
|
"learning_rate": 8.858524227625498e-07, |
|
"loss": 1.1644, |
|
"step": 1009 |
|
}, |
|
{ |
|
"epoch": 0.8699397071490095, |
|
"grad_norm": 0.8249273896217346, |
|
"learning_rate": 8.744071779795171e-07, |
|
"loss": 1.1637, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.8708010335917312, |
|
"grad_norm": 0.9221320152282715, |
|
"learning_rate": 8.630329700586481e-07, |
|
"loss": 1.1803, |
|
"step": 1011 |
|
}, |
|
{ |
|
"epoch": 0.8716623600344531, |
|
"grad_norm": 0.8558515906333923, |
|
"learning_rate": 8.517298875408253e-07, |
|
"loss": 1.1779, |
|
"step": 1012 |
|
}, |
|
{ |
|
"epoch": 0.8725236864771748, |
|
"grad_norm": 0.8073273301124573, |
|
"learning_rate": 8.40498018413266e-07, |
|
"loss": 1.1561, |
|
"step": 1013 |
|
}, |
|
{ |
|
"epoch": 0.8733850129198967, |
|
"grad_norm": 0.7956110835075378, |
|
"learning_rate": 8.293374501088358e-07, |
|
"loss": 1.1593, |
|
"step": 1014 |
|
}, |
|
{ |
|
"epoch": 0.8742463393626184, |
|
"grad_norm": 0.8590007424354553, |
|
"learning_rate": 8.182482695053728e-07, |
|
"loss": 1.1656, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.8751076658053403, |
|
"grad_norm": 0.867030143737793, |
|
"learning_rate": 8.072305629250033e-07, |
|
"loss": 1.1469, |
|
"step": 1016 |
|
}, |
|
{ |
|
"epoch": 0.875968992248062, |
|
"grad_norm": 0.8188326358795166, |
|
"learning_rate": 7.962844161334748e-07, |
|
"loss": 1.1654, |
|
"step": 1017 |
|
}, |
|
{ |
|
"epoch": 0.8768303186907838, |
|
"grad_norm": 0.8768230676651001, |
|
"learning_rate": 7.854099143394933e-07, |
|
"loss": 1.1844, |
|
"step": 1018 |
|
}, |
|
{ |
|
"epoch": 0.8776916451335056, |
|
"grad_norm": 0.840861976146698, |
|
"learning_rate": 7.746071421940482e-07, |
|
"loss": 1.1787, |
|
"step": 1019 |
|
}, |
|
{ |
|
"epoch": 0.8785529715762274, |
|
"grad_norm": 0.8956863880157471, |
|
"learning_rate": 7.638761837897612e-07, |
|
"loss": 1.1662, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.8794142980189492, |
|
"grad_norm": 0.867888867855072, |
|
"learning_rate": 7.532171226602336e-07, |
|
"loss": 1.1606, |
|
"step": 1021 |
|
}, |
|
{ |
|
"epoch": 0.880275624461671, |
|
"grad_norm": 0.8646805286407471, |
|
"learning_rate": 7.426300417793919e-07, |
|
"loss": 1.1673, |
|
"step": 1022 |
|
}, |
|
{ |
|
"epoch": 0.8811369509043928, |
|
"grad_norm": 0.8361663818359375, |
|
"learning_rate": 7.321150235608398e-07, |
|
"loss": 1.1581, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 0.8819982773471146, |
|
"grad_norm": 0.8646005988121033, |
|
"learning_rate": 7.21672149857221e-07, |
|
"loss": 1.2194, |
|
"step": 1024 |
|
}, |
|
{ |
|
"epoch": 0.8828596037898363, |
|
"grad_norm": 0.829197883605957, |
|
"learning_rate": 7.113015019595793e-07, |
|
"loss": 1.1772, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.8837209302325582, |
|
"grad_norm": 0.8020138144493103, |
|
"learning_rate": 7.010031605967316e-07, |
|
"loss": 1.1722, |
|
"step": 1026 |
|
}, |
|
{ |
|
"epoch": 0.8845822566752799, |
|
"grad_norm": 0.9305247664451599, |
|
"learning_rate": 6.907772059346285e-07, |
|
"loss": 1.1753, |
|
"step": 1027 |
|
}, |
|
{ |
|
"epoch": 0.8854435831180018, |
|
"grad_norm": 0.8150957226753235, |
|
"learning_rate": 6.806237175757457e-07, |
|
"loss": 1.1573, |
|
"step": 1028 |
|
}, |
|
{ |
|
"epoch": 0.8863049095607235, |
|
"grad_norm": 0.8378347158432007, |
|
"learning_rate": 6.705427745584469e-07, |
|
"loss": 1.1683, |
|
"step": 1029 |
|
}, |
|
{ |
|
"epoch": 0.8871662360034454, |
|
"grad_norm": 0.8095634579658508, |
|
"learning_rate": 6.605344553563775e-07, |
|
"loss": 1.1681, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.8880275624461671, |
|
"grad_norm": 0.8543710708618164, |
|
"learning_rate": 6.505988378778616e-07, |
|
"loss": 1.1875, |
|
"step": 1031 |
|
}, |
|
{ |
|
"epoch": 0.8888888888888888, |
|
"grad_norm": 0.9036997556686401, |
|
"learning_rate": 6.407359994652773e-07, |
|
"loss": 1.218, |
|
"step": 1032 |
|
}, |
|
{ |
|
"epoch": 0.8897502153316107, |
|
"grad_norm": 0.8567870855331421, |
|
"learning_rate": 6.309460168944692e-07, |
|
"loss": 1.1726, |
|
"step": 1033 |
|
}, |
|
{ |
|
"epoch": 0.8906115417743324, |
|
"grad_norm": 0.827688455581665, |
|
"learning_rate": 6.212289663741477e-07, |
|
"loss": 1.181, |
|
"step": 1034 |
|
}, |
|
{ |
|
"epoch": 0.8914728682170543, |
|
"grad_norm": 0.8630082011222839, |
|
"learning_rate": 6.11584923545292e-07, |
|
"loss": 1.1624, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.892334194659776, |
|
"grad_norm": 0.8756182789802551, |
|
"learning_rate": 6.020139634805622e-07, |
|
"loss": 1.164, |
|
"step": 1036 |
|
}, |
|
{ |
|
"epoch": 0.8931955211024979, |
|
"grad_norm": 0.8342011570930481, |
|
"learning_rate": 5.925161606837182e-07, |
|
"loss": 1.1552, |
|
"step": 1037 |
|
}, |
|
{ |
|
"epoch": 0.8940568475452196, |
|
"grad_norm": 0.8205728530883789, |
|
"learning_rate": 5.830915890890366e-07, |
|
"loss": 1.1652, |
|
"step": 1038 |
|
}, |
|
{ |
|
"epoch": 0.8949181739879414, |
|
"grad_norm": 0.8547372221946716, |
|
"learning_rate": 5.737403220607374e-07, |
|
"loss": 1.1837, |
|
"step": 1039 |
|
}, |
|
{ |
|
"epoch": 0.8957795004306632, |
|
"grad_norm": 0.8426522612571716, |
|
"learning_rate": 5.644624323924108e-07, |
|
"loss": 1.2164, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.896640826873385, |
|
"grad_norm": 0.8302751183509827, |
|
"learning_rate": 5.552579923064516e-07, |
|
"loss": 1.1782, |
|
"step": 1041 |
|
}, |
|
{ |
|
"epoch": 0.8975021533161068, |
|
"grad_norm": 0.8574025630950928, |
|
"learning_rate": 5.461270734534973e-07, |
|
"loss": 1.1599, |
|
"step": 1042 |
|
}, |
|
{ |
|
"epoch": 0.8983634797588286, |
|
"grad_norm": 0.8473777770996094, |
|
"learning_rate": 5.370697469118713e-07, |
|
"loss": 1.1562, |
|
"step": 1043 |
|
}, |
|
{ |
|
"epoch": 0.8992248062015504, |
|
"grad_norm": 0.8828932046890259, |
|
"learning_rate": 5.28086083187025e-07, |
|
"loss": 1.1839, |
|
"step": 1044 |
|
}, |
|
{ |
|
"epoch": 0.9000861326442722, |
|
"grad_norm": 0.9018280506134033, |
|
"learning_rate": 5.191761522109939e-07, |
|
"loss": 1.1884, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.9009474590869939, |
|
"grad_norm": 0.8722638487815857, |
|
"learning_rate": 5.10340023341851e-07, |
|
"loss": 1.1788, |
|
"step": 1046 |
|
}, |
|
{ |
|
"epoch": 0.9018087855297158, |
|
"grad_norm": 0.8644421696662903, |
|
"learning_rate": 5.015777653631693e-07, |
|
"loss": 1.1753, |
|
"step": 1047 |
|
}, |
|
{ |
|
"epoch": 0.9026701119724375, |
|
"grad_norm": 0.892296552658081, |
|
"learning_rate": 4.928894464834843e-07, |
|
"loss": 1.1831, |
|
"step": 1048 |
|
}, |
|
{ |
|
"epoch": 0.9035314384151594, |
|
"grad_norm": 0.8241122364997864, |
|
"learning_rate": 4.84275134335761e-07, |
|
"loss": 1.1732, |
|
"step": 1049 |
|
}, |
|
{ |
|
"epoch": 0.9043927648578811, |
|
"grad_norm": 0.8443476557731628, |
|
"learning_rate": 4.757348959768704e-07, |
|
"loss": 1.2135, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.905254091300603, |
|
"grad_norm": 0.8274143934249878, |
|
"learning_rate": 4.67268797887066e-07, |
|
"loss": 1.1765, |
|
"step": 1051 |
|
}, |
|
{ |
|
"epoch": 0.9061154177433247, |
|
"grad_norm": 0.8572297096252441, |
|
"learning_rate": 4.5887690596946975e-07, |
|
"loss": 1.1602, |
|
"step": 1052 |
|
}, |
|
{ |
|
"epoch": 0.9069767441860465, |
|
"grad_norm": 0.7972581386566162, |
|
"learning_rate": 4.5055928554955665e-07, |
|
"loss": 1.1749, |
|
"step": 1053 |
|
}, |
|
{ |
|
"epoch": 0.9078380706287683, |
|
"grad_norm": 0.8930532932281494, |
|
"learning_rate": 4.4231600137464305e-07, |
|
"loss": 1.16, |
|
"step": 1054 |
|
}, |
|
{ |
|
"epoch": 0.9086993970714901, |
|
"grad_norm": 0.866405189037323, |
|
"learning_rate": 4.341471176133838e-07, |
|
"loss": 1.1627, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.9095607235142119, |
|
"grad_norm": 0.8310356736183167, |
|
"learning_rate": 4.2605269785528037e-07, |
|
"loss": 1.1672, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 0.9104220499569337, |
|
"grad_norm": 0.8165515661239624, |
|
"learning_rate": 4.1803280511017564e-07, |
|
"loss": 1.2008, |
|
"step": 1057 |
|
}, |
|
{ |
|
"epoch": 0.9112833763996555, |
|
"grad_norm": 0.9229612946510315, |
|
"learning_rate": 4.100875018077688e-07, |
|
"loss": 1.1724, |
|
"step": 1058 |
|
}, |
|
{ |
|
"epoch": 0.9121447028423773, |
|
"grad_norm": 0.8320847749710083, |
|
"learning_rate": 4.0221684979712906e-07, |
|
"loss": 1.167, |
|
"step": 1059 |
|
}, |
|
{ |
|
"epoch": 0.9130060292850991, |
|
"grad_norm": 0.8543427586555481, |
|
"learning_rate": 3.9442091034621156e-07, |
|
"loss": 1.1767, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.9138673557278209, |
|
"grad_norm": 0.9095268249511719, |
|
"learning_rate": 3.8669974414138553e-07, |
|
"loss": 1.1781, |
|
"step": 1061 |
|
}, |
|
{ |
|
"epoch": 0.9147286821705426, |
|
"grad_norm": 0.8602164387702942, |
|
"learning_rate": 3.7905341128695484e-07, |
|
"loss": 1.1204, |
|
"step": 1062 |
|
}, |
|
{ |
|
"epoch": 0.9155900086132644, |
|
"grad_norm": 0.8160679936408997, |
|
"learning_rate": 3.7148197130469574e-07, |
|
"loss": 1.1729, |
|
"step": 1063 |
|
}, |
|
{ |
|
"epoch": 0.9164513350559862, |
|
"grad_norm": 0.8283565640449524, |
|
"learning_rate": 3.639854831333911e-07, |
|
"loss": 1.1956, |
|
"step": 1064 |
|
}, |
|
{ |
|
"epoch": 0.917312661498708, |
|
"grad_norm": 0.8678607940673828, |
|
"learning_rate": 3.5656400512837365e-07, |
|
"loss": 1.1625, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.9181739879414298, |
|
"grad_norm": 0.9065176844596863, |
|
"learning_rate": 3.4921759506106876e-07, |
|
"loss": 1.1178, |
|
"step": 1066 |
|
}, |
|
{ |
|
"epoch": 0.9190353143841516, |
|
"grad_norm": 0.8343356251716614, |
|
"learning_rate": 3.4194631011854827e-07, |
|
"loss": 1.1498, |
|
"step": 1067 |
|
}, |
|
{ |
|
"epoch": 0.9198966408268734, |
|
"grad_norm": 0.840149998664856, |
|
"learning_rate": 3.347502069030795e-07, |
|
"loss": 1.148, |
|
"step": 1068 |
|
}, |
|
{ |
|
"epoch": 0.9207579672695951, |
|
"grad_norm": 1.0363847017288208, |
|
"learning_rate": 3.2762934143169333e-07, |
|
"loss": 1.1507, |
|
"step": 1069 |
|
}, |
|
{ |
|
"epoch": 0.921619293712317, |
|
"grad_norm": 0.8863590955734253, |
|
"learning_rate": 3.205837691357405e-07, |
|
"loss": 1.1516, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.9224806201550387, |
|
"grad_norm": 0.8128695487976074, |
|
"learning_rate": 3.136135448604594e-07, |
|
"loss": 1.1545, |
|
"step": 1071 |
|
}, |
|
{ |
|
"epoch": 0.9233419465977606, |
|
"grad_norm": 0.793180525302887, |
|
"learning_rate": 3.067187228645618e-07, |
|
"loss": 1.1439, |
|
"step": 1072 |
|
}, |
|
{ |
|
"epoch": 0.9242032730404823, |
|
"grad_norm": 0.890590488910675, |
|
"learning_rate": 2.9989935681979165e-07, |
|
"loss": 1.1357, |
|
"step": 1073 |
|
}, |
|
{ |
|
"epoch": 0.9250645994832042, |
|
"grad_norm": 0.840506374835968, |
|
"learning_rate": 2.931554998105235e-07, |
|
"loss": 1.1369, |
|
"step": 1074 |
|
}, |
|
{ |
|
"epoch": 0.9259259259259259, |
|
"grad_norm": 0.9699150919914246, |
|
"learning_rate": 2.8648720433334e-07, |
|
"loss": 1.1735, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.9267872523686477, |
|
"grad_norm": 0.9378847479820251, |
|
"learning_rate": 2.798945222966265e-07, |
|
"loss": 1.1793, |
|
"step": 1076 |
|
}, |
|
{ |
|
"epoch": 0.9276485788113695, |
|
"grad_norm": 0.9439303278923035, |
|
"learning_rate": 2.733775050201626e-07, |
|
"loss": 1.145, |
|
"step": 1077 |
|
}, |
|
{ |
|
"epoch": 0.9285099052540913, |
|
"grad_norm": 0.9288724660873413, |
|
"learning_rate": 2.6693620323473556e-07, |
|
"loss": 1.1739, |
|
"step": 1078 |
|
}, |
|
{ |
|
"epoch": 0.9293712316968131, |
|
"grad_norm": 0.8163039088249207, |
|
"learning_rate": 2.605706670817276e-07, |
|
"loss": 1.1643, |
|
"step": 1079 |
|
}, |
|
{ |
|
"epoch": 0.9302325581395349, |
|
"grad_norm": 0.824459433555603, |
|
"learning_rate": 2.5428094611273825e-07, |
|
"loss": 1.1689, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.9310938845822567, |
|
"grad_norm": 0.8641778826713562, |
|
"learning_rate": 2.480670892891934e-07, |
|
"loss": 1.1534, |
|
"step": 1081 |
|
}, |
|
{ |
|
"epoch": 0.9319552110249785, |
|
"grad_norm": 0.8537282943725586, |
|
"learning_rate": 2.419291449819683e-07, |
|
"loss": 1.195, |
|
"step": 1082 |
|
}, |
|
{ |
|
"epoch": 0.9328165374677002, |
|
"grad_norm": 0.9170420169830322, |
|
"learning_rate": 2.358671609710017e-07, |
|
"loss": 1.1937, |
|
"step": 1083 |
|
}, |
|
{ |
|
"epoch": 0.9336778639104221, |
|
"grad_norm": 0.973427951335907, |
|
"learning_rate": 2.298811844449389e-07, |
|
"loss": 1.1538, |
|
"step": 1084 |
|
}, |
|
{ |
|
"epoch": 0.9345391903531438, |
|
"grad_norm": 0.9759657382965088, |
|
"learning_rate": 2.2397126200074837e-07, |
|
"loss": 1.171, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.9354005167958657, |
|
"grad_norm": 0.8713492751121521, |
|
"learning_rate": 2.1813743964336998e-07, |
|
"loss": 1.1599, |
|
"step": 1086 |
|
}, |
|
{ |
|
"epoch": 0.9362618432385874, |
|
"grad_norm": 0.8439943790435791, |
|
"learning_rate": 2.1237976278535522e-07, |
|
"loss": 1.1923, |
|
"step": 1087 |
|
}, |
|
{ |
|
"epoch": 0.9371231696813093, |
|
"grad_norm": 0.8235443234443665, |
|
"learning_rate": 2.066982762465075e-07, |
|
"loss": 1.207, |
|
"step": 1088 |
|
}, |
|
{ |
|
"epoch": 0.937984496124031, |
|
"grad_norm": 0.8441301584243774, |
|
"learning_rate": 2.0109302425354139e-07, |
|
"loss": 1.1855, |
|
"step": 1089 |
|
}, |
|
{ |
|
"epoch": 0.9388458225667528, |
|
"grad_norm": 0.8680086731910706, |
|
"learning_rate": 1.9556405043973158e-07, |
|
"loss": 1.1622, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.9397071490094746, |
|
"grad_norm": 1.0325438976287842, |
|
"learning_rate": 1.901113978445801e-07, |
|
"loss": 1.1902, |
|
"step": 1091 |
|
}, |
|
{ |
|
"epoch": 0.9405684754521964, |
|
"grad_norm": 0.908032238483429, |
|
"learning_rate": 1.8473510891347412e-07, |
|
"loss": 1.1998, |
|
"step": 1092 |
|
}, |
|
{ |
|
"epoch": 0.9414298018949182, |
|
"grad_norm": 0.8270619511604309, |
|
"learning_rate": 1.794352254973597e-07, |
|
"loss": 1.1663, |
|
"step": 1093 |
|
}, |
|
{ |
|
"epoch": 0.94229112833764, |
|
"grad_norm": 0.8700262308120728, |
|
"learning_rate": 1.742117888524153e-07, |
|
"loss": 1.1732, |
|
"step": 1094 |
|
}, |
|
{ |
|
"epoch": 0.9431524547803618, |
|
"grad_norm": 0.8367083668708801, |
|
"learning_rate": 1.6906483963973207e-07, |
|
"loss": 1.1595, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.9440137812230835, |
|
"grad_norm": 0.9255692958831787, |
|
"learning_rate": 1.6399441792499305e-07, |
|
"loss": 1.1828, |
|
"step": 1096 |
|
}, |
|
{ |
|
"epoch": 0.9448751076658053, |
|
"grad_norm": 0.8711833357810974, |
|
"learning_rate": 1.590005631781666e-07, |
|
"loss": 1.189, |
|
"step": 1097 |
|
}, |
|
{ |
|
"epoch": 0.9457364341085271, |
|
"grad_norm": 0.8770515322685242, |
|
"learning_rate": 1.5408331427319345e-07, |
|
"loss": 1.1903, |
|
"step": 1098 |
|
}, |
|
{ |
|
"epoch": 0.9465977605512489, |
|
"grad_norm": 0.8291750550270081, |
|
"learning_rate": 1.4924270948769027e-07, |
|
"loss": 1.1713, |
|
"step": 1099 |
|
}, |
|
{ |
|
"epoch": 0.9474590869939707, |
|
"grad_norm": 0.9927818179130554, |
|
"learning_rate": 1.4447878650264867e-07, |
|
"loss": 1.2139, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.9483204134366925, |
|
"grad_norm": 0.8271976709365845, |
|
"learning_rate": 1.3979158240213787e-07, |
|
"loss": 1.1788, |
|
"step": 1101 |
|
}, |
|
{ |
|
"epoch": 0.9491817398794143, |
|
"grad_norm": 0.8746073842048645, |
|
"learning_rate": 1.3518113367302356e-07, |
|
"loss": 1.149, |
|
"step": 1102 |
|
}, |
|
{ |
|
"epoch": 0.9500430663221361, |
|
"grad_norm": 0.8062018752098083, |
|
"learning_rate": 1.3064747620468053e-07, |
|
"loss": 1.1709, |
|
"step": 1103 |
|
}, |
|
{ |
|
"epoch": 0.9509043927648578, |
|
"grad_norm": 0.8557183146476746, |
|
"learning_rate": 1.2619064528871185e-07, |
|
"loss": 1.1872, |
|
"step": 1104 |
|
}, |
|
{ |
|
"epoch": 0.9517657192075797, |
|
"grad_norm": 0.8311978578567505, |
|
"learning_rate": 1.218106756186743e-07, |
|
"loss": 1.1717, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.9526270456503014, |
|
"grad_norm": 0.8271847367286682, |
|
"learning_rate": 1.1750760128981131e-07, |
|
"loss": 1.1682, |
|
"step": 1106 |
|
}, |
|
{ |
|
"epoch": 0.9534883720930233, |
|
"grad_norm": 0.8393845558166504, |
|
"learning_rate": 1.132814557987838e-07, |
|
"loss": 1.1971, |
|
"step": 1107 |
|
}, |
|
{ |
|
"epoch": 0.954349698535745, |
|
"grad_norm": 0.8846034407615662, |
|
"learning_rate": 1.0913227204341292e-07, |
|
"loss": 1.1751, |
|
"step": 1108 |
|
}, |
|
{ |
|
"epoch": 0.9552110249784669, |
|
"grad_norm": 0.7971451282501221, |
|
"learning_rate": 1.0506008232242348e-07, |
|
"loss": 1.1872, |
|
"step": 1109 |
|
}, |
|
{ |
|
"epoch": 0.9560723514211886, |
|
"grad_norm": 0.8897345066070557, |
|
"learning_rate": 1.010649183351875e-07, |
|
"loss": 1.1562, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.9569336778639104, |
|
"grad_norm": 0.8979521989822388, |
|
"learning_rate": 9.714681118148329e-08, |
|
"loss": 1.1412, |
|
"step": 1111 |
|
}, |
|
{ |
|
"epoch": 0.9577950043066322, |
|
"grad_norm": 0.836390495300293, |
|
"learning_rate": 9.330579136125117e-08, |
|
"loss": 1.1585, |
|
"step": 1112 |
|
}, |
|
{ |
|
"epoch": 0.958656330749354, |
|
"grad_norm": 0.8353316187858582, |
|
"learning_rate": 8.95418887743571e-08, |
|
"loss": 1.1722, |
|
"step": 1113 |
|
}, |
|
{ |
|
"epoch": 0.9595176571920758, |
|
"grad_norm": 0.9047684073448181, |
|
"learning_rate": 8.58551327203594e-08, |
|
"loss": 1.1509, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.9603789836347976, |
|
"grad_norm": 0.8520369529724121, |
|
"learning_rate": 8.224555189827565e-08, |
|
"loss": 1.1899, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.9612403100775194, |
|
"grad_norm": 0.8386793732643127, |
|
"learning_rate": 7.871317440637072e-08, |
|
"loss": 1.1716, |
|
"step": 1116 |
|
}, |
|
{ |
|
"epoch": 0.9621016365202412, |
|
"grad_norm": 0.819486677646637, |
|
"learning_rate": 7.525802774192791e-08, |
|
"loss": 1.1616, |
|
"step": 1117 |
|
}, |
|
{ |
|
"epoch": 0.9629629629629629, |
|
"grad_norm": 0.8136334419250488, |
|
"learning_rate": 7.188013880103817e-08, |
|
"loss": 1.1653, |
|
"step": 1118 |
|
}, |
|
{ |
|
"epoch": 0.9638242894056848, |
|
"grad_norm": 0.8472459316253662, |
|
"learning_rate": 6.857953387839012e-08, |
|
"loss": 1.1666, |
|
"step": 1119 |
|
}, |
|
{ |
|
"epoch": 0.9646856158484065, |
|
"grad_norm": 0.924691915512085, |
|
"learning_rate": 6.535623866706698e-08, |
|
"loss": 1.1448, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.9655469422911284, |
|
"grad_norm": 0.7984655499458313, |
|
"learning_rate": 6.22102782583478e-08, |
|
"loss": 1.1922, |
|
"step": 1121 |
|
}, |
|
{ |
|
"epoch": 0.9664082687338501, |
|
"grad_norm": 0.9025313854217529, |
|
"learning_rate": 5.9141677141506536e-08, |
|
"loss": 1.1607, |
|
"step": 1122 |
|
}, |
|
{ |
|
"epoch": 0.967269595176572, |
|
"grad_norm": 0.8492075204849243, |
|
"learning_rate": 5.61504592036255e-08, |
|
"loss": 1.1702, |
|
"step": 1123 |
|
}, |
|
{ |
|
"epoch": 0.9681309216192937, |
|
"grad_norm": 0.8583032488822937, |
|
"learning_rate": 5.323664772941217e-08, |
|
"loss": 1.1547, |
|
"step": 1124 |
|
}, |
|
{ |
|
"epoch": 0.9689922480620154, |
|
"grad_norm": 0.8614605665206909, |
|
"learning_rate": 5.0400265401009395e-08, |
|
"loss": 1.2079, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.9698535745047373, |
|
"grad_norm": 0.7994558215141296, |
|
"learning_rate": 4.76413342978288e-08, |
|
"loss": 1.1677, |
|
"step": 1126 |
|
}, |
|
{ |
|
"epoch": 0.970714900947459, |
|
"grad_norm": 0.8343400955200195, |
|
"learning_rate": 4.4959875896370965e-08, |
|
"loss": 1.1701, |
|
"step": 1127 |
|
}, |
|
{ |
|
"epoch": 0.9715762273901809, |
|
"grad_norm": 0.8843151330947876, |
|
"learning_rate": 4.2355911070062205e-08, |
|
"loss": 1.1443, |
|
"step": 1128 |
|
}, |
|
{ |
|
"epoch": 0.9724375538329026, |
|
"grad_norm": 0.8479021191596985, |
|
"learning_rate": 3.982946008909139e-08, |
|
"loss": 1.1963, |
|
"step": 1129 |
|
}, |
|
{ |
|
"epoch": 0.9732988802756245, |
|
"grad_norm": 0.834740936756134, |
|
"learning_rate": 3.738054262025226e-08, |
|
"loss": 1.1788, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.9741602067183462, |
|
"grad_norm": 0.8662601709365845, |
|
"learning_rate": 3.500917772679025e-08, |
|
"loss": 1.1637, |
|
"step": 1131 |
|
}, |
|
{ |
|
"epoch": 0.9750215331610681, |
|
"grad_norm": 1.2002208232879639, |
|
"learning_rate": 3.271538386825257e-08, |
|
"loss": 1.1841, |
|
"step": 1132 |
|
}, |
|
{ |
|
"epoch": 0.9758828596037898, |
|
"grad_norm": 0.8810811638832092, |
|
"learning_rate": 3.049917890034837e-08, |
|
"loss": 1.1845, |
|
"step": 1133 |
|
}, |
|
{ |
|
"epoch": 0.9767441860465116, |
|
"grad_norm": 0.8231048583984375, |
|
"learning_rate": 2.8360580074804355e-08, |
|
"loss": 1.1378, |
|
"step": 1134 |
|
}, |
|
{ |
|
"epoch": 0.9776055124892334, |
|
"grad_norm": 0.8419777750968933, |
|
"learning_rate": 2.6299604039237147e-08, |
|
"loss": 1.1737, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.9784668389319552, |
|
"grad_norm": 0.8421429991722107, |
|
"learning_rate": 2.43162668370156e-08, |
|
"loss": 1.1403, |
|
"step": 1136 |
|
}, |
|
{ |
|
"epoch": 0.979328165374677, |
|
"grad_norm": 0.8512549996376038, |
|
"learning_rate": 2.2410583907142012e-08, |
|
"loss": 1.1771, |
|
"step": 1137 |
|
}, |
|
{ |
|
"epoch": 0.9801894918173988, |
|
"grad_norm": 0.8299239873886108, |
|
"learning_rate": 2.0582570084132224e-08, |
|
"loss": 1.1864, |
|
"step": 1138 |
|
}, |
|
{ |
|
"epoch": 0.9810508182601206, |
|
"grad_norm": 0.8365635871887207, |
|
"learning_rate": 1.883223959789571e-08, |
|
"loss": 1.1988, |
|
"step": 1139 |
|
}, |
|
{ |
|
"epoch": 0.9819121447028424, |
|
"grad_norm": 0.8790571689605713, |
|
"learning_rate": 1.7159606073627875e-08, |
|
"loss": 1.1613, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.9827734711455641, |
|
"grad_norm": 0.8469464778900146, |
|
"learning_rate": 1.5564682531702402e-08, |
|
"loss": 1.1488, |
|
"step": 1141 |
|
}, |
|
{ |
|
"epoch": 0.983634797588286, |
|
"grad_norm": 0.9134259223937988, |
|
"learning_rate": 1.4047481387573503e-08, |
|
"loss": 1.1662, |
|
"step": 1142 |
|
}, |
|
{ |
|
"epoch": 0.9844961240310077, |
|
"grad_norm": 0.8976234793663025, |
|
"learning_rate": 1.2608014451672701e-08, |
|
"loss": 1.2034, |
|
"step": 1143 |
|
}, |
|
{ |
|
"epoch": 0.9853574504737296, |
|
"grad_norm": 1.0268605947494507, |
|
"learning_rate": 1.1246292929325552e-08, |
|
"loss": 1.1368, |
|
"step": 1144 |
|
}, |
|
{ |
|
"epoch": 0.9862187769164513, |
|
"grad_norm": 0.8144389986991882, |
|
"learning_rate": 9.96232742065506e-09, |
|
"loss": 1.1789, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.9870801033591732, |
|
"grad_norm": 0.9007599353790283, |
|
"learning_rate": 8.756127920505065e-09, |
|
"loss": 1.1993, |
|
"step": 1146 |
|
}, |
|
{ |
|
"epoch": 0.9879414298018949, |
|
"grad_norm": 0.8165833950042725, |
|
"learning_rate": 7.627703818363642e-09, |
|
"loss": 1.17, |
|
"step": 1147 |
|
}, |
|
{ |
|
"epoch": 0.9888027562446167, |
|
"grad_norm": 0.9031330347061157, |
|
"learning_rate": 6.577063898285385e-09, |
|
"loss": 1.1961, |
|
"step": 1148 |
|
}, |
|
{ |
|
"epoch": 0.9896640826873385, |
|
"grad_norm": 0.8834061026573181, |
|
"learning_rate": 5.604216338824797e-09, |
|
"loss": 1.2096, |
|
"step": 1149 |
|
}, |
|
{ |
|
"epoch": 0.9905254091300603, |
|
"grad_norm": 0.8041283488273621, |
|
"learning_rate": 4.7091687129718896e-09, |
|
"loss": 1.1761, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.9913867355727821, |
|
"grad_norm": 0.9003199338912964, |
|
"learning_rate": 3.891927988098898e-09, |
|
"loss": 1.206, |
|
"step": 1151 |
|
}, |
|
{ |
|
"epoch": 0.9922480620155039, |
|
"grad_norm": 0.82999587059021, |
|
"learning_rate": 3.1525005258969953e-09, |
|
"loss": 1.1214, |
|
"step": 1152 |
|
}, |
|
{ |
|
"epoch": 0.9931093884582257, |
|
"grad_norm": 0.8493615984916687, |
|
"learning_rate": 2.490892082331886e-09, |
|
"loss": 1.2059, |
|
"step": 1153 |
|
}, |
|
{ |
|
"epoch": 0.9939707149009475, |
|
"grad_norm": 0.8477335572242737, |
|
"learning_rate": 1.907107807600506e-09, |
|
"loss": 1.1827, |
|
"step": 1154 |
|
}, |
|
{ |
|
"epoch": 0.9948320413436692, |
|
"grad_norm": 0.8850772380828857, |
|
"learning_rate": 1.4011522460866122e-09, |
|
"loss": 1.1666, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.9956933677863911, |
|
"grad_norm": 0.8809195756912231, |
|
"learning_rate": 9.730293363297006e-10, |
|
"loss": 1.1649, |
|
"step": 1156 |
|
}, |
|
{ |
|
"epoch": 0.9965546942291128, |
|
"grad_norm": 0.891930103302002, |
|
"learning_rate": 6.227424109883639e-10, |
|
"loss": 1.1778, |
|
"step": 1157 |
|
}, |
|
{ |
|
"epoch": 0.9974160206718347, |
|
"grad_norm": 0.8007426261901855, |
|
"learning_rate": 3.502941968225315e-10, |
|
"loss": 1.1696, |
|
"step": 1158 |
|
}, |
|
{ |
|
"epoch": 0.9982773471145564, |
|
"grad_norm": 0.8914885520935059, |
|
"learning_rate": 1.5568681466682223e-10, |
|
"loss": 1.1262, |
|
"step": 1159 |
|
}, |
|
{ |
|
"epoch": 0.9991386735572783, |
|
"grad_norm": 0.8544787168502808, |
|
"learning_rate": 3.8921779411671414e-11, |
|
"loss": 1.1751, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.9441701769828796, |
|
"learning_rate": 0.0, |
|
"loss": 1.1453, |
|
"step": 1161 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 1161, |
|
"total_flos": 2.598284398506174e+19, |
|
"train_loss": 1.285096526043257, |
|
"train_runtime": 25375.9482, |
|
"train_samples_per_second": 35.131, |
|
"train_steps_per_second": 0.046 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 1161, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 3000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.598284398506174e+19, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|