|
{ |
|
"best_metric": 0.890067458152771, |
|
"best_model_checkpoint": "autotrain-okjuj-buuqs/checkpoint-804", |
|
"epoch": 6.0, |
|
"eval_steps": 500, |
|
"global_step": 804, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.014925373134328358, |
|
"grad_norm": 3.468337297439575, |
|
"learning_rate": 4.975124378109453e-07, |
|
"loss": 3.5512, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.029850746268656716, |
|
"grad_norm": 3.3904953002929688, |
|
"learning_rate": 9.950248756218907e-07, |
|
"loss": 3.5495, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.04477611940298507, |
|
"grad_norm": 8.210745811462402, |
|
"learning_rate": 1.4925373134328358e-06, |
|
"loss": 3.5323, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.05970149253731343, |
|
"grad_norm": 3.321380376815796, |
|
"learning_rate": 1.9900497512437813e-06, |
|
"loss": 3.5311, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.07462686567164178, |
|
"grad_norm": 3.103438377380371, |
|
"learning_rate": 2.4875621890547264e-06, |
|
"loss": 3.5312, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.08955223880597014, |
|
"grad_norm": 3.2276968955993652, |
|
"learning_rate": 2.9850746268656716e-06, |
|
"loss": 3.5472, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.1044776119402985, |
|
"grad_norm": 3.511948585510254, |
|
"learning_rate": 3.482587064676617e-06, |
|
"loss": 3.5238, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.11940298507462686, |
|
"grad_norm": 3.459820508956909, |
|
"learning_rate": 3.980099502487563e-06, |
|
"loss": 3.5457, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.13432835820895522, |
|
"grad_norm": 3.187225103378296, |
|
"learning_rate": 4.477611940298508e-06, |
|
"loss": 3.5064, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.14925373134328357, |
|
"grad_norm": 3.3685667514801025, |
|
"learning_rate": 4.975124378109453e-06, |
|
"loss": 3.5019, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.16417910447761194, |
|
"grad_norm": 3.525916576385498, |
|
"learning_rate": 5.472636815920398e-06, |
|
"loss": 3.5202, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.1791044776119403, |
|
"grad_norm": 3.199352979660034, |
|
"learning_rate": 5.970149253731343e-06, |
|
"loss": 3.4878, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.19402985074626866, |
|
"grad_norm": 3.151981830596924, |
|
"learning_rate": 6.467661691542288e-06, |
|
"loss": 3.5254, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.208955223880597, |
|
"grad_norm": 3.353118896484375, |
|
"learning_rate": 6.965174129353234e-06, |
|
"loss": 3.5179, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.22388059701492538, |
|
"grad_norm": 5.352255344390869, |
|
"learning_rate": 7.4626865671641785e-06, |
|
"loss": 3.4951, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.23880597014925373, |
|
"grad_norm": 4.2452712059021, |
|
"learning_rate": 7.960199004975125e-06, |
|
"loss": 3.5144, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.2537313432835821, |
|
"grad_norm": 3.4914658069610596, |
|
"learning_rate": 8.45771144278607e-06, |
|
"loss": 3.4755, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.26865671641791045, |
|
"grad_norm": 3.413722038269043, |
|
"learning_rate": 8.955223880597016e-06, |
|
"loss": 3.5059, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.2835820895522388, |
|
"grad_norm": 3.5784690380096436, |
|
"learning_rate": 9.45273631840796e-06, |
|
"loss": 3.503, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.29850746268656714, |
|
"grad_norm": 3.3770291805267334, |
|
"learning_rate": 9.950248756218906e-06, |
|
"loss": 3.4968, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.31343283582089554, |
|
"grad_norm": 3.709470748901367, |
|
"learning_rate": 1.0447761194029851e-05, |
|
"loss": 3.5453, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.3283582089552239, |
|
"grad_norm": 3.529064655303955, |
|
"learning_rate": 1.0945273631840796e-05, |
|
"loss": 3.4461, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.34328358208955223, |
|
"grad_norm": 3.4094252586364746, |
|
"learning_rate": 1.1442786069651743e-05, |
|
"loss": 3.483, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.3582089552238806, |
|
"grad_norm": 3.385557174682617, |
|
"learning_rate": 1.1940298507462686e-05, |
|
"loss": 3.4522, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.373134328358209, |
|
"grad_norm": 3.4028918743133545, |
|
"learning_rate": 1.2437810945273633e-05, |
|
"loss": 3.4814, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.3880597014925373, |
|
"grad_norm": 3.482884168624878, |
|
"learning_rate": 1.2935323383084577e-05, |
|
"loss": 3.4917, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.40298507462686567, |
|
"grad_norm": 3.4329397678375244, |
|
"learning_rate": 1.3432835820895523e-05, |
|
"loss": 3.4945, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.417910447761194, |
|
"grad_norm": 3.6114747524261475, |
|
"learning_rate": 1.3930348258706468e-05, |
|
"loss": 3.3991, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.43283582089552236, |
|
"grad_norm": 3.743497848510742, |
|
"learning_rate": 1.4427860696517415e-05, |
|
"loss": 3.4314, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.44776119402985076, |
|
"grad_norm": 13.538641929626465, |
|
"learning_rate": 1.4925373134328357e-05, |
|
"loss": 3.4495, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.4626865671641791, |
|
"grad_norm": 3.536511182785034, |
|
"learning_rate": 1.5422885572139304e-05, |
|
"loss": 3.4579, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.47761194029850745, |
|
"grad_norm": 4.389700889587402, |
|
"learning_rate": 1.592039800995025e-05, |
|
"loss": 3.3908, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.4925373134328358, |
|
"grad_norm": 4.821465969085693, |
|
"learning_rate": 1.6417910447761194e-05, |
|
"loss": 3.4305, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.5074626865671642, |
|
"grad_norm": 3.5865774154663086, |
|
"learning_rate": 1.691542288557214e-05, |
|
"loss": 3.4297, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.5223880597014925, |
|
"grad_norm": 3.3377368450164795, |
|
"learning_rate": 1.7412935323383088e-05, |
|
"loss": 3.5506, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.5373134328358209, |
|
"grad_norm": 3.887446880340576, |
|
"learning_rate": 1.791044776119403e-05, |
|
"loss": 3.4623, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.5522388059701493, |
|
"grad_norm": 3.546586275100708, |
|
"learning_rate": 1.8407960199004975e-05, |
|
"loss": 3.4059, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.5671641791044776, |
|
"grad_norm": 3.8084304332733154, |
|
"learning_rate": 1.890547263681592e-05, |
|
"loss": 3.3708, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.582089552238806, |
|
"grad_norm": 3.5616061687469482, |
|
"learning_rate": 1.9402985074626868e-05, |
|
"loss": 3.4297, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.5970149253731343, |
|
"grad_norm": 3.7048122882843018, |
|
"learning_rate": 1.990049751243781e-05, |
|
"loss": 3.3541, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.6119402985074627, |
|
"grad_norm": 3.7918601036071777, |
|
"learning_rate": 2.0398009950248755e-05, |
|
"loss": 3.3734, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.6268656716417911, |
|
"grad_norm": 4.018467903137207, |
|
"learning_rate": 2.0895522388059702e-05, |
|
"loss": 3.3347, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.6417910447761194, |
|
"grad_norm": 3.6919217109680176, |
|
"learning_rate": 2.139303482587065e-05, |
|
"loss": 3.4747, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.6567164179104478, |
|
"grad_norm": 3.553607940673828, |
|
"learning_rate": 2.1890547263681592e-05, |
|
"loss": 3.346, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.6716417910447762, |
|
"grad_norm": 3.717123508453369, |
|
"learning_rate": 2.238805970149254e-05, |
|
"loss": 3.3681, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.6865671641791045, |
|
"grad_norm": 4.329309463500977, |
|
"learning_rate": 2.2885572139303486e-05, |
|
"loss": 3.2841, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.7014925373134329, |
|
"grad_norm": 3.9699785709381104, |
|
"learning_rate": 2.338308457711443e-05, |
|
"loss": 3.3339, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.7164179104477612, |
|
"grad_norm": 3.9800264835357666, |
|
"learning_rate": 2.3880597014925373e-05, |
|
"loss": 3.272, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.7313432835820896, |
|
"grad_norm": 3.881544589996338, |
|
"learning_rate": 2.437810945273632e-05, |
|
"loss": 3.3739, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.746268656716418, |
|
"grad_norm": 3.7055130004882812, |
|
"learning_rate": 2.4875621890547266e-05, |
|
"loss": 3.2847, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.7611940298507462, |
|
"grad_norm": 4.061646461486816, |
|
"learning_rate": 2.537313432835821e-05, |
|
"loss": 3.2573, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.7761194029850746, |
|
"grad_norm": 4.569900035858154, |
|
"learning_rate": 2.5870646766169153e-05, |
|
"loss": 3.2594, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.7910447761194029, |
|
"grad_norm": 4.053166389465332, |
|
"learning_rate": 2.6368159203980103e-05, |
|
"loss": 3.2739, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.8059701492537313, |
|
"grad_norm": 5.338119029998779, |
|
"learning_rate": 2.6865671641791047e-05, |
|
"loss": 3.1965, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.8208955223880597, |
|
"grad_norm": 4.987647533416748, |
|
"learning_rate": 2.736318407960199e-05, |
|
"loss": 3.1951, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.835820895522388, |
|
"grad_norm": 4.582657814025879, |
|
"learning_rate": 2.7860696517412937e-05, |
|
"loss": 3.1758, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.8507462686567164, |
|
"grad_norm": 4.130568027496338, |
|
"learning_rate": 2.835820895522388e-05, |
|
"loss": 3.2563, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.8656716417910447, |
|
"grad_norm": 3.6177594661712646, |
|
"learning_rate": 2.885572139303483e-05, |
|
"loss": 3.1394, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.8805970149253731, |
|
"grad_norm": 4.4392242431640625, |
|
"learning_rate": 2.935323383084577e-05, |
|
"loss": 2.9928, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.8955223880597015, |
|
"grad_norm": 4.2055439949035645, |
|
"learning_rate": 2.9850746268656714e-05, |
|
"loss": 3.0482, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.9104477611940298, |
|
"grad_norm": 4.438642978668213, |
|
"learning_rate": 3.0348258706467664e-05, |
|
"loss": 3.024, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.9253731343283582, |
|
"grad_norm": 4.41525936126709, |
|
"learning_rate": 3.084577114427861e-05, |
|
"loss": 3.1619, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.9402985074626866, |
|
"grad_norm": 4.188882827758789, |
|
"learning_rate": 3.1343283582089554e-05, |
|
"loss": 3.2697, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.9552238805970149, |
|
"grad_norm": 3.903090000152588, |
|
"learning_rate": 3.18407960199005e-05, |
|
"loss": 3.0515, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.9701492537313433, |
|
"grad_norm": 4.03048849105835, |
|
"learning_rate": 3.233830845771145e-05, |
|
"loss": 2.9312, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.9850746268656716, |
|
"grad_norm": 3.842494249343872, |
|
"learning_rate": 3.283582089552239e-05, |
|
"loss": 3.0714, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 3.9151153564453125, |
|
"learning_rate": 3.3333333333333335e-05, |
|
"loss": 2.8809, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.375, |
|
"eval_f1_macro": 0.2794185541385505, |
|
"eval_f1_micro": 0.375, |
|
"eval_f1_weighted": 0.29179203240000456, |
|
"eval_loss": 2.9781627655029297, |
|
"eval_precision_macro": 0.26769010075416033, |
|
"eval_precision_micro": 0.375, |
|
"eval_precision_weighted": 0.28345060265850164, |
|
"eval_recall_macro": 0.3689075630252101, |
|
"eval_recall_micro": 0.375, |
|
"eval_recall_weighted": 0.375, |
|
"eval_runtime": 38.3444, |
|
"eval_samples_per_second": 5.216, |
|
"eval_steps_per_second": 0.183, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 1.0149253731343284, |
|
"grad_norm": 3.894672155380249, |
|
"learning_rate": 3.383084577114428e-05, |
|
"loss": 2.8126, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 1.0298507462686568, |
|
"grad_norm": 4.255545139312744, |
|
"learning_rate": 3.432835820895522e-05, |
|
"loss": 2.9291, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 1.044776119402985, |
|
"grad_norm": 4.275138854980469, |
|
"learning_rate": 3.4825870646766175e-05, |
|
"loss": 2.8673, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.0597014925373134, |
|
"grad_norm": 4.796700477600098, |
|
"learning_rate": 3.5323383084577115e-05, |
|
"loss": 2.8431, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 1.0746268656716418, |
|
"grad_norm": 4.217311859130859, |
|
"learning_rate": 3.582089552238806e-05, |
|
"loss": 2.8104, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 1.0895522388059702, |
|
"grad_norm": 4.417973041534424, |
|
"learning_rate": 3.631840796019901e-05, |
|
"loss": 2.7942, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 1.1044776119402986, |
|
"grad_norm": 3.8580806255340576, |
|
"learning_rate": 3.681592039800995e-05, |
|
"loss": 2.6514, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 1.1194029850746268, |
|
"grad_norm": 4.099730014801025, |
|
"learning_rate": 3.73134328358209e-05, |
|
"loss": 2.693, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.1343283582089552, |
|
"grad_norm": 4.598573684692383, |
|
"learning_rate": 3.781094527363184e-05, |
|
"loss": 2.7205, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 1.1492537313432836, |
|
"grad_norm": 4.880782604217529, |
|
"learning_rate": 3.830845771144278e-05, |
|
"loss": 2.8268, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 1.164179104477612, |
|
"grad_norm": 3.9455466270446777, |
|
"learning_rate": 3.8805970149253736e-05, |
|
"loss": 2.5482, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 1.1791044776119404, |
|
"grad_norm": 3.5259761810302734, |
|
"learning_rate": 3.9303482587064676e-05, |
|
"loss": 2.4929, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 1.1940298507462686, |
|
"grad_norm": 4.351325035095215, |
|
"learning_rate": 3.980099502487562e-05, |
|
"loss": 2.6261, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.208955223880597, |
|
"grad_norm": 3.8126604557037354, |
|
"learning_rate": 4.029850746268657e-05, |
|
"loss": 2.3924, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 1.2238805970149254, |
|
"grad_norm": 5.261939525604248, |
|
"learning_rate": 4.079601990049751e-05, |
|
"loss": 2.3364, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 1.2388059701492538, |
|
"grad_norm": 5.208514213562012, |
|
"learning_rate": 4.1293532338308464e-05, |
|
"loss": 2.3356, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 1.2537313432835822, |
|
"grad_norm": 4.369601249694824, |
|
"learning_rate": 4.1791044776119404e-05, |
|
"loss": 2.4191, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 1.2686567164179103, |
|
"grad_norm": 5.281370639801025, |
|
"learning_rate": 4.228855721393035e-05, |
|
"loss": 2.4544, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.2835820895522387, |
|
"grad_norm": 4.296084880828857, |
|
"learning_rate": 4.27860696517413e-05, |
|
"loss": 2.4446, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 1.2985074626865671, |
|
"grad_norm": 3.7815287113189697, |
|
"learning_rate": 4.328358208955224e-05, |
|
"loss": 2.3687, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 1.3134328358208955, |
|
"grad_norm": 7.161779403686523, |
|
"learning_rate": 4.3781094527363184e-05, |
|
"loss": 2.67, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 1.328358208955224, |
|
"grad_norm": 4.957999229431152, |
|
"learning_rate": 4.427860696517413e-05, |
|
"loss": 2.4644, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 1.3432835820895521, |
|
"grad_norm": 4.455275058746338, |
|
"learning_rate": 4.477611940298508e-05, |
|
"loss": 2.4389, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.3582089552238805, |
|
"grad_norm": 5.357694625854492, |
|
"learning_rate": 4.5273631840796025e-05, |
|
"loss": 2.1849, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 1.373134328358209, |
|
"grad_norm": 6.346301078796387, |
|
"learning_rate": 4.577114427860697e-05, |
|
"loss": 2.3134, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 1.3880597014925373, |
|
"grad_norm": 6.189851760864258, |
|
"learning_rate": 4.626865671641791e-05, |
|
"loss": 2.322, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 1.4029850746268657, |
|
"grad_norm": 3.729904890060425, |
|
"learning_rate": 4.676616915422886e-05, |
|
"loss": 2.2053, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 1.417910447761194, |
|
"grad_norm": 4.210063457489014, |
|
"learning_rate": 4.7263681592039805e-05, |
|
"loss": 1.9863, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.4328358208955223, |
|
"grad_norm": 4.447649002075195, |
|
"learning_rate": 4.7761194029850745e-05, |
|
"loss": 2.0302, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 1.4477611940298507, |
|
"grad_norm": 5.010406970977783, |
|
"learning_rate": 4.82587064676617e-05, |
|
"loss": 2.379, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 1.462686567164179, |
|
"grad_norm": 4.134596824645996, |
|
"learning_rate": 4.875621890547264e-05, |
|
"loss": 2.1104, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 1.4776119402985075, |
|
"grad_norm": 3.678053140640259, |
|
"learning_rate": 4.9253731343283586e-05, |
|
"loss": 2.0437, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 1.4925373134328357, |
|
"grad_norm": 6.969429016113281, |
|
"learning_rate": 4.975124378109453e-05, |
|
"loss": 2.1688, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.5074626865671643, |
|
"grad_norm": 5.489563941955566, |
|
"learning_rate": 4.997236042012162e-05, |
|
"loss": 1.9494, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 1.5223880597014925, |
|
"grad_norm": 3.8776581287384033, |
|
"learning_rate": 4.991708126036485e-05, |
|
"loss": 1.7246, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 1.537313432835821, |
|
"grad_norm": 3.915743350982666, |
|
"learning_rate": 4.9861802100608076e-05, |
|
"loss": 2.1206, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 1.5522388059701493, |
|
"grad_norm": 10.498429298400879, |
|
"learning_rate": 4.9806522940851304e-05, |
|
"loss": 1.8089, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 1.5671641791044775, |
|
"grad_norm": 5.074433326721191, |
|
"learning_rate": 4.975124378109453e-05, |
|
"loss": 2.2841, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.582089552238806, |
|
"grad_norm": 16.675743103027344, |
|
"learning_rate": 4.969596462133776e-05, |
|
"loss": 1.9404, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 1.5970149253731343, |
|
"grad_norm": 4.692626953125, |
|
"learning_rate": 4.964068546158099e-05, |
|
"loss": 1.9563, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 1.6119402985074627, |
|
"grad_norm": 4.603738307952881, |
|
"learning_rate": 4.958540630182422e-05, |
|
"loss": 2.4004, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 1.626865671641791, |
|
"grad_norm": 29.3167781829834, |
|
"learning_rate": 4.9530127142067445e-05, |
|
"loss": 2.1461, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 1.6417910447761193, |
|
"grad_norm": 4.929641246795654, |
|
"learning_rate": 4.947484798231067e-05, |
|
"loss": 1.6009, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.6567164179104479, |
|
"grad_norm": 7.109385967254639, |
|
"learning_rate": 4.94195688225539e-05, |
|
"loss": 1.884, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 1.671641791044776, |
|
"grad_norm": 14.075770378112793, |
|
"learning_rate": 4.936428966279712e-05, |
|
"loss": 1.6543, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 1.6865671641791045, |
|
"grad_norm": 4.957328796386719, |
|
"learning_rate": 4.930901050304035e-05, |
|
"loss": 1.6223, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 1.7014925373134329, |
|
"grad_norm": 6.562300205230713, |
|
"learning_rate": 4.9253731343283586e-05, |
|
"loss": 1.4456, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 1.716417910447761, |
|
"grad_norm": 5.4999237060546875, |
|
"learning_rate": 4.9198452183526814e-05, |
|
"loss": 1.7814, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.7313432835820897, |
|
"grad_norm": 5.5423736572265625, |
|
"learning_rate": 4.914317302377004e-05, |
|
"loss": 1.884, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 1.7462686567164178, |
|
"grad_norm": 5.507659435272217, |
|
"learning_rate": 4.908789386401327e-05, |
|
"loss": 1.8302, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 1.7611940298507462, |
|
"grad_norm": 7.983649730682373, |
|
"learning_rate": 4.90326147042565e-05, |
|
"loss": 1.9323, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 1.7761194029850746, |
|
"grad_norm": 4.843520641326904, |
|
"learning_rate": 4.8977335544499726e-05, |
|
"loss": 1.6411, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 1.7910447761194028, |
|
"grad_norm": 4.717188835144043, |
|
"learning_rate": 4.8922056384742954e-05, |
|
"loss": 1.6714, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.8059701492537314, |
|
"grad_norm": 4.654526233673096, |
|
"learning_rate": 4.886677722498618e-05, |
|
"loss": 1.6478, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 1.8208955223880596, |
|
"grad_norm": 5.104541301727295, |
|
"learning_rate": 4.881149806522941e-05, |
|
"loss": 1.6125, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 1.835820895522388, |
|
"grad_norm": 4.356812000274658, |
|
"learning_rate": 4.875621890547264e-05, |
|
"loss": 1.9471, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 1.8507462686567164, |
|
"grad_norm": 4.166391849517822, |
|
"learning_rate": 4.870093974571587e-05, |
|
"loss": 1.4783, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 1.8656716417910446, |
|
"grad_norm": 6.2490057945251465, |
|
"learning_rate": 4.8645660585959095e-05, |
|
"loss": 1.8157, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.8805970149253732, |
|
"grad_norm": 5.395322322845459, |
|
"learning_rate": 4.859038142620232e-05, |
|
"loss": 1.8702, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 1.8955223880597014, |
|
"grad_norm": 4.278140544891357, |
|
"learning_rate": 4.853510226644555e-05, |
|
"loss": 1.5588, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 1.9104477611940298, |
|
"grad_norm": 5.143709182739258, |
|
"learning_rate": 4.847982310668878e-05, |
|
"loss": 1.5783, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 1.9253731343283582, |
|
"grad_norm": 8.840173721313477, |
|
"learning_rate": 4.842454394693201e-05, |
|
"loss": 1.5149, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 1.9402985074626866, |
|
"grad_norm": 5.001150608062744, |
|
"learning_rate": 4.836926478717524e-05, |
|
"loss": 1.5529, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.955223880597015, |
|
"grad_norm": 4.259793281555176, |
|
"learning_rate": 4.831398562741847e-05, |
|
"loss": 1.7178, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 1.9701492537313432, |
|
"grad_norm": 3.2455430030822754, |
|
"learning_rate": 4.82587064676617e-05, |
|
"loss": 1.26, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 1.9850746268656716, |
|
"grad_norm": 3.899599313735962, |
|
"learning_rate": 4.820342730790493e-05, |
|
"loss": 1.5229, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 6.097535610198975, |
|
"learning_rate": 4.814814814814815e-05, |
|
"loss": 1.6973, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.63, |
|
"eval_f1_macro": 0.569843507632617, |
|
"eval_f1_micro": 0.63, |
|
"eval_f1_weighted": 0.5873017265672064, |
|
"eval_loss": 1.6746174097061157, |
|
"eval_precision_macro": 0.5998444366091426, |
|
"eval_precision_micro": 0.63, |
|
"eval_precision_weighted": 0.623490675990676, |
|
"eval_recall_macro": 0.6212885154061624, |
|
"eval_recall_micro": 0.63, |
|
"eval_recall_weighted": 0.63, |
|
"eval_runtime": 35.3294, |
|
"eval_samples_per_second": 5.661, |
|
"eval_steps_per_second": 0.198, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 2.014925373134328, |
|
"grad_norm": 2.5601863861083984, |
|
"learning_rate": 4.8092868988391376e-05, |
|
"loss": 1.0241, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.029850746268657, |
|
"grad_norm": 4.754575252532959, |
|
"learning_rate": 4.8037589828634604e-05, |
|
"loss": 1.5737, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 2.044776119402985, |
|
"grad_norm": 4.370357990264893, |
|
"learning_rate": 4.798231066887783e-05, |
|
"loss": 1.4961, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 2.0597014925373136, |
|
"grad_norm": 8.790416717529297, |
|
"learning_rate": 4.792703150912106e-05, |
|
"loss": 1.6075, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 2.074626865671642, |
|
"grad_norm": 4.466557502746582, |
|
"learning_rate": 4.787175234936429e-05, |
|
"loss": 1.4832, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 2.08955223880597, |
|
"grad_norm": 3.5914487838745117, |
|
"learning_rate": 4.781647318960752e-05, |
|
"loss": 1.1338, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.1044776119402986, |
|
"grad_norm": 3.8444533348083496, |
|
"learning_rate": 4.7761194029850745e-05, |
|
"loss": 1.2921, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 2.1194029850746268, |
|
"grad_norm": 7.032942295074463, |
|
"learning_rate": 4.770591487009397e-05, |
|
"loss": 1.5322, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 2.1343283582089554, |
|
"grad_norm": 6.563926696777344, |
|
"learning_rate": 4.76506357103372e-05, |
|
"loss": 1.2677, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 2.1492537313432836, |
|
"grad_norm": 3.413158655166626, |
|
"learning_rate": 4.7595356550580436e-05, |
|
"loss": 1.2462, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 2.1641791044776117, |
|
"grad_norm": 4.282972812652588, |
|
"learning_rate": 4.7540077390823664e-05, |
|
"loss": 1.1632, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.1791044776119404, |
|
"grad_norm": 11.003588676452637, |
|
"learning_rate": 4.748479823106689e-05, |
|
"loss": 1.2687, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 2.1940298507462686, |
|
"grad_norm": 4.725019931793213, |
|
"learning_rate": 4.742951907131012e-05, |
|
"loss": 1.2352, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 2.208955223880597, |
|
"grad_norm": 3.6079459190368652, |
|
"learning_rate": 4.737423991155335e-05, |
|
"loss": 1.4994, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 2.2238805970149254, |
|
"grad_norm": 3.576037645339966, |
|
"learning_rate": 4.731896075179658e-05, |
|
"loss": 1.1007, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 2.2388059701492535, |
|
"grad_norm": 3.0327272415161133, |
|
"learning_rate": 4.7263681592039805e-05, |
|
"loss": 1.0624, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.253731343283582, |
|
"grad_norm": 5.710484504699707, |
|
"learning_rate": 4.720840243228303e-05, |
|
"loss": 1.2262, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 2.2686567164179103, |
|
"grad_norm": 3.059814453125, |
|
"learning_rate": 4.715312327252626e-05, |
|
"loss": 1.0292, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 2.283582089552239, |
|
"grad_norm": 4.122474193572998, |
|
"learning_rate": 4.709784411276949e-05, |
|
"loss": 1.1731, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 2.298507462686567, |
|
"grad_norm": 8.998103141784668, |
|
"learning_rate": 4.704256495301272e-05, |
|
"loss": 1.5075, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 2.3134328358208958, |
|
"grad_norm": 3.947554349899292, |
|
"learning_rate": 4.6987285793255946e-05, |
|
"loss": 1.1681, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.328358208955224, |
|
"grad_norm": 6.467041492462158, |
|
"learning_rate": 4.693200663349917e-05, |
|
"loss": 1.2688, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 2.343283582089552, |
|
"grad_norm": 6.714489936828613, |
|
"learning_rate": 4.68767274737424e-05, |
|
"loss": 1.5061, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 2.3582089552238807, |
|
"grad_norm": 4.802642822265625, |
|
"learning_rate": 4.682144831398563e-05, |
|
"loss": 1.0498, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 2.373134328358209, |
|
"grad_norm": 3.7944350242614746, |
|
"learning_rate": 4.676616915422886e-05, |
|
"loss": 1.1943, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 2.388059701492537, |
|
"grad_norm": 6.245696067810059, |
|
"learning_rate": 4.6710889994472086e-05, |
|
"loss": 0.9827, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.4029850746268657, |
|
"grad_norm": 3.385528326034546, |
|
"learning_rate": 4.6655610834715315e-05, |
|
"loss": 1.2411, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 2.417910447761194, |
|
"grad_norm": 4.5617828369140625, |
|
"learning_rate": 4.660033167495854e-05, |
|
"loss": 0.947, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 2.4328358208955225, |
|
"grad_norm": 6.033749580383301, |
|
"learning_rate": 4.654505251520177e-05, |
|
"loss": 1.1135, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 2.4477611940298507, |
|
"grad_norm": 4.091092586517334, |
|
"learning_rate": 4.6489773355445e-05, |
|
"loss": 1.2219, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 2.4626865671641793, |
|
"grad_norm": 4.028707027435303, |
|
"learning_rate": 4.643449419568823e-05, |
|
"loss": 0.8016, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.4776119402985075, |
|
"grad_norm": 13.044775009155273, |
|
"learning_rate": 4.6379215035931455e-05, |
|
"loss": 1.0265, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 2.4925373134328357, |
|
"grad_norm": 3.4753003120422363, |
|
"learning_rate": 4.632393587617468e-05, |
|
"loss": 1.1795, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 2.5074626865671643, |
|
"grad_norm": 10.754464149475098, |
|
"learning_rate": 4.626865671641791e-05, |
|
"loss": 1.0397, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 2.5223880597014925, |
|
"grad_norm": 5.801613807678223, |
|
"learning_rate": 4.621337755666114e-05, |
|
"loss": 1.0849, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 2.5373134328358207, |
|
"grad_norm": 7.185355186462402, |
|
"learning_rate": 4.615809839690437e-05, |
|
"loss": 1.3039, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 2.5522388059701493, |
|
"grad_norm": 5.9482622146606445, |
|
"learning_rate": 4.6102819237147596e-05, |
|
"loss": 0.9354, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 2.5671641791044775, |
|
"grad_norm": 5.785240173339844, |
|
"learning_rate": 4.6047540077390824e-05, |
|
"loss": 1.1711, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 2.582089552238806, |
|
"grad_norm": 4.588739395141602, |
|
"learning_rate": 4.599226091763406e-05, |
|
"loss": 1.2397, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 2.5970149253731343, |
|
"grad_norm": 9.58119010925293, |
|
"learning_rate": 4.593698175787729e-05, |
|
"loss": 1.0885, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 2.611940298507463, |
|
"grad_norm": 7.116174221038818, |
|
"learning_rate": 4.5881702598120515e-05, |
|
"loss": 1.4472, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 2.626865671641791, |
|
"grad_norm": 4.783851146697998, |
|
"learning_rate": 4.582642343836374e-05, |
|
"loss": 1.0596, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 2.6417910447761193, |
|
"grad_norm": 8.053864479064941, |
|
"learning_rate": 4.577114427860697e-05, |
|
"loss": 1.1109, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 2.656716417910448, |
|
"grad_norm": 7.039917945861816, |
|
"learning_rate": 4.571586511885019e-05, |
|
"loss": 1.0317, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 2.671641791044776, |
|
"grad_norm": 10.328742980957031, |
|
"learning_rate": 4.566058595909342e-05, |
|
"loss": 1.3927, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 2.6865671641791042, |
|
"grad_norm": 6.044647693634033, |
|
"learning_rate": 4.560530679933665e-05, |
|
"loss": 1.6115, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 2.701492537313433, |
|
"grad_norm": 5.099875450134277, |
|
"learning_rate": 4.555002763957988e-05, |
|
"loss": 0.7492, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 2.716417910447761, |
|
"grad_norm": 5.763055324554443, |
|
"learning_rate": 4.5494748479823105e-05, |
|
"loss": 1.2244, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 2.7313432835820897, |
|
"grad_norm": 7.4969987869262695, |
|
"learning_rate": 4.5439469320066333e-05, |
|
"loss": 0.9924, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 2.746268656716418, |
|
"grad_norm": 11.879778861999512, |
|
"learning_rate": 4.538419016030956e-05, |
|
"loss": 1.3022, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 2.7611940298507465, |
|
"grad_norm": 2.6563034057617188, |
|
"learning_rate": 4.532891100055279e-05, |
|
"loss": 1.0468, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 2.7761194029850746, |
|
"grad_norm": 4.352832317352295, |
|
"learning_rate": 4.5273631840796025e-05, |
|
"loss": 0.7805, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 2.791044776119403, |
|
"grad_norm": 5.752820014953613, |
|
"learning_rate": 4.521835268103925e-05, |
|
"loss": 1.0971, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 2.8059701492537314, |
|
"grad_norm": 6.028427600860596, |
|
"learning_rate": 4.516307352128248e-05, |
|
"loss": 0.9887, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 2.8208955223880596, |
|
"grad_norm": 5.541314601898193, |
|
"learning_rate": 4.510779436152571e-05, |
|
"loss": 1.0044, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 2.835820895522388, |
|
"grad_norm": 6.708818435668945, |
|
"learning_rate": 4.505251520176894e-05, |
|
"loss": 0.9756, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 2.8507462686567164, |
|
"grad_norm": 5.191244602203369, |
|
"learning_rate": 4.4997236042012165e-05, |
|
"loss": 1.0194, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 2.8656716417910446, |
|
"grad_norm": 5.064216613769531, |
|
"learning_rate": 4.4941956882255393e-05, |
|
"loss": 1.1003, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 2.8805970149253732, |
|
"grad_norm": 4.351062774658203, |
|
"learning_rate": 4.488667772249862e-05, |
|
"loss": 1.0402, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 2.8955223880597014, |
|
"grad_norm": 9.046034812927246, |
|
"learning_rate": 4.483139856274185e-05, |
|
"loss": 1.0788, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 2.91044776119403, |
|
"grad_norm": 9.672553062438965, |
|
"learning_rate": 4.477611940298508e-05, |
|
"loss": 1.1218, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 2.925373134328358, |
|
"grad_norm": 6.787928581237793, |
|
"learning_rate": 4.4720840243228306e-05, |
|
"loss": 0.9067, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 2.9402985074626864, |
|
"grad_norm": 7.9477152824401855, |
|
"learning_rate": 4.4665561083471534e-05, |
|
"loss": 1.0954, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 2.955223880597015, |
|
"grad_norm": 6.69403600692749, |
|
"learning_rate": 4.461028192371476e-05, |
|
"loss": 1.2409, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 2.970149253731343, |
|
"grad_norm": 6.419997692108154, |
|
"learning_rate": 4.455500276395799e-05, |
|
"loss": 0.7133, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 2.9850746268656714, |
|
"grad_norm": 6.382747173309326, |
|
"learning_rate": 4.449972360420122e-05, |
|
"loss": 1.3368, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 6.085859298706055, |
|
"learning_rate": 4.4444444444444447e-05, |
|
"loss": 0.8309, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.7, |
|
"eval_f1_macro": 0.6561413388812769, |
|
"eval_f1_micro": 0.7, |
|
"eval_f1_weighted": 0.6696798888246258, |
|
"eval_loss": 1.1434708833694458, |
|
"eval_precision_macro": 0.6706204141575659, |
|
"eval_precision_micro": 0.7, |
|
"eval_precision_weighted": 0.6840218575284365, |
|
"eval_recall_macro": 0.6900560224089635, |
|
"eval_recall_micro": 0.7, |
|
"eval_recall_weighted": 0.7, |
|
"eval_runtime": 36.7029, |
|
"eval_samples_per_second": 5.449, |
|
"eval_steps_per_second": 0.191, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 3.014925373134328, |
|
"grad_norm": 8.274943351745605, |
|
"learning_rate": 4.4389165284687675e-05, |
|
"loss": 0.728, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 3.029850746268657, |
|
"grad_norm": 3.896862745285034, |
|
"learning_rate": 4.43338861249309e-05, |
|
"loss": 0.8955, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 3.044776119402985, |
|
"grad_norm": 3.3480184078216553, |
|
"learning_rate": 4.427860696517413e-05, |
|
"loss": 0.84, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 3.0597014925373136, |
|
"grad_norm": 9.12331485748291, |
|
"learning_rate": 4.422332780541736e-05, |
|
"loss": 0.8026, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 3.074626865671642, |
|
"grad_norm": 6.628831386566162, |
|
"learning_rate": 4.416804864566059e-05, |
|
"loss": 1.0466, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 3.08955223880597, |
|
"grad_norm": 2.638613224029541, |
|
"learning_rate": 4.4112769485903815e-05, |
|
"loss": 0.6834, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 3.1044776119402986, |
|
"grad_norm": 3.1941006183624268, |
|
"learning_rate": 4.4057490326147043e-05, |
|
"loss": 0.9236, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 3.1194029850746268, |
|
"grad_norm": 4.337372779846191, |
|
"learning_rate": 4.400221116639027e-05, |
|
"loss": 0.8447, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 3.1343283582089554, |
|
"grad_norm": 8.465522766113281, |
|
"learning_rate": 4.39469320066335e-05, |
|
"loss": 0.8768, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 3.1492537313432836, |
|
"grad_norm": 4.252687454223633, |
|
"learning_rate": 4.389165284687673e-05, |
|
"loss": 0.7148, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 3.1641791044776117, |
|
"grad_norm": 5.22479772567749, |
|
"learning_rate": 4.3836373687119956e-05, |
|
"loss": 0.7375, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 3.1791044776119404, |
|
"grad_norm": 3.702615737915039, |
|
"learning_rate": 4.3781094527363184e-05, |
|
"loss": 0.8488, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 3.1940298507462686, |
|
"grad_norm": 3.2345383167266846, |
|
"learning_rate": 4.372581536760641e-05, |
|
"loss": 0.7861, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 3.208955223880597, |
|
"grad_norm": 7.831496238708496, |
|
"learning_rate": 4.367053620784965e-05, |
|
"loss": 1.117, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 3.2238805970149254, |
|
"grad_norm": 2.40031361579895, |
|
"learning_rate": 4.3615257048092875e-05, |
|
"loss": 0.7079, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 3.2388059701492535, |
|
"grad_norm": 4.2407050132751465, |
|
"learning_rate": 4.3559977888336103e-05, |
|
"loss": 0.7778, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 3.253731343283582, |
|
"grad_norm": 3.885312795639038, |
|
"learning_rate": 4.350469872857933e-05, |
|
"loss": 1.0815, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 3.2686567164179103, |
|
"grad_norm": 3.0180788040161133, |
|
"learning_rate": 4.344941956882256e-05, |
|
"loss": 0.6417, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 3.283582089552239, |
|
"grad_norm": 6.578612804412842, |
|
"learning_rate": 4.339414040906579e-05, |
|
"loss": 0.9152, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 3.298507462686567, |
|
"grad_norm": 4.349058151245117, |
|
"learning_rate": 4.3338861249309016e-05, |
|
"loss": 0.9936, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 3.3134328358208958, |
|
"grad_norm": 7.187576770782471, |
|
"learning_rate": 4.328358208955224e-05, |
|
"loss": 0.9757, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 3.328358208955224, |
|
"grad_norm": 3.981358528137207, |
|
"learning_rate": 4.3228302929795465e-05, |
|
"loss": 0.6527, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 3.343283582089552, |
|
"grad_norm": 4.734504222869873, |
|
"learning_rate": 4.3173023770038694e-05, |
|
"loss": 0.917, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 3.3582089552238807, |
|
"grad_norm": 7.110349655151367, |
|
"learning_rate": 4.311774461028192e-05, |
|
"loss": 0.7495, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 3.373134328358209, |
|
"grad_norm": 4.323439121246338, |
|
"learning_rate": 4.306246545052515e-05, |
|
"loss": 0.7577, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 3.388059701492537, |
|
"grad_norm": 6.730025768280029, |
|
"learning_rate": 4.300718629076838e-05, |
|
"loss": 1.2703, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 3.4029850746268657, |
|
"grad_norm": 10.35930347442627, |
|
"learning_rate": 4.2951907131011606e-05, |
|
"loss": 0.4539, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 3.417910447761194, |
|
"grad_norm": 3.9190192222595215, |
|
"learning_rate": 4.289662797125484e-05, |
|
"loss": 0.8863, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 3.4328358208955225, |
|
"grad_norm": 6.045052528381348, |
|
"learning_rate": 4.284134881149807e-05, |
|
"loss": 0.8221, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 3.4477611940298507, |
|
"grad_norm": 9.53062629699707, |
|
"learning_rate": 4.27860696517413e-05, |
|
"loss": 0.9652, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 3.4626865671641793, |
|
"grad_norm": 5.4133710861206055, |
|
"learning_rate": 4.2730790491984525e-05, |
|
"loss": 0.6179, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 3.4776119402985075, |
|
"grad_norm": 5.08922004699707, |
|
"learning_rate": 4.2675511332227754e-05, |
|
"loss": 0.8349, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 3.4925373134328357, |
|
"grad_norm": 9.440414428710938, |
|
"learning_rate": 4.262023217247098e-05, |
|
"loss": 0.7352, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 3.5074626865671643, |
|
"grad_norm": 5.086453437805176, |
|
"learning_rate": 4.256495301271421e-05, |
|
"loss": 0.6674, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 3.5223880597014925, |
|
"grad_norm": 5.283850193023682, |
|
"learning_rate": 4.250967385295744e-05, |
|
"loss": 0.9129, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 3.5373134328358207, |
|
"grad_norm": 6.586909770965576, |
|
"learning_rate": 4.2454394693200666e-05, |
|
"loss": 0.4512, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 3.5522388059701493, |
|
"grad_norm": 6.328141212463379, |
|
"learning_rate": 4.2399115533443894e-05, |
|
"loss": 0.9986, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 3.5671641791044775, |
|
"grad_norm": 11.822186470031738, |
|
"learning_rate": 4.234383637368712e-05, |
|
"loss": 0.6238, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 3.582089552238806, |
|
"grad_norm": 12.303596496582031, |
|
"learning_rate": 4.228855721393035e-05, |
|
"loss": 0.666, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 3.5970149253731343, |
|
"grad_norm": 10.991721153259277, |
|
"learning_rate": 4.223327805417358e-05, |
|
"loss": 0.8176, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 3.611940298507463, |
|
"grad_norm": 4.936182022094727, |
|
"learning_rate": 4.217799889441681e-05, |
|
"loss": 0.6964, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 3.626865671641791, |
|
"grad_norm": 4.455245494842529, |
|
"learning_rate": 4.2122719734660035e-05, |
|
"loss": 0.8681, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 3.6417910447761193, |
|
"grad_norm": 4.1669206619262695, |
|
"learning_rate": 4.206744057490326e-05, |
|
"loss": 0.8556, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 3.656716417910448, |
|
"grad_norm": 2.757246732711792, |
|
"learning_rate": 4.201216141514649e-05, |
|
"loss": 0.6557, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 3.671641791044776, |
|
"grad_norm": 4.065201282501221, |
|
"learning_rate": 4.195688225538972e-05, |
|
"loss": 0.7899, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 3.6865671641791042, |
|
"grad_norm": 5.666295528411865, |
|
"learning_rate": 4.190160309563295e-05, |
|
"loss": 0.7589, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 3.701492537313433, |
|
"grad_norm": 4.929719924926758, |
|
"learning_rate": 4.1846323935876176e-05, |
|
"loss": 0.6709, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 3.716417910447761, |
|
"grad_norm": 6.261268138885498, |
|
"learning_rate": 4.1791044776119404e-05, |
|
"loss": 0.4849, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 3.7313432835820897, |
|
"grad_norm": 1.9559329748153687, |
|
"learning_rate": 4.173576561636263e-05, |
|
"loss": 0.4897, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 3.746268656716418, |
|
"grad_norm": 3.6171419620513916, |
|
"learning_rate": 4.168048645660586e-05, |
|
"loss": 0.4541, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 3.7611940298507465, |
|
"grad_norm": 5.1620988845825195, |
|
"learning_rate": 4.162520729684909e-05, |
|
"loss": 0.4952, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 3.7761194029850746, |
|
"grad_norm": 6.901045322418213, |
|
"learning_rate": 4.1569928137092316e-05, |
|
"loss": 0.8728, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 3.791044776119403, |
|
"grad_norm": 6.784924507141113, |
|
"learning_rate": 4.1514648977335544e-05, |
|
"loss": 1.0379, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 3.8059701492537314, |
|
"grad_norm": 3.829332113265991, |
|
"learning_rate": 4.145936981757877e-05, |
|
"loss": 0.8488, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 3.8208955223880596, |
|
"grad_norm": 8.228715896606445, |
|
"learning_rate": 4.1404090657822e-05, |
|
"loss": 0.4737, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 3.835820895522388, |
|
"grad_norm": 7.089122295379639, |
|
"learning_rate": 4.134881149806523e-05, |
|
"loss": 0.8644, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 3.8507462686567164, |
|
"grad_norm": 6.432941913604736, |
|
"learning_rate": 4.1293532338308464e-05, |
|
"loss": 0.5153, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 3.8656716417910446, |
|
"grad_norm": 8.107361793518066, |
|
"learning_rate": 4.123825317855169e-05, |
|
"loss": 0.6265, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 3.8805970149253732, |
|
"grad_norm": 4.366127014160156, |
|
"learning_rate": 4.118297401879492e-05, |
|
"loss": 0.9007, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 3.8955223880597014, |
|
"grad_norm": 3.596374988555908, |
|
"learning_rate": 4.112769485903815e-05, |
|
"loss": 0.7653, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 3.91044776119403, |
|
"grad_norm": 6.282411098480225, |
|
"learning_rate": 4.1072415699281376e-05, |
|
"loss": 0.8757, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 3.925373134328358, |
|
"grad_norm": 6.818746566772461, |
|
"learning_rate": 4.1017136539524604e-05, |
|
"loss": 0.7569, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 3.9402985074626864, |
|
"grad_norm": 6.6983113288879395, |
|
"learning_rate": 4.096185737976783e-05, |
|
"loss": 0.79, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 3.955223880597015, |
|
"grad_norm": 3.532710075378418, |
|
"learning_rate": 4.090657822001106e-05, |
|
"loss": 0.6981, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 3.970149253731343, |
|
"grad_norm": 6.195641040802002, |
|
"learning_rate": 4.085129906025428e-05, |
|
"loss": 0.8261, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 3.9850746268656714, |
|
"grad_norm": 8.55378246307373, |
|
"learning_rate": 4.079601990049751e-05, |
|
"loss": 0.7682, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 7.6657891273498535, |
|
"learning_rate": 4.074074074074074e-05, |
|
"loss": 0.6755, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.74, |
|
"eval_f1_macro": 0.6978798162621692, |
|
"eval_f1_micro": 0.74, |
|
"eval_f1_weighted": 0.7101990093240095, |
|
"eval_loss": 0.93646240234375, |
|
"eval_precision_macro": 0.7079046770223241, |
|
"eval_precision_micro": 0.74, |
|
"eval_precision_weighted": 0.719476911976912, |
|
"eval_recall_macro": 0.7278711484593837, |
|
"eval_recall_micro": 0.74, |
|
"eval_recall_weighted": 0.74, |
|
"eval_runtime": 40.7905, |
|
"eval_samples_per_second": 4.903, |
|
"eval_steps_per_second": 0.172, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 4.014925373134329, |
|
"grad_norm": 3.920659065246582, |
|
"learning_rate": 4.0685461580983966e-05, |
|
"loss": 0.5528, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 4.029850746268656, |
|
"grad_norm": 5.147222995758057, |
|
"learning_rate": 4.0630182421227194e-05, |
|
"loss": 1.0055, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 4.044776119402985, |
|
"grad_norm": 2.260542869567871, |
|
"learning_rate": 4.057490326147043e-05, |
|
"loss": 0.6422, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 4.059701492537314, |
|
"grad_norm": 5.286315441131592, |
|
"learning_rate": 4.051962410171366e-05, |
|
"loss": 0.5399, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 4.074626865671641, |
|
"grad_norm": 8.412203788757324, |
|
"learning_rate": 4.0464344941956886e-05, |
|
"loss": 0.6129, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 4.08955223880597, |
|
"grad_norm": 6.18348503112793, |
|
"learning_rate": 4.0409065782200114e-05, |
|
"loss": 0.5879, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 4.104477611940299, |
|
"grad_norm": 2.4661457538604736, |
|
"learning_rate": 4.035378662244334e-05, |
|
"loss": 0.5941, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 4.119402985074627, |
|
"grad_norm": 4.261096477508545, |
|
"learning_rate": 4.029850746268657e-05, |
|
"loss": 0.9773, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 4.134328358208955, |
|
"grad_norm": 6.717421054840088, |
|
"learning_rate": 4.02432283029298e-05, |
|
"loss": 0.7255, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 4.149253731343284, |
|
"grad_norm": 4.847568511962891, |
|
"learning_rate": 4.0187949143173026e-05, |
|
"loss": 0.6345, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 4.164179104477612, |
|
"grad_norm": 3.6066529750823975, |
|
"learning_rate": 4.0132669983416254e-05, |
|
"loss": 0.4784, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 4.17910447761194, |
|
"grad_norm": 1.8522593975067139, |
|
"learning_rate": 4.007739082365948e-05, |
|
"loss": 0.4678, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 4.1940298507462686, |
|
"grad_norm": 7.491433620452881, |
|
"learning_rate": 4.002211166390271e-05, |
|
"loss": 0.6955, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 4.208955223880597, |
|
"grad_norm": 3.9711689949035645, |
|
"learning_rate": 3.996683250414594e-05, |
|
"loss": 0.7432, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 4.223880597014926, |
|
"grad_norm": 3.5602264404296875, |
|
"learning_rate": 3.991155334438917e-05, |
|
"loss": 0.6618, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 4.2388059701492535, |
|
"grad_norm": 3.383795976638794, |
|
"learning_rate": 3.9856274184632395e-05, |
|
"loss": 0.6079, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 4.253731343283582, |
|
"grad_norm": 2.320563554763794, |
|
"learning_rate": 3.980099502487562e-05, |
|
"loss": 0.5248, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 4.268656716417911, |
|
"grad_norm": 3.110429048538208, |
|
"learning_rate": 3.974571586511885e-05, |
|
"loss": 0.3517, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 4.2835820895522385, |
|
"grad_norm": 6.524468421936035, |
|
"learning_rate": 3.9690436705362086e-05, |
|
"loss": 0.6481, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 4.298507462686567, |
|
"grad_norm": 3.3383541107177734, |
|
"learning_rate": 3.9635157545605314e-05, |
|
"loss": 0.6297, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 4.313432835820896, |
|
"grad_norm": 1.686207890510559, |
|
"learning_rate": 3.9579878385848536e-05, |
|
"loss": 0.5142, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 4.3283582089552235, |
|
"grad_norm": 9.0740385055542, |
|
"learning_rate": 3.9524599226091764e-05, |
|
"loss": 0.478, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 4.343283582089552, |
|
"grad_norm": 8.83989143371582, |
|
"learning_rate": 3.946932006633499e-05, |
|
"loss": 0.7348, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 4.358208955223881, |
|
"grad_norm": 9.292248725891113, |
|
"learning_rate": 3.941404090657822e-05, |
|
"loss": 0.6026, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 4.373134328358209, |
|
"grad_norm": 4.79518461227417, |
|
"learning_rate": 3.935876174682145e-05, |
|
"loss": 0.818, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 4.388059701492537, |
|
"grad_norm": 6.292643070220947, |
|
"learning_rate": 3.9303482587064676e-05, |
|
"loss": 0.5522, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 4.402985074626866, |
|
"grad_norm": 4.075753688812256, |
|
"learning_rate": 3.9248203427307905e-05, |
|
"loss": 0.5259, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 4.417910447761194, |
|
"grad_norm": 2.985786199569702, |
|
"learning_rate": 3.919292426755113e-05, |
|
"loss": 0.4218, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 4.432835820895522, |
|
"grad_norm": 4.570261478424072, |
|
"learning_rate": 3.913764510779436e-05, |
|
"loss": 0.4735, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 4.447761194029851, |
|
"grad_norm": 2.8057727813720703, |
|
"learning_rate": 3.908236594803759e-05, |
|
"loss": 0.4241, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 4.462686567164179, |
|
"grad_norm": 7.44076681137085, |
|
"learning_rate": 3.902708678828082e-05, |
|
"loss": 0.4751, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 4.477611940298507, |
|
"grad_norm": 4.970695972442627, |
|
"learning_rate": 3.8971807628524045e-05, |
|
"loss": 0.5297, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 4.492537313432836, |
|
"grad_norm": 3.2911581993103027, |
|
"learning_rate": 3.891652846876728e-05, |
|
"loss": 0.4756, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 4.507462686567164, |
|
"grad_norm": 6.148906230926514, |
|
"learning_rate": 3.886124930901051e-05, |
|
"loss": 0.5251, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 4.522388059701493, |
|
"grad_norm": 4.234598636627197, |
|
"learning_rate": 3.8805970149253736e-05, |
|
"loss": 0.3342, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 4.537313432835821, |
|
"grad_norm": 4.786794662475586, |
|
"learning_rate": 3.8750690989496964e-05, |
|
"loss": 0.6528, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 4.552238805970149, |
|
"grad_norm": 8.322453498840332, |
|
"learning_rate": 3.869541182974019e-05, |
|
"loss": 0.5658, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 4.567164179104478, |
|
"grad_norm": 6.955239772796631, |
|
"learning_rate": 3.864013266998342e-05, |
|
"loss": 0.5496, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 4.582089552238806, |
|
"grad_norm": 4.89069938659668, |
|
"learning_rate": 3.858485351022665e-05, |
|
"loss": 0.4191, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 4.597014925373134, |
|
"grad_norm": 9.381403923034668, |
|
"learning_rate": 3.852957435046988e-05, |
|
"loss": 0.5993, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 4.611940298507463, |
|
"grad_norm": 4.977263450622559, |
|
"learning_rate": 3.8474295190713105e-05, |
|
"loss": 0.569, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 4.6268656716417915, |
|
"grad_norm": 3.183396816253662, |
|
"learning_rate": 3.841901603095633e-05, |
|
"loss": 0.5183, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 4.641791044776119, |
|
"grad_norm": 3.1676368713378906, |
|
"learning_rate": 3.8363736871199555e-05, |
|
"loss": 0.4647, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 4.656716417910448, |
|
"grad_norm": 9.969623565673828, |
|
"learning_rate": 3.830845771144278e-05, |
|
"loss": 0.4468, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 4.6716417910447765, |
|
"grad_norm": 11.773579597473145, |
|
"learning_rate": 3.825317855168601e-05, |
|
"loss": 0.477, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 4.686567164179104, |
|
"grad_norm": 3.361043691635132, |
|
"learning_rate": 3.8197899391929246e-05, |
|
"loss": 0.4533, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 4.701492537313433, |
|
"grad_norm": 6.481675148010254, |
|
"learning_rate": 3.8142620232172474e-05, |
|
"loss": 0.5998, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 4.7164179104477615, |
|
"grad_norm": 10.961061477661133, |
|
"learning_rate": 3.80873410724157e-05, |
|
"loss": 0.4246, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 4.731343283582089, |
|
"grad_norm": 11.879002571105957, |
|
"learning_rate": 3.803206191265893e-05, |
|
"loss": 0.4113, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 4.746268656716418, |
|
"grad_norm": 6.598226547241211, |
|
"learning_rate": 3.797678275290216e-05, |
|
"loss": 0.4986, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 4.7611940298507465, |
|
"grad_norm": 9.366608619689941, |
|
"learning_rate": 3.7921503593145386e-05, |
|
"loss": 0.3439, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 4.776119402985074, |
|
"grad_norm": 11.350964546203613, |
|
"learning_rate": 3.7866224433388615e-05, |
|
"loss": 0.8422, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 4.791044776119403, |
|
"grad_norm": 8.099052429199219, |
|
"learning_rate": 3.781094527363184e-05, |
|
"loss": 0.687, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 4.8059701492537314, |
|
"grad_norm": 5.031801223754883, |
|
"learning_rate": 3.775566611387507e-05, |
|
"loss": 0.484, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 4.82089552238806, |
|
"grad_norm": 3.5851833820343018, |
|
"learning_rate": 3.77003869541183e-05, |
|
"loss": 0.6037, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 4.835820895522388, |
|
"grad_norm": 14.049162864685059, |
|
"learning_rate": 3.764510779436153e-05, |
|
"loss": 0.5521, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 4.850746268656716, |
|
"grad_norm": 5.086363315582275, |
|
"learning_rate": 3.7589828634604755e-05, |
|
"loss": 0.5173, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 4.865671641791045, |
|
"grad_norm": 2.833794355392456, |
|
"learning_rate": 3.753454947484798e-05, |
|
"loss": 0.7762, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 4.880597014925373, |
|
"grad_norm": 2.7670750617980957, |
|
"learning_rate": 3.747927031509121e-05, |
|
"loss": 0.5592, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 4.895522388059701, |
|
"grad_norm": 2.6060335636138916, |
|
"learning_rate": 3.742399115533444e-05, |
|
"loss": 0.2484, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 4.91044776119403, |
|
"grad_norm": 5.139571666717529, |
|
"learning_rate": 3.736871199557767e-05, |
|
"loss": 0.5157, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 4.925373134328359, |
|
"grad_norm": 3.047414779663086, |
|
"learning_rate": 3.73134328358209e-05, |
|
"loss": 0.4967, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 4.940298507462686, |
|
"grad_norm": 7.455503940582275, |
|
"learning_rate": 3.725815367606413e-05, |
|
"loss": 0.4034, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 4.955223880597015, |
|
"grad_norm": 12.100015640258789, |
|
"learning_rate": 3.720287451630736e-05, |
|
"loss": 0.5245, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 4.970149253731344, |
|
"grad_norm": 3.1610209941864014, |
|
"learning_rate": 3.714759535655058e-05, |
|
"loss": 0.6381, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 4.985074626865671, |
|
"grad_norm": 3.781731367111206, |
|
"learning_rate": 3.709231619679381e-05, |
|
"loss": 0.5727, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"grad_norm": 1.7648042440414429, |
|
"learning_rate": 3.7037037037037037e-05, |
|
"loss": 0.3107, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_accuracy": 0.725, |
|
"eval_f1_macro": 0.67604038118744, |
|
"eval_f1_micro": 0.725, |
|
"eval_f1_weighted": 0.6927145770895771, |
|
"eval_loss": 1.0104252099990845, |
|
"eval_precision_macro": 0.7111726508785332, |
|
"eval_precision_micro": 0.725, |
|
"eval_precision_weighted": 0.7247546897546897, |
|
"eval_recall_macro": 0.7092436974789916, |
|
"eval_recall_micro": 0.725, |
|
"eval_recall_weighted": 0.725, |
|
"eval_runtime": 39.3596, |
|
"eval_samples_per_second": 5.081, |
|
"eval_steps_per_second": 0.178, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 5.014925373134329, |
|
"grad_norm": 1.6856744289398193, |
|
"learning_rate": 3.6981757877280265e-05, |
|
"loss": 0.289, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 5.029850746268656, |
|
"grad_norm": 2.8615660667419434, |
|
"learning_rate": 3.692647871752349e-05, |
|
"loss": 0.4746, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 5.044776119402985, |
|
"grad_norm": 6.883438587188721, |
|
"learning_rate": 3.687119955776672e-05, |
|
"loss": 0.6167, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 5.059701492537314, |
|
"grad_norm": 7.769782543182373, |
|
"learning_rate": 3.681592039800995e-05, |
|
"loss": 0.3471, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 5.074626865671641, |
|
"grad_norm": 3.5327107906341553, |
|
"learning_rate": 3.676064123825318e-05, |
|
"loss": 0.4453, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 5.08955223880597, |
|
"grad_norm": 3.9162540435791016, |
|
"learning_rate": 3.6705362078496405e-05, |
|
"loss": 0.4471, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 5.104477611940299, |
|
"grad_norm": 2.9827005863189697, |
|
"learning_rate": 3.6650082918739633e-05, |
|
"loss": 0.344, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 5.119402985074627, |
|
"grad_norm": 5.053178310394287, |
|
"learning_rate": 3.659480375898287e-05, |
|
"loss": 0.3973, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 5.134328358208955, |
|
"grad_norm": 5.1265716552734375, |
|
"learning_rate": 3.6539524599226097e-05, |
|
"loss": 0.3783, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 5.149253731343284, |
|
"grad_norm": 3.368522882461548, |
|
"learning_rate": 3.6484245439469325e-05, |
|
"loss": 0.5253, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 5.164179104477612, |
|
"grad_norm": 5.44252872467041, |
|
"learning_rate": 3.642896627971255e-05, |
|
"loss": 0.54, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 5.17910447761194, |
|
"grad_norm": 7.313326358795166, |
|
"learning_rate": 3.637368711995578e-05, |
|
"loss": 0.3734, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 5.1940298507462686, |
|
"grad_norm": 3.947737693786621, |
|
"learning_rate": 3.631840796019901e-05, |
|
"loss": 0.4204, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 5.208955223880597, |
|
"grad_norm": 2.3334567546844482, |
|
"learning_rate": 3.626312880044224e-05, |
|
"loss": 0.3622, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 5.223880597014926, |
|
"grad_norm": 4.052284240722656, |
|
"learning_rate": 3.6207849640685465e-05, |
|
"loss": 0.2778, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 5.2388059701492535, |
|
"grad_norm": 11.430601119995117, |
|
"learning_rate": 3.6152570480928693e-05, |
|
"loss": 0.6734, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 5.253731343283582, |
|
"grad_norm": 5.606107234954834, |
|
"learning_rate": 3.609729132117192e-05, |
|
"loss": 0.4028, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 5.268656716417911, |
|
"grad_norm": 13.239790916442871, |
|
"learning_rate": 3.604201216141515e-05, |
|
"loss": 0.6932, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 5.2835820895522385, |
|
"grad_norm": 2.727085590362549, |
|
"learning_rate": 3.598673300165838e-05, |
|
"loss": 0.6325, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 5.298507462686567, |
|
"grad_norm": 4.920370101928711, |
|
"learning_rate": 3.59314538419016e-05, |
|
"loss": 0.215, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 5.313432835820896, |
|
"grad_norm": 3.629782199859619, |
|
"learning_rate": 3.5876174682144834e-05, |
|
"loss": 0.5572, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 5.3283582089552235, |
|
"grad_norm": 5.6093668937683105, |
|
"learning_rate": 3.582089552238806e-05, |
|
"loss": 0.7121, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 5.343283582089552, |
|
"grad_norm": 6.661886215209961, |
|
"learning_rate": 3.576561636263129e-05, |
|
"loss": 0.4479, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 5.358208955223881, |
|
"grad_norm": 10.350639343261719, |
|
"learning_rate": 3.571033720287452e-05, |
|
"loss": 0.5578, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 5.373134328358209, |
|
"grad_norm": 6.086325168609619, |
|
"learning_rate": 3.565505804311775e-05, |
|
"loss": 0.4269, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 5.388059701492537, |
|
"grad_norm": 6.263159275054932, |
|
"learning_rate": 3.5599778883360975e-05, |
|
"loss": 0.5466, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 5.402985074626866, |
|
"grad_norm": 6.2910990715026855, |
|
"learning_rate": 3.55444997236042e-05, |
|
"loss": 0.3659, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 5.417910447761194, |
|
"grad_norm": 4.90096378326416, |
|
"learning_rate": 3.548922056384743e-05, |
|
"loss": 0.6204, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 5.432835820895522, |
|
"grad_norm": 3.6040308475494385, |
|
"learning_rate": 3.543394140409066e-05, |
|
"loss": 0.5589, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 5.447761194029851, |
|
"grad_norm": 14.020445823669434, |
|
"learning_rate": 3.537866224433389e-05, |
|
"loss": 0.5483, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 5.462686567164179, |
|
"grad_norm": 2.3177435398101807, |
|
"learning_rate": 3.5323383084577115e-05, |
|
"loss": 0.3787, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 5.477611940298507, |
|
"grad_norm": 2.891357183456421, |
|
"learning_rate": 3.5268103924820344e-05, |
|
"loss": 0.347, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 5.492537313432836, |
|
"grad_norm": 11.490731239318848, |
|
"learning_rate": 3.521282476506357e-05, |
|
"loss": 0.5475, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 5.507462686567164, |
|
"grad_norm": 2.3620495796203613, |
|
"learning_rate": 3.51575456053068e-05, |
|
"loss": 0.4033, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 5.522388059701493, |
|
"grad_norm": 3.9230000972747803, |
|
"learning_rate": 3.510226644555003e-05, |
|
"loss": 0.4246, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 5.537313432835821, |
|
"grad_norm": 3.205672025680542, |
|
"learning_rate": 3.5046987285793256e-05, |
|
"loss": 0.4462, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 5.552238805970149, |
|
"grad_norm": 5.685773849487305, |
|
"learning_rate": 3.499170812603649e-05, |
|
"loss": 0.379, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 5.567164179104478, |
|
"grad_norm": 5.982370376586914, |
|
"learning_rate": 3.493642896627972e-05, |
|
"loss": 0.4962, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 5.582089552238806, |
|
"grad_norm": 13.228385925292969, |
|
"learning_rate": 3.488114980652295e-05, |
|
"loss": 0.3345, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 5.597014925373134, |
|
"grad_norm": 5.765232086181641, |
|
"learning_rate": 3.4825870646766175e-05, |
|
"loss": 0.4973, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 5.611940298507463, |
|
"grad_norm": 1.701096534729004, |
|
"learning_rate": 3.4770591487009404e-05, |
|
"loss": 0.3608, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 5.6268656716417915, |
|
"grad_norm": 14.079666137695312, |
|
"learning_rate": 3.4715312327252625e-05, |
|
"loss": 0.5854, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 5.641791044776119, |
|
"grad_norm": 3.6387109756469727, |
|
"learning_rate": 3.466003316749585e-05, |
|
"loss": 0.2784, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 5.656716417910448, |
|
"grad_norm": 1.6772111654281616, |
|
"learning_rate": 3.460475400773908e-05, |
|
"loss": 0.2205, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 5.6716417910447765, |
|
"grad_norm": 5.0109405517578125, |
|
"learning_rate": 3.454947484798231e-05, |
|
"loss": 0.4189, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 5.686567164179104, |
|
"grad_norm": 4.438418865203857, |
|
"learning_rate": 3.449419568822554e-05, |
|
"loss": 0.1976, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 5.701492537313433, |
|
"grad_norm": 4.297623157501221, |
|
"learning_rate": 3.4438916528468766e-05, |
|
"loss": 0.3569, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 5.7164179104477615, |
|
"grad_norm": 7.163605213165283, |
|
"learning_rate": 3.4383637368711994e-05, |
|
"loss": 0.2972, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 5.731343283582089, |
|
"grad_norm": 3.6532185077667236, |
|
"learning_rate": 3.432835820895522e-05, |
|
"loss": 0.1953, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 5.746268656716418, |
|
"grad_norm": 10.143959045410156, |
|
"learning_rate": 3.427307904919845e-05, |
|
"loss": 0.4292, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 5.7611940298507465, |
|
"grad_norm": 7.471317768096924, |
|
"learning_rate": 3.4217799889441685e-05, |
|
"loss": 0.472, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 5.776119402985074, |
|
"grad_norm": 3.14628529548645, |
|
"learning_rate": 3.416252072968491e-05, |
|
"loss": 0.4863, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 5.791044776119403, |
|
"grad_norm": 2.2698168754577637, |
|
"learning_rate": 3.410724156992814e-05, |
|
"loss": 0.2093, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 5.8059701492537314, |
|
"grad_norm": 3.6853973865509033, |
|
"learning_rate": 3.405196241017137e-05, |
|
"loss": 0.5749, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 5.82089552238806, |
|
"grad_norm": 9.027390480041504, |
|
"learning_rate": 3.39966832504146e-05, |
|
"loss": 0.8224, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 5.835820895522388, |
|
"grad_norm": 4.245992660522461, |
|
"learning_rate": 3.3941404090657825e-05, |
|
"loss": 0.2923, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 5.850746268656716, |
|
"grad_norm": 10.722375869750977, |
|
"learning_rate": 3.3886124930901054e-05, |
|
"loss": 0.7797, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 5.865671641791045, |
|
"grad_norm": 4.297248363494873, |
|
"learning_rate": 3.383084577114428e-05, |
|
"loss": 0.5619, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 5.880597014925373, |
|
"grad_norm": 4.837419509887695, |
|
"learning_rate": 3.377556661138751e-05, |
|
"loss": 0.3356, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 5.895522388059701, |
|
"grad_norm": 1.2375975847244263, |
|
"learning_rate": 3.372028745163074e-05, |
|
"loss": 0.1544, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 5.91044776119403, |
|
"grad_norm": 12.756596565246582, |
|
"learning_rate": 3.3665008291873966e-05, |
|
"loss": 0.3809, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 5.925373134328359, |
|
"grad_norm": 3.4860129356384277, |
|
"learning_rate": 3.3609729132117194e-05, |
|
"loss": 0.3144, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 5.940298507462686, |
|
"grad_norm": 7.757724761962891, |
|
"learning_rate": 3.355444997236042e-05, |
|
"loss": 0.3695, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 5.955223880597015, |
|
"grad_norm": 2.4491782188415527, |
|
"learning_rate": 3.349917081260365e-05, |
|
"loss": 0.275, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 5.970149253731344, |
|
"grad_norm": 3.097658395767212, |
|
"learning_rate": 3.344389165284688e-05, |
|
"loss": 0.5223, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 5.985074626865671, |
|
"grad_norm": 3.9302120208740234, |
|
"learning_rate": 3.338861249309011e-05, |
|
"loss": 0.2202, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"grad_norm": 4.732783794403076, |
|
"learning_rate": 3.3333333333333335e-05, |
|
"loss": 0.4422, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_accuracy": 0.76, |
|
"eval_f1_macro": 0.7178212262554823, |
|
"eval_f1_micro": 0.76, |
|
"eval_f1_weighted": 0.7341365415630122, |
|
"eval_loss": 0.890067458152771, |
|
"eval_precision_macro": 0.7468954248366012, |
|
"eval_precision_micro": 0.76, |
|
"eval_precision_weighted": 0.766861111111111, |
|
"eval_recall_macro": 0.7448179271708684, |
|
"eval_recall_micro": 0.76, |
|
"eval_recall_weighted": 0.76, |
|
"eval_runtime": 37.7852, |
|
"eval_samples_per_second": 5.293, |
|
"eval_steps_per_second": 0.185, |
|
"step": 804 |
|
} |
|
], |
|
"logging_steps": 2, |
|
"max_steps": 2010, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 15, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.01 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 844829150146560.0, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|