| { | |
| "best_metric": 0.7045454545454546, | |
| "best_model_checkpoint": "squarerun/checkpoint-435", | |
| "epoch": 45.0, | |
| "eval_steps": 500, | |
| "global_step": 1305, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.06896551724137931, | |
| "grad_norm": 2.82375431060791, | |
| "learning_rate": 1.5267175572519084e-06, | |
| "loss": 1.9171, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.13793103448275862, | |
| "grad_norm": 1.8254162073135376, | |
| "learning_rate": 3.053435114503817e-06, | |
| "loss": 1.9196, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.20689655172413793, | |
| "grad_norm": 1.877601146697998, | |
| "learning_rate": 4.580152671755725e-06, | |
| "loss": 1.9126, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.27586206896551724, | |
| "grad_norm": 1.842034935951233, | |
| "learning_rate": 6.106870229007634e-06, | |
| "loss": 1.9303, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.3448275862068966, | |
| "grad_norm": 2.6617424488067627, | |
| "learning_rate": 7.633587786259543e-06, | |
| "loss": 1.9303, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.41379310344827586, | |
| "grad_norm": 1.9283453226089478, | |
| "learning_rate": 9.16030534351145e-06, | |
| "loss": 1.957, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.4827586206896552, | |
| "grad_norm": 1.7121816873550415, | |
| "learning_rate": 1.0687022900763359e-05, | |
| "loss": 1.9194, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.5517241379310345, | |
| "grad_norm": 1.5719634294509888, | |
| "learning_rate": 1.2213740458015267e-05, | |
| "loss": 1.9612, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.6206896551724138, | |
| "grad_norm": 1.857325553894043, | |
| "learning_rate": 1.3740458015267178e-05, | |
| "loss": 1.8854, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.6896551724137931, | |
| "grad_norm": 2.044182538986206, | |
| "learning_rate": 1.5267175572519086e-05, | |
| "loss": 1.9012, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.7586206896551724, | |
| "grad_norm": 1.0480695962905884, | |
| "learning_rate": 1.6793893129770993e-05, | |
| "loss": 1.8995, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.8275862068965517, | |
| "grad_norm": 1.2127116918563843, | |
| "learning_rate": 1.83206106870229e-05, | |
| "loss": 1.8923, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.896551724137931, | |
| "grad_norm": 2.0899465084075928, | |
| "learning_rate": 1.984732824427481e-05, | |
| "loss": 1.8761, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.9655172413793104, | |
| "grad_norm": 1.5972628593444824, | |
| "learning_rate": 2.1374045801526718e-05, | |
| "loss": 1.903, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_accuracy": 0.17424242424242425, | |
| "eval_f1_macro": 0.06584113560857748, | |
| "eval_f1_micro": 0.17424242424242425, | |
| "eval_f1_weighted": 0.09004420526619257, | |
| "eval_loss": 1.8867504596710205, | |
| "eval_precision_macro": 0.050235478806907374, | |
| "eval_precision_micro": 0.17424242424242425, | |
| "eval_precision_weighted": 0.0693056943056943, | |
| "eval_recall_macro": 0.1293121693121693, | |
| "eval_recall_micro": 0.17424242424242425, | |
| "eval_recall_weighted": 0.17424242424242425, | |
| "eval_runtime": 2.3295, | |
| "eval_samples_per_second": 56.664, | |
| "eval_steps_per_second": 7.298, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 1.0344827586206897, | |
| "grad_norm": 2.178473711013794, | |
| "learning_rate": 2.2900763358778628e-05, | |
| "loss": 1.891, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 1.103448275862069, | |
| "grad_norm": 1.646786093711853, | |
| "learning_rate": 2.4427480916030535e-05, | |
| "loss": 1.8301, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 1.1724137931034484, | |
| "grad_norm": 1.4257752895355225, | |
| "learning_rate": 2.5954198473282442e-05, | |
| "loss": 1.8346, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 1.2413793103448276, | |
| "grad_norm": 1.3215000629425049, | |
| "learning_rate": 2.7480916030534355e-05, | |
| "loss": 1.905, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 1.3103448275862069, | |
| "grad_norm": 1.763136625289917, | |
| "learning_rate": 2.900763358778626e-05, | |
| "loss": 1.8771, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 1.3793103448275863, | |
| "grad_norm": 1.7233166694641113, | |
| "learning_rate": 3.053435114503817e-05, | |
| "loss": 1.8977, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 1.4482758620689655, | |
| "grad_norm": 1.3865277767181396, | |
| "learning_rate": 3.2061068702290076e-05, | |
| "loss": 1.8815, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 1.5172413793103448, | |
| "grad_norm": 2.227480411529541, | |
| "learning_rate": 3.358778625954199e-05, | |
| "loss": 1.8876, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 1.5862068965517242, | |
| "grad_norm": 1.27533757686615, | |
| "learning_rate": 3.511450381679389e-05, | |
| "loss": 1.9272, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 1.6551724137931034, | |
| "grad_norm": 1.4631495475769043, | |
| "learning_rate": 3.66412213740458e-05, | |
| "loss": 1.8779, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 1.7241379310344827, | |
| "grad_norm": 1.4583027362823486, | |
| "learning_rate": 3.816793893129771e-05, | |
| "loss": 1.9327, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 1.793103448275862, | |
| "grad_norm": 1.856370210647583, | |
| "learning_rate": 3.969465648854962e-05, | |
| "loss": 1.84, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 1.8620689655172413, | |
| "grad_norm": 2.055604934692383, | |
| "learning_rate": 4.122137404580153e-05, | |
| "loss": 1.9563, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 1.9310344827586206, | |
| "grad_norm": 2.059593439102173, | |
| "learning_rate": 4.2748091603053435e-05, | |
| "loss": 1.8515, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 1.1090753078460693, | |
| "learning_rate": 4.4274809160305345e-05, | |
| "loss": 1.8662, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_accuracy": 0.2196969696969697, | |
| "eval_f1_macro": 0.07539682539682539, | |
| "eval_f1_micro": 0.2196969696969697, | |
| "eval_f1_weighted": 0.10037878787878787, | |
| "eval_loss": 1.8739551305770874, | |
| "eval_precision_macro": 0.06031746031746032, | |
| "eval_precision_micro": 0.2196969696969697, | |
| "eval_precision_weighted": 0.07727272727272727, | |
| "eval_recall_macro": 0.15797430083144368, | |
| "eval_recall_micro": 0.2196969696969697, | |
| "eval_recall_weighted": 0.2196969696969697, | |
| "eval_runtime": 2.3196, | |
| "eval_samples_per_second": 56.907, | |
| "eval_steps_per_second": 7.329, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 2.0689655172413794, | |
| "grad_norm": 1.6586414575576782, | |
| "learning_rate": 4.5801526717557256e-05, | |
| "loss": 1.9037, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 2.1379310344827585, | |
| "grad_norm": 1.2022308111190796, | |
| "learning_rate": 4.7328244274809166e-05, | |
| "loss": 1.8325, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 2.206896551724138, | |
| "grad_norm": 2.1608262062072754, | |
| "learning_rate": 4.885496183206107e-05, | |
| "loss": 1.8745, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 2.2758620689655173, | |
| "grad_norm": 2.147564172744751, | |
| "learning_rate": 5.038167938931297e-05, | |
| "loss": 1.9165, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 2.344827586206897, | |
| "grad_norm": 1.5065761804580688, | |
| "learning_rate": 5.1908396946564884e-05, | |
| "loss": 1.7814, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 2.413793103448276, | |
| "grad_norm": 1.8903454542160034, | |
| "learning_rate": 5.3435114503816794e-05, | |
| "loss": 1.8222, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 2.4827586206896552, | |
| "grad_norm": 1.8370472192764282, | |
| "learning_rate": 5.496183206106871e-05, | |
| "loss": 1.8853, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 2.5517241379310347, | |
| "grad_norm": 1.6297768354415894, | |
| "learning_rate": 5.648854961832062e-05, | |
| "loss": 1.8725, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 2.6206896551724137, | |
| "grad_norm": 1.4402002096176147, | |
| "learning_rate": 5.801526717557252e-05, | |
| "loss": 1.8109, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 2.689655172413793, | |
| "grad_norm": 2.1347367763519287, | |
| "learning_rate": 5.954198473282443e-05, | |
| "loss": 1.8819, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 2.7586206896551726, | |
| "grad_norm": 1.732499599456787, | |
| "learning_rate": 6.106870229007635e-05, | |
| "loss": 1.9278, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 2.8275862068965516, | |
| "grad_norm": 1.4573041200637817, | |
| "learning_rate": 6.259541984732826e-05, | |
| "loss": 1.8427, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 2.896551724137931, | |
| "grad_norm": 1.4883620738983154, | |
| "learning_rate": 6.412213740458015e-05, | |
| "loss": 1.9268, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 2.9655172413793105, | |
| "grad_norm": 1.8515830039978027, | |
| "learning_rate": 6.564885496183206e-05, | |
| "loss": 1.9291, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_accuracy": 0.20454545454545456, | |
| "eval_f1_macro": 0.04851752021563342, | |
| "eval_f1_micro": 0.20454545454545456, | |
| "eval_f1_weighted": 0.06946826758147512, | |
| "eval_loss": 1.8862061500549316, | |
| "eval_precision_macro": 0.029220779220779224, | |
| "eval_precision_micro": 0.20454545454545456, | |
| "eval_precision_weighted": 0.04183884297520662, | |
| "eval_recall_macro": 0.14285714285714285, | |
| "eval_recall_micro": 0.20454545454545456, | |
| "eval_recall_weighted": 0.20454545454545456, | |
| "eval_runtime": 2.3118, | |
| "eval_samples_per_second": 57.1, | |
| "eval_steps_per_second": 7.354, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 3.0344827586206895, | |
| "grad_norm": 2.0376968383789062, | |
| "learning_rate": 6.717557251908397e-05, | |
| "loss": 1.9533, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 3.103448275862069, | |
| "grad_norm": 1.5388916730880737, | |
| "learning_rate": 6.870229007633588e-05, | |
| "loss": 1.8628, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 3.1724137931034484, | |
| "grad_norm": 1.5677577257156372, | |
| "learning_rate": 7.022900763358778e-05, | |
| "loss": 1.9064, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 3.2413793103448274, | |
| "grad_norm": 2.0428049564361572, | |
| "learning_rate": 7.175572519083969e-05, | |
| "loss": 1.8747, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 3.310344827586207, | |
| "grad_norm": 2.1180825233459473, | |
| "learning_rate": 7.32824427480916e-05, | |
| "loss": 1.8842, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 3.3793103448275863, | |
| "grad_norm": 1.6793889999389648, | |
| "learning_rate": 7.480916030534351e-05, | |
| "loss": 1.8041, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 3.4482758620689653, | |
| "grad_norm": 1.4385968446731567, | |
| "learning_rate": 7.633587786259542e-05, | |
| "loss": 1.7604, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 3.5172413793103448, | |
| "grad_norm": 2.31620717048645, | |
| "learning_rate": 7.786259541984733e-05, | |
| "loss": 1.8775, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 3.586206896551724, | |
| "grad_norm": 1.4493972063064575, | |
| "learning_rate": 7.938931297709924e-05, | |
| "loss": 1.7543, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 3.655172413793103, | |
| "grad_norm": 1.9932034015655518, | |
| "learning_rate": 8.091603053435115e-05, | |
| "loss": 1.8983, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 3.7241379310344827, | |
| "grad_norm": 1.7643533945083618, | |
| "learning_rate": 8.244274809160306e-05, | |
| "loss": 1.8041, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 3.793103448275862, | |
| "grad_norm": 2.046786069869995, | |
| "learning_rate": 8.396946564885496e-05, | |
| "loss": 1.8857, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 3.862068965517241, | |
| "grad_norm": 2.543879985809326, | |
| "learning_rate": 8.549618320610687e-05, | |
| "loss": 1.9249, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 3.9310344827586206, | |
| "grad_norm": 2.2258007526397705, | |
| "learning_rate": 8.702290076335878e-05, | |
| "loss": 1.8182, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "grad_norm": 1.7140247821807861, | |
| "learning_rate": 8.854961832061069e-05, | |
| "loss": 1.7838, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_accuracy": 0.26515151515151514, | |
| "eval_f1_macro": 0.11706349206349206, | |
| "eval_f1_micro": 0.26515151515151514, | |
| "eval_f1_weighted": 0.14741161616161616, | |
| "eval_loss": 1.8126838207244873, | |
| "eval_precision_macro": 0.10915750915750914, | |
| "eval_precision_micro": 0.26515151515151514, | |
| "eval_precision_weighted": 0.13205128205128203, | |
| "eval_recall_macro": 0.19727891156462585, | |
| "eval_recall_micro": 0.26515151515151514, | |
| "eval_recall_weighted": 0.26515151515151514, | |
| "eval_runtime": 2.3137, | |
| "eval_samples_per_second": 57.052, | |
| "eval_steps_per_second": 7.348, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 4.068965517241379, | |
| "grad_norm": 2.673126220703125, | |
| "learning_rate": 9.007633587786259e-05, | |
| "loss": 1.7886, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 4.137931034482759, | |
| "grad_norm": 1.7165743112564087, | |
| "learning_rate": 9.160305343511451e-05, | |
| "loss": 1.7106, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 4.206896551724138, | |
| "grad_norm": 1.7721195220947266, | |
| "learning_rate": 9.312977099236642e-05, | |
| "loss": 1.7189, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 4.275862068965517, | |
| "grad_norm": 2.3345961570739746, | |
| "learning_rate": 9.465648854961833e-05, | |
| "loss": 1.6691, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 4.344827586206897, | |
| "grad_norm": 4.940219402313232, | |
| "learning_rate": 9.618320610687024e-05, | |
| "loss": 1.661, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 4.413793103448276, | |
| "grad_norm": 1.7643016576766968, | |
| "learning_rate": 9.770992366412214e-05, | |
| "loss": 1.6656, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 4.482758620689655, | |
| "grad_norm": 2.4930827617645264, | |
| "learning_rate": 9.923664122137405e-05, | |
| "loss": 1.7436, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 4.551724137931035, | |
| "grad_norm": 3.006988763809204, | |
| "learning_rate": 9.991482112436117e-05, | |
| "loss": 1.7548, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 4.620689655172414, | |
| "grad_norm": 2.7570700645446777, | |
| "learning_rate": 9.974446337308349e-05, | |
| "loss": 1.8445, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 4.689655172413794, | |
| "grad_norm": 3.1985325813293457, | |
| "learning_rate": 9.957410562180581e-05, | |
| "loss": 1.837, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 4.758620689655173, | |
| "grad_norm": 1.967631220817566, | |
| "learning_rate": 9.940374787052812e-05, | |
| "loss": 1.7548, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 4.827586206896552, | |
| "grad_norm": 2.0813746452331543, | |
| "learning_rate": 9.923339011925044e-05, | |
| "loss": 1.7141, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 4.896551724137931, | |
| "grad_norm": 1.9808541536331177, | |
| "learning_rate": 9.906303236797275e-05, | |
| "loss": 1.6873, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 4.9655172413793105, | |
| "grad_norm": 1.347161889076233, | |
| "learning_rate": 9.889267461669507e-05, | |
| "loss": 1.7113, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_accuracy": 0.3484848484848485, | |
| "eval_f1_macro": 0.2133114591830188, | |
| "eval_f1_micro": 0.3484848484848485, | |
| "eval_f1_weighted": 0.2592243730692438, | |
| "eval_loss": 1.6978838443756104, | |
| "eval_precision_macro": 0.31892757102841135, | |
| "eval_precision_micro": 0.3484848484848485, | |
| "eval_precision_weighted": 0.363105848399966, | |
| "eval_recall_macro": 0.28217687074829934, | |
| "eval_recall_micro": 0.3484848484848485, | |
| "eval_recall_weighted": 0.3484848484848485, | |
| "eval_runtime": 2.3205, | |
| "eval_samples_per_second": 56.883, | |
| "eval_steps_per_second": 7.326, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 5.0344827586206895, | |
| "grad_norm": 2.642209529876709, | |
| "learning_rate": 9.872231686541739e-05, | |
| "loss": 1.9708, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 5.103448275862069, | |
| "grad_norm": 2.9190890789031982, | |
| "learning_rate": 9.85519591141397e-05, | |
| "loss": 1.6693, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 5.172413793103448, | |
| "grad_norm": 2.941225290298462, | |
| "learning_rate": 9.838160136286202e-05, | |
| "loss": 1.6507, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 5.241379310344827, | |
| "grad_norm": 2.5989174842834473, | |
| "learning_rate": 9.821124361158434e-05, | |
| "loss": 1.4749, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 5.310344827586207, | |
| "grad_norm": 2.6868557929992676, | |
| "learning_rate": 9.804088586030665e-05, | |
| "loss": 1.5171, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 5.379310344827586, | |
| "grad_norm": 2.383096218109131, | |
| "learning_rate": 9.787052810902897e-05, | |
| "loss": 1.5249, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 5.448275862068965, | |
| "grad_norm": 2.130948543548584, | |
| "learning_rate": 9.770017035775129e-05, | |
| "loss": 1.5954, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 5.517241379310345, | |
| "grad_norm": 3.0674099922180176, | |
| "learning_rate": 9.75298126064736e-05, | |
| "loss": 1.4673, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 5.586206896551724, | |
| "grad_norm": 2.2742505073547363, | |
| "learning_rate": 9.735945485519592e-05, | |
| "loss": 1.6837, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 5.655172413793103, | |
| "grad_norm": 2.3131611347198486, | |
| "learning_rate": 9.718909710391823e-05, | |
| "loss": 1.4481, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 5.724137931034483, | |
| "grad_norm": 2.722858190536499, | |
| "learning_rate": 9.701873935264055e-05, | |
| "loss": 1.4517, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 5.793103448275862, | |
| "grad_norm": 2.8864994049072266, | |
| "learning_rate": 9.684838160136287e-05, | |
| "loss": 1.4883, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 5.862068965517241, | |
| "grad_norm": 3.632519483566284, | |
| "learning_rate": 9.667802385008518e-05, | |
| "loss": 1.4962, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 5.931034482758621, | |
| "grad_norm": 3.605977773666382, | |
| "learning_rate": 9.65076660988075e-05, | |
| "loss": 1.6613, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "grad_norm": 2.918729305267334, | |
| "learning_rate": 9.633730834752982e-05, | |
| "loss": 1.6459, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_accuracy": 0.3939393939393939, | |
| "eval_f1_macro": 0.2713789610398824, | |
| "eval_f1_micro": 0.3939393939393939, | |
| "eval_f1_weighted": 0.3224658104540408, | |
| "eval_loss": 1.5577471256256104, | |
| "eval_precision_macro": 0.4295932138037401, | |
| "eval_precision_micro": 0.3939393939393939, | |
| "eval_precision_weighted": 0.4530558622663886, | |
| "eval_recall_macro": 0.3197883597883598, | |
| "eval_recall_micro": 0.3939393939393939, | |
| "eval_recall_weighted": 0.3939393939393939, | |
| "eval_runtime": 2.3137, | |
| "eval_samples_per_second": 57.052, | |
| "eval_steps_per_second": 7.348, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 6.068965517241379, | |
| "grad_norm": 3.8798978328704834, | |
| "learning_rate": 9.616695059625213e-05, | |
| "loss": 1.349, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 6.137931034482759, | |
| "grad_norm": 2.2578699588775635, | |
| "learning_rate": 9.599659284497445e-05, | |
| "loss": 1.3423, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 6.206896551724138, | |
| "grad_norm": 3.3518385887145996, | |
| "learning_rate": 9.582623509369678e-05, | |
| "loss": 1.4951, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 6.275862068965517, | |
| "grad_norm": 2.808967351913452, | |
| "learning_rate": 9.565587734241908e-05, | |
| "loss": 1.601, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 6.344827586206897, | |
| "grad_norm": 2.637211322784424, | |
| "learning_rate": 9.54855195911414e-05, | |
| "loss": 1.5365, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 6.413793103448276, | |
| "grad_norm": 3.144742727279663, | |
| "learning_rate": 9.531516183986373e-05, | |
| "loss": 1.3149, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 6.482758620689655, | |
| "grad_norm": 3.037153482437134, | |
| "learning_rate": 9.514480408858603e-05, | |
| "loss": 1.3832, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 6.551724137931035, | |
| "grad_norm": 2.0758683681488037, | |
| "learning_rate": 9.497444633730836e-05, | |
| "loss": 1.3602, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 6.620689655172414, | |
| "grad_norm": 2.5833723545074463, | |
| "learning_rate": 9.480408858603066e-05, | |
| "loss": 1.4422, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 6.689655172413794, | |
| "grad_norm": 2.2444944381713867, | |
| "learning_rate": 9.463373083475299e-05, | |
| "loss": 1.3833, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 6.758620689655173, | |
| "grad_norm": 2.526270627975464, | |
| "learning_rate": 9.446337308347531e-05, | |
| "loss": 1.6393, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 6.827586206896552, | |
| "grad_norm": 2.5092673301696777, | |
| "learning_rate": 9.429301533219762e-05, | |
| "loss": 1.3183, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 6.896551724137931, | |
| "grad_norm": 2.4997923374176025, | |
| "learning_rate": 9.412265758091994e-05, | |
| "loss": 1.1796, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 6.9655172413793105, | |
| "grad_norm": 2.6789186000823975, | |
| "learning_rate": 9.395229982964226e-05, | |
| "loss": 1.4829, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_accuracy": 0.5227272727272727, | |
| "eval_f1_macro": 0.4068715903691273, | |
| "eval_f1_micro": 0.5227272727272727, | |
| "eval_f1_weighted": 0.46112149344907966, | |
| "eval_loss": 1.381365180015564, | |
| "eval_precision_macro": 0.37857508594835787, | |
| "eval_precision_micro": 0.5227272727272727, | |
| "eval_precision_weighted": 0.42157170786203046, | |
| "eval_recall_macro": 0.4510506424792139, | |
| "eval_recall_micro": 0.5227272727272727, | |
| "eval_recall_weighted": 0.5227272727272727, | |
| "eval_runtime": 2.3191, | |
| "eval_samples_per_second": 56.919, | |
| "eval_steps_per_second": 7.331, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 7.0344827586206895, | |
| "grad_norm": 2.8184642791748047, | |
| "learning_rate": 9.378194207836457e-05, | |
| "loss": 1.0758, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 7.103448275862069, | |
| "grad_norm": 3.6033294200897217, | |
| "learning_rate": 9.361158432708689e-05, | |
| "loss": 1.2313, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 7.172413793103448, | |
| "grad_norm": 3.935368299484253, | |
| "learning_rate": 9.344122657580921e-05, | |
| "loss": 1.5238, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 7.241379310344827, | |
| "grad_norm": 2.5644354820251465, | |
| "learning_rate": 9.327086882453152e-05, | |
| "loss": 1.4117, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 7.310344827586207, | |
| "grad_norm": 3.393187999725342, | |
| "learning_rate": 9.310051107325384e-05, | |
| "loss": 1.5217, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 7.379310344827586, | |
| "grad_norm": 2.552579879760742, | |
| "learning_rate": 9.293015332197616e-05, | |
| "loss": 1.145, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 7.448275862068965, | |
| "grad_norm": 2.249608039855957, | |
| "learning_rate": 9.275979557069847e-05, | |
| "loss": 1.3357, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 7.517241379310345, | |
| "grad_norm": 2.600088357925415, | |
| "learning_rate": 9.258943781942079e-05, | |
| "loss": 1.1833, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 7.586206896551724, | |
| "grad_norm": 2.311519145965576, | |
| "learning_rate": 9.24190800681431e-05, | |
| "loss": 1.0488, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 7.655172413793103, | |
| "grad_norm": 2.6178081035614014, | |
| "learning_rate": 9.224872231686542e-05, | |
| "loss": 1.2106, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 7.724137931034483, | |
| "grad_norm": 3.762129545211792, | |
| "learning_rate": 9.207836456558774e-05, | |
| "loss": 1.1438, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 7.793103448275862, | |
| "grad_norm": 3.4986026287078857, | |
| "learning_rate": 9.190800681431005e-05, | |
| "loss": 1.4612, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 7.862068965517241, | |
| "grad_norm": 2.58974552154541, | |
| "learning_rate": 9.173764906303237e-05, | |
| "loss": 0.9742, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 7.931034482758621, | |
| "grad_norm": 2.3729593753814697, | |
| "learning_rate": 9.156729131175469e-05, | |
| "loss": 1.3228, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "grad_norm": 2.6781985759735107, | |
| "learning_rate": 9.1396933560477e-05, | |
| "loss": 1.2847, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_accuracy": 0.45454545454545453, | |
| "eval_f1_macro": 0.3675035848097526, | |
| "eval_f1_micro": 0.45454545454545453, | |
| "eval_f1_weighted": 0.4175505688962959, | |
| "eval_loss": 1.3783015012741089, | |
| "eval_precision_macro": 0.49916764667595237, | |
| "eval_precision_micro": 0.45454545454545453, | |
| "eval_precision_weighted": 0.570213988501515, | |
| "eval_recall_macro": 0.4079591836734694, | |
| "eval_recall_micro": 0.45454545454545453, | |
| "eval_recall_weighted": 0.45454545454545453, | |
| "eval_runtime": 2.4636, | |
| "eval_samples_per_second": 53.581, | |
| "eval_steps_per_second": 6.901, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 8.068965517241379, | |
| "grad_norm": 3.757275104522705, | |
| "learning_rate": 9.122657580919932e-05, | |
| "loss": 1.0559, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 8.137931034482758, | |
| "grad_norm": 2.25482439994812, | |
| "learning_rate": 9.105621805792164e-05, | |
| "loss": 1.181, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 8.206896551724139, | |
| "grad_norm": 3.7089550495147705, | |
| "learning_rate": 9.088586030664395e-05, | |
| "loss": 1.3496, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 8.275862068965518, | |
| "grad_norm": 2.729207754135132, | |
| "learning_rate": 9.071550255536627e-05, | |
| "loss": 1.1984, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 8.344827586206897, | |
| "grad_norm": 2.9847185611724854, | |
| "learning_rate": 9.054514480408858e-05, | |
| "loss": 1.1024, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 8.413793103448276, | |
| "grad_norm": 3.846353530883789, | |
| "learning_rate": 9.03747870528109e-05, | |
| "loss": 1.1308, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 8.482758620689655, | |
| "grad_norm": 3.132477283477783, | |
| "learning_rate": 9.020442930153323e-05, | |
| "loss": 1.2066, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 8.551724137931034, | |
| "grad_norm": 3.4364778995513916, | |
| "learning_rate": 9.003407155025553e-05, | |
| "loss": 1.1284, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 8.620689655172415, | |
| "grad_norm": 3.002615451812744, | |
| "learning_rate": 8.986371379897785e-05, | |
| "loss": 1.0977, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 8.689655172413794, | |
| "grad_norm": 2.997836112976074, | |
| "learning_rate": 8.969335604770018e-05, | |
| "loss": 0.977, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 8.758620689655173, | |
| "grad_norm": 4.079631805419922, | |
| "learning_rate": 8.952299829642248e-05, | |
| "loss": 1.3108, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 8.827586206896552, | |
| "grad_norm": 4.065947532653809, | |
| "learning_rate": 8.93526405451448e-05, | |
| "loss": 1.1677, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 8.89655172413793, | |
| "grad_norm": 4.696402549743652, | |
| "learning_rate": 8.918228279386713e-05, | |
| "loss": 1.1538, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 8.96551724137931, | |
| "grad_norm": 2.8291256427764893, | |
| "learning_rate": 8.901192504258944e-05, | |
| "loss": 0.7746, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "eval_accuracy": 0.5757575757575758, | |
| "eval_f1_macro": 0.4579442109215706, | |
| "eval_f1_micro": 0.5757575757575758, | |
| "eval_f1_weighted": 0.5297653032901569, | |
| "eval_loss": 1.1536388397216797, | |
| "eval_precision_macro": 0.5301086941953815, | |
| "eval_precision_micro": 0.5757575757575758, | |
| "eval_precision_weighted": 0.589626978576458, | |
| "eval_recall_macro": 0.4852532123960695, | |
| "eval_recall_micro": 0.5757575757575758, | |
| "eval_recall_weighted": 0.5757575757575758, | |
| "eval_runtime": 2.3555, | |
| "eval_samples_per_second": 56.04, | |
| "eval_steps_per_second": 7.217, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 9.03448275862069, | |
| "grad_norm": 3.098844289779663, | |
| "learning_rate": 8.884156729131176e-05, | |
| "loss": 1.0718, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 9.10344827586207, | |
| "grad_norm": 2.9349565505981445, | |
| "learning_rate": 8.867120954003408e-05, | |
| "loss": 0.9468, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 9.172413793103448, | |
| "grad_norm": 4.122389316558838, | |
| "learning_rate": 8.850085178875639e-05, | |
| "loss": 1.1746, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 9.241379310344827, | |
| "grad_norm": 2.8826913833618164, | |
| "learning_rate": 8.833049403747871e-05, | |
| "loss": 0.8535, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 9.310344827586206, | |
| "grad_norm": 3.3379621505737305, | |
| "learning_rate": 8.816013628620102e-05, | |
| "loss": 0.9954, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 9.379310344827585, | |
| "grad_norm": 5.267432689666748, | |
| "learning_rate": 8.798977853492334e-05, | |
| "loss": 0.9242, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 9.448275862068966, | |
| "grad_norm": 3.121428966522217, | |
| "learning_rate": 8.781942078364566e-05, | |
| "loss": 1.0053, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 9.517241379310345, | |
| "grad_norm": 3.526177167892456, | |
| "learning_rate": 8.764906303236797e-05, | |
| "loss": 0.8233, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 9.586206896551724, | |
| "grad_norm": 2.7245631217956543, | |
| "learning_rate": 8.747870528109029e-05, | |
| "loss": 0.9055, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 9.655172413793103, | |
| "grad_norm": 1.7660030126571655, | |
| "learning_rate": 8.730834752981261e-05, | |
| "loss": 0.5839, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 9.724137931034482, | |
| "grad_norm": 3.6391730308532715, | |
| "learning_rate": 8.713798977853492e-05, | |
| "loss": 1.1208, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 9.793103448275861, | |
| "grad_norm": 4.08744478225708, | |
| "learning_rate": 8.696763202725724e-05, | |
| "loss": 0.8893, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 9.862068965517242, | |
| "grad_norm": 4.032042503356934, | |
| "learning_rate": 8.679727427597956e-05, | |
| "loss": 0.973, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 9.931034482758621, | |
| "grad_norm": 5.962705135345459, | |
| "learning_rate": 8.662691652470187e-05, | |
| "loss": 1.0092, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "grad_norm": 5.076080799102783, | |
| "learning_rate": 8.645655877342419e-05, | |
| "loss": 1.0172, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_accuracy": 0.5909090909090909, | |
| "eval_f1_macro": 0.47002148227712137, | |
| "eval_f1_micro": 0.5909090909090909, | |
| "eval_f1_weighted": 0.5365472393103972, | |
| "eval_loss": 1.221096396446228, | |
| "eval_precision_macro": 0.572196223316913, | |
| "eval_precision_micro": 0.5909090909090909, | |
| "eval_precision_weighted": 0.6399457941483804, | |
| "eval_recall_macro": 0.518238851095994, | |
| "eval_recall_micro": 0.5909090909090909, | |
| "eval_recall_weighted": 0.5909090909090909, | |
| "eval_runtime": 2.3184, | |
| "eval_samples_per_second": 56.936, | |
| "eval_steps_per_second": 7.333, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 10.068965517241379, | |
| "grad_norm": 4.521097183227539, | |
| "learning_rate": 8.628620102214651e-05, | |
| "loss": 0.9744, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 10.137931034482758, | |
| "grad_norm": 2.6411921977996826, | |
| "learning_rate": 8.611584327086882e-05, | |
| "loss": 0.7116, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 10.206896551724139, | |
| "grad_norm": 4.236311912536621, | |
| "learning_rate": 8.594548551959114e-05, | |
| "loss": 1.016, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 10.275862068965518, | |
| "grad_norm": 4.348973751068115, | |
| "learning_rate": 8.577512776831345e-05, | |
| "loss": 0.7783, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 10.344827586206897, | |
| "grad_norm": 2.6386284828186035, | |
| "learning_rate": 8.560477001703579e-05, | |
| "loss": 1.0324, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 10.413793103448276, | |
| "grad_norm": 3.886791706085205, | |
| "learning_rate": 8.543441226575811e-05, | |
| "loss": 0.8633, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 10.482758620689655, | |
| "grad_norm": 5.126855373382568, | |
| "learning_rate": 8.526405451448042e-05, | |
| "loss": 0.9599, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 10.551724137931034, | |
| "grad_norm": 3.6911346912384033, | |
| "learning_rate": 8.509369676320274e-05, | |
| "loss": 1.0509, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 10.620689655172415, | |
| "grad_norm": 3.802255392074585, | |
| "learning_rate": 8.492333901192505e-05, | |
| "loss": 0.8213, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 10.689655172413794, | |
| "grad_norm": 5.074642181396484, | |
| "learning_rate": 8.475298126064737e-05, | |
| "loss": 1.0766, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 10.758620689655173, | |
| "grad_norm": 4.2655720710754395, | |
| "learning_rate": 8.458262350936969e-05, | |
| "loss": 0.9113, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 10.827586206896552, | |
| "grad_norm": 3.382858991622925, | |
| "learning_rate": 8.4412265758092e-05, | |
| "loss": 0.7092, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 10.89655172413793, | |
| "grad_norm": 4.771972179412842, | |
| "learning_rate": 8.424190800681432e-05, | |
| "loss": 0.547, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 10.96551724137931, | |
| "grad_norm": 4.0019378662109375, | |
| "learning_rate": 8.407155025553664e-05, | |
| "loss": 0.7865, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "eval_accuracy": 0.6136363636363636, | |
| "eval_f1_macro": 0.5281776060910356, | |
| "eval_f1_micro": 0.6136363636363636, | |
| "eval_f1_weighted": 0.596118348928358, | |
| "eval_loss": 1.1357340812683105, | |
| "eval_precision_macro": 0.5342261904761905, | |
| "eval_precision_micro": 0.6136363636363636, | |
| "eval_precision_weighted": 0.6009311868686869, | |
| "eval_recall_macro": 0.5431594860166289, | |
| "eval_recall_micro": 0.6136363636363636, | |
| "eval_recall_weighted": 0.6136363636363636, | |
| "eval_runtime": 2.3136, | |
| "eval_samples_per_second": 57.054, | |
| "eval_steps_per_second": 7.348, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 11.03448275862069, | |
| "grad_norm": 3.1186282634735107, | |
| "learning_rate": 8.390119250425895e-05, | |
| "loss": 0.6251, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 11.10344827586207, | |
| "grad_norm": 4.061969757080078, | |
| "learning_rate": 8.373083475298127e-05, | |
| "loss": 0.709, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 11.172413793103448, | |
| "grad_norm": 3.080505132675171, | |
| "learning_rate": 8.356047700170359e-05, | |
| "loss": 0.6536, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 11.241379310344827, | |
| "grad_norm": 4.702264785766602, | |
| "learning_rate": 8.33901192504259e-05, | |
| "loss": 0.764, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 11.310344827586206, | |
| "grad_norm": 3.9699347019195557, | |
| "learning_rate": 8.321976149914822e-05, | |
| "loss": 0.7632, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 11.379310344827585, | |
| "grad_norm": 3.679388999938965, | |
| "learning_rate": 8.304940374787054e-05, | |
| "loss": 1.0315, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 11.448275862068966, | |
| "grad_norm": 3.994912624359131, | |
| "learning_rate": 8.287904599659285e-05, | |
| "loss": 1.0187, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 11.517241379310345, | |
| "grad_norm": 5.398199558258057, | |
| "learning_rate": 8.270868824531517e-05, | |
| "loss": 0.5242, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 11.586206896551724, | |
| "grad_norm": 3.0399880409240723, | |
| "learning_rate": 8.253833049403748e-05, | |
| "loss": 0.7834, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 11.655172413793103, | |
| "grad_norm": 3.8799502849578857, | |
| "learning_rate": 8.23679727427598e-05, | |
| "loss": 0.7336, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 11.724137931034482, | |
| "grad_norm": 2.1970696449279785, | |
| "learning_rate": 8.219761499148212e-05, | |
| "loss": 0.6121, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 11.793103448275861, | |
| "grad_norm": 4.472761631011963, | |
| "learning_rate": 8.202725724020443e-05, | |
| "loss": 0.8251, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 11.862068965517242, | |
| "grad_norm": 2.3058536052703857, | |
| "learning_rate": 8.185689948892675e-05, | |
| "loss": 0.7287, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 11.931034482758621, | |
| "grad_norm": 4.462862014770508, | |
| "learning_rate": 8.168654173764907e-05, | |
| "loss": 0.7826, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "grad_norm": 4.2772440910339355, | |
| "learning_rate": 8.151618398637138e-05, | |
| "loss": 0.8335, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "eval_accuracy": 0.6060606060606061, | |
| "eval_f1_macro": 0.5314848082363927, | |
| "eval_f1_micro": 0.6060606060606061, | |
| "eval_f1_weighted": 0.6017190744361463, | |
| "eval_loss": 1.152970552444458, | |
| "eval_precision_macro": 0.536459399813437, | |
| "eval_precision_micro": 0.6060606060606061, | |
| "eval_precision_weighted": 0.6208833437094308, | |
| "eval_recall_macro": 0.5489417989417988, | |
| "eval_recall_micro": 0.6060606060606061, | |
| "eval_recall_weighted": 0.6060606060606061, | |
| "eval_runtime": 2.3168, | |
| "eval_samples_per_second": 56.976, | |
| "eval_steps_per_second": 7.338, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 12.068965517241379, | |
| "grad_norm": 3.217472791671753, | |
| "learning_rate": 8.13458262350937e-05, | |
| "loss": 0.6884, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 12.137931034482758, | |
| "grad_norm": 3.676238536834717, | |
| "learning_rate": 8.117546848381603e-05, | |
| "loss": 0.6886, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 12.206896551724139, | |
| "grad_norm": 3.851684093475342, | |
| "learning_rate": 8.100511073253833e-05, | |
| "loss": 0.8202, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 12.275862068965518, | |
| "grad_norm": 3.2188940048217773, | |
| "learning_rate": 8.083475298126066e-05, | |
| "loss": 0.6026, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 12.344827586206897, | |
| "grad_norm": 3.6800220012664795, | |
| "learning_rate": 8.066439522998296e-05, | |
| "loss": 0.5762, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 12.413793103448276, | |
| "grad_norm": 4.217052459716797, | |
| "learning_rate": 8.049403747870528e-05, | |
| "loss": 0.824, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 12.482758620689655, | |
| "grad_norm": 4.007477760314941, | |
| "learning_rate": 8.03236797274276e-05, | |
| "loss": 0.5009, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 12.551724137931034, | |
| "grad_norm": 3.780930995941162, | |
| "learning_rate": 8.015332197614991e-05, | |
| "loss": 0.6427, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 12.620689655172415, | |
| "grad_norm": 3.340984582901001, | |
| "learning_rate": 7.998296422487224e-05, | |
| "loss": 0.6936, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 12.689655172413794, | |
| "grad_norm": 6.2322258949279785, | |
| "learning_rate": 7.981260647359456e-05, | |
| "loss": 0.7349, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 12.758620689655173, | |
| "grad_norm": 4.197908878326416, | |
| "learning_rate": 7.964224872231687e-05, | |
| "loss": 1.0689, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 12.827586206896552, | |
| "grad_norm": 3.51385498046875, | |
| "learning_rate": 7.947189097103919e-05, | |
| "loss": 0.4522, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 12.89655172413793, | |
| "grad_norm": 4.907188415527344, | |
| "learning_rate": 7.930153321976151e-05, | |
| "loss": 0.6371, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 12.96551724137931, | |
| "grad_norm": 4.427891731262207, | |
| "learning_rate": 7.913117546848382e-05, | |
| "loss": 0.6959, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "eval_accuracy": 0.6666666666666666, | |
| "eval_f1_macro": 0.5637755814516842, | |
| "eval_f1_micro": 0.6666666666666666, | |
| "eval_f1_weighted": 0.6451496154313497, | |
| "eval_loss": 1.1307237148284912, | |
| "eval_precision_macro": 0.5912295422147639, | |
| "eval_precision_micro": 0.6666666666666666, | |
| "eval_precision_weighted": 0.6614738231117541, | |
| "eval_recall_macro": 0.577278911564626, | |
| "eval_recall_micro": 0.6666666666666666, | |
| "eval_recall_weighted": 0.6666666666666666, | |
| "eval_runtime": 2.3277, | |
| "eval_samples_per_second": 56.707, | |
| "eval_steps_per_second": 7.303, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 13.03448275862069, | |
| "grad_norm": 1.9292876720428467, | |
| "learning_rate": 7.896081771720614e-05, | |
| "loss": 0.5996, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 13.10344827586207, | |
| "grad_norm": 4.09467887878418, | |
| "learning_rate": 7.879045996592846e-05, | |
| "loss": 0.5993, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 13.172413793103448, | |
| "grad_norm": 2.3077073097229004, | |
| "learning_rate": 7.862010221465077e-05, | |
| "loss": 0.6907, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 13.241379310344827, | |
| "grad_norm": 4.537380218505859, | |
| "learning_rate": 7.844974446337309e-05, | |
| "loss": 0.6659, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 13.310344827586206, | |
| "grad_norm": 2.6575474739074707, | |
| "learning_rate": 7.82793867120954e-05, | |
| "loss": 0.4833, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 13.379310344827585, | |
| "grad_norm": 4.1183929443359375, | |
| "learning_rate": 7.810902896081772e-05, | |
| "loss": 0.6274, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 13.448275862068966, | |
| "grad_norm": 2.1658565998077393, | |
| "learning_rate": 7.793867120954004e-05, | |
| "loss": 0.6313, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 13.517241379310345, | |
| "grad_norm": 2.3575236797332764, | |
| "learning_rate": 7.776831345826235e-05, | |
| "loss": 0.5313, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 13.586206896551724, | |
| "grad_norm": 2.720118999481201, | |
| "learning_rate": 7.759795570698467e-05, | |
| "loss": 0.3753, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 13.655172413793103, | |
| "grad_norm": 4.6414475440979, | |
| "learning_rate": 7.742759795570699e-05, | |
| "loss": 0.8126, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 13.724137931034482, | |
| "grad_norm": 4.100179195404053, | |
| "learning_rate": 7.72572402044293e-05, | |
| "loss": 0.6283, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 13.793103448275861, | |
| "grad_norm": 2.20033860206604, | |
| "learning_rate": 7.708688245315162e-05, | |
| "loss": 0.6954, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 13.862068965517242, | |
| "grad_norm": 4.777364253997803, | |
| "learning_rate": 7.691652470187394e-05, | |
| "loss": 0.6302, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 13.931034482758621, | |
| "grad_norm": 3.3690972328186035, | |
| "learning_rate": 7.674616695059625e-05, | |
| "loss": 0.8916, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "grad_norm": 5.667202472686768, | |
| "learning_rate": 7.657580919931857e-05, | |
| "loss": 0.5864, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "eval_accuracy": 0.5984848484848485, | |
| "eval_f1_macro": 0.521144758673933, | |
| "eval_f1_micro": 0.5984848484848485, | |
| "eval_f1_weighted": 0.5893727222006053, | |
| "eval_loss": 1.1956861019134521, | |
| "eval_precision_macro": 0.5536575030457123, | |
| "eval_precision_micro": 0.5984848484848485, | |
| "eval_precision_weighted": 0.6274957584746239, | |
| "eval_recall_macro": 0.5389191232048375, | |
| "eval_recall_micro": 0.5984848484848485, | |
| "eval_recall_weighted": 0.5984848484848485, | |
| "eval_runtime": 2.3168, | |
| "eval_samples_per_second": 56.975, | |
| "eval_steps_per_second": 7.338, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 14.068965517241379, | |
| "grad_norm": 6.122737884521484, | |
| "learning_rate": 7.64054514480409e-05, | |
| "loss": 0.5025, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 14.137931034482758, | |
| "grad_norm": 2.439298629760742, | |
| "learning_rate": 7.62350936967632e-05, | |
| "loss": 0.4959, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 14.206896551724139, | |
| "grad_norm": 3.6299631595611572, | |
| "learning_rate": 7.606473594548552e-05, | |
| "loss": 0.5056, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 14.275862068965518, | |
| "grad_norm": 4.751211166381836, | |
| "learning_rate": 7.589437819420783e-05, | |
| "loss": 0.7691, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 14.344827586206897, | |
| "grad_norm": 4.157205104827881, | |
| "learning_rate": 7.572402044293015e-05, | |
| "loss": 0.5365, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 14.413793103448276, | |
| "grad_norm": 4.521823406219482, | |
| "learning_rate": 7.555366269165248e-05, | |
| "loss": 0.4214, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 14.482758620689655, | |
| "grad_norm": 1.9958875179290771, | |
| "learning_rate": 7.538330494037478e-05, | |
| "loss": 0.3496, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 14.551724137931034, | |
| "grad_norm": 3.086988925933838, | |
| "learning_rate": 7.52129471890971e-05, | |
| "loss": 0.6931, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 14.620689655172415, | |
| "grad_norm": 5.060652256011963, | |
| "learning_rate": 7.504258943781943e-05, | |
| "loss": 0.6347, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 14.689655172413794, | |
| "grad_norm": 2.613652229309082, | |
| "learning_rate": 7.487223168654173e-05, | |
| "loss": 0.5666, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 14.758620689655173, | |
| "grad_norm": 2.4513535499572754, | |
| "learning_rate": 7.470187393526406e-05, | |
| "loss": 0.4125, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 14.827586206896552, | |
| "grad_norm": 2.7038187980651855, | |
| "learning_rate": 7.453151618398638e-05, | |
| "loss": 0.4961, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 14.89655172413793, | |
| "grad_norm": 4.4557085037231445, | |
| "learning_rate": 7.436115843270869e-05, | |
| "loss": 0.4675, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 14.96551724137931, | |
| "grad_norm": 2.6344242095947266, | |
| "learning_rate": 7.419080068143101e-05, | |
| "loss": 0.6145, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "eval_accuracy": 0.7045454545454546, | |
| "eval_f1_macro": 0.6086358317016884, | |
| "eval_f1_micro": 0.7045454545454546, | |
| "eval_f1_weighted": 0.6833139383291237, | |
| "eval_loss": 0.995664656162262, | |
| "eval_precision_macro": 0.6164373139373139, | |
| "eval_precision_micro": 0.7045454545454546, | |
| "eval_precision_weighted": 0.6790800235497206, | |
| "eval_recall_macro": 0.6160166288737718, | |
| "eval_recall_micro": 0.7045454545454546, | |
| "eval_recall_weighted": 0.7045454545454546, | |
| "eval_runtime": 2.3352, | |
| "eval_samples_per_second": 56.526, | |
| "eval_steps_per_second": 7.28, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 15.03448275862069, | |
| "grad_norm": 2.6183178424835205, | |
| "learning_rate": 7.402044293015332e-05, | |
| "loss": 0.5516, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 15.10344827586207, | |
| "grad_norm": 1.3235374689102173, | |
| "learning_rate": 7.385008517887564e-05, | |
| "loss": 0.4201, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 15.172413793103448, | |
| "grad_norm": 0.8775862455368042, | |
| "learning_rate": 7.367972742759796e-05, | |
| "loss": 0.3188, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 15.241379310344827, | |
| "grad_norm": 1.7227734327316284, | |
| "learning_rate": 7.350936967632027e-05, | |
| "loss": 0.3833, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 15.310344827586206, | |
| "grad_norm": 4.330628871917725, | |
| "learning_rate": 7.333901192504259e-05, | |
| "loss": 0.5839, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 15.379310344827585, | |
| "grad_norm": 3.0039358139038086, | |
| "learning_rate": 7.316865417376491e-05, | |
| "loss": 0.6361, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 15.448275862068966, | |
| "grad_norm": 3.6399760246276855, | |
| "learning_rate": 7.299829642248722e-05, | |
| "loss": 0.7401, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 15.517241379310345, | |
| "grad_norm": 4.385819435119629, | |
| "learning_rate": 7.282793867120954e-05, | |
| "loss": 0.3982, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 15.586206896551724, | |
| "grad_norm": 3.1038174629211426, | |
| "learning_rate": 7.265758091993186e-05, | |
| "loss": 0.5036, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 15.655172413793103, | |
| "grad_norm": 3.780447244644165, | |
| "learning_rate": 7.248722316865417e-05, | |
| "loss": 0.3955, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 15.724137931034482, | |
| "grad_norm": 3.7337963581085205, | |
| "learning_rate": 7.231686541737649e-05, | |
| "loss": 0.4673, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 15.793103448275861, | |
| "grad_norm": 2.2187936305999756, | |
| "learning_rate": 7.214650766609881e-05, | |
| "loss": 0.3166, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 15.862068965517242, | |
| "grad_norm": 5.6369476318359375, | |
| "learning_rate": 7.197614991482112e-05, | |
| "loss": 0.5263, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 15.931034482758621, | |
| "grad_norm": 2.084303855895996, | |
| "learning_rate": 7.180579216354344e-05, | |
| "loss": 0.3393, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "grad_norm": 4.383190631866455, | |
| "learning_rate": 7.163543441226575e-05, | |
| "loss": 0.5632, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "eval_accuracy": 0.5984848484848485, | |
| "eval_f1_macro": 0.5112417598080162, | |
| "eval_f1_micro": 0.5984848484848485, | |
| "eval_f1_weighted": 0.5780904923018929, | |
| "eval_loss": 1.230219841003418, | |
| "eval_precision_macro": 0.5219010225654743, | |
| "eval_precision_micro": 0.5984848484848485, | |
| "eval_precision_weighted": 0.5853223415062738, | |
| "eval_recall_macro": 0.5235525321239607, | |
| "eval_recall_micro": 0.5984848484848485, | |
| "eval_recall_weighted": 0.5984848484848485, | |
| "eval_runtime": 2.3147, | |
| "eval_samples_per_second": 57.027, | |
| "eval_steps_per_second": 7.344, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 16.06896551724138, | |
| "grad_norm": 2.8568015098571777, | |
| "learning_rate": 7.146507666098807e-05, | |
| "loss": 0.453, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 16.137931034482758, | |
| "grad_norm": 3.751786231994629, | |
| "learning_rate": 7.12947189097104e-05, | |
| "loss": 0.4951, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 16.20689655172414, | |
| "grad_norm": 3.3349523544311523, | |
| "learning_rate": 7.112436115843271e-05, | |
| "loss": 0.3434, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 16.275862068965516, | |
| "grad_norm": 0.8290226459503174, | |
| "learning_rate": 7.095400340715504e-05, | |
| "loss": 0.2894, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 16.344827586206897, | |
| "grad_norm": 2.8990588188171387, | |
| "learning_rate": 7.078364565587734e-05, | |
| "loss": 0.3078, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 16.413793103448278, | |
| "grad_norm": 3.8644514083862305, | |
| "learning_rate": 7.061328790459967e-05, | |
| "loss": 0.5261, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 16.482758620689655, | |
| "grad_norm": 3.9147894382476807, | |
| "learning_rate": 7.044293015332199e-05, | |
| "loss": 0.4895, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 16.551724137931036, | |
| "grad_norm": 6.574154376983643, | |
| "learning_rate": 7.02725724020443e-05, | |
| "loss": 0.4676, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 16.620689655172413, | |
| "grad_norm": 2.001779079437256, | |
| "learning_rate": 7.010221465076662e-05, | |
| "loss": 0.379, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 16.689655172413794, | |
| "grad_norm": 3.2596888542175293, | |
| "learning_rate": 6.993185689948894e-05, | |
| "loss": 0.4082, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 16.75862068965517, | |
| "grad_norm": 3.3688106536865234, | |
| "learning_rate": 6.976149914821125e-05, | |
| "loss": 0.2854, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 16.82758620689655, | |
| "grad_norm": 4.847587585449219, | |
| "learning_rate": 6.959114139693357e-05, | |
| "loss": 0.465, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 16.896551724137932, | |
| "grad_norm": 1.7020337581634521, | |
| "learning_rate": 6.942078364565589e-05, | |
| "loss": 0.2685, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 16.96551724137931, | |
| "grad_norm": 4.529393672943115, | |
| "learning_rate": 6.92504258943782e-05, | |
| "loss": 0.3392, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "eval_accuracy": 0.6287878787878788, | |
| "eval_f1_macro": 0.5334567769861888, | |
| "eval_f1_micro": 0.6287878787878788, | |
| "eval_f1_weighted": 0.6043366300496426, | |
| "eval_loss": 1.192486047744751, | |
| "eval_precision_macro": 0.5902809125181222, | |
| "eval_precision_micro": 0.6287878787878788, | |
| "eval_precision_weighted": 0.6434826072733183, | |
| "eval_recall_macro": 0.535479969765684, | |
| "eval_recall_micro": 0.6287878787878788, | |
| "eval_recall_weighted": 0.6287878787878788, | |
| "eval_runtime": 2.312, | |
| "eval_samples_per_second": 57.094, | |
| "eval_steps_per_second": 7.353, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 17.03448275862069, | |
| "grad_norm": 5.458866596221924, | |
| "learning_rate": 6.908006814310052e-05, | |
| "loss": 0.4056, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 17.103448275862068, | |
| "grad_norm": 4.587505340576172, | |
| "learning_rate": 6.890971039182284e-05, | |
| "loss": 0.2633, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 17.17241379310345, | |
| "grad_norm": 2.994865894317627, | |
| "learning_rate": 6.873935264054515e-05, | |
| "loss": 0.1985, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 17.24137931034483, | |
| "grad_norm": 2.954942464828491, | |
| "learning_rate": 6.856899488926747e-05, | |
| "loss": 0.2269, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 17.310344827586206, | |
| "grad_norm": 3.353323459625244, | |
| "learning_rate": 6.839863713798978e-05, | |
| "loss": 0.2764, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 17.379310344827587, | |
| "grad_norm": 3.8564820289611816, | |
| "learning_rate": 6.82282793867121e-05, | |
| "loss": 0.4139, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 17.448275862068964, | |
| "grad_norm": 6.136475086212158, | |
| "learning_rate": 6.805792163543442e-05, | |
| "loss": 0.4473, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 17.517241379310345, | |
| "grad_norm": 4.821500301361084, | |
| "learning_rate": 6.788756388415673e-05, | |
| "loss": 0.4126, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 17.586206896551722, | |
| "grad_norm": 6.1974382400512695, | |
| "learning_rate": 6.771720613287905e-05, | |
| "loss": 0.6224, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 17.655172413793103, | |
| "grad_norm": 0.697867214679718, | |
| "learning_rate": 6.754684838160137e-05, | |
| "loss": 0.2629, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 17.724137931034484, | |
| "grad_norm": 6.468963623046875, | |
| "learning_rate": 6.737649063032368e-05, | |
| "loss": 0.4406, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 17.79310344827586, | |
| "grad_norm": 5.978774547576904, | |
| "learning_rate": 6.7206132879046e-05, | |
| "loss": 0.3807, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 17.862068965517242, | |
| "grad_norm": 2.750267744064331, | |
| "learning_rate": 6.703577512776832e-05, | |
| "loss": 0.3092, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 17.93103448275862, | |
| "grad_norm": 3.745131492614746, | |
| "learning_rate": 6.686541737649063e-05, | |
| "loss": 0.5316, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "grad_norm": 1.9382257461547852, | |
| "learning_rate": 6.669505962521295e-05, | |
| "loss": 0.2998, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "eval_accuracy": 0.6363636363636364, | |
| "eval_f1_macro": 0.5543597280552314, | |
| "eval_f1_micro": 0.6363636363636364, | |
| "eval_f1_weighted": 0.625135726951402, | |
| "eval_loss": 1.1443867683410645, | |
| "eval_precision_macro": 0.5519855144855145, | |
| "eval_precision_micro": 0.6363636363636364, | |
| "eval_precision_weighted": 0.6248355495514586, | |
| "eval_recall_macro": 0.566984126984127, | |
| "eval_recall_micro": 0.6363636363636364, | |
| "eval_recall_weighted": 0.6363636363636364, | |
| "eval_runtime": 2.3094, | |
| "eval_samples_per_second": 57.157, | |
| "eval_steps_per_second": 7.361, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 18.06896551724138, | |
| "grad_norm": 3.785149335861206, | |
| "learning_rate": 6.652470187393528e-05, | |
| "loss": 0.285, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 18.137931034482758, | |
| "grad_norm": 1.203855276107788, | |
| "learning_rate": 6.635434412265758e-05, | |
| "loss": 0.2495, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 18.20689655172414, | |
| "grad_norm": 1.5303699970245361, | |
| "learning_rate": 6.61839863713799e-05, | |
| "loss": 0.2783, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 18.275862068965516, | |
| "grad_norm": 0.9707581996917725, | |
| "learning_rate": 6.601362862010221e-05, | |
| "loss": 0.3239, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 18.344827586206897, | |
| "grad_norm": 5.368850231170654, | |
| "learning_rate": 6.584327086882453e-05, | |
| "loss": 0.3304, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 18.413793103448278, | |
| "grad_norm": 2.649095296859741, | |
| "learning_rate": 6.567291311754686e-05, | |
| "loss": 0.2108, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 18.482758620689655, | |
| "grad_norm": 6.152352333068848, | |
| "learning_rate": 6.550255536626916e-05, | |
| "loss": 0.4266, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 18.551724137931036, | |
| "grad_norm": 2.5450921058654785, | |
| "learning_rate": 6.533219761499149e-05, | |
| "loss": 0.3052, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 18.620689655172413, | |
| "grad_norm": 6.302542209625244, | |
| "learning_rate": 6.516183986371381e-05, | |
| "loss": 0.2785, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 18.689655172413794, | |
| "grad_norm": 1.360953688621521, | |
| "learning_rate": 6.499148211243612e-05, | |
| "loss": 0.3293, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 18.75862068965517, | |
| "grad_norm": 1.7876797914505005, | |
| "learning_rate": 6.482112436115844e-05, | |
| "loss": 0.445, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 18.82758620689655, | |
| "grad_norm": 2.4594149589538574, | |
| "learning_rate": 6.465076660988076e-05, | |
| "loss": 0.2731, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 18.896551724137932, | |
| "grad_norm": 0.9318203330039978, | |
| "learning_rate": 6.448040885860307e-05, | |
| "loss": 0.2172, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 18.96551724137931, | |
| "grad_norm": 1.2655283212661743, | |
| "learning_rate": 6.431005110732539e-05, | |
| "loss": 0.2706, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "eval_accuracy": 0.6439393939393939, | |
| "eval_f1_macro": 0.5578558652329144, | |
| "eval_f1_micro": 0.6439393939393939, | |
| "eval_f1_weighted": 0.630778897377258, | |
| "eval_loss": 1.1071969270706177, | |
| "eval_precision_macro": 0.578964121533781, | |
| "eval_precision_micro": 0.6439393939393939, | |
| "eval_precision_weighted": 0.6404139552007551, | |
| "eval_recall_macro": 0.5571126228269084, | |
| "eval_recall_micro": 0.6439393939393939, | |
| "eval_recall_weighted": 0.6439393939393939, | |
| "eval_runtime": 2.3243, | |
| "eval_samples_per_second": 56.79, | |
| "eval_steps_per_second": 7.314, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 19.03448275862069, | |
| "grad_norm": 2.0673553943634033, | |
| "learning_rate": 6.41396933560477e-05, | |
| "loss": 0.2022, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 19.103448275862068, | |
| "grad_norm": 2.8956522941589355, | |
| "learning_rate": 6.396933560477002e-05, | |
| "loss": 0.2003, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 19.17241379310345, | |
| "grad_norm": 4.309751987457275, | |
| "learning_rate": 6.379897785349234e-05, | |
| "loss": 0.1702, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 19.24137931034483, | |
| "grad_norm": 3.166416645050049, | |
| "learning_rate": 6.362862010221465e-05, | |
| "loss": 0.2946, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 19.310344827586206, | |
| "grad_norm": 3.587289810180664, | |
| "learning_rate": 6.345826235093697e-05, | |
| "loss": 0.2177, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 19.379310344827587, | |
| "grad_norm": 1.5400705337524414, | |
| "learning_rate": 6.328790459965929e-05, | |
| "loss": 0.1854, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 19.448275862068964, | |
| "grad_norm": 5.357079029083252, | |
| "learning_rate": 6.31175468483816e-05, | |
| "loss": 0.3174, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 19.517241379310345, | |
| "grad_norm": 6.901614189147949, | |
| "learning_rate": 6.294718909710392e-05, | |
| "loss": 0.2479, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 19.586206896551722, | |
| "grad_norm": 0.9026078581809998, | |
| "learning_rate": 6.277683134582624e-05, | |
| "loss": 0.1429, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 19.655172413793103, | |
| "grad_norm": 1.1682796478271484, | |
| "learning_rate": 6.260647359454855e-05, | |
| "loss": 0.196, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 19.724137931034484, | |
| "grad_norm": 5.2128095626831055, | |
| "learning_rate": 6.243611584327087e-05, | |
| "loss": 0.4045, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 19.79310344827586, | |
| "grad_norm": 1.158043622970581, | |
| "learning_rate": 6.22657580919932e-05, | |
| "loss": 0.2009, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 19.862068965517242, | |
| "grad_norm": 3.482198715209961, | |
| "learning_rate": 6.20954003407155e-05, | |
| "loss": 0.3158, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 19.93103448275862, | |
| "grad_norm": 5.8238749504089355, | |
| "learning_rate": 6.192504258943782e-05, | |
| "loss": 0.2716, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "grad_norm": 0.5987008810043335, | |
| "learning_rate": 6.175468483816013e-05, | |
| "loss": 0.2012, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "eval_accuracy": 0.6212121212121212, | |
| "eval_f1_macro": 0.5277531023715308, | |
| "eval_f1_micro": 0.6212121212121212, | |
| "eval_f1_weighted": 0.6012226561830858, | |
| "eval_loss": 1.1352587938308716, | |
| "eval_precision_macro": 0.5433209647495362, | |
| "eval_precision_micro": 0.6212121212121212, | |
| "eval_precision_weighted": 0.6063147710874984, | |
| "eval_recall_macro": 0.5345502645502646, | |
| "eval_recall_micro": 0.6212121212121212, | |
| "eval_recall_weighted": 0.6212121212121212, | |
| "eval_runtime": 2.3114, | |
| "eval_samples_per_second": 57.108, | |
| "eval_steps_per_second": 7.355, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 20.06896551724138, | |
| "grad_norm": 2.505438804626465, | |
| "learning_rate": 6.158432708688245e-05, | |
| "loss": 0.224, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 20.137931034482758, | |
| "grad_norm": 2.0416412353515625, | |
| "learning_rate": 6.141396933560477e-05, | |
| "loss": 0.2043, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 20.20689655172414, | |
| "grad_norm": 1.9071308374404907, | |
| "learning_rate": 6.124361158432708e-05, | |
| "loss": 0.2274, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 20.275862068965516, | |
| "grad_norm": 0.7224813103675842, | |
| "learning_rate": 6.10732538330494e-05, | |
| "loss": 0.2217, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 20.344827586206897, | |
| "grad_norm": 2.6817595958709717, | |
| "learning_rate": 6.090289608177172e-05, | |
| "loss": 0.1622, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 20.413793103448278, | |
| "grad_norm": 1.356693983078003, | |
| "learning_rate": 6.073253833049404e-05, | |
| "loss": 0.2742, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 20.482758620689655, | |
| "grad_norm": 1.0252429246902466, | |
| "learning_rate": 6.0562180579216355e-05, | |
| "loss": 0.2045, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 20.551724137931036, | |
| "grad_norm": 4.123605251312256, | |
| "learning_rate": 6.039182282793867e-05, | |
| "loss": 0.2, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 20.620689655172413, | |
| "grad_norm": 6.3188252449035645, | |
| "learning_rate": 6.0221465076660985e-05, | |
| "loss": 0.4003, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 20.689655172413794, | |
| "grad_norm": 0.8833699226379395, | |
| "learning_rate": 6.0051107325383306e-05, | |
| "loss": 0.1515, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 20.75862068965517, | |
| "grad_norm": 4.920917510986328, | |
| "learning_rate": 5.988074957410562e-05, | |
| "loss": 0.2139, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 20.82758620689655, | |
| "grad_norm": 4.224968910217285, | |
| "learning_rate": 5.9710391822827936e-05, | |
| "loss": 0.3153, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 20.896551724137932, | |
| "grad_norm": 3.5032029151916504, | |
| "learning_rate": 5.954003407155026e-05, | |
| "loss": 0.3067, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 20.96551724137931, | |
| "grad_norm": 10.811616897583008, | |
| "learning_rate": 5.936967632027257e-05, | |
| "loss": 0.532, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 21.0, | |
| "eval_accuracy": 0.6212121212121212, | |
| "eval_f1_macro": 0.5421176283730086, | |
| "eval_f1_micro": 0.6212121212121212, | |
| "eval_f1_weighted": 0.6079439132523782, | |
| "eval_loss": 1.2502676248550415, | |
| "eval_precision_macro": 0.5651044277360067, | |
| "eval_precision_micro": 0.6212121212121212, | |
| "eval_precision_weighted": 0.6253389154704945, | |
| "eval_recall_macro": 0.5488057445200303, | |
| "eval_recall_micro": 0.6212121212121212, | |
| "eval_recall_weighted": 0.6212121212121212, | |
| "eval_runtime": 2.3136, | |
| "eval_samples_per_second": 57.055, | |
| "eval_steps_per_second": 7.348, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 21.03448275862069, | |
| "grad_norm": 3.4482898712158203, | |
| "learning_rate": 5.919931856899489e-05, | |
| "loss": 0.153, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 21.103448275862068, | |
| "grad_norm": 7.616995811462402, | |
| "learning_rate": 5.90289608177172e-05, | |
| "loss": 0.2165, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 21.17241379310345, | |
| "grad_norm": 2.2934651374816895, | |
| "learning_rate": 5.8858603066439524e-05, | |
| "loss": 0.1314, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 21.24137931034483, | |
| "grad_norm": 3.9521491527557373, | |
| "learning_rate": 5.868824531516184e-05, | |
| "loss": 0.2012, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 21.310344827586206, | |
| "grad_norm": 5.459672451019287, | |
| "learning_rate": 5.851788756388415e-05, | |
| "loss": 0.1458, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 21.379310344827587, | |
| "grad_norm": 4.220061302185059, | |
| "learning_rate": 5.8347529812606475e-05, | |
| "loss": 0.386, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 21.448275862068964, | |
| "grad_norm": 5.224494934082031, | |
| "learning_rate": 5.817717206132879e-05, | |
| "loss": 0.3038, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 21.517241379310345, | |
| "grad_norm": 1.2199459075927734, | |
| "learning_rate": 5.8006814310051104e-05, | |
| "loss": 0.1402, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 21.586206896551722, | |
| "grad_norm": 0.805578887462616, | |
| "learning_rate": 5.783645655877342e-05, | |
| "loss": 0.1285, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 21.655172413793103, | |
| "grad_norm": 1.9446711540222168, | |
| "learning_rate": 5.766609880749574e-05, | |
| "loss": 0.2583, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 21.724137931034484, | |
| "grad_norm": 4.205049514770508, | |
| "learning_rate": 5.7495741056218056e-05, | |
| "loss": 0.222, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 21.79310344827586, | |
| "grad_norm": 6.98620080947876, | |
| "learning_rate": 5.732538330494037e-05, | |
| "loss": 0.2813, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 21.862068965517242, | |
| "grad_norm": 5.716063499450684, | |
| "learning_rate": 5.715502555366269e-05, | |
| "loss": 0.219, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 21.93103448275862, | |
| "grad_norm": 5.608608722686768, | |
| "learning_rate": 5.6984667802385014e-05, | |
| "loss": 0.1247, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "grad_norm": 1.143032431602478, | |
| "learning_rate": 5.6814310051107335e-05, | |
| "loss": 0.0963, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "eval_accuracy": 0.6287878787878788, | |
| "eval_f1_macro": 0.5702385325783071, | |
| "eval_f1_micro": 0.6287878787878788, | |
| "eval_f1_weighted": 0.6226690485974845, | |
| "eval_loss": 1.2203162908554077, | |
| "eval_precision_macro": 0.5806972789115646, | |
| "eval_precision_micro": 0.6287878787878788, | |
| "eval_precision_weighted": 0.6326839826839827, | |
| "eval_recall_macro": 0.5744897959183674, | |
| "eval_recall_micro": 0.6287878787878788, | |
| "eval_recall_weighted": 0.6287878787878788, | |
| "eval_runtime": 2.3213, | |
| "eval_samples_per_second": 56.865, | |
| "eval_steps_per_second": 7.323, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 22.06896551724138, | |
| "grad_norm": 5.11508321762085, | |
| "learning_rate": 5.664395229982965e-05, | |
| "loss": 0.1649, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 22.137931034482758, | |
| "grad_norm": 0.7132185697555542, | |
| "learning_rate": 5.6473594548551965e-05, | |
| "loss": 0.1236, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 22.20689655172414, | |
| "grad_norm": 0.6817318201065063, | |
| "learning_rate": 5.6303236797274287e-05, | |
| "loss": 0.0845, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 22.275862068965516, | |
| "grad_norm": 2.2783679962158203, | |
| "learning_rate": 5.61328790459966e-05, | |
| "loss": 0.092, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 22.344827586206897, | |
| "grad_norm": 3.2513129711151123, | |
| "learning_rate": 5.5962521294718916e-05, | |
| "loss": 0.1328, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 22.413793103448278, | |
| "grad_norm": 6.269516468048096, | |
| "learning_rate": 5.579216354344123e-05, | |
| "loss": 0.2076, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 22.482758620689655, | |
| "grad_norm": 0.8975334763526917, | |
| "learning_rate": 5.562180579216355e-05, | |
| "loss": 0.1214, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 22.551724137931036, | |
| "grad_norm": 0.4030892848968506, | |
| "learning_rate": 5.545144804088587e-05, | |
| "loss": 0.1049, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 22.620689655172413, | |
| "grad_norm": 1.1087042093276978, | |
| "learning_rate": 5.528109028960818e-05, | |
| "loss": 0.1951, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 22.689655172413794, | |
| "grad_norm": 3.3451485633850098, | |
| "learning_rate": 5.5110732538330504e-05, | |
| "loss": 0.1705, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 22.75862068965517, | |
| "grad_norm": 7.313813209533691, | |
| "learning_rate": 5.494037478705282e-05, | |
| "loss": 0.1902, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 22.82758620689655, | |
| "grad_norm": 1.9803084135055542, | |
| "learning_rate": 5.4770017035775133e-05, | |
| "loss": 0.1722, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 22.896551724137932, | |
| "grad_norm": 3.1250195503234863, | |
| "learning_rate": 5.459965928449745e-05, | |
| "loss": 0.137, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 22.96551724137931, | |
| "grad_norm": 4.113572120666504, | |
| "learning_rate": 5.442930153321977e-05, | |
| "loss": 0.1076, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 23.0, | |
| "eval_accuracy": 0.6136363636363636, | |
| "eval_f1_macro": 0.5216254460440507, | |
| "eval_f1_micro": 0.6136363636363636, | |
| "eval_f1_weighted": 0.5893711827578106, | |
| "eval_loss": 1.379845142364502, | |
| "eval_precision_macro": 0.5338904716851017, | |
| "eval_precision_micro": 0.6136363636363636, | |
| "eval_precision_weighted": 0.5970973452876414, | |
| "eval_recall_macro": 0.5370143613000756, | |
| "eval_recall_micro": 0.6136363636363636, | |
| "eval_recall_weighted": 0.6136363636363636, | |
| "eval_runtime": 2.3195, | |
| "eval_samples_per_second": 56.908, | |
| "eval_steps_per_second": 7.329, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 23.03448275862069, | |
| "grad_norm": 1.8101320266723633, | |
| "learning_rate": 5.4258943781942085e-05, | |
| "loss": 0.1341, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 23.103448275862068, | |
| "grad_norm": 0.7451447248458862, | |
| "learning_rate": 5.40885860306644e-05, | |
| "loss": 0.0805, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 23.17241379310345, | |
| "grad_norm": 0.24270519614219666, | |
| "learning_rate": 5.391822827938672e-05, | |
| "loss": 0.0853, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 23.24137931034483, | |
| "grad_norm": 3.7437033653259277, | |
| "learning_rate": 5.3747870528109036e-05, | |
| "loss": 0.1616, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 23.310344827586206, | |
| "grad_norm": 6.19563102722168, | |
| "learning_rate": 5.357751277683135e-05, | |
| "loss": 0.1714, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 23.379310344827587, | |
| "grad_norm": 3.476529359817505, | |
| "learning_rate": 5.3407155025553665e-05, | |
| "loss": 0.2492, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 23.448275862068964, | |
| "grad_norm": 13.964451789855957, | |
| "learning_rate": 5.323679727427599e-05, | |
| "loss": 0.2977, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 23.517241379310345, | |
| "grad_norm": 0.46169885993003845, | |
| "learning_rate": 5.30664395229983e-05, | |
| "loss": 0.0657, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 23.586206896551722, | |
| "grad_norm": 0.8046391010284424, | |
| "learning_rate": 5.289608177172062e-05, | |
| "loss": 0.0722, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 23.655172413793103, | |
| "grad_norm": 3.1619741916656494, | |
| "learning_rate": 5.272572402044293e-05, | |
| "loss": 0.3397, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 23.724137931034484, | |
| "grad_norm": 4.315649032592773, | |
| "learning_rate": 5.255536626916525e-05, | |
| "loss": 0.198, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 23.79310344827586, | |
| "grad_norm": 2.2377915382385254, | |
| "learning_rate": 5.238500851788757e-05, | |
| "loss": 0.1962, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 23.862068965517242, | |
| "grad_norm": 1.14326012134552, | |
| "learning_rate": 5.221465076660988e-05, | |
| "loss": 0.1153, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 23.93103448275862, | |
| "grad_norm": 1.239844799041748, | |
| "learning_rate": 5.2044293015332204e-05, | |
| "loss": 0.0955, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "grad_norm": 0.43866583704948425, | |
| "learning_rate": 5.187393526405452e-05, | |
| "loss": 0.1773, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "eval_accuracy": 0.6287878787878788, | |
| "eval_f1_macro": 0.5421531023295404, | |
| "eval_f1_micro": 0.6287878787878788, | |
| "eval_f1_weighted": 0.6169479734448103, | |
| "eval_loss": 1.312862515449524, | |
| "eval_precision_macro": 0.5581038009609438, | |
| "eval_precision_micro": 0.6287878787878788, | |
| "eval_precision_weighted": 0.6253027275754549, | |
| "eval_recall_macro": 0.5452985638699924, | |
| "eval_recall_micro": 0.6287878787878788, | |
| "eval_recall_weighted": 0.6287878787878788, | |
| "eval_runtime": 2.32, | |
| "eval_samples_per_second": 56.896, | |
| "eval_steps_per_second": 7.328, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 24.06896551724138, | |
| "grad_norm": 0.3895813822746277, | |
| "learning_rate": 5.1703577512776834e-05, | |
| "loss": 0.0867, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 24.137931034482758, | |
| "grad_norm": 0.483073353767395, | |
| "learning_rate": 5.153321976149915e-05, | |
| "loss": 0.0728, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 24.20689655172414, | |
| "grad_norm": 0.414039671421051, | |
| "learning_rate": 5.136286201022147e-05, | |
| "loss": 0.0723, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 24.275862068965516, | |
| "grad_norm": 1.4860070943832397, | |
| "learning_rate": 5.1192504258943785e-05, | |
| "loss": 0.0724, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 24.344827586206897, | |
| "grad_norm": 3.1332881450653076, | |
| "learning_rate": 5.10221465076661e-05, | |
| "loss": 0.2335, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 24.413793103448278, | |
| "grad_norm": 0.35537058115005493, | |
| "learning_rate": 5.085178875638842e-05, | |
| "loss": 0.0597, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 24.482758620689655, | |
| "grad_norm": 0.21754102408885956, | |
| "learning_rate": 5.0681431005110736e-05, | |
| "loss": 0.0556, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 24.551724137931036, | |
| "grad_norm": 0.2571113109588623, | |
| "learning_rate": 5.051107325383305e-05, | |
| "loss": 0.0597, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 24.620689655172413, | |
| "grad_norm": 4.0861124992370605, | |
| "learning_rate": 5.0340715502555366e-05, | |
| "loss": 0.0977, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 24.689655172413794, | |
| "grad_norm": 5.946489334106445, | |
| "learning_rate": 5.017035775127769e-05, | |
| "loss": 0.1114, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 24.75862068965517, | |
| "grad_norm": 0.2896263301372528, | |
| "learning_rate": 5e-05, | |
| "loss": 0.065, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 24.82758620689655, | |
| "grad_norm": 1.2867406606674194, | |
| "learning_rate": 4.982964224872232e-05, | |
| "loss": 0.0753, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 24.896551724137932, | |
| "grad_norm": 0.8423182964324951, | |
| "learning_rate": 4.965928449744464e-05, | |
| "loss": 0.1063, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 24.96551724137931, | |
| "grad_norm": 0.9143354892730713, | |
| "learning_rate": 4.9488926746166954e-05, | |
| "loss": 0.0598, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "eval_accuracy": 0.6515151515151515, | |
| "eval_f1_macro": 0.5632663632663633, | |
| "eval_f1_micro": 0.6515151515151515, | |
| "eval_f1_weighted": 0.6381401426855972, | |
| "eval_loss": 1.285531759262085, | |
| "eval_precision_macro": 0.5845804988662131, | |
| "eval_precision_micro": 0.6515151515151515, | |
| "eval_precision_weighted": 0.6562049062049062, | |
| "eval_recall_macro": 0.5713227513227512, | |
| "eval_recall_micro": 0.6515151515151515, | |
| "eval_recall_weighted": 0.6515151515151515, | |
| "eval_runtime": 2.3186, | |
| "eval_samples_per_second": 56.931, | |
| "eval_steps_per_second": 7.332, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 25.03448275862069, | |
| "grad_norm": 2.1182336807250977, | |
| "learning_rate": 4.931856899488927e-05, | |
| "loss": 0.1474, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 25.103448275862068, | |
| "grad_norm": 4.869687080383301, | |
| "learning_rate": 4.914821124361158e-05, | |
| "loss": 0.0703, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 25.17241379310345, | |
| "grad_norm": 0.5248813629150391, | |
| "learning_rate": 4.8977853492333905e-05, | |
| "loss": 0.0831, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 25.24137931034483, | |
| "grad_norm": 5.952985763549805, | |
| "learning_rate": 4.880749574105622e-05, | |
| "loss": 0.1589, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 25.310344827586206, | |
| "grad_norm": 0.14842434227466583, | |
| "learning_rate": 4.8637137989778534e-05, | |
| "loss": 0.0472, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 25.379310344827587, | |
| "grad_norm": 1.5915170907974243, | |
| "learning_rate": 4.8466780238500856e-05, | |
| "loss": 0.0788, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 25.448275862068964, | |
| "grad_norm": 0.9509754180908203, | |
| "learning_rate": 4.829642248722317e-05, | |
| "loss": 0.1258, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 25.517241379310345, | |
| "grad_norm": 0.5396161079406738, | |
| "learning_rate": 4.8126064735945486e-05, | |
| "loss": 0.0547, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 25.586206896551722, | |
| "grad_norm": 0.6499558091163635, | |
| "learning_rate": 4.79557069846678e-05, | |
| "loss": 0.0606, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 25.655172413793103, | |
| "grad_norm": 0.24467584490776062, | |
| "learning_rate": 4.778534923339012e-05, | |
| "loss": 0.0465, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 25.724137931034484, | |
| "grad_norm": 1.15547513961792, | |
| "learning_rate": 4.761499148211244e-05, | |
| "loss": 0.0808, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 25.79310344827586, | |
| "grad_norm": 5.311452388763428, | |
| "learning_rate": 4.744463373083475e-05, | |
| "loss": 0.2962, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 25.862068965517242, | |
| "grad_norm": 0.5441858172416687, | |
| "learning_rate": 4.727427597955707e-05, | |
| "loss": 0.0454, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 25.93103448275862, | |
| "grad_norm": 6.218692302703857, | |
| "learning_rate": 4.710391822827939e-05, | |
| "loss": 0.2303, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "grad_norm": 1.4410537481307983, | |
| "learning_rate": 4.69335604770017e-05, | |
| "loss": 0.0632, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "eval_accuracy": 0.6590909090909091, | |
| "eval_f1_macro": 0.6414174726494373, | |
| "eval_f1_micro": 0.6590909090909091, | |
| "eval_f1_weighted": 0.6643352984418467, | |
| "eval_loss": 1.3154751062393188, | |
| "eval_precision_macro": 0.6524587989650101, | |
| "eval_precision_micro": 0.6590909090909091, | |
| "eval_precision_weighted": 0.6925259605348538, | |
| "eval_recall_macro": 0.6584580498866214, | |
| "eval_recall_micro": 0.6590909090909091, | |
| "eval_recall_weighted": 0.6590909090909091, | |
| "eval_runtime": 2.3313, | |
| "eval_samples_per_second": 56.62, | |
| "eval_steps_per_second": 7.292, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 26.06896551724138, | |
| "grad_norm": 0.5304362177848816, | |
| "learning_rate": 4.676320272572402e-05, | |
| "loss": 0.0479, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 26.137931034482758, | |
| "grad_norm": 2.6431336402893066, | |
| "learning_rate": 4.659284497444634e-05, | |
| "loss": 0.1618, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 26.20689655172414, | |
| "grad_norm": 0.23049412667751312, | |
| "learning_rate": 4.642248722316866e-05, | |
| "loss": 0.0411, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 26.275862068965516, | |
| "grad_norm": 2.1315579414367676, | |
| "learning_rate": 4.6252129471890976e-05, | |
| "loss": 0.0493, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 26.344827586206897, | |
| "grad_norm": 0.13269948959350586, | |
| "learning_rate": 4.608177172061329e-05, | |
| "loss": 0.0437, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 26.413793103448278, | |
| "grad_norm": 6.615429401397705, | |
| "learning_rate": 4.591141396933561e-05, | |
| "loss": 0.1554, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 26.482758620689655, | |
| "grad_norm": 0.19145725667476654, | |
| "learning_rate": 4.574105621805793e-05, | |
| "loss": 0.0791, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 26.551724137931036, | |
| "grad_norm": 0.16384397447109222, | |
| "learning_rate": 4.557069846678024e-05, | |
| "loss": 0.0493, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 26.620689655172413, | |
| "grad_norm": 9.984197616577148, | |
| "learning_rate": 4.5400340715502557e-05, | |
| "loss": 0.1315, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 26.689655172413794, | |
| "grad_norm": 0.5026840567588806, | |
| "learning_rate": 4.522998296422488e-05, | |
| "loss": 0.1321, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 26.75862068965517, | |
| "grad_norm": 1.3563400506973267, | |
| "learning_rate": 4.505962521294719e-05, | |
| "loss": 0.062, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 26.82758620689655, | |
| "grad_norm": 0.9698802828788757, | |
| "learning_rate": 4.488926746166951e-05, | |
| "loss": 0.2455, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 26.896551724137932, | |
| "grad_norm": 0.633067786693573, | |
| "learning_rate": 4.471890971039183e-05, | |
| "loss": 0.2064, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 26.96551724137931, | |
| "grad_norm": 2.6660585403442383, | |
| "learning_rate": 4.4548551959114144e-05, | |
| "loss": 0.0644, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 27.0, | |
| "eval_accuracy": 0.6439393939393939, | |
| "eval_f1_macro": 0.5587869494846239, | |
| "eval_f1_micro": 0.6439393939393939, | |
| "eval_f1_weighted": 0.6314992501039013, | |
| "eval_loss": 1.3210983276367188, | |
| "eval_precision_macro": 0.5744595234850107, | |
| "eval_precision_micro": 0.6439393939393939, | |
| "eval_precision_weighted": 0.6356558291102383, | |
| "eval_recall_macro": 0.559546485260771, | |
| "eval_recall_micro": 0.6439393939393939, | |
| "eval_recall_weighted": 0.6439393939393939, | |
| "eval_runtime": 2.313, | |
| "eval_samples_per_second": 57.068, | |
| "eval_steps_per_second": 7.35, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 27.03448275862069, | |
| "grad_norm": 1.29978346824646, | |
| "learning_rate": 4.437819420783646e-05, | |
| "loss": 0.0725, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 27.103448275862068, | |
| "grad_norm": 0.27794161438941956, | |
| "learning_rate": 4.4207836456558774e-05, | |
| "loss": 0.0392, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 27.17241379310345, | |
| "grad_norm": 2.312304735183716, | |
| "learning_rate": 4.4037478705281095e-05, | |
| "loss": 0.0548, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 27.24137931034483, | |
| "grad_norm": 0.5731463432312012, | |
| "learning_rate": 4.386712095400341e-05, | |
| "loss": 0.0438, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 27.310344827586206, | |
| "grad_norm": 0.17765074968338013, | |
| "learning_rate": 4.3696763202725725e-05, | |
| "loss": 0.0423, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 27.379310344827587, | |
| "grad_norm": 7.317860126495361, | |
| "learning_rate": 4.352640545144805e-05, | |
| "loss": 0.1978, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 27.448275862068964, | |
| "grad_norm": 0.3569648265838623, | |
| "learning_rate": 4.335604770017036e-05, | |
| "loss": 0.0662, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 27.517241379310345, | |
| "grad_norm": 0.5732914805412292, | |
| "learning_rate": 4.3185689948892676e-05, | |
| "loss": 0.0781, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 27.586206896551722, | |
| "grad_norm": 10.172492980957031, | |
| "learning_rate": 4.301533219761499e-05, | |
| "loss": 0.1123, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 27.655172413793103, | |
| "grad_norm": 0.29396459460258484, | |
| "learning_rate": 4.284497444633731e-05, | |
| "loss": 0.0433, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 27.724137931034484, | |
| "grad_norm": 0.17927728593349457, | |
| "learning_rate": 4.267461669505963e-05, | |
| "loss": 0.075, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 27.79310344827586, | |
| "grad_norm": 0.27556878328323364, | |
| "learning_rate": 4.250425894378194e-05, | |
| "loss": 0.0457, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 27.862068965517242, | |
| "grad_norm": 1.3352630138397217, | |
| "learning_rate": 4.2333901192504264e-05, | |
| "loss": 0.0662, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 27.93103448275862, | |
| "grad_norm": 0.1427721381187439, | |
| "learning_rate": 4.216354344122658e-05, | |
| "loss": 0.055, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "grad_norm": 6.838861465454102, | |
| "learning_rate": 4.1993185689948894e-05, | |
| "loss": 0.1495, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "eval_accuracy": 0.6363636363636364, | |
| "eval_f1_macro": 0.5539424598362265, | |
| "eval_f1_micro": 0.6363636363636364, | |
| "eval_f1_weighted": 0.6244950107428886, | |
| "eval_loss": 1.4195839166641235, | |
| "eval_precision_macro": 0.5650270364556079, | |
| "eval_precision_micro": 0.6363636363636364, | |
| "eval_precision_weighted": 0.6269573019573019, | |
| "eval_recall_macro": 0.5555555555555556, | |
| "eval_recall_micro": 0.6363636363636364, | |
| "eval_recall_weighted": 0.6363636363636364, | |
| "eval_runtime": 2.3019, | |
| "eval_samples_per_second": 57.344, | |
| "eval_steps_per_second": 7.385, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 28.06896551724138, | |
| "grad_norm": 0.4316509962081909, | |
| "learning_rate": 4.182282793867121e-05, | |
| "loss": 0.0652, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 28.137931034482758, | |
| "grad_norm": 0.5784430503845215, | |
| "learning_rate": 4.165247018739353e-05, | |
| "loss": 0.0404, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 28.20689655172414, | |
| "grad_norm": 0.12339165061712265, | |
| "learning_rate": 4.1482112436115845e-05, | |
| "loss": 0.0342, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 28.275862068965516, | |
| "grad_norm": 9.590283393859863, | |
| "learning_rate": 4.131175468483816e-05, | |
| "loss": 0.09, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 28.344827586206897, | |
| "grad_norm": 5.919569492340088, | |
| "learning_rate": 4.1141396933560474e-05, | |
| "loss": 0.1994, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 28.413793103448278, | |
| "grad_norm": 0.18216995894908905, | |
| "learning_rate": 4.0971039182282796e-05, | |
| "loss": 0.0384, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 28.482758620689655, | |
| "grad_norm": 0.36890658736228943, | |
| "learning_rate": 4.080068143100511e-05, | |
| "loss": 0.0993, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 28.551724137931036, | |
| "grad_norm": 0.1248667761683464, | |
| "learning_rate": 4.0630323679727426e-05, | |
| "loss": 0.0398, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 28.620689655172413, | |
| "grad_norm": 0.8702082633972168, | |
| "learning_rate": 4.045996592844975e-05, | |
| "loss": 0.0409, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 28.689655172413794, | |
| "grad_norm": 0.16621750593185425, | |
| "learning_rate": 4.028960817717206e-05, | |
| "loss": 0.0386, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 28.75862068965517, | |
| "grad_norm": 2.846982717514038, | |
| "learning_rate": 4.011925042589438e-05, | |
| "loss": 0.0558, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 28.82758620689655, | |
| "grad_norm": 0.125545471906662, | |
| "learning_rate": 3.994889267461669e-05, | |
| "loss": 0.0348, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 28.896551724137932, | |
| "grad_norm": 0.17237234115600586, | |
| "learning_rate": 3.977853492333901e-05, | |
| "loss": 0.038, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 28.96551724137931, | |
| "grad_norm": 0.919359028339386, | |
| "learning_rate": 3.960817717206133e-05, | |
| "loss": 0.0413, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "eval_accuracy": 0.6136363636363636, | |
| "eval_f1_macro": 0.5377616118217622, | |
| "eval_f1_micro": 0.6136363636363636, | |
| "eval_f1_weighted": 0.6102201958063201, | |
| "eval_loss": 1.4026981592178345, | |
| "eval_precision_macro": 0.5405016036594984, | |
| "eval_precision_micro": 0.6136363636363636, | |
| "eval_precision_weighted": 0.6100277712717904, | |
| "eval_recall_macro": 0.5379591836734694, | |
| "eval_recall_micro": 0.6136363636363636, | |
| "eval_recall_weighted": 0.6136363636363636, | |
| "eval_runtime": 2.3087, | |
| "eval_samples_per_second": 57.176, | |
| "eval_steps_per_second": 7.364, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 29.03448275862069, | |
| "grad_norm": 0.18945536017417908, | |
| "learning_rate": 3.943781942078364e-05, | |
| "loss": 0.0357, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 29.103448275862068, | |
| "grad_norm": 0.18712615966796875, | |
| "learning_rate": 3.9267461669505964e-05, | |
| "loss": 0.037, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 29.17241379310345, | |
| "grad_norm": 0.21111753582954407, | |
| "learning_rate": 3.9097103918228286e-05, | |
| "loss": 0.0404, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 29.24137931034483, | |
| "grad_norm": 0.13927434384822845, | |
| "learning_rate": 3.89267461669506e-05, | |
| "loss": 0.0379, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 29.310344827586206, | |
| "grad_norm": 0.3238166272640228, | |
| "learning_rate": 3.8756388415672916e-05, | |
| "loss": 0.0359, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 29.379310344827587, | |
| "grad_norm": 0.4770512878894806, | |
| "learning_rate": 3.858603066439524e-05, | |
| "loss": 0.0624, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 29.448275862068964, | |
| "grad_norm": 12.409387588500977, | |
| "learning_rate": 3.841567291311755e-05, | |
| "loss": 0.1086, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 29.517241379310345, | |
| "grad_norm": 0.1156606376171112, | |
| "learning_rate": 3.824531516183987e-05, | |
| "loss": 0.0345, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 29.586206896551722, | |
| "grad_norm": 0.11645504087209702, | |
| "learning_rate": 3.807495741056218e-05, | |
| "loss": 0.1033, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 29.655172413793103, | |
| "grad_norm": 0.11167976260185242, | |
| "learning_rate": 3.79045996592845e-05, | |
| "loss": 0.042, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 29.724137931034484, | |
| "grad_norm": 0.1662347912788391, | |
| "learning_rate": 3.773424190800682e-05, | |
| "loss": 0.0336, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 29.79310344827586, | |
| "grad_norm": 0.13040611147880554, | |
| "learning_rate": 3.756388415672913e-05, | |
| "loss": 0.0335, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 29.862068965517242, | |
| "grad_norm": 2.854461193084717, | |
| "learning_rate": 3.7393526405451454e-05, | |
| "loss": 0.0475, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 29.93103448275862, | |
| "grad_norm": 0.21949215233325958, | |
| "learning_rate": 3.722316865417377e-05, | |
| "loss": 0.0622, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "grad_norm": 0.12153808027505875, | |
| "learning_rate": 3.7052810902896084e-05, | |
| "loss": 0.0323, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "eval_accuracy": 0.6363636363636364, | |
| "eval_f1_macro": 0.5640770593951866, | |
| "eval_f1_micro": 0.6363636363636364, | |
| "eval_f1_weighted": 0.6329471182412358, | |
| "eval_loss": 1.4301841259002686, | |
| "eval_precision_macro": 0.5688986640599544, | |
| "eval_precision_micro": 0.6363636363636364, | |
| "eval_precision_weighted": 0.6429970526378151, | |
| "eval_recall_macro": 0.5711866969009826, | |
| "eval_recall_micro": 0.6363636363636364, | |
| "eval_recall_weighted": 0.6363636363636364, | |
| "eval_runtime": 2.3127, | |
| "eval_samples_per_second": 57.076, | |
| "eval_steps_per_second": 7.351, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 30.06896551724138, | |
| "grad_norm": 10.862177848815918, | |
| "learning_rate": 3.68824531516184e-05, | |
| "loss": 0.0783, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 30.137931034482758, | |
| "grad_norm": 0.12320105731487274, | |
| "learning_rate": 3.671209540034072e-05, | |
| "loss": 0.0354, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 30.20689655172414, | |
| "grad_norm": 0.14022612571716309, | |
| "learning_rate": 3.6541737649063035e-05, | |
| "loss": 0.0315, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 30.275862068965516, | |
| "grad_norm": 2.5534183979034424, | |
| "learning_rate": 3.637137989778535e-05, | |
| "loss": 0.0403, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 30.344827586206897, | |
| "grad_norm": 0.13770627975463867, | |
| "learning_rate": 3.6201022146507665e-05, | |
| "loss": 0.0552, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 30.413793103448278, | |
| "grad_norm": 0.3032757639884949, | |
| "learning_rate": 3.6030664395229987e-05, | |
| "loss": 0.0618, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 30.482758620689655, | |
| "grad_norm": 9.125004768371582, | |
| "learning_rate": 3.58603066439523e-05, | |
| "loss": 0.0986, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 30.551724137931036, | |
| "grad_norm": 0.39664426445961, | |
| "learning_rate": 3.5689948892674616e-05, | |
| "loss": 0.0303, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 30.620689655172413, | |
| "grad_norm": 0.8094646334648132, | |
| "learning_rate": 3.551959114139694e-05, | |
| "loss": 0.0347, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 30.689655172413794, | |
| "grad_norm": 0.10856208205223083, | |
| "learning_rate": 3.534923339011925e-05, | |
| "loss": 0.0661, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 30.75862068965517, | |
| "grad_norm": 0.09460315108299255, | |
| "learning_rate": 3.517887563884157e-05, | |
| "loss": 0.0547, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 30.82758620689655, | |
| "grad_norm": 2.669100284576416, | |
| "learning_rate": 3.500851788756388e-05, | |
| "loss": 0.0503, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 30.896551724137932, | |
| "grad_norm": 0.212706059217453, | |
| "learning_rate": 3.4838160136286204e-05, | |
| "loss": 0.0637, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 30.96551724137931, | |
| "grad_norm": 0.41738227009773254, | |
| "learning_rate": 3.466780238500852e-05, | |
| "loss": 0.0452, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 31.0, | |
| "eval_accuracy": 0.6515151515151515, | |
| "eval_f1_macro": 0.5705624622730447, | |
| "eval_f1_micro": 0.6515151515151515, | |
| "eval_f1_weighted": 0.6412393051725056, | |
| "eval_loss": 1.457693338394165, | |
| "eval_precision_macro": 0.5835023111808825, | |
| "eval_precision_micro": 0.6515151515151515, | |
| "eval_precision_weighted": 0.647774968087468, | |
| "eval_recall_macro": 0.5738321995464852, | |
| "eval_recall_micro": 0.6515151515151515, | |
| "eval_recall_weighted": 0.6515151515151515, | |
| "eval_runtime": 2.3087, | |
| "eval_samples_per_second": 57.175, | |
| "eval_steps_per_second": 7.363, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 31.03448275862069, | |
| "grad_norm": 0.5998754501342773, | |
| "learning_rate": 3.4497444633730833e-05, | |
| "loss": 0.0327, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 31.103448275862068, | |
| "grad_norm": 0.10836658626794815, | |
| "learning_rate": 3.4327086882453155e-05, | |
| "loss": 0.0317, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 31.17241379310345, | |
| "grad_norm": 0.3383055329322815, | |
| "learning_rate": 3.415672913117547e-05, | |
| "loss": 0.1491, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 31.24137931034483, | |
| "grad_norm": 0.13469484448432922, | |
| "learning_rate": 3.3986371379897785e-05, | |
| "loss": 0.0619, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 31.310344827586206, | |
| "grad_norm": 0.15399225056171417, | |
| "learning_rate": 3.38160136286201e-05, | |
| "loss": 0.0576, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 31.379310344827587, | |
| "grad_norm": 0.2508799731731415, | |
| "learning_rate": 3.364565587734242e-05, | |
| "loss": 0.0283, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 31.448275862068964, | |
| "grad_norm": 0.09723825007677078, | |
| "learning_rate": 3.3475298126064736e-05, | |
| "loss": 0.0289, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 31.517241379310345, | |
| "grad_norm": 0.11320387572050095, | |
| "learning_rate": 3.330494037478705e-05, | |
| "loss": 0.041, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 31.586206896551722, | |
| "grad_norm": 0.09735345840454102, | |
| "learning_rate": 3.313458262350937e-05, | |
| "loss": 0.0296, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 31.655172413793103, | |
| "grad_norm": 0.1624167114496231, | |
| "learning_rate": 3.296422487223169e-05, | |
| "loss": 0.0284, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 31.724137931034484, | |
| "grad_norm": 0.15523399412631989, | |
| "learning_rate": 3.2793867120954e-05, | |
| "loss": 0.0316, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 31.79310344827586, | |
| "grad_norm": 0.6281076073646545, | |
| "learning_rate": 3.262350936967632e-05, | |
| "loss": 0.0301, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 31.862068965517242, | |
| "grad_norm": 0.10810775309801102, | |
| "learning_rate": 3.245315161839864e-05, | |
| "loss": 0.1419, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 31.93103448275862, | |
| "grad_norm": 0.09449290484189987, | |
| "learning_rate": 3.228279386712095e-05, | |
| "loss": 0.1476, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "grad_norm": 0.11448001116514206, | |
| "learning_rate": 3.2112436115843275e-05, | |
| "loss": 0.0285, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "eval_accuracy": 0.6439393939393939, | |
| "eval_f1_macro": 0.5596847319551131, | |
| "eval_f1_micro": 0.6439393939393939, | |
| "eval_f1_weighted": 0.6300303506636783, | |
| "eval_loss": 1.422427773475647, | |
| "eval_precision_macro": 0.5618261922409388, | |
| "eval_precision_micro": 0.6439393939393939, | |
| "eval_precision_weighted": 0.6250287835331823, | |
| "eval_recall_macro": 0.5657142857142857, | |
| "eval_recall_micro": 0.6439393939393939, | |
| "eval_recall_weighted": 0.6439393939393939, | |
| "eval_runtime": 2.3052, | |
| "eval_samples_per_second": 57.263, | |
| "eval_steps_per_second": 7.375, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 32.06896551724138, | |
| "grad_norm": 0.09805119037628174, | |
| "learning_rate": 3.194207836456559e-05, | |
| "loss": 0.0305, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 32.13793103448276, | |
| "grad_norm": 0.6063154339790344, | |
| "learning_rate": 3.177172061328791e-05, | |
| "loss": 0.0317, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 32.206896551724135, | |
| "grad_norm": 0.10352925211191177, | |
| "learning_rate": 3.1601362862010226e-05, | |
| "loss": 0.0272, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 32.275862068965516, | |
| "grad_norm": 12.472454071044922, | |
| "learning_rate": 3.143100511073254e-05, | |
| "loss": 0.1287, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 32.3448275862069, | |
| "grad_norm": 2.9856293201446533, | |
| "learning_rate": 3.1260647359454856e-05, | |
| "loss": 0.0404, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 32.41379310344828, | |
| "grad_norm": 0.22983527183532715, | |
| "learning_rate": 3.109028960817718e-05, | |
| "loss": 0.0285, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 32.48275862068966, | |
| "grad_norm": 0.1563962697982788, | |
| "learning_rate": 3.091993185689949e-05, | |
| "loss": 0.0262, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 32.55172413793103, | |
| "grad_norm": 0.10210330039262772, | |
| "learning_rate": 3.074957410562181e-05, | |
| "loss": 0.0286, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 32.62068965517241, | |
| "grad_norm": 0.10098184645175934, | |
| "learning_rate": 3.057921635434413e-05, | |
| "loss": 0.0273, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 32.689655172413794, | |
| "grad_norm": 0.19312156736850739, | |
| "learning_rate": 3.0408858603066443e-05, | |
| "loss": 0.0272, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 32.758620689655174, | |
| "grad_norm": 0.8814791440963745, | |
| "learning_rate": 3.0238500851788758e-05, | |
| "loss": 0.0335, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 32.827586206896555, | |
| "grad_norm": 0.38704535365104675, | |
| "learning_rate": 3.0068143100511076e-05, | |
| "loss": 0.03, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 32.89655172413793, | |
| "grad_norm": 0.36842238903045654, | |
| "learning_rate": 2.989778534923339e-05, | |
| "loss": 0.0466, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 32.96551724137931, | |
| "grad_norm": 0.12097663432359695, | |
| "learning_rate": 2.972742759795571e-05, | |
| "loss": 0.0241, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 33.0, | |
| "eval_accuracy": 0.6363636363636364, | |
| "eval_f1_macro": 0.5541615185524713, | |
| "eval_f1_micro": 0.6363636363636364, | |
| "eval_f1_weighted": 0.6252499952281971, | |
| "eval_loss": 1.4512511491775513, | |
| "eval_precision_macro": 0.5700210921075582, | |
| "eval_precision_micro": 0.6363636363636364, | |
| "eval_precision_weighted": 0.6309470561115298, | |
| "eval_recall_macro": 0.5532501889644746, | |
| "eval_recall_micro": 0.6363636363636364, | |
| "eval_recall_weighted": 0.6363636363636364, | |
| "eval_runtime": 2.3075, | |
| "eval_samples_per_second": 57.205, | |
| "eval_steps_per_second": 7.367, | |
| "step": 957 | |
| }, | |
| { | |
| "epoch": 33.03448275862069, | |
| "grad_norm": 0.1205965206027031, | |
| "learning_rate": 2.9557069846678027e-05, | |
| "loss": 0.0294, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 33.10344827586207, | |
| "grad_norm": 0.09090107679367065, | |
| "learning_rate": 2.9386712095400342e-05, | |
| "loss": 0.0296, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 33.172413793103445, | |
| "grad_norm": 0.08838998526334763, | |
| "learning_rate": 2.921635434412266e-05, | |
| "loss": 0.026, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 33.241379310344826, | |
| "grad_norm": 0.2968568801879883, | |
| "learning_rate": 2.9045996592844975e-05, | |
| "loss": 0.0405, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 33.310344827586206, | |
| "grad_norm": 0.11646933853626251, | |
| "learning_rate": 2.8875638841567293e-05, | |
| "loss": 0.0276, | |
| "step": 966 | |
| }, | |
| { | |
| "epoch": 33.37931034482759, | |
| "grad_norm": 0.09644908457994461, | |
| "learning_rate": 2.8705281090289608e-05, | |
| "loss": 0.0257, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 33.44827586206897, | |
| "grad_norm": 0.258109986782074, | |
| "learning_rate": 2.8534923339011926e-05, | |
| "loss": 0.0279, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 33.51724137931034, | |
| "grad_norm": 1.4819831848144531, | |
| "learning_rate": 2.8364565587734245e-05, | |
| "loss": 0.0681, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 33.58620689655172, | |
| "grad_norm": 0.12361834198236465, | |
| "learning_rate": 2.819420783645656e-05, | |
| "loss": 0.03, | |
| "step": 974 | |
| }, | |
| { | |
| "epoch": 33.6551724137931, | |
| "grad_norm": 0.12781710922718048, | |
| "learning_rate": 2.8023850085178878e-05, | |
| "loss": 0.025, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 33.724137931034484, | |
| "grad_norm": 0.10559111833572388, | |
| "learning_rate": 2.7853492333901192e-05, | |
| "loss": 0.0261, | |
| "step": 978 | |
| }, | |
| { | |
| "epoch": 33.793103448275865, | |
| "grad_norm": 0.08984573930501938, | |
| "learning_rate": 2.768313458262351e-05, | |
| "loss": 0.0244, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 33.86206896551724, | |
| "grad_norm": 0.08784312754869461, | |
| "learning_rate": 2.7512776831345826e-05, | |
| "loss": 0.0283, | |
| "step": 982 | |
| }, | |
| { | |
| "epoch": 33.93103448275862, | |
| "grad_norm": 0.09832680970430374, | |
| "learning_rate": 2.7342419080068144e-05, | |
| "loss": 0.0278, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 34.0, | |
| "grad_norm": 0.07693750411272049, | |
| "learning_rate": 2.717206132879046e-05, | |
| "loss": 0.0224, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 34.0, | |
| "eval_accuracy": 0.6742424242424242, | |
| "eval_f1_macro": 0.5794868578213456, | |
| "eval_f1_micro": 0.6742424242424242, | |
| "eval_f1_weighted": 0.65445260813248, | |
| "eval_loss": 1.4700735807418823, | |
| "eval_precision_macro": 0.5855705141419428, | |
| "eval_precision_micro": 0.6742424242424242, | |
| "eval_precision_weighted": 0.652266504539232, | |
| "eval_recall_macro": 0.5901965230536659, | |
| "eval_recall_micro": 0.6742424242424242, | |
| "eval_recall_weighted": 0.6742424242424242, | |
| "eval_runtime": 2.4834, | |
| "eval_samples_per_second": 53.154, | |
| "eval_steps_per_second": 6.846, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 34.06896551724138, | |
| "grad_norm": 0.08656442165374756, | |
| "learning_rate": 2.7001703577512777e-05, | |
| "loss": 0.0231, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 34.13793103448276, | |
| "grad_norm": 0.10006435215473175, | |
| "learning_rate": 2.6831345826235095e-05, | |
| "loss": 0.0233, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 34.206896551724135, | |
| "grad_norm": 0.10205938667058945, | |
| "learning_rate": 2.666098807495741e-05, | |
| "loss": 0.0269, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 34.275862068965516, | |
| "grad_norm": 0.14712445437908173, | |
| "learning_rate": 2.6490630323679728e-05, | |
| "loss": 0.0263, | |
| "step": 994 | |
| }, | |
| { | |
| "epoch": 34.3448275862069, | |
| "grad_norm": 0.476298987865448, | |
| "learning_rate": 2.6320272572402043e-05, | |
| "loss": 0.0529, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 34.41379310344828, | |
| "grad_norm": 0.08625518530607224, | |
| "learning_rate": 2.614991482112436e-05, | |
| "loss": 0.0242, | |
| "step": 998 | |
| }, | |
| { | |
| "epoch": 34.48275862068966, | |
| "grad_norm": 0.1275705099105835, | |
| "learning_rate": 2.5979557069846676e-05, | |
| "loss": 0.0264, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 34.55172413793103, | |
| "grad_norm": 0.07826410233974457, | |
| "learning_rate": 2.5809199318568994e-05, | |
| "loss": 0.0227, | |
| "step": 1002 | |
| }, | |
| { | |
| "epoch": 34.62068965517241, | |
| "grad_norm": 0.3700888454914093, | |
| "learning_rate": 2.5638841567291312e-05, | |
| "loss": 0.0303, | |
| "step": 1004 | |
| }, | |
| { | |
| "epoch": 34.689655172413794, | |
| "grad_norm": 0.15820983052253723, | |
| "learning_rate": 2.5468483816013627e-05, | |
| "loss": 0.0253, | |
| "step": 1006 | |
| }, | |
| { | |
| "epoch": 34.758620689655174, | |
| "grad_norm": 0.11389223486185074, | |
| "learning_rate": 2.5298126064735945e-05, | |
| "loss": 0.0247, | |
| "step": 1008 | |
| }, | |
| { | |
| "epoch": 34.827586206896555, | |
| "grad_norm": 0.08679822832345963, | |
| "learning_rate": 2.512776831345826e-05, | |
| "loss": 0.0273, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 34.89655172413793, | |
| "grad_norm": 0.47994059324264526, | |
| "learning_rate": 2.495741056218058e-05, | |
| "loss": 0.0523, | |
| "step": 1012 | |
| }, | |
| { | |
| "epoch": 34.96551724137931, | |
| "grad_norm": 0.08631986379623413, | |
| "learning_rate": 2.4787052810902896e-05, | |
| "loss": 0.0228, | |
| "step": 1014 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "eval_accuracy": 0.6590909090909091, | |
| "eval_f1_macro": 0.5772129682250426, | |
| "eval_f1_micro": 0.6590909090909091, | |
| "eval_f1_weighted": 0.6488555790270587, | |
| "eval_loss": 1.4697364568710327, | |
| "eval_precision_macro": 0.5869569578494528, | |
| "eval_precision_micro": 0.6590909090909091, | |
| "eval_precision_weighted": 0.649717549185095, | |
| "eval_recall_macro": 0.577384731670446, | |
| "eval_recall_micro": 0.6590909090909091, | |
| "eval_recall_weighted": 0.6590909090909091, | |
| "eval_runtime": 2.3103, | |
| "eval_samples_per_second": 57.135, | |
| "eval_steps_per_second": 7.358, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 35.03448275862069, | |
| "grad_norm": 0.16488300263881683, | |
| "learning_rate": 2.4616695059625215e-05, | |
| "loss": 0.0273, | |
| "step": 1016 | |
| }, | |
| { | |
| "epoch": 35.10344827586207, | |
| "grad_norm": 0.5682753920555115, | |
| "learning_rate": 2.444633730834753e-05, | |
| "loss": 0.0278, | |
| "step": 1018 | |
| }, | |
| { | |
| "epoch": 35.172413793103445, | |
| "grad_norm": 0.4378216564655304, | |
| "learning_rate": 2.4275979557069848e-05, | |
| "loss": 0.0508, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 35.241379310344826, | |
| "grad_norm": 0.14175258576869965, | |
| "learning_rate": 2.4105621805792162e-05, | |
| "loss": 0.0235, | |
| "step": 1022 | |
| }, | |
| { | |
| "epoch": 35.310344827586206, | |
| "grad_norm": 0.08108456432819366, | |
| "learning_rate": 2.393526405451448e-05, | |
| "loss": 0.0218, | |
| "step": 1024 | |
| }, | |
| { | |
| "epoch": 35.37931034482759, | |
| "grad_norm": 0.08066286891698837, | |
| "learning_rate": 2.37649063032368e-05, | |
| "loss": 0.0212, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 35.44827586206897, | |
| "grad_norm": 0.14611414074897766, | |
| "learning_rate": 2.3594548551959114e-05, | |
| "loss": 0.0252, | |
| "step": 1028 | |
| }, | |
| { | |
| "epoch": 35.51724137931034, | |
| "grad_norm": 2.4684391021728516, | |
| "learning_rate": 2.3424190800681432e-05, | |
| "loss": 0.0292, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 35.58620689655172, | |
| "grad_norm": 0.43363311886787415, | |
| "learning_rate": 2.3253833049403747e-05, | |
| "loss": 0.0469, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 35.6551724137931, | |
| "grad_norm": 0.0832495167851448, | |
| "learning_rate": 2.3083475298126068e-05, | |
| "loss": 0.0259, | |
| "step": 1034 | |
| }, | |
| { | |
| "epoch": 35.724137931034484, | |
| "grad_norm": 1.9775596857070923, | |
| "learning_rate": 2.2913117546848383e-05, | |
| "loss": 0.0303, | |
| "step": 1036 | |
| }, | |
| { | |
| "epoch": 35.793103448275865, | |
| "grad_norm": 0.08929072320461273, | |
| "learning_rate": 2.27427597955707e-05, | |
| "loss": 0.0238, | |
| "step": 1038 | |
| }, | |
| { | |
| "epoch": 35.86206896551724, | |
| "grad_norm": 0.0937528908252716, | |
| "learning_rate": 2.2572402044293016e-05, | |
| "loss": 0.0258, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 35.93103448275862, | |
| "grad_norm": 0.07003093510866165, | |
| "learning_rate": 2.2402044293015334e-05, | |
| "loss": 0.0218, | |
| "step": 1042 | |
| }, | |
| { | |
| "epoch": 36.0, | |
| "grad_norm": 0.11906813085079193, | |
| "learning_rate": 2.223168654173765e-05, | |
| "loss": 0.0231, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 36.0, | |
| "eval_accuracy": 0.6590909090909091, | |
| "eval_f1_macro": 0.5744938259848144, | |
| "eval_f1_micro": 0.6590909090909091, | |
| "eval_f1_weighted": 0.6491010450687431, | |
| "eval_loss": 1.5315077304840088, | |
| "eval_precision_macro": 0.5783361246596541, | |
| "eval_precision_micro": 0.6590909090909091, | |
| "eval_precision_weighted": 0.648294506666833, | |
| "eval_recall_macro": 0.5788133030990174, | |
| "eval_recall_micro": 0.6590909090909091, | |
| "eval_recall_weighted": 0.6590909090909091, | |
| "eval_runtime": 2.314, | |
| "eval_samples_per_second": 57.045, | |
| "eval_steps_per_second": 7.347, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 36.06896551724138, | |
| "grad_norm": 1.3290965557098389, | |
| "learning_rate": 2.2061328790459967e-05, | |
| "loss": 0.028, | |
| "step": 1046 | |
| }, | |
| { | |
| "epoch": 36.13793103448276, | |
| "grad_norm": 0.1269829422235489, | |
| "learning_rate": 2.1890971039182286e-05, | |
| "loss": 0.0232, | |
| "step": 1048 | |
| }, | |
| { | |
| "epoch": 36.206896551724135, | |
| "grad_norm": 0.14340566098690033, | |
| "learning_rate": 2.17206132879046e-05, | |
| "loss": 0.0223, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 36.275862068965516, | |
| "grad_norm": 4.602632999420166, | |
| "learning_rate": 2.155025553662692e-05, | |
| "loss": 0.1708, | |
| "step": 1052 | |
| }, | |
| { | |
| "epoch": 36.3448275862069, | |
| "grad_norm": 0.09221786260604858, | |
| "learning_rate": 2.1379897785349233e-05, | |
| "loss": 0.0343, | |
| "step": 1054 | |
| }, | |
| { | |
| "epoch": 36.41379310344828, | |
| "grad_norm": 0.11092764139175415, | |
| "learning_rate": 2.120954003407155e-05, | |
| "loss": 0.0253, | |
| "step": 1056 | |
| }, | |
| { | |
| "epoch": 36.48275862068966, | |
| "grad_norm": 0.09004894644021988, | |
| "learning_rate": 2.1039182282793866e-05, | |
| "loss": 0.0223, | |
| "step": 1058 | |
| }, | |
| { | |
| "epoch": 36.55172413793103, | |
| "grad_norm": 0.08809762448072433, | |
| "learning_rate": 2.0868824531516185e-05, | |
| "loss": 0.0229, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 36.62068965517241, | |
| "grad_norm": 0.08181875944137573, | |
| "learning_rate": 2.0698466780238503e-05, | |
| "loss": 0.0224, | |
| "step": 1062 | |
| }, | |
| { | |
| "epoch": 36.689655172413794, | |
| "grad_norm": 0.11032210290431976, | |
| "learning_rate": 2.0528109028960818e-05, | |
| "loss": 0.022, | |
| "step": 1064 | |
| }, | |
| { | |
| "epoch": 36.758620689655174, | |
| "grad_norm": 0.07693520188331604, | |
| "learning_rate": 2.0357751277683136e-05, | |
| "loss": 0.0206, | |
| "step": 1066 | |
| }, | |
| { | |
| "epoch": 36.827586206896555, | |
| "grad_norm": 0.07412053644657135, | |
| "learning_rate": 2.018739352640545e-05, | |
| "loss": 0.0212, | |
| "step": 1068 | |
| }, | |
| { | |
| "epoch": 36.89655172413793, | |
| "grad_norm": 4.513420104980469, | |
| "learning_rate": 2.001703577512777e-05, | |
| "loss": 0.0467, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 36.96551724137931, | |
| "grad_norm": 0.08167243748903275, | |
| "learning_rate": 1.9846678023850084e-05, | |
| "loss": 0.0457, | |
| "step": 1072 | |
| }, | |
| { | |
| "epoch": 37.0, | |
| "eval_accuracy": 0.6439393939393939, | |
| "eval_f1_macro": 0.5532375287737503, | |
| "eval_f1_micro": 0.6439393939393939, | |
| "eval_f1_weighted": 0.6276644798060326, | |
| "eval_loss": 1.520954966545105, | |
| "eval_precision_macro": 0.5641351058808365, | |
| "eval_precision_micro": 0.6439393939393939, | |
| "eval_precision_weighted": 0.6317489292950739, | |
| "eval_recall_macro": 0.5605820105820106, | |
| "eval_recall_micro": 0.6439393939393939, | |
| "eval_recall_weighted": 0.6439393939393939, | |
| "eval_runtime": 2.3121, | |
| "eval_samples_per_second": 57.091, | |
| "eval_steps_per_second": 7.353, | |
| "step": 1073 | |
| }, | |
| { | |
| "epoch": 37.03448275862069, | |
| "grad_norm": 0.3906444311141968, | |
| "learning_rate": 1.9676320272572402e-05, | |
| "loss": 0.0419, | |
| "step": 1074 | |
| }, | |
| { | |
| "epoch": 37.10344827586207, | |
| "grad_norm": 0.36884042620658875, | |
| "learning_rate": 1.950596252129472e-05, | |
| "loss": 0.0413, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 37.172413793103445, | |
| "grad_norm": 0.07680459320545197, | |
| "learning_rate": 1.9335604770017038e-05, | |
| "loss": 0.02, | |
| "step": 1078 | |
| }, | |
| { | |
| "epoch": 37.241379310344826, | |
| "grad_norm": 0.07265330106019974, | |
| "learning_rate": 1.9165247018739353e-05, | |
| "loss": 0.0228, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 37.310344827586206, | |
| "grad_norm": 0.15032625198364258, | |
| "learning_rate": 1.899488926746167e-05, | |
| "loss": 0.0223, | |
| "step": 1082 | |
| }, | |
| { | |
| "epoch": 37.37931034482759, | |
| "grad_norm": 1.7580552101135254, | |
| "learning_rate": 1.882453151618399e-05, | |
| "loss": 0.0259, | |
| "step": 1084 | |
| }, | |
| { | |
| "epoch": 37.44827586206897, | |
| "grad_norm": 0.07701607793569565, | |
| "learning_rate": 1.8654173764906304e-05, | |
| "loss": 0.0211, | |
| "step": 1086 | |
| }, | |
| { | |
| "epoch": 37.51724137931034, | |
| "grad_norm": 0.13334742188453674, | |
| "learning_rate": 1.8483816013628622e-05, | |
| "loss": 0.0281, | |
| "step": 1088 | |
| }, | |
| { | |
| "epoch": 37.58620689655172, | |
| "grad_norm": 0.08021260797977448, | |
| "learning_rate": 1.8313458262350937e-05, | |
| "loss": 0.0199, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 37.6551724137931, | |
| "grad_norm": 0.06848295032978058, | |
| "learning_rate": 1.8143100511073255e-05, | |
| "loss": 0.021, | |
| "step": 1092 | |
| }, | |
| { | |
| "epoch": 37.724137931034484, | |
| "grad_norm": 0.09606592357158661, | |
| "learning_rate": 1.797274275979557e-05, | |
| "loss": 0.0498, | |
| "step": 1094 | |
| }, | |
| { | |
| "epoch": 37.793103448275865, | |
| "grad_norm": 0.09107226878404617, | |
| "learning_rate": 1.780238500851789e-05, | |
| "loss": 0.022, | |
| "step": 1096 | |
| }, | |
| { | |
| "epoch": 37.86206896551724, | |
| "grad_norm": 0.07415739446878433, | |
| "learning_rate": 1.7632027257240207e-05, | |
| "loss": 0.02, | |
| "step": 1098 | |
| }, | |
| { | |
| "epoch": 37.93103448275862, | |
| "grad_norm": 0.10790450870990753, | |
| "learning_rate": 1.746166950596252e-05, | |
| "loss": 0.0219, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 38.0, | |
| "grad_norm": 0.07875606417655945, | |
| "learning_rate": 1.729131175468484e-05, | |
| "loss": 0.0197, | |
| "step": 1102 | |
| }, | |
| { | |
| "epoch": 38.0, | |
| "eval_accuracy": 0.6515151515151515, | |
| "eval_f1_macro": 0.5636123898806141, | |
| "eval_f1_micro": 0.6515151515151515, | |
| "eval_f1_weighted": 0.6385926013380662, | |
| "eval_loss": 1.4955607652664185, | |
| "eval_precision_macro": 0.5590277777777778, | |
| "eval_precision_micro": 0.6515151515151515, | |
| "eval_precision_weighted": 0.6296033249158249, | |
| "eval_recall_macro": 0.5713756613756613, | |
| "eval_recall_micro": 0.6515151515151515, | |
| "eval_recall_weighted": 0.6515151515151515, | |
| "eval_runtime": 2.3054, | |
| "eval_samples_per_second": 57.256, | |
| "eval_steps_per_second": 7.374, | |
| "step": 1102 | |
| }, | |
| { | |
| "epoch": 38.06896551724138, | |
| "grad_norm": 0.10304208844900131, | |
| "learning_rate": 1.7120954003407155e-05, | |
| "loss": 0.0228, | |
| "step": 1104 | |
| }, | |
| { | |
| "epoch": 38.13793103448276, | |
| "grad_norm": 0.07519730180501938, | |
| "learning_rate": 1.6950596252129473e-05, | |
| "loss": 0.0199, | |
| "step": 1106 | |
| }, | |
| { | |
| "epoch": 38.206896551724135, | |
| "grad_norm": 0.07089941203594208, | |
| "learning_rate": 1.6780238500851788e-05, | |
| "loss": 0.0216, | |
| "step": 1108 | |
| }, | |
| { | |
| "epoch": 38.275862068965516, | |
| "grad_norm": 0.06765555590391159, | |
| "learning_rate": 1.6609880749574106e-05, | |
| "loss": 0.043, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 38.3448275862069, | |
| "grad_norm": 0.07931018620729446, | |
| "learning_rate": 1.643952299829642e-05, | |
| "loss": 0.0237, | |
| "step": 1112 | |
| }, | |
| { | |
| "epoch": 38.41379310344828, | |
| "grad_norm": 0.46399685740470886, | |
| "learning_rate": 1.626916524701874e-05, | |
| "loss": 0.0493, | |
| "step": 1114 | |
| }, | |
| { | |
| "epoch": 38.48275862068966, | |
| "grad_norm": 0.07535680383443832, | |
| "learning_rate": 1.6098807495741057e-05, | |
| "loss": 0.0206, | |
| "step": 1116 | |
| }, | |
| { | |
| "epoch": 38.55172413793103, | |
| "grad_norm": 0.07040387392044067, | |
| "learning_rate": 1.5928449744463375e-05, | |
| "loss": 0.0207, | |
| "step": 1118 | |
| }, | |
| { | |
| "epoch": 38.62068965517241, | |
| "grad_norm": 0.06553125381469727, | |
| "learning_rate": 1.5758091993185693e-05, | |
| "loss": 0.0203, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 38.689655172413794, | |
| "grad_norm": 0.10302378237247467, | |
| "learning_rate": 1.5587734241908008e-05, | |
| "loss": 0.02, | |
| "step": 1122 | |
| }, | |
| { | |
| "epoch": 38.758620689655174, | |
| "grad_norm": 4.081165790557861, | |
| "learning_rate": 1.5417376490630326e-05, | |
| "loss": 0.0323, | |
| "step": 1124 | |
| }, | |
| { | |
| "epoch": 38.827586206896555, | |
| "grad_norm": 0.435150682926178, | |
| "learning_rate": 1.5247018739352641e-05, | |
| "loss": 0.044, | |
| "step": 1126 | |
| }, | |
| { | |
| "epoch": 38.89655172413793, | |
| "grad_norm": 0.07893764227628708, | |
| "learning_rate": 1.507666098807496e-05, | |
| "loss": 0.0192, | |
| "step": 1128 | |
| }, | |
| { | |
| "epoch": 38.96551724137931, | |
| "grad_norm": 0.0766209065914154, | |
| "learning_rate": 1.4906303236797276e-05, | |
| "loss": 0.0219, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 39.0, | |
| "eval_accuracy": 0.6590909090909091, | |
| "eval_f1_macro": 0.5981099582690218, | |
| "eval_f1_micro": 0.6590909090909091, | |
| "eval_f1_weighted": 0.6540264969624213, | |
| "eval_loss": 1.4909534454345703, | |
| "eval_precision_macro": 0.6063481453187336, | |
| "eval_precision_micro": 0.6590909090909091, | |
| "eval_precision_weighted": 0.6553739264300761, | |
| "eval_recall_macro": 0.5969538926681783, | |
| "eval_recall_micro": 0.6590909090909091, | |
| "eval_recall_weighted": 0.6590909090909091, | |
| "eval_runtime": 2.3146, | |
| "eval_samples_per_second": 57.029, | |
| "eval_steps_per_second": 7.345, | |
| "step": 1131 | |
| }, | |
| { | |
| "epoch": 39.03448275862069, | |
| "grad_norm": 0.0769796147942543, | |
| "learning_rate": 1.4735945485519592e-05, | |
| "loss": 0.0194, | |
| "step": 1132 | |
| }, | |
| { | |
| "epoch": 39.10344827586207, | |
| "grad_norm": 0.07249811291694641, | |
| "learning_rate": 1.4565587734241909e-05, | |
| "loss": 0.0198, | |
| "step": 1134 | |
| }, | |
| { | |
| "epoch": 39.172413793103445, | |
| "grad_norm": 0.07959507405757904, | |
| "learning_rate": 1.4395229982964225e-05, | |
| "loss": 0.022, | |
| "step": 1136 | |
| }, | |
| { | |
| "epoch": 39.241379310344826, | |
| "grad_norm": 0.44026121497154236, | |
| "learning_rate": 1.4224872231686542e-05, | |
| "loss": 0.0457, | |
| "step": 1138 | |
| }, | |
| { | |
| "epoch": 39.310344827586206, | |
| "grad_norm": 0.072527214884758, | |
| "learning_rate": 1.4054514480408858e-05, | |
| "loss": 0.0208, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 39.37931034482759, | |
| "grad_norm": 0.22599418461322784, | |
| "learning_rate": 1.3884156729131177e-05, | |
| "loss": 0.0199, | |
| "step": 1142 | |
| }, | |
| { | |
| "epoch": 39.44827586206897, | |
| "grad_norm": 0.07740206271409988, | |
| "learning_rate": 1.3713798977853493e-05, | |
| "loss": 0.0195, | |
| "step": 1144 | |
| }, | |
| { | |
| "epoch": 39.51724137931034, | |
| "grad_norm": 0.07508723437786102, | |
| "learning_rate": 1.354344122657581e-05, | |
| "loss": 0.0188, | |
| "step": 1146 | |
| }, | |
| { | |
| "epoch": 39.58620689655172, | |
| "grad_norm": 0.07625211030244827, | |
| "learning_rate": 1.3373083475298126e-05, | |
| "loss": 0.0231, | |
| "step": 1148 | |
| }, | |
| { | |
| "epoch": 39.6551724137931, | |
| "grad_norm": 0.0625428780913353, | |
| "learning_rate": 1.3202725724020443e-05, | |
| "loss": 0.0184, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 39.724137931034484, | |
| "grad_norm": 0.06486790627241135, | |
| "learning_rate": 1.303236797274276e-05, | |
| "loss": 0.0185, | |
| "step": 1152 | |
| }, | |
| { | |
| "epoch": 39.793103448275865, | |
| "grad_norm": 0.09792531281709671, | |
| "learning_rate": 1.2862010221465076e-05, | |
| "loss": 0.0223, | |
| "step": 1154 | |
| }, | |
| { | |
| "epoch": 39.86206896551724, | |
| "grad_norm": 0.5140519142150879, | |
| "learning_rate": 1.2691652470187392e-05, | |
| "loss": 0.052, | |
| "step": 1156 | |
| }, | |
| { | |
| "epoch": 39.93103448275862, | |
| "grad_norm": 0.10591938346624374, | |
| "learning_rate": 1.252129471890971e-05, | |
| "loss": 0.0198, | |
| "step": 1158 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "grad_norm": 0.08880186080932617, | |
| "learning_rate": 1.2350936967632027e-05, | |
| "loss": 0.0212, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "eval_accuracy": 0.6515151515151515, | |
| "eval_f1_macro": 0.5911823960353372, | |
| "eval_f1_micro": 0.6515151515151515, | |
| "eval_f1_weighted": 0.6461534176607706, | |
| "eval_loss": 1.5049692392349243, | |
| "eval_precision_macro": 0.5996689473237461, | |
| "eval_precision_micro": 0.6515151515151515, | |
| "eval_precision_weighted": 0.6472166962333723, | |
| "eval_recall_macro": 0.5898110355253213, | |
| "eval_recall_micro": 0.6515151515151515, | |
| "eval_recall_weighted": 0.6515151515151515, | |
| "eval_runtime": 2.3355, | |
| "eval_samples_per_second": 56.519, | |
| "eval_steps_per_second": 7.279, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 40.06896551724138, | |
| "grad_norm": 0.06325220316648483, | |
| "learning_rate": 1.2180579216354345e-05, | |
| "loss": 0.0191, | |
| "step": 1162 | |
| }, | |
| { | |
| "epoch": 40.13793103448276, | |
| "grad_norm": 0.0681692361831665, | |
| "learning_rate": 1.2010221465076662e-05, | |
| "loss": 0.0181, | |
| "step": 1164 | |
| }, | |
| { | |
| "epoch": 40.206896551724135, | |
| "grad_norm": 0.14537657797336578, | |
| "learning_rate": 1.1839863713798978e-05, | |
| "loss": 0.0198, | |
| "step": 1166 | |
| }, | |
| { | |
| "epoch": 40.275862068965516, | |
| "grad_norm": 0.060531288385391235, | |
| "learning_rate": 1.1669505962521295e-05, | |
| "loss": 0.0177, | |
| "step": 1168 | |
| }, | |
| { | |
| "epoch": 40.3448275862069, | |
| "grad_norm": 0.07529687881469727, | |
| "learning_rate": 1.1499148211243613e-05, | |
| "loss": 0.0212, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 40.41379310344828, | |
| "grad_norm": 0.37628257274627686, | |
| "learning_rate": 1.132879045996593e-05, | |
| "loss": 0.0398, | |
| "step": 1172 | |
| }, | |
| { | |
| "epoch": 40.48275862068966, | |
| "grad_norm": 0.15280403196811676, | |
| "learning_rate": 1.1158432708688246e-05, | |
| "loss": 0.0208, | |
| "step": 1174 | |
| }, | |
| { | |
| "epoch": 40.55172413793103, | |
| "grad_norm": 0.06886907666921616, | |
| "learning_rate": 1.0988074957410562e-05, | |
| "loss": 0.0213, | |
| "step": 1176 | |
| }, | |
| { | |
| "epoch": 40.62068965517241, | |
| "grad_norm": 0.481803834438324, | |
| "learning_rate": 1.0817717206132879e-05, | |
| "loss": 0.0478, | |
| "step": 1178 | |
| }, | |
| { | |
| "epoch": 40.689655172413794, | |
| "grad_norm": 0.06786196678876877, | |
| "learning_rate": 1.0647359454855197e-05, | |
| "loss": 0.02, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 40.758620689655174, | |
| "grad_norm": 0.07694634050130844, | |
| "learning_rate": 1.0477001703577514e-05, | |
| "loss": 0.0196, | |
| "step": 1182 | |
| }, | |
| { | |
| "epoch": 40.827586206896555, | |
| "grad_norm": 0.06034290790557861, | |
| "learning_rate": 1.030664395229983e-05, | |
| "loss": 0.0209, | |
| "step": 1184 | |
| }, | |
| { | |
| "epoch": 40.89655172413793, | |
| "grad_norm": 0.06212831288576126, | |
| "learning_rate": 1.0136286201022147e-05, | |
| "loss": 0.0189, | |
| "step": 1186 | |
| }, | |
| { | |
| "epoch": 40.96551724137931, | |
| "grad_norm": 0.07599721103906631, | |
| "learning_rate": 9.965928449744463e-06, | |
| "loss": 0.0212, | |
| "step": 1188 | |
| }, | |
| { | |
| "epoch": 41.0, | |
| "eval_accuracy": 0.6590909090909091, | |
| "eval_f1_macro": 0.5976809954751132, | |
| "eval_f1_micro": 0.6590909090909091, | |
| "eval_f1_weighted": 0.6537291752365282, | |
| "eval_loss": 1.5091404914855957, | |
| "eval_precision_macro": 0.6079748081296069, | |
| "eval_precision_micro": 0.6590909090909091, | |
| "eval_precision_weighted": 0.6557976985643745, | |
| "eval_recall_macro": 0.5955253212396069, | |
| "eval_recall_micro": 0.6590909090909091, | |
| "eval_recall_weighted": 0.6590909090909091, | |
| "eval_runtime": 2.3163, | |
| "eval_samples_per_second": 56.988, | |
| "eval_steps_per_second": 7.339, | |
| "step": 1189 | |
| }, | |
| { | |
| "epoch": 41.03448275862069, | |
| "grad_norm": 0.07553800195455551, | |
| "learning_rate": 9.795570698466781e-06, | |
| "loss": 0.0202, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 41.10344827586207, | |
| "grad_norm": 0.08795283734798431, | |
| "learning_rate": 9.625212947189098e-06, | |
| "loss": 0.0205, | |
| "step": 1192 | |
| }, | |
| { | |
| "epoch": 41.172413793103445, | |
| "grad_norm": 0.08366087824106216, | |
| "learning_rate": 9.454855195911414e-06, | |
| "loss": 0.0202, | |
| "step": 1194 | |
| }, | |
| { | |
| "epoch": 41.241379310344826, | |
| "grad_norm": 0.5255555510520935, | |
| "learning_rate": 9.28449744463373e-06, | |
| "loss": 0.0414, | |
| "step": 1196 | |
| }, | |
| { | |
| "epoch": 41.310344827586206, | |
| "grad_norm": 0.37184038758277893, | |
| "learning_rate": 9.114139693356049e-06, | |
| "loss": 0.0384, | |
| "step": 1198 | |
| }, | |
| { | |
| "epoch": 41.37931034482759, | |
| "grad_norm": 0.07240649312734604, | |
| "learning_rate": 8.943781942078366e-06, | |
| "loss": 0.0203, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 41.44827586206897, | |
| "grad_norm": 0.07140922546386719, | |
| "learning_rate": 8.773424190800682e-06, | |
| "loss": 0.0179, | |
| "step": 1202 | |
| }, | |
| { | |
| "epoch": 41.51724137931034, | |
| "grad_norm": 1.3324774503707886, | |
| "learning_rate": 8.603066439522999e-06, | |
| "loss": 0.025, | |
| "step": 1204 | |
| }, | |
| { | |
| "epoch": 41.58620689655172, | |
| "grad_norm": 0.0719638466835022, | |
| "learning_rate": 8.432708688245315e-06, | |
| "loss": 0.0196, | |
| "step": 1206 | |
| }, | |
| { | |
| "epoch": 41.6551724137931, | |
| "grad_norm": 0.06998977810144424, | |
| "learning_rate": 8.262350936967632e-06, | |
| "loss": 0.0208, | |
| "step": 1208 | |
| }, | |
| { | |
| "epoch": 41.724137931034484, | |
| "grad_norm": 0.06412280350923538, | |
| "learning_rate": 8.091993185689948e-06, | |
| "loss": 0.0207, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 41.793103448275865, | |
| "grad_norm": 0.06555226445198059, | |
| "learning_rate": 7.921635434412266e-06, | |
| "loss": 0.0202, | |
| "step": 1212 | |
| }, | |
| { | |
| "epoch": 41.86206896551724, | |
| "grad_norm": 0.06935158371925354, | |
| "learning_rate": 7.751277683134583e-06, | |
| "loss": 0.018, | |
| "step": 1214 | |
| }, | |
| { | |
| "epoch": 41.93103448275862, | |
| "grad_norm": 0.1250106245279312, | |
| "learning_rate": 7.5809199318569e-06, | |
| "loss": 0.019, | |
| "step": 1216 | |
| }, | |
| { | |
| "epoch": 42.0, | |
| "grad_norm": 0.06154052913188934, | |
| "learning_rate": 7.410562180579217e-06, | |
| "loss": 0.0202, | |
| "step": 1218 | |
| }, | |
| { | |
| "epoch": 42.0, | |
| "eval_accuracy": 0.6515151515151515, | |
| "eval_f1_macro": 0.5654648815113407, | |
| "eval_f1_micro": 0.6515151515151515, | |
| "eval_f1_weighted": 0.6411283473548656, | |
| "eval_loss": 1.4960678815841675, | |
| "eval_precision_macro": 0.5707833560774738, | |
| "eval_precision_micro": 0.6515151515151515, | |
| "eval_precision_weighted": 0.6411209251895527, | |
| "eval_recall_macro": 0.5694935752078608, | |
| "eval_recall_micro": 0.6515151515151515, | |
| "eval_recall_weighted": 0.6515151515151515, | |
| "eval_runtime": 2.3136, | |
| "eval_samples_per_second": 57.055, | |
| "eval_steps_per_second": 7.348, | |
| "step": 1218 | |
| }, | |
| { | |
| "epoch": 42.06896551724138, | |
| "grad_norm": 0.10921443998813629, | |
| "learning_rate": 7.240204429301534e-06, | |
| "loss": 0.0216, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 42.13793103448276, | |
| "grad_norm": 0.08505412936210632, | |
| "learning_rate": 7.0698466780238505e-06, | |
| "loss": 0.0181, | |
| "step": 1222 | |
| }, | |
| { | |
| "epoch": 42.206896551724135, | |
| "grad_norm": 0.06461058557033539, | |
| "learning_rate": 6.899488926746167e-06, | |
| "loss": 0.0437, | |
| "step": 1224 | |
| }, | |
| { | |
| "epoch": 42.275862068965516, | |
| "grad_norm": 0.0641995221376419, | |
| "learning_rate": 6.7291311754684835e-06, | |
| "loss": 0.0181, | |
| "step": 1226 | |
| }, | |
| { | |
| "epoch": 42.3448275862069, | |
| "grad_norm": 0.07062099874019623, | |
| "learning_rate": 6.558773424190801e-06, | |
| "loss": 0.0195, | |
| "step": 1228 | |
| }, | |
| { | |
| "epoch": 42.41379310344828, | |
| "grad_norm": 0.06934709846973419, | |
| "learning_rate": 6.388415672913117e-06, | |
| "loss": 0.0197, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 42.48275862068966, | |
| "grad_norm": 0.07409219443798065, | |
| "learning_rate": 6.218057921635435e-06, | |
| "loss": 0.0195, | |
| "step": 1232 | |
| }, | |
| { | |
| "epoch": 42.55172413793103, | |
| "grad_norm": 0.07015547901391983, | |
| "learning_rate": 6.047700170357751e-06, | |
| "loss": 0.0478, | |
| "step": 1234 | |
| }, | |
| { | |
| "epoch": 42.62068965517241, | |
| "grad_norm": 0.06233002990484238, | |
| "learning_rate": 5.877342419080068e-06, | |
| "loss": 0.0181, | |
| "step": 1236 | |
| }, | |
| { | |
| "epoch": 42.689655172413794, | |
| "grad_norm": 0.07605413347482681, | |
| "learning_rate": 5.706984667802386e-06, | |
| "loss": 0.0195, | |
| "step": 1238 | |
| }, | |
| { | |
| "epoch": 42.758620689655174, | |
| "grad_norm": 0.06724195927381516, | |
| "learning_rate": 5.5366269165247025e-06, | |
| "loss": 0.0197, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 42.827586206896555, | |
| "grad_norm": 0.0727328211069107, | |
| "learning_rate": 5.366269165247019e-06, | |
| "loss": 0.0199, | |
| "step": 1242 | |
| }, | |
| { | |
| "epoch": 42.89655172413793, | |
| "grad_norm": 0.061808932572603226, | |
| "learning_rate": 5.1959114139693355e-06, | |
| "loss": 0.0181, | |
| "step": 1244 | |
| }, | |
| { | |
| "epoch": 42.96551724137931, | |
| "grad_norm": 0.0680130198597908, | |
| "learning_rate": 5.025553662691653e-06, | |
| "loss": 0.0216, | |
| "step": 1246 | |
| }, | |
| { | |
| "epoch": 43.0, | |
| "eval_accuracy": 0.6515151515151515, | |
| "eval_f1_macro": 0.5654648815113407, | |
| "eval_f1_micro": 0.6515151515151515, | |
| "eval_f1_weighted": 0.6411283473548656, | |
| "eval_loss": 1.491672396659851, | |
| "eval_precision_macro": 0.5707833560774738, | |
| "eval_precision_micro": 0.6515151515151515, | |
| "eval_precision_weighted": 0.6411209251895527, | |
| "eval_recall_macro": 0.5694935752078608, | |
| "eval_recall_micro": 0.6515151515151515, | |
| "eval_recall_weighted": 0.6515151515151515, | |
| "eval_runtime": 2.3103, | |
| "eval_samples_per_second": 57.136, | |
| "eval_steps_per_second": 7.358, | |
| "step": 1247 | |
| }, | |
| { | |
| "epoch": 43.03448275862069, | |
| "grad_norm": 0.06223515793681145, | |
| "learning_rate": 4.85519591141397e-06, | |
| "loss": 0.0179, | |
| "step": 1248 | |
| }, | |
| { | |
| "epoch": 43.10344827586207, | |
| "grad_norm": 0.0637139230966568, | |
| "learning_rate": 4.684838160136287e-06, | |
| "loss": 0.0196, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 43.172413793103445, | |
| "grad_norm": 0.06648910790681839, | |
| "learning_rate": 4.514480408858603e-06, | |
| "loss": 0.0376, | |
| "step": 1252 | |
| }, | |
| { | |
| "epoch": 43.241379310344826, | |
| "grad_norm": 0.08096196502447128, | |
| "learning_rate": 4.34412265758092e-06, | |
| "loss": 0.045, | |
| "step": 1254 | |
| }, | |
| { | |
| "epoch": 43.310344827586206, | |
| "grad_norm": 0.0885854959487915, | |
| "learning_rate": 4.173764906303237e-06, | |
| "loss": 0.0206, | |
| "step": 1256 | |
| }, | |
| { | |
| "epoch": 43.37931034482759, | |
| "grad_norm": 0.0679527223110199, | |
| "learning_rate": 4.0034071550255544e-06, | |
| "loss": 0.0185, | |
| "step": 1258 | |
| }, | |
| { | |
| "epoch": 43.44827586206897, | |
| "grad_norm": 0.07362633943557739, | |
| "learning_rate": 3.833049403747871e-06, | |
| "loss": 0.02, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 43.51724137931034, | |
| "grad_norm": 0.0694928914308548, | |
| "learning_rate": 3.6626916524701875e-06, | |
| "loss": 0.0196, | |
| "step": 1262 | |
| }, | |
| { | |
| "epoch": 43.58620689655172, | |
| "grad_norm": 0.06688851863145828, | |
| "learning_rate": 3.4923339011925044e-06, | |
| "loss": 0.0179, | |
| "step": 1264 | |
| }, | |
| { | |
| "epoch": 43.6551724137931, | |
| "grad_norm": 0.11848729103803635, | |
| "learning_rate": 3.321976149914821e-06, | |
| "loss": 0.0219, | |
| "step": 1266 | |
| }, | |
| { | |
| "epoch": 43.724137931034484, | |
| "grad_norm": 0.9978265762329102, | |
| "learning_rate": 3.151618398637138e-06, | |
| "loss": 0.0194, | |
| "step": 1268 | |
| }, | |
| { | |
| "epoch": 43.793103448275865, | |
| "grad_norm": 0.06428433954715729, | |
| "learning_rate": 2.9812606473594548e-06, | |
| "loss": 0.0182, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 43.86206896551724, | |
| "grad_norm": 0.20028172433376312, | |
| "learning_rate": 2.810902896081772e-06, | |
| "loss": 0.0195, | |
| "step": 1272 | |
| }, | |
| { | |
| "epoch": 43.93103448275862, | |
| "grad_norm": 0.07131318747997284, | |
| "learning_rate": 2.6405451448040886e-06, | |
| "loss": 0.0183, | |
| "step": 1274 | |
| }, | |
| { | |
| "epoch": 44.0, | |
| "grad_norm": 0.13960784673690796, | |
| "learning_rate": 2.4701873935264056e-06, | |
| "loss": 0.0199, | |
| "step": 1276 | |
| }, | |
| { | |
| "epoch": 44.0, | |
| "eval_accuracy": 0.6515151515151515, | |
| "eval_f1_macro": 0.5674399966365533, | |
| "eval_f1_micro": 0.6515151515151515, | |
| "eval_f1_weighted": 0.6423054871360109, | |
| "eval_loss": 1.485514760017395, | |
| "eval_precision_macro": 0.5694260552951398, | |
| "eval_precision_micro": 0.6515151515151515, | |
| "eval_precision_weighted": 0.6400870203536964, | |
| "eval_recall_macro": 0.5717157974300832, | |
| "eval_recall_micro": 0.6515151515151515, | |
| "eval_recall_weighted": 0.6515151515151515, | |
| "eval_runtime": 2.3079, | |
| "eval_samples_per_second": 57.194, | |
| "eval_steps_per_second": 7.366, | |
| "step": 1276 | |
| }, | |
| { | |
| "epoch": 44.06896551724138, | |
| "grad_norm": 0.09822448343038559, | |
| "learning_rate": 2.2998296422487225e-06, | |
| "loss": 0.0178, | |
| "step": 1278 | |
| }, | |
| { | |
| "epoch": 44.13793103448276, | |
| "grad_norm": 0.0626293197274208, | |
| "learning_rate": 2.1294718909710394e-06, | |
| "loss": 0.0185, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 44.206896551724135, | |
| "grad_norm": 0.08184079080820084, | |
| "learning_rate": 1.9591141396933563e-06, | |
| "loss": 0.0178, | |
| "step": 1282 | |
| }, | |
| { | |
| "epoch": 44.275862068965516, | |
| "grad_norm": 0.06804870814085007, | |
| "learning_rate": 1.788756388415673e-06, | |
| "loss": 0.0178, | |
| "step": 1284 | |
| }, | |
| { | |
| "epoch": 44.3448275862069, | |
| "grad_norm": 0.07142641395330429, | |
| "learning_rate": 1.6183986371379898e-06, | |
| "loss": 0.0215, | |
| "step": 1286 | |
| }, | |
| { | |
| "epoch": 44.41379310344828, | |
| "grad_norm": 0.06901556253433228, | |
| "learning_rate": 1.4480408858603067e-06, | |
| "loss": 0.0186, | |
| "step": 1288 | |
| }, | |
| { | |
| "epoch": 44.48275862068966, | |
| "grad_norm": 0.07577712833881378, | |
| "learning_rate": 1.2776831345826234e-06, | |
| "loss": 0.0207, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 44.55172413793103, | |
| "grad_norm": 0.0839504599571228, | |
| "learning_rate": 1.1073253833049404e-06, | |
| "loss": 0.0213, | |
| "step": 1292 | |
| }, | |
| { | |
| "epoch": 44.62068965517241, | |
| "grad_norm": 0.11831686645746231, | |
| "learning_rate": 9.369676320272573e-07, | |
| "loss": 0.0206, | |
| "step": 1294 | |
| }, | |
| { | |
| "epoch": 44.689655172413794, | |
| "grad_norm": 0.4428146183490753, | |
| "learning_rate": 7.666098807495741e-07, | |
| "loss": 0.0697, | |
| "step": 1296 | |
| }, | |
| { | |
| "epoch": 44.758620689655174, | |
| "grad_norm": 0.05690554529428482, | |
| "learning_rate": 5.962521294718911e-07, | |
| "loss": 0.0182, | |
| "step": 1298 | |
| }, | |
| { | |
| "epoch": 44.827586206896555, | |
| "grad_norm": 0.06728281080722809, | |
| "learning_rate": 4.2589437819420783e-07, | |
| "loss": 0.0169, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 44.89655172413793, | |
| "grad_norm": 0.06453272700309753, | |
| "learning_rate": 2.555366269165247e-07, | |
| "loss": 0.0189, | |
| "step": 1302 | |
| }, | |
| { | |
| "epoch": 44.96551724137931, | |
| "grad_norm": 0.06359589844942093, | |
| "learning_rate": 8.517887563884157e-08, | |
| "loss": 0.027, | |
| "step": 1304 | |
| }, | |
| { | |
| "epoch": 45.0, | |
| "eval_accuracy": 0.6515151515151515, | |
| "eval_f1_macro": 0.5674399966365533, | |
| "eval_f1_micro": 0.6515151515151515, | |
| "eval_f1_weighted": 0.6423054871360109, | |
| "eval_loss": 1.4831533432006836, | |
| "eval_precision_macro": 0.5694260552951398, | |
| "eval_precision_micro": 0.6515151515151515, | |
| "eval_precision_weighted": 0.6400870203536964, | |
| "eval_recall_macro": 0.5717157974300832, | |
| "eval_recall_micro": 0.6515151515151515, | |
| "eval_recall_weighted": 0.6515151515151515, | |
| "eval_runtime": 2.3557, | |
| "eval_samples_per_second": 56.034, | |
| "eval_steps_per_second": 7.217, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 45.0, | |
| "step": 1305, | |
| "total_flos": 1.611130661124526e+18, | |
| "train_loss": 0.5009927850430724, | |
| "train_runtime": 1277.117, | |
| "train_samples_per_second": 16.279, | |
| "train_steps_per_second": 1.022 | |
| } | |
| ], | |
| "logging_steps": 2, | |
| "max_steps": 1305, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 45, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.611130661124526e+18, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |