{ "best_metric": 0.9781976744186046, "best_model_checkpoint": "mobilenet_v2_1.0_224-finetuned-plantdisease/checkpoint-2614", "epoch": 19.965576592082616, "eval_steps": 500, "global_step": 2900, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.06884681583476764, "grad_norm": 5.361344337463379, "learning_rate": 1.724137931034483e-06, "loss": 2.7576, "step": 10 }, { "epoch": 0.13769363166953527, "grad_norm": 5.03896427154541, "learning_rate": 3.448275862068966e-06, "loss": 2.7085, "step": 20 }, { "epoch": 0.20654044750430292, "grad_norm": 5.099004745483398, "learning_rate": 5.172413793103448e-06, "loss": 2.7048, "step": 30 }, { "epoch": 0.27538726333907054, "grad_norm": 5.294215202331543, "learning_rate": 6.896551724137932e-06, "loss": 2.6637, "step": 40 }, { "epoch": 0.3442340791738382, "grad_norm": 5.142210483551025, "learning_rate": 8.620689655172414e-06, "loss": 2.6127, "step": 50 }, { "epoch": 0.41308089500860584, "grad_norm": 5.095492839813232, "learning_rate": 1.0344827586206897e-05, "loss": 2.5458, "step": 60 }, { "epoch": 0.4819277108433735, "grad_norm": 4.894885063171387, "learning_rate": 1.206896551724138e-05, "loss": 2.4968, "step": 70 }, { "epoch": 0.5507745266781411, "grad_norm": 4.9816365242004395, "learning_rate": 1.3793103448275863e-05, "loss": 2.4015, "step": 80 }, { "epoch": 0.6196213425129088, "grad_norm": 4.671796798706055, "learning_rate": 1.5517241379310346e-05, "loss": 2.3308, "step": 90 }, { "epoch": 0.6884681583476764, "grad_norm": 4.635168552398682, "learning_rate": 1.7241379310344828e-05, "loss": 2.2212, "step": 100 }, { "epoch": 0.7573149741824441, "grad_norm": 4.575173854827881, "learning_rate": 1.896551724137931e-05, "loss": 2.1272, "step": 110 }, { "epoch": 0.8261617900172117, "grad_norm": 4.4043660163879395, "learning_rate": 2.0689655172413793e-05, "loss": 2.0085, "step": 120 }, { "epoch": 0.8950086058519794, "grad_norm": 4.405401229858398, "learning_rate": 2.2413793103448276e-05, "loss": 1.9442, "step": 130 }, { "epoch": 0.963855421686747, "grad_norm": 4.29350471496582, "learning_rate": 2.413793103448276e-05, "loss": 1.7982, "step": 140 }, { "epoch": 0.9982788296041308, "eval_accuracy": 0.40358527131782945, "eval_loss": 1.9824707508087158, "eval_runtime": 8.0421, "eval_samples_per_second": 256.65, "eval_steps_per_second": 8.082, "step": 145 }, { "epoch": 1.0327022375215147, "grad_norm": 4.268815040588379, "learning_rate": 2.5862068965517244e-05, "loss": 1.7064, "step": 150 }, { "epoch": 1.1015490533562822, "grad_norm": 4.16851806640625, "learning_rate": 2.7586206896551727e-05, "loss": 1.5684, "step": 160 }, { "epoch": 1.1703958691910499, "grad_norm": 4.322150230407715, "learning_rate": 2.9310344827586206e-05, "loss": 1.4844, "step": 170 }, { "epoch": 1.2392426850258176, "grad_norm": 3.958456039428711, "learning_rate": 3.103448275862069e-05, "loss": 1.3968, "step": 180 }, { "epoch": 1.3080895008605853, "grad_norm": 3.7061522006988525, "learning_rate": 3.275862068965517e-05, "loss": 1.2972, "step": 190 }, { "epoch": 1.3769363166953528, "grad_norm": 3.8323745727539062, "learning_rate": 3.4482758620689657e-05, "loss": 1.1544, "step": 200 }, { "epoch": 1.4457831325301205, "grad_norm": 3.52496600151062, "learning_rate": 3.620689655172414e-05, "loss": 1.0645, "step": 210 }, { "epoch": 1.5146299483648882, "grad_norm": 3.744525194168091, "learning_rate": 3.793103448275862e-05, "loss": 1.0238, "step": 220 }, { "epoch": 1.5834767641996557, "grad_norm": 3.494678497314453, "learning_rate": 3.965517241379311e-05, "loss": 0.9237, "step": 230 }, { "epoch": 1.6523235800344234, "grad_norm": 3.6277663707733154, "learning_rate": 4.1379310344827587e-05, "loss": 0.881, "step": 240 }, { "epoch": 1.721170395869191, "grad_norm": 2.9900975227355957, "learning_rate": 4.3103448275862066e-05, "loss": 0.7883, "step": 250 }, { "epoch": 1.7900172117039586, "grad_norm": 3.553596258163452, "learning_rate": 4.482758620689655e-05, "loss": 0.7713, "step": 260 }, { "epoch": 1.8588640275387265, "grad_norm": 3.206496477127075, "learning_rate": 4.655172413793104e-05, "loss": 0.7063, "step": 270 }, { "epoch": 1.927710843373494, "grad_norm": 3.446793556213379, "learning_rate": 4.827586206896552e-05, "loss": 0.6517, "step": 280 }, { "epoch": 1.9965576592082617, "grad_norm": 3.7154784202575684, "learning_rate": 5e-05, "loss": 0.6137, "step": 290 }, { "epoch": 1.9965576592082617, "eval_accuracy": 0.6414728682170543, "eval_loss": 1.1130067110061646, "eval_runtime": 7.6055, "eval_samples_per_second": 271.384, "eval_steps_per_second": 8.546, "step": 290 }, { "epoch": 2.0654044750430294, "grad_norm": 3.542126417160034, "learning_rate": 4.980842911877395e-05, "loss": 0.6058, "step": 300 }, { "epoch": 2.134251290877797, "grad_norm": 3.6798830032348633, "learning_rate": 4.96168582375479e-05, "loss": 0.5967, "step": 310 }, { "epoch": 2.2030981067125643, "grad_norm": 3.2142186164855957, "learning_rate": 4.9425287356321845e-05, "loss": 0.5376, "step": 320 }, { "epoch": 2.2719449225473323, "grad_norm": 3.1692137718200684, "learning_rate": 4.9233716475095786e-05, "loss": 0.5135, "step": 330 }, { "epoch": 2.3407917383820998, "grad_norm": 2.785217523574829, "learning_rate": 4.904214559386973e-05, "loss": 0.4794, "step": 340 }, { "epoch": 2.4096385542168672, "grad_norm": 3.252037763595581, "learning_rate": 4.885057471264368e-05, "loss": 0.5012, "step": 350 }, { "epoch": 2.478485370051635, "grad_norm": 2.9605207443237305, "learning_rate": 4.865900383141763e-05, "loss": 0.4613, "step": 360 }, { "epoch": 2.5473321858864026, "grad_norm": 3.263587474822998, "learning_rate": 4.846743295019157e-05, "loss": 0.4191, "step": 370 }, { "epoch": 2.6161790017211706, "grad_norm": 3.321648120880127, "learning_rate": 4.827586206896552e-05, "loss": 0.4206, "step": 380 }, { "epoch": 2.685025817555938, "grad_norm": 2.7976036071777344, "learning_rate": 4.8084291187739464e-05, "loss": 0.462, "step": 390 }, { "epoch": 2.7538726333907055, "grad_norm": 3.0656702518463135, "learning_rate": 4.789272030651341e-05, "loss": 0.41, "step": 400 }, { "epoch": 2.8227194492254735, "grad_norm": 3.332609176635742, "learning_rate": 4.770114942528736e-05, "loss": 0.4012, "step": 410 }, { "epoch": 2.891566265060241, "grad_norm": 3.5820395946502686, "learning_rate": 4.7509578544061307e-05, "loss": 0.3905, "step": 420 }, { "epoch": 2.960413080895009, "grad_norm": 3.3419365882873535, "learning_rate": 4.7318007662835254e-05, "loss": 0.4176, "step": 430 }, { "epoch": 2.9948364888123926, "eval_accuracy": 0.8468992248062015, "eval_loss": 0.48874419927597046, "eval_runtime": 7.9016, "eval_samples_per_second": 261.213, "eval_steps_per_second": 8.226, "step": 435 }, { "epoch": 3.0292598967297764, "grad_norm": 3.0542681217193604, "learning_rate": 4.7126436781609195e-05, "loss": 0.4024, "step": 440 }, { "epoch": 3.098106712564544, "grad_norm": 2.760667562484741, "learning_rate": 4.693486590038315e-05, "loss": 0.3706, "step": 450 }, { "epoch": 3.1669535283993113, "grad_norm": 3.2174575328826904, "learning_rate": 4.674329501915709e-05, "loss": 0.3509, "step": 460 }, { "epoch": 3.2358003442340793, "grad_norm": 3.6685619354248047, "learning_rate": 4.655172413793104e-05, "loss": 0.3613, "step": 470 }, { "epoch": 3.3046471600688467, "grad_norm": 3.4590651988983154, "learning_rate": 4.6360153256704985e-05, "loss": 0.346, "step": 480 }, { "epoch": 3.3734939759036147, "grad_norm": 3.0743446350097656, "learning_rate": 4.616858237547893e-05, "loss": 0.3356, "step": 490 }, { "epoch": 3.442340791738382, "grad_norm": 3.5646955966949463, "learning_rate": 4.597701149425287e-05, "loss": 0.3404, "step": 500 }, { "epoch": 3.5111876075731496, "grad_norm": 3.2671332359313965, "learning_rate": 4.578544061302682e-05, "loss": 0.3419, "step": 510 }, { "epoch": 3.580034423407917, "grad_norm": 2.7561419010162354, "learning_rate": 4.559386973180077e-05, "loss": 0.3306, "step": 520 }, { "epoch": 3.648881239242685, "grad_norm": 3.1499881744384766, "learning_rate": 4.5402298850574716e-05, "loss": 0.3575, "step": 530 }, { "epoch": 3.7177280550774525, "grad_norm": 2.576195240020752, "learning_rate": 4.5210727969348656e-05, "loss": 0.2954, "step": 540 }, { "epoch": 3.7865748709122204, "grad_norm": 2.685713529586792, "learning_rate": 4.501915708812261e-05, "loss": 0.2971, "step": 550 }, { "epoch": 3.855421686746988, "grad_norm": 3.1608219146728516, "learning_rate": 4.482758620689655e-05, "loss": 0.3057, "step": 560 }, { "epoch": 3.9242685025817554, "grad_norm": 3.3374826908111572, "learning_rate": 4.46360153256705e-05, "loss": 0.2815, "step": 570 }, { "epoch": 3.9931153184165233, "grad_norm": 3.0209603309631348, "learning_rate": 4.4444444444444447e-05, "loss": 0.3107, "step": 580 }, { "epoch": 4.0, "eval_accuracy": 0.8943798449612403, "eval_loss": 0.34142178297042847, "eval_runtime": 7.9804, "eval_samples_per_second": 258.634, "eval_steps_per_second": 8.145, "step": 581 }, { "epoch": 4.061962134251291, "grad_norm": 2.7664377689361572, "learning_rate": 4.4252873563218394e-05, "loss": 0.2689, "step": 590 }, { "epoch": 4.130808950086059, "grad_norm": 2.80204439163208, "learning_rate": 4.406130268199234e-05, "loss": 0.3003, "step": 600 }, { "epoch": 4.199655765920826, "grad_norm": 3.0250327587127686, "learning_rate": 4.386973180076628e-05, "loss": 0.2768, "step": 610 }, { "epoch": 4.268502581755594, "grad_norm": 3.8367512226104736, "learning_rate": 4.367816091954024e-05, "loss": 0.2582, "step": 620 }, { "epoch": 4.337349397590361, "grad_norm": 3.3538951873779297, "learning_rate": 4.348659003831418e-05, "loss": 0.2849, "step": 630 }, { "epoch": 4.406196213425129, "grad_norm": 3.2220115661621094, "learning_rate": 4.3295019157088125e-05, "loss": 0.2744, "step": 640 }, { "epoch": 4.475043029259897, "grad_norm": 3.196542263031006, "learning_rate": 4.3103448275862066e-05, "loss": 0.2744, "step": 650 }, { "epoch": 4.5438898450946645, "grad_norm": 3.07588791847229, "learning_rate": 4.291187739463602e-05, "loss": 0.2766, "step": 660 }, { "epoch": 4.612736660929432, "grad_norm": 3.722421169281006, "learning_rate": 4.272030651340996e-05, "loss": 0.2699, "step": 670 }, { "epoch": 4.6815834767641995, "grad_norm": 2.5164051055908203, "learning_rate": 4.252873563218391e-05, "loss": 0.2401, "step": 680 }, { "epoch": 4.750430292598967, "grad_norm": 3.1823108196258545, "learning_rate": 4.2337164750957856e-05, "loss": 0.2293, "step": 690 }, { "epoch": 4.8192771084337345, "grad_norm": 3.4586544036865234, "learning_rate": 4.21455938697318e-05, "loss": 0.27, "step": 700 }, { "epoch": 4.888123924268503, "grad_norm": 3.251978635787964, "learning_rate": 4.195402298850575e-05, "loss": 0.2551, "step": 710 }, { "epoch": 4.95697074010327, "grad_norm": 2.606138229370117, "learning_rate": 4.17624521072797e-05, "loss": 0.2255, "step": 720 }, { "epoch": 4.998278829604131, "eval_accuracy": 0.9123062015503876, "eval_loss": 0.27319762110710144, "eval_runtime": 7.8624, "eval_samples_per_second": 262.515, "eval_steps_per_second": 8.267, "step": 726 }, { "epoch": 5.025817555938038, "grad_norm": 2.684196949005127, "learning_rate": 4.1570881226053646e-05, "loss": 0.2487, "step": 730 }, { "epoch": 5.094664371772805, "grad_norm": 2.931312084197998, "learning_rate": 4.1379310344827587e-05, "loss": 0.2314, "step": 740 }, { "epoch": 5.163511187607573, "grad_norm": 3.0064966678619385, "learning_rate": 4.1187739463601534e-05, "loss": 0.2293, "step": 750 }, { "epoch": 5.232358003442341, "grad_norm": 3.3251922130584717, "learning_rate": 4.099616858237548e-05, "loss": 0.2388, "step": 760 }, { "epoch": 5.301204819277109, "grad_norm": 1.7967256307601929, "learning_rate": 4.080459770114943e-05, "loss": 0.2169, "step": 770 }, { "epoch": 5.370051635111876, "grad_norm": 3.74770450592041, "learning_rate": 4.061302681992337e-05, "loss": 0.2096, "step": 780 }, { "epoch": 5.438898450946644, "grad_norm": 3.072739601135254, "learning_rate": 4.0421455938697324e-05, "loss": 0.2047, "step": 790 }, { "epoch": 5.507745266781411, "grad_norm": 2.96258807182312, "learning_rate": 4.0229885057471265e-05, "loss": 0.2313, "step": 800 }, { "epoch": 5.576592082616179, "grad_norm": 2.5839080810546875, "learning_rate": 4.003831417624521e-05, "loss": 0.2112, "step": 810 }, { "epoch": 5.645438898450947, "grad_norm": 3.1343889236450195, "learning_rate": 3.984674329501916e-05, "loss": 0.2482, "step": 820 }, { "epoch": 5.714285714285714, "grad_norm": 2.240661382675171, "learning_rate": 3.965517241379311e-05, "loss": 0.2146, "step": 830 }, { "epoch": 5.783132530120482, "grad_norm": 3.1076242923736572, "learning_rate": 3.9463601532567055e-05, "loss": 0.2234, "step": 840 }, { "epoch": 5.851979345955249, "grad_norm": 2.9466423988342285, "learning_rate": 3.9272030651340996e-05, "loss": 0.2301, "step": 850 }, { "epoch": 5.920826161790017, "grad_norm": 3.1047613620758057, "learning_rate": 3.908045977011495e-05, "loss": 0.2202, "step": 860 }, { "epoch": 5.989672977624785, "grad_norm": 2.182670831680298, "learning_rate": 3.888888888888889e-05, "loss": 0.1833, "step": 870 }, { "epoch": 5.9965576592082614, "eval_accuracy": 0.7582364341085271, "eval_loss": 0.7461967468261719, "eval_runtime": 7.912, "eval_samples_per_second": 260.87, "eval_steps_per_second": 8.215, "step": 871 }, { "epoch": 6.058519793459553, "grad_norm": 2.8595638275146484, "learning_rate": 3.869731800766284e-05, "loss": 0.198, "step": 880 }, { "epoch": 6.12736660929432, "grad_norm": 1.812150478363037, "learning_rate": 3.850574712643678e-05, "loss": 0.17, "step": 890 }, { "epoch": 6.196213425129088, "grad_norm": 2.8384931087493896, "learning_rate": 3.831417624521073e-05, "loss": 0.2387, "step": 900 }, { "epoch": 6.265060240963855, "grad_norm": 4.172893047332764, "learning_rate": 3.8122605363984674e-05, "loss": 0.2202, "step": 910 }, { "epoch": 6.333907056798623, "grad_norm": 4.461463928222656, "learning_rate": 3.793103448275862e-05, "loss": 0.2501, "step": 920 }, { "epoch": 6.402753872633391, "grad_norm": 3.430365562438965, "learning_rate": 3.773946360153257e-05, "loss": 0.185, "step": 930 }, { "epoch": 6.4716006884681585, "grad_norm": 3.3742434978485107, "learning_rate": 3.7547892720306517e-05, "loss": 0.2166, "step": 940 }, { "epoch": 6.540447504302926, "grad_norm": 3.7096402645111084, "learning_rate": 3.735632183908046e-05, "loss": 0.234, "step": 950 }, { "epoch": 6.6092943201376935, "grad_norm": 2.182164192199707, "learning_rate": 3.716475095785441e-05, "loss": 0.1852, "step": 960 }, { "epoch": 6.678141135972461, "grad_norm": 2.585125684738159, "learning_rate": 3.697318007662835e-05, "loss": 0.1947, "step": 970 }, { "epoch": 6.746987951807229, "grad_norm": 2.5588345527648926, "learning_rate": 3.67816091954023e-05, "loss": 0.211, "step": 980 }, { "epoch": 6.815834767641997, "grad_norm": 3.0795257091522217, "learning_rate": 3.659003831417625e-05, "loss": 0.1967, "step": 990 }, { "epoch": 6.884681583476764, "grad_norm": 3.7196614742279053, "learning_rate": 3.6398467432950195e-05, "loss": 0.1836, "step": 1000 }, { "epoch": 6.953528399311532, "grad_norm": 2.3857243061065674, "learning_rate": 3.620689655172414e-05, "loss": 0.2062, "step": 1010 }, { "epoch": 6.994836488812393, "eval_accuracy": 0.8803294573643411, "eval_loss": 0.3770907521247864, "eval_runtime": 7.9957, "eval_samples_per_second": 258.14, "eval_steps_per_second": 8.129, "step": 1016 }, { "epoch": 7.022375215146299, "grad_norm": 2.5683236122131348, "learning_rate": 3.601532567049808e-05, "loss": 0.1754, "step": 1020 }, { "epoch": 7.091222030981067, "grad_norm": 2.679636001586914, "learning_rate": 3.582375478927204e-05, "loss": 0.1675, "step": 1030 }, { "epoch": 7.160068846815835, "grad_norm": 2.5351173877716064, "learning_rate": 3.563218390804598e-05, "loss": 0.1565, "step": 1040 }, { "epoch": 7.228915662650603, "grad_norm": 2.9072484970092773, "learning_rate": 3.5440613026819926e-05, "loss": 0.1813, "step": 1050 }, { "epoch": 7.29776247848537, "grad_norm": 2.5954692363739014, "learning_rate": 3.5249042145593867e-05, "loss": 0.2118, "step": 1060 }, { "epoch": 7.366609294320138, "grad_norm": 2.7799787521362305, "learning_rate": 3.505747126436782e-05, "loss": 0.1744, "step": 1070 }, { "epoch": 7.435456110154905, "grad_norm": 3.0674917697906494, "learning_rate": 3.486590038314176e-05, "loss": 0.1922, "step": 1080 }, { "epoch": 7.504302925989673, "grad_norm": 3.5969247817993164, "learning_rate": 3.467432950191571e-05, "loss": 0.1794, "step": 1090 }, { "epoch": 7.573149741824441, "grad_norm": 2.6146047115325928, "learning_rate": 3.4482758620689657e-05, "loss": 0.1682, "step": 1100 }, { "epoch": 7.641996557659208, "grad_norm": 2.785433053970337, "learning_rate": 3.4291187739463604e-05, "loss": 0.1901, "step": 1110 }, { "epoch": 7.710843373493976, "grad_norm": 2.3648314476013184, "learning_rate": 3.409961685823755e-05, "loss": 0.1521, "step": 1120 }, { "epoch": 7.779690189328743, "grad_norm": 2.3298792839050293, "learning_rate": 3.390804597701149e-05, "loss": 0.1901, "step": 1130 }, { "epoch": 7.848537005163511, "grad_norm": 2.2863216400146484, "learning_rate": 3.371647509578545e-05, "loss": 0.1706, "step": 1140 }, { "epoch": 7.917383820998279, "grad_norm": 3.0322484970092773, "learning_rate": 3.352490421455939e-05, "loss": 0.1612, "step": 1150 }, { "epoch": 7.986230636833047, "grad_norm": 2.785752534866333, "learning_rate": 3.3333333333333335e-05, "loss": 0.1657, "step": 1160 }, { "epoch": 8.0, "eval_accuracy": 0.8541666666666666, "eval_loss": 0.47179850935935974, "eval_runtime": 7.8842, "eval_samples_per_second": 261.788, "eval_steps_per_second": 8.244, "step": 1162 }, { "epoch": 8.055077452667813, "grad_norm": 3.0953948497772217, "learning_rate": 3.314176245210728e-05, "loss": 0.1848, "step": 1170 }, { "epoch": 8.123924268502583, "grad_norm": 2.7531402111053467, "learning_rate": 3.295019157088123e-05, "loss": 0.1451, "step": 1180 }, { "epoch": 8.19277108433735, "grad_norm": 3.849405288696289, "learning_rate": 3.275862068965517e-05, "loss": 0.1682, "step": 1190 }, { "epoch": 8.261617900172118, "grad_norm": 2.8468093872070312, "learning_rate": 3.256704980842912e-05, "loss": 0.1868, "step": 1200 }, { "epoch": 8.330464716006885, "grad_norm": 3.0676023960113525, "learning_rate": 3.2375478927203066e-05, "loss": 0.1745, "step": 1210 }, { "epoch": 8.399311531841652, "grad_norm": 2.8244659900665283, "learning_rate": 3.218390804597701e-05, "loss": 0.1783, "step": 1220 }, { "epoch": 8.46815834767642, "grad_norm": 3.3981621265411377, "learning_rate": 3.1992337164750954e-05, "loss": 0.1509, "step": 1230 }, { "epoch": 8.537005163511187, "grad_norm": 3.2597496509552, "learning_rate": 3.180076628352491e-05, "loss": 0.1463, "step": 1240 }, { "epoch": 8.605851979345955, "grad_norm": 2.98665452003479, "learning_rate": 3.160919540229885e-05, "loss": 0.1854, "step": 1250 }, { "epoch": 8.674698795180722, "grad_norm": 3.56426739692688, "learning_rate": 3.1417624521072797e-05, "loss": 0.1801, "step": 1260 }, { "epoch": 8.74354561101549, "grad_norm": 3.350994110107422, "learning_rate": 3.1226053639846744e-05, "loss": 0.146, "step": 1270 }, { "epoch": 8.812392426850257, "grad_norm": 2.637911081314087, "learning_rate": 3.103448275862069e-05, "loss": 0.1759, "step": 1280 }, { "epoch": 8.881239242685027, "grad_norm": 3.2409827709198, "learning_rate": 3.084291187739464e-05, "loss": 0.1897, "step": 1290 }, { "epoch": 8.950086058519794, "grad_norm": 2.5662424564361572, "learning_rate": 3.065134099616858e-05, "loss": 0.1427, "step": 1300 }, { "epoch": 8.99827882960413, "eval_accuracy": 0.8473837209302325, "eval_loss": 0.4902307689189911, "eval_runtime": 7.7511, "eval_samples_per_second": 266.284, "eval_steps_per_second": 8.386, "step": 1307 }, { "epoch": 9.018932874354562, "grad_norm": 2.6389100551605225, "learning_rate": 3.045977011494253e-05, "loss": 0.1577, "step": 1310 }, { "epoch": 9.087779690189329, "grad_norm": 2.561455249786377, "learning_rate": 3.0268199233716475e-05, "loss": 0.1686, "step": 1320 }, { "epoch": 9.156626506024097, "grad_norm": 2.796520233154297, "learning_rate": 3.0076628352490422e-05, "loss": 0.1623, "step": 1330 }, { "epoch": 9.225473321858864, "grad_norm": 3.3344945907592773, "learning_rate": 2.988505747126437e-05, "loss": 0.1684, "step": 1340 }, { "epoch": 9.294320137693632, "grad_norm": 2.494370698928833, "learning_rate": 2.9693486590038317e-05, "loss": 0.1679, "step": 1350 }, { "epoch": 9.363166953528399, "grad_norm": 2.9866812229156494, "learning_rate": 2.950191570881226e-05, "loss": 0.168, "step": 1360 }, { "epoch": 9.432013769363166, "grad_norm": 2.8670761585235596, "learning_rate": 2.9310344827586206e-05, "loss": 0.1587, "step": 1370 }, { "epoch": 9.500860585197934, "grad_norm": 2.6785264015197754, "learning_rate": 2.9118773946360157e-05, "loss": 0.1487, "step": 1380 }, { "epoch": 9.569707401032701, "grad_norm": 3.9908454418182373, "learning_rate": 2.89272030651341e-05, "loss": 0.1913, "step": 1390 }, { "epoch": 9.638554216867469, "grad_norm": 3.4217581748962402, "learning_rate": 2.8735632183908045e-05, "loss": 0.1568, "step": 1400 }, { "epoch": 9.707401032702238, "grad_norm": 3.4280409812927246, "learning_rate": 2.8544061302681996e-05, "loss": 0.1335, "step": 1410 }, { "epoch": 9.776247848537006, "grad_norm": 3.57127046585083, "learning_rate": 2.835249042145594e-05, "loss": 0.1508, "step": 1420 }, { "epoch": 9.845094664371773, "grad_norm": 2.238633394241333, "learning_rate": 2.8160919540229884e-05, "loss": 0.1878, "step": 1430 }, { "epoch": 9.91394148020654, "grad_norm": 2.7804625034332275, "learning_rate": 2.796934865900383e-05, "loss": 0.1497, "step": 1440 }, { "epoch": 9.982788296041308, "grad_norm": 3.1482419967651367, "learning_rate": 2.777777777777778e-05, "loss": 0.1598, "step": 1450 }, { "epoch": 9.996557659208262, "eval_accuracy": 0.9273255813953488, "eval_loss": 0.22291144728660583, "eval_runtime": 7.6884, "eval_samples_per_second": 268.456, "eval_steps_per_second": 8.454, "step": 1452 }, { "epoch": 10.051635111876076, "grad_norm": 2.5692484378814697, "learning_rate": 2.7586206896551727e-05, "loss": 0.1397, "step": 1460 }, { "epoch": 10.120481927710843, "grad_norm": 1.686840295791626, "learning_rate": 2.739463601532567e-05, "loss": 0.1246, "step": 1470 }, { "epoch": 10.18932874354561, "grad_norm": 2.7757647037506104, "learning_rate": 2.720306513409962e-05, "loss": 0.14, "step": 1480 }, { "epoch": 10.258175559380378, "grad_norm": 2.688798189163208, "learning_rate": 2.7011494252873566e-05, "loss": 0.1386, "step": 1490 }, { "epoch": 10.327022375215146, "grad_norm": 3.0354042053222656, "learning_rate": 2.681992337164751e-05, "loss": 0.1585, "step": 1500 }, { "epoch": 10.395869191049915, "grad_norm": 3.36676287651062, "learning_rate": 2.662835249042146e-05, "loss": 0.1368, "step": 1510 }, { "epoch": 10.464716006884682, "grad_norm": 2.3574771881103516, "learning_rate": 2.6436781609195405e-05, "loss": 0.1559, "step": 1520 }, { "epoch": 10.53356282271945, "grad_norm": 2.6529464721679688, "learning_rate": 2.624521072796935e-05, "loss": 0.154, "step": 1530 }, { "epoch": 10.602409638554217, "grad_norm": 3.449251174926758, "learning_rate": 2.6053639846743293e-05, "loss": 0.1401, "step": 1540 }, { "epoch": 10.671256454388985, "grad_norm": 2.449934482574463, "learning_rate": 2.5862068965517244e-05, "loss": 0.1408, "step": 1550 }, { "epoch": 10.740103270223752, "grad_norm": 2.8734378814697266, "learning_rate": 2.5670498084291188e-05, "loss": 0.1445, "step": 1560 }, { "epoch": 10.80895008605852, "grad_norm": 2.4971940517425537, "learning_rate": 2.5478927203065132e-05, "loss": 0.1643, "step": 1570 }, { "epoch": 10.877796901893287, "grad_norm": 3.7385921478271484, "learning_rate": 2.5287356321839083e-05, "loss": 0.1521, "step": 1580 }, { "epoch": 10.946643717728055, "grad_norm": 2.7152345180511475, "learning_rate": 2.5095785440613027e-05, "loss": 0.1504, "step": 1590 }, { "epoch": 10.994836488812393, "eval_accuracy": 0.8972868217054264, "eval_loss": 0.3021294176578522, "eval_runtime": 7.4857, "eval_samples_per_second": 275.726, "eval_steps_per_second": 8.683, "step": 1597 }, { "epoch": 11.015490533562822, "grad_norm": 3.032806158065796, "learning_rate": 2.4904214559386975e-05, "loss": 0.1366, "step": 1600 }, { "epoch": 11.08433734939759, "grad_norm": 2.85953950881958, "learning_rate": 2.4712643678160922e-05, "loss": 0.1433, "step": 1610 }, { "epoch": 11.153184165232357, "grad_norm": 2.1525700092315674, "learning_rate": 2.4521072796934867e-05, "loss": 0.1455, "step": 1620 }, { "epoch": 11.222030981067126, "grad_norm": 2.248845100402832, "learning_rate": 2.4329501915708814e-05, "loss": 0.1269, "step": 1630 }, { "epoch": 11.290877796901894, "grad_norm": 2.333122968673706, "learning_rate": 2.413793103448276e-05, "loss": 0.156, "step": 1640 }, { "epoch": 11.359724612736661, "grad_norm": 2.1817004680633545, "learning_rate": 2.3946360153256706e-05, "loss": 0.1423, "step": 1650 }, { "epoch": 11.428571428571429, "grad_norm": 2.1537561416625977, "learning_rate": 2.3754789272030653e-05, "loss": 0.1327, "step": 1660 }, { "epoch": 11.497418244406196, "grad_norm": 1.6727231740951538, "learning_rate": 2.3563218390804597e-05, "loss": 0.1272, "step": 1670 }, { "epoch": 11.566265060240964, "grad_norm": 3.15240216255188, "learning_rate": 2.3371647509578545e-05, "loss": 0.1474, "step": 1680 }, { "epoch": 11.635111876075731, "grad_norm": 2.031428813934326, "learning_rate": 2.3180076628352492e-05, "loss": 0.1341, "step": 1690 }, { "epoch": 11.703958691910499, "grad_norm": 2.5997140407562256, "learning_rate": 2.2988505747126437e-05, "loss": 0.153, "step": 1700 }, { "epoch": 11.772805507745266, "grad_norm": 2.405869483947754, "learning_rate": 2.2796934865900384e-05, "loss": 0.1434, "step": 1710 }, { "epoch": 11.841652323580034, "grad_norm": 2.8463237285614014, "learning_rate": 2.2605363984674328e-05, "loss": 0.1545, "step": 1720 }, { "epoch": 11.910499139414803, "grad_norm": 2.8656628131866455, "learning_rate": 2.2413793103448276e-05, "loss": 0.1412, "step": 1730 }, { "epoch": 11.97934595524957, "grad_norm": 2.9610302448272705, "learning_rate": 2.2222222222222223e-05, "loss": 0.1456, "step": 1740 }, { "epoch": 12.0, "eval_accuracy": 0.9224806201550387, "eval_loss": 0.24217405915260315, "eval_runtime": 7.2188, "eval_samples_per_second": 285.921, "eval_steps_per_second": 9.004, "step": 1743 }, { "epoch": 12.048192771084338, "grad_norm": 2.278578042984009, "learning_rate": 2.203065134099617e-05, "loss": 0.1079, "step": 1750 }, { "epoch": 12.117039586919105, "grad_norm": 2.1894547939300537, "learning_rate": 2.183908045977012e-05, "loss": 0.1298, "step": 1760 }, { "epoch": 12.185886402753873, "grad_norm": 2.7255043983459473, "learning_rate": 2.1647509578544062e-05, "loss": 0.1502, "step": 1770 }, { "epoch": 12.25473321858864, "grad_norm": 1.9726896286010742, "learning_rate": 2.145593869731801e-05, "loss": 0.1589, "step": 1780 }, { "epoch": 12.323580034423408, "grad_norm": 2.095195770263672, "learning_rate": 2.1264367816091954e-05, "loss": 0.1345, "step": 1790 }, { "epoch": 12.392426850258175, "grad_norm": 2.9772017002105713, "learning_rate": 2.10727969348659e-05, "loss": 0.1537, "step": 1800 }, { "epoch": 12.461273666092943, "grad_norm": 1.4328383207321167, "learning_rate": 2.088122605363985e-05, "loss": 0.1274, "step": 1810 }, { "epoch": 12.53012048192771, "grad_norm": 2.386387586593628, "learning_rate": 2.0689655172413793e-05, "loss": 0.1333, "step": 1820 }, { "epoch": 12.598967297762478, "grad_norm": 1.9341486692428589, "learning_rate": 2.049808429118774e-05, "loss": 0.1397, "step": 1830 }, { "epoch": 12.667814113597245, "grad_norm": 2.633357048034668, "learning_rate": 2.0306513409961685e-05, "loss": 0.1151, "step": 1840 }, { "epoch": 12.736660929432015, "grad_norm": 3.059253215789795, "learning_rate": 2.0114942528735632e-05, "loss": 0.119, "step": 1850 }, { "epoch": 12.805507745266782, "grad_norm": 1.629744291305542, "learning_rate": 1.992337164750958e-05, "loss": 0.1194, "step": 1860 }, { "epoch": 12.87435456110155, "grad_norm": 3.2468857765197754, "learning_rate": 1.9731800766283527e-05, "loss": 0.1567, "step": 1870 }, { "epoch": 12.943201376936317, "grad_norm": 2.7284107208251953, "learning_rate": 1.9540229885057475e-05, "loss": 0.119, "step": 1880 }, { "epoch": 12.99827882960413, "eval_accuracy": 0.9021317829457365, "eval_loss": 0.2836342751979828, "eval_runtime": 7.2838, "eval_samples_per_second": 283.368, "eval_steps_per_second": 8.924, "step": 1888 }, { "epoch": 13.012048192771084, "grad_norm": 3.377997875213623, "learning_rate": 1.934865900383142e-05, "loss": 0.119, "step": 1890 }, { "epoch": 13.080895008605852, "grad_norm": 2.6370134353637695, "learning_rate": 1.9157088122605367e-05, "loss": 0.1394, "step": 1900 }, { "epoch": 13.14974182444062, "grad_norm": 2.5573413372039795, "learning_rate": 1.896551724137931e-05, "loss": 0.1475, "step": 1910 }, { "epoch": 13.218588640275387, "grad_norm": 2.2910749912261963, "learning_rate": 1.8773946360153258e-05, "loss": 0.1332, "step": 1920 }, { "epoch": 13.287435456110154, "grad_norm": 2.3592233657836914, "learning_rate": 1.8582375478927206e-05, "loss": 0.1373, "step": 1930 }, { "epoch": 13.356282271944922, "grad_norm": 2.7366604804992676, "learning_rate": 1.839080459770115e-05, "loss": 0.129, "step": 1940 }, { "epoch": 13.42512908777969, "grad_norm": 2.452705144882202, "learning_rate": 1.8199233716475097e-05, "loss": 0.1183, "step": 1950 }, { "epoch": 13.493975903614459, "grad_norm": 3.3416621685028076, "learning_rate": 1.800766283524904e-05, "loss": 0.154, "step": 1960 }, { "epoch": 13.562822719449226, "grad_norm": 3.04327654838562, "learning_rate": 1.781609195402299e-05, "loss": 0.1264, "step": 1970 }, { "epoch": 13.631669535283994, "grad_norm": 2.6300888061523438, "learning_rate": 1.7624521072796933e-05, "loss": 0.1646, "step": 1980 }, { "epoch": 13.700516351118761, "grad_norm": 3.809067487716675, "learning_rate": 1.743295019157088e-05, "loss": 0.1363, "step": 1990 }, { "epoch": 13.769363166953529, "grad_norm": 3.2865700721740723, "learning_rate": 1.7241379310344828e-05, "loss": 0.1145, "step": 2000 }, { "epoch": 13.838209982788296, "grad_norm": 2.2611865997314453, "learning_rate": 1.7049808429118776e-05, "loss": 0.116, "step": 2010 }, { "epoch": 13.907056798623064, "grad_norm": 2.3190841674804688, "learning_rate": 1.6858237547892723e-05, "loss": 0.1105, "step": 2020 }, { "epoch": 13.975903614457831, "grad_norm": 3.105708360671997, "learning_rate": 1.6666666666666667e-05, "loss": 0.114, "step": 2030 }, { "epoch": 13.996557659208262, "eval_accuracy": 0.9292635658914729, "eval_loss": 0.20378853380680084, "eval_runtime": 7.3214, "eval_samples_per_second": 281.914, "eval_steps_per_second": 8.878, "step": 2033 }, { "epoch": 14.044750430292599, "grad_norm": 2.9859328269958496, "learning_rate": 1.6475095785440615e-05, "loss": 0.1319, "step": 2040 }, { "epoch": 14.113597246127366, "grad_norm": 4.717568874359131, "learning_rate": 1.628352490421456e-05, "loss": 0.1426, "step": 2050 }, { "epoch": 14.182444061962133, "grad_norm": 3.3271358013153076, "learning_rate": 1.6091954022988507e-05, "loss": 0.1458, "step": 2060 }, { "epoch": 14.251290877796903, "grad_norm": 3.017382860183716, "learning_rate": 1.5900383141762454e-05, "loss": 0.1576, "step": 2070 }, { "epoch": 14.32013769363167, "grad_norm": 3.2462213039398193, "learning_rate": 1.5708812260536398e-05, "loss": 0.1438, "step": 2080 }, { "epoch": 14.388984509466438, "grad_norm": 3.2721612453460693, "learning_rate": 1.5517241379310346e-05, "loss": 0.1632, "step": 2090 }, { "epoch": 14.457831325301205, "grad_norm": 2.183432102203369, "learning_rate": 1.532567049808429e-05, "loss": 0.1071, "step": 2100 }, { "epoch": 14.526678141135973, "grad_norm": 2.744126319885254, "learning_rate": 1.5134099616858237e-05, "loss": 0.1234, "step": 2110 }, { "epoch": 14.59552495697074, "grad_norm": 1.959697961807251, "learning_rate": 1.4942528735632185e-05, "loss": 0.1224, "step": 2120 }, { "epoch": 14.664371772805508, "grad_norm": 3.4430899620056152, "learning_rate": 1.475095785440613e-05, "loss": 0.1174, "step": 2130 }, { "epoch": 14.733218588640275, "grad_norm": 4.045992374420166, "learning_rate": 1.4559386973180078e-05, "loss": 0.1315, "step": 2140 }, { "epoch": 14.802065404475043, "grad_norm": 2.5135693550109863, "learning_rate": 1.4367816091954022e-05, "loss": 0.111, "step": 2150 }, { "epoch": 14.87091222030981, "grad_norm": 3.5096664428710938, "learning_rate": 1.417624521072797e-05, "loss": 0.1429, "step": 2160 }, { "epoch": 14.939759036144578, "grad_norm": 2.6633498668670654, "learning_rate": 1.3984674329501916e-05, "loss": 0.1378, "step": 2170 }, { "epoch": 14.994836488812393, "eval_accuracy": 0.9239341085271318, "eval_loss": 0.2173481285572052, "eval_runtime": 7.5899, "eval_samples_per_second": 271.94, "eval_steps_per_second": 8.564, "step": 2178 }, { "epoch": 15.008605851979347, "grad_norm": 2.4078171253204346, "learning_rate": 1.3793103448275863e-05, "loss": 0.1133, "step": 2180 }, { "epoch": 15.077452667814114, "grad_norm": 3.3953912258148193, "learning_rate": 1.360153256704981e-05, "loss": 0.1486, "step": 2190 }, { "epoch": 15.146299483648882, "grad_norm": 2.9477121829986572, "learning_rate": 1.3409961685823755e-05, "loss": 0.1097, "step": 2200 }, { "epoch": 15.21514629948365, "grad_norm": 2.8032114505767822, "learning_rate": 1.3218390804597702e-05, "loss": 0.136, "step": 2210 }, { "epoch": 15.283993115318417, "grad_norm": 3.5413339138031006, "learning_rate": 1.3026819923371647e-05, "loss": 0.1434, "step": 2220 }, { "epoch": 15.352839931153184, "grad_norm": 2.4817466735839844, "learning_rate": 1.2835249042145594e-05, "loss": 0.1178, "step": 2230 }, { "epoch": 15.421686746987952, "grad_norm": 2.3475193977355957, "learning_rate": 1.2643678160919542e-05, "loss": 0.0938, "step": 2240 }, { "epoch": 15.49053356282272, "grad_norm": 3.2757270336151123, "learning_rate": 1.2452107279693487e-05, "loss": 0.159, "step": 2250 }, { "epoch": 15.559380378657487, "grad_norm": 2.7539010047912598, "learning_rate": 1.2260536398467433e-05, "loss": 0.1351, "step": 2260 }, { "epoch": 15.628227194492254, "grad_norm": 2.8831663131713867, "learning_rate": 1.206896551724138e-05, "loss": 0.1341, "step": 2270 }, { "epoch": 15.697074010327022, "grad_norm": 3.276233434677124, "learning_rate": 1.1877394636015327e-05, "loss": 0.1348, "step": 2280 }, { "epoch": 15.76592082616179, "grad_norm": 2.2141618728637695, "learning_rate": 1.1685823754789272e-05, "loss": 0.1166, "step": 2290 }, { "epoch": 15.834767641996558, "grad_norm": 2.0663790702819824, "learning_rate": 1.1494252873563218e-05, "loss": 0.0998, "step": 2300 }, { "epoch": 15.903614457831326, "grad_norm": 2.9078729152679443, "learning_rate": 1.1302681992337164e-05, "loss": 0.1267, "step": 2310 }, { "epoch": 15.972461273666093, "grad_norm": 2.513744831085205, "learning_rate": 1.1111111111111112e-05, "loss": 0.1249, "step": 2320 }, { "epoch": 16.0, "eval_accuracy": 0.9186046511627907, "eval_loss": 0.24669910967350006, "eval_runtime": 7.8261, "eval_samples_per_second": 263.732, "eval_steps_per_second": 8.306, "step": 2324 }, { "epoch": 16.04130808950086, "grad_norm": 2.510911226272583, "learning_rate": 1.091954022988506e-05, "loss": 0.1052, "step": 2330 }, { "epoch": 16.110154905335627, "grad_norm": 2.5557456016540527, "learning_rate": 1.0727969348659005e-05, "loss": 0.1293, "step": 2340 }, { "epoch": 16.179001721170398, "grad_norm": 2.11968994140625, "learning_rate": 1.053639846743295e-05, "loss": 0.1139, "step": 2350 }, { "epoch": 16.247848537005165, "grad_norm": 2.4394736289978027, "learning_rate": 1.0344827586206897e-05, "loss": 0.1206, "step": 2360 }, { "epoch": 16.316695352839933, "grad_norm": 2.758610963821411, "learning_rate": 1.0153256704980842e-05, "loss": 0.1098, "step": 2370 }, { "epoch": 16.3855421686747, "grad_norm": 2.410712480545044, "learning_rate": 9.96168582375479e-06, "loss": 0.1052, "step": 2380 }, { "epoch": 16.454388984509468, "grad_norm": 2.830049991607666, "learning_rate": 9.770114942528738e-06, "loss": 0.11, "step": 2390 }, { "epoch": 16.523235800344235, "grad_norm": 2.921481132507324, "learning_rate": 9.578544061302683e-06, "loss": 0.1038, "step": 2400 }, { "epoch": 16.592082616179002, "grad_norm": 3.5350148677825928, "learning_rate": 9.386973180076629e-06, "loss": 0.1447, "step": 2410 }, { "epoch": 16.66092943201377, "grad_norm": 3.2687947750091553, "learning_rate": 9.195402298850575e-06, "loss": 0.119, "step": 2420 }, { "epoch": 16.729776247848537, "grad_norm": 2.843630790710449, "learning_rate": 9.00383141762452e-06, "loss": 0.1361, "step": 2430 }, { "epoch": 16.798623063683305, "grad_norm": 2.6418073177337646, "learning_rate": 8.812260536398467e-06, "loss": 0.1125, "step": 2440 }, { "epoch": 16.867469879518072, "grad_norm": 2.408013343811035, "learning_rate": 8.620689655172414e-06, "loss": 0.1354, "step": 2450 }, { "epoch": 16.93631669535284, "grad_norm": 3.4269073009490967, "learning_rate": 8.429118773946362e-06, "loss": 0.1504, "step": 2460 }, { "epoch": 16.998278829604132, "eval_accuracy": 0.9253875968992248, "eval_loss": 0.232245072722435, "eval_runtime": 7.9776, "eval_samples_per_second": 258.725, "eval_steps_per_second": 8.148, "step": 2469 }, { "epoch": 17.005163511187607, "grad_norm": 3.9189724922180176, "learning_rate": 8.237547892720307e-06, "loss": 0.1346, "step": 2470 }, { "epoch": 17.074010327022375, "grad_norm": 3.447214126586914, "learning_rate": 8.045977011494253e-06, "loss": 0.1254, "step": 2480 }, { "epoch": 17.142857142857142, "grad_norm": 1.5260127782821655, "learning_rate": 7.854406130268199e-06, "loss": 0.1264, "step": 2490 }, { "epoch": 17.21170395869191, "grad_norm": 2.7160017490386963, "learning_rate": 7.662835249042145e-06, "loss": 0.1124, "step": 2500 }, { "epoch": 17.280550774526677, "grad_norm": 2.530102491378784, "learning_rate": 7.4712643678160925e-06, "loss": 0.1116, "step": 2510 }, { "epoch": 17.349397590361445, "grad_norm": 2.009098768234253, "learning_rate": 7.279693486590039e-06, "loss": 0.0977, "step": 2520 }, { "epoch": 17.418244406196212, "grad_norm": 2.3196418285369873, "learning_rate": 7.088122605363985e-06, "loss": 0.1527, "step": 2530 }, { "epoch": 17.48709122203098, "grad_norm": 2.2051141262054443, "learning_rate": 6.896551724137932e-06, "loss": 0.1285, "step": 2540 }, { "epoch": 17.555938037865747, "grad_norm": 2.9338431358337402, "learning_rate": 6.7049808429118775e-06, "loss": 0.1182, "step": 2550 }, { "epoch": 17.624784853700515, "grad_norm": 2.8327407836914062, "learning_rate": 6.513409961685823e-06, "loss": 0.0985, "step": 2560 }, { "epoch": 17.693631669535286, "grad_norm": 2.8627865314483643, "learning_rate": 6.321839080459771e-06, "loss": 0.1185, "step": 2570 }, { "epoch": 17.762478485370053, "grad_norm": 2.3446600437164307, "learning_rate": 6.130268199233717e-06, "loss": 0.1024, "step": 2580 }, { "epoch": 17.83132530120482, "grad_norm": 3.2858073711395264, "learning_rate": 5.938697318007663e-06, "loss": 0.1213, "step": 2590 }, { "epoch": 17.900172117039588, "grad_norm": 2.4796440601348877, "learning_rate": 5.747126436781609e-06, "loss": 0.0974, "step": 2600 }, { "epoch": 17.969018932874356, "grad_norm": 1.4947516918182373, "learning_rate": 5.555555555555556e-06, "loss": 0.0972, "step": 2610 }, { "epoch": 17.99655765920826, "eval_accuracy": 0.9781976744186046, "eval_loss": 0.08412329852581024, "eval_runtime": 8.8996, "eval_samples_per_second": 231.921, "eval_steps_per_second": 7.304, "step": 2614 }, { "epoch": 18.037865748709123, "grad_norm": 2.7059268951416016, "learning_rate": 5.3639846743295025e-06, "loss": 0.1074, "step": 2620 }, { "epoch": 18.10671256454389, "grad_norm": 2.0837604999542236, "learning_rate": 5.172413793103448e-06, "loss": 0.106, "step": 2630 }, { "epoch": 18.175559380378658, "grad_norm": 3.52249813079834, "learning_rate": 4.980842911877395e-06, "loss": 0.1475, "step": 2640 }, { "epoch": 18.244406196213426, "grad_norm": 2.8017690181732178, "learning_rate": 4.789272030651342e-06, "loss": 0.128, "step": 2650 }, { "epoch": 18.313253012048193, "grad_norm": 1.7369544506072998, "learning_rate": 4.5977011494252875e-06, "loss": 0.0977, "step": 2660 }, { "epoch": 18.38209982788296, "grad_norm": 2.63714599609375, "learning_rate": 4.406130268199233e-06, "loss": 0.1067, "step": 2670 }, { "epoch": 18.450946643717728, "grad_norm": 2.7819557189941406, "learning_rate": 4.214559386973181e-06, "loss": 0.1364, "step": 2680 }, { "epoch": 18.519793459552496, "grad_norm": 2.259345769882202, "learning_rate": 4.022988505747127e-06, "loss": 0.0957, "step": 2690 }, { "epoch": 18.588640275387263, "grad_norm": 2.4775497913360596, "learning_rate": 3.8314176245210725e-06, "loss": 0.1273, "step": 2700 }, { "epoch": 18.65748709122203, "grad_norm": 2.889547824859619, "learning_rate": 3.6398467432950196e-06, "loss": 0.1293, "step": 2710 }, { "epoch": 18.726333907056798, "grad_norm": 2.8670730590820312, "learning_rate": 3.448275862068966e-06, "loss": 0.1279, "step": 2720 }, { "epoch": 18.795180722891565, "grad_norm": 1.9103004932403564, "learning_rate": 3.2567049808429117e-06, "loss": 0.1292, "step": 2730 }, { "epoch": 18.864027538726333, "grad_norm": 1.6978858709335327, "learning_rate": 3.0651340996168583e-06, "loss": 0.1182, "step": 2740 }, { "epoch": 18.9328743545611, "grad_norm": 3.6564149856567383, "learning_rate": 2.8735632183908046e-06, "loss": 0.1293, "step": 2750 }, { "epoch": 18.994836488812393, "eval_accuracy": 0.9467054263565892, "eval_loss": 0.15116389095783234, "eval_runtime": 7.8213, "eval_samples_per_second": 263.893, "eval_steps_per_second": 8.311, "step": 2759 }, { "epoch": 19.001721170395868, "grad_norm": 3.7127535343170166, "learning_rate": 2.6819923371647512e-06, "loss": 0.1027, "step": 2760 }, { "epoch": 19.070567986230635, "grad_norm": 2.170724868774414, "learning_rate": 2.4904214559386975e-06, "loss": 0.0993, "step": 2770 }, { "epoch": 19.139414802065403, "grad_norm": 3.131479501724243, "learning_rate": 2.2988505747126437e-06, "loss": 0.1378, "step": 2780 }, { "epoch": 19.20826161790017, "grad_norm": 1.942002534866333, "learning_rate": 2.1072796934865904e-06, "loss": 0.1163, "step": 2790 }, { "epoch": 19.27710843373494, "grad_norm": 2.8317081928253174, "learning_rate": 1.9157088122605362e-06, "loss": 0.1467, "step": 2800 }, { "epoch": 19.34595524956971, "grad_norm": 1.791185736656189, "learning_rate": 1.724137931034483e-06, "loss": 0.1139, "step": 2810 }, { "epoch": 19.414802065404476, "grad_norm": 2.639554023742676, "learning_rate": 1.5325670498084292e-06, "loss": 0.1139, "step": 2820 }, { "epoch": 19.483648881239244, "grad_norm": 1.9929583072662354, "learning_rate": 1.3409961685823756e-06, "loss": 0.1093, "step": 2830 }, { "epoch": 19.55249569707401, "grad_norm": 2.7852931022644043, "learning_rate": 1.1494252873563219e-06, "loss": 0.1305, "step": 2840 }, { "epoch": 19.62134251290878, "grad_norm": 3.228508710861206, "learning_rate": 9.578544061302681e-07, "loss": 0.1236, "step": 2850 }, { "epoch": 19.690189328743546, "grad_norm": 3.5496957302093506, "learning_rate": 7.662835249042146e-07, "loss": 0.1357, "step": 2860 }, { "epoch": 19.759036144578314, "grad_norm": 1.8435531854629517, "learning_rate": 5.747126436781609e-07, "loss": 0.1134, "step": 2870 }, { "epoch": 19.82788296041308, "grad_norm": 3.3433897495269775, "learning_rate": 3.831417624521073e-07, "loss": 0.1472, "step": 2880 }, { "epoch": 19.89672977624785, "grad_norm": 2.541207790374756, "learning_rate": 1.9157088122605365e-07, "loss": 0.1057, "step": 2890 }, { "epoch": 19.965576592082616, "grad_norm": 3.152078151702881, "learning_rate": 0.0, "loss": 0.1072, "step": 2900 }, { "epoch": 19.965576592082616, "eval_accuracy": 0.9447674418604651, "eval_loss": 0.16629785299301147, "eval_runtime": 7.3483, "eval_samples_per_second": 280.88, "eval_steps_per_second": 8.846, "step": 2900 }, { "epoch": 19.965576592082616, "step": 2900, "total_flos": 9.813099581032366e+17, "train_loss": 0.3346556580683281, "train_runtime": 3422.7405, "train_samples_per_second": 108.533, "train_steps_per_second": 0.847 } ], "logging_steps": 10, "max_steps": 2900, "num_input_tokens_seen": 0, "num_train_epochs": 20, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 9.813099581032366e+17, "train_batch_size": 32, "trial_name": null, "trial_params": null }