{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "global_step": 5534, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "compression_loss": 91.35626220703125, "epoch": 0.0, "learning_rate": 3e-05, "loss": 92.1804, "step": 1 }, { "compression_loss": 89.19490051269531, "epoch": 0.0, "learning_rate": 3e-05, "loss": 90.1993, "step": 2 }, { "compression_loss": 89.62368774414062, "epoch": 0.0, "learning_rate": 2.9994578966389592e-05, "loss": 90.4198, "step": 3 }, { "compression_loss": 87.4991455078125, "epoch": 0.0, "learning_rate": 2.9989157932779183e-05, "loss": 88.5779, "step": 4 }, { "compression_loss": 86.41934204101562, "epoch": 0.0, "learning_rate": 2.9983736899168775e-05, "loss": 87.0004, "step": 5 }, { "compression_loss": 86.34823608398438, "epoch": 0.0, "learning_rate": 2.9978315865558366e-05, "loss": 86.9579, "step": 6 }, { "compression_loss": 85.31929016113281, "epoch": 0.0, "learning_rate": 2.997289483194796e-05, "loss": 85.8473, "step": 7 }, { "compression_loss": 85.520263671875, "epoch": 0.0, "learning_rate": 2.9967473798337552e-05, "loss": 86.5099, "step": 8 }, { "compression_loss": 85.39549255371094, "epoch": 0.0, "learning_rate": 2.9962052764727143e-05, "loss": 86.1724, "step": 9 }, { "compression_loss": 85.04629516601562, "epoch": 0.0, "learning_rate": 2.9956631731116734e-05, "loss": 85.8961, "step": 10 }, { "compression_loss": 84.53693389892578, "epoch": 0.0, "learning_rate": 2.9951210697506326e-05, "loss": 85.15, "step": 11 }, { "compression_loss": 85.87057495117188, "epoch": 0.0, "learning_rate": 2.9945789663895917e-05, "loss": 86.5898, "step": 12 }, { "compression_loss": 84.54175567626953, "epoch": 0.0, "learning_rate": 2.9940368630285508e-05, "loss": 85.0405, "step": 13 }, { "compression_loss": 83.59049987792969, "epoch": 0.01, "learning_rate": 2.99349475966751e-05, "loss": 84.2668, "step": 14 }, { "compression_loss": 82.98521423339844, "epoch": 0.01, "learning_rate": 2.9929526563064694e-05, "loss": 83.2961, "step": 15 }, { "compression_loss": 87.25186920166016, "epoch": 0.01, "learning_rate": 2.9924105529454285e-05, "loss": 87.8979, "step": 16 }, { "compression_loss": 85.9286117553711, "epoch": 0.01, "learning_rate": 2.9918684495843877e-05, "loss": 87.0484, "step": 17 }, { "compression_loss": 83.7164077758789, "epoch": 0.01, "learning_rate": 2.9913263462233468e-05, "loss": 84.3355, "step": 18 }, { "compression_loss": 84.54891967773438, "epoch": 0.01, "learning_rate": 2.990784242862306e-05, "loss": 84.8745, "step": 19 }, { "compression_loss": 84.94427490234375, "epoch": 0.01, "learning_rate": 2.990242139501265e-05, "loss": 85.5586, "step": 20 }, { "compression_loss": 87.55119323730469, "epoch": 0.01, "learning_rate": 2.9897000361402242e-05, "loss": 88.2214, "step": 21 }, { "compression_loss": 85.30451202392578, "epoch": 0.01, "learning_rate": 2.9891579327791833e-05, "loss": 85.8437, "step": 22 }, { "compression_loss": 84.00923919677734, "epoch": 0.01, "learning_rate": 2.9886158294181428e-05, "loss": 84.3815, "step": 23 }, { "compression_loss": 84.09503173828125, "epoch": 0.01, "learning_rate": 2.9886158294181428e-05, "loss": 84.8233, "step": 24 }, { "compression_loss": 83.22807312011719, "epoch": 0.01, "learning_rate": 2.9880737260571016e-05, "loss": 83.7779, "step": 25 }, { "compression_loss": 83.84465789794922, "epoch": 0.01, "learning_rate": 2.9875316226960607e-05, "loss": 84.6001, "step": 26 }, { "compression_loss": 85.7331771850586, "epoch": 0.01, "learning_rate": 2.9869895193350198e-05, "loss": 86.1989, "step": 27 }, { "compression_loss": 83.74343872070312, "epoch": 0.01, "learning_rate": 2.986447415973979e-05, "loss": 84.2601, "step": 28 }, { "compression_loss": 85.95801544189453, "epoch": 0.01, "learning_rate": 2.985905312612938e-05, "loss": 86.9184, "step": 29 }, { "compression_loss": 85.94073486328125, "epoch": 0.01, "learning_rate": 2.9853632092518972e-05, "loss": 86.8876, "step": 30 }, { "compression_loss": 85.01473236083984, "epoch": 0.01, "learning_rate": 2.9848211058908563e-05, "loss": 85.4962, "step": 31 }, { "compression_loss": 84.75965881347656, "epoch": 0.01, "learning_rate": 2.9842790025298158e-05, "loss": 85.2387, "step": 32 }, { "compression_loss": 85.44303894042969, "epoch": 0.01, "learning_rate": 2.983736899168775e-05, "loss": 86.3569, "step": 33 }, { "compression_loss": 85.07192993164062, "epoch": 0.01, "learning_rate": 2.983194795807734e-05, "loss": 85.4714, "step": 34 }, { "compression_loss": 83.8150405883789, "epoch": 0.01, "learning_rate": 2.982652692446693e-05, "loss": 84.1965, "step": 35 }, { "compression_loss": 85.31266021728516, "epoch": 0.01, "learning_rate": 2.9821105890856523e-05, "loss": 85.7802, "step": 36 }, { "compression_loss": 86.22575378417969, "epoch": 0.01, "learning_rate": 2.9815684857246114e-05, "loss": 86.7469, "step": 37 }, { "compression_loss": 84.20779418945312, "epoch": 0.01, "learning_rate": 2.9810263823635705e-05, "loss": 84.7825, "step": 38 }, { "compression_loss": 86.74066925048828, "epoch": 0.01, "learning_rate": 2.9804842790025297e-05, "loss": 87.517, "step": 39 }, { "compression_loss": 84.11668395996094, "epoch": 0.01, "learning_rate": 2.979942175641489e-05, "loss": 84.7449, "step": 40 }, { "compression_loss": 84.99517822265625, "epoch": 0.01, "learning_rate": 2.9794000722804483e-05, "loss": 85.9993, "step": 41 }, { "compression_loss": 85.35609436035156, "epoch": 0.02, "learning_rate": 2.9788579689194074e-05, "loss": 86.0265, "step": 42 }, { "compression_loss": 85.41738891601562, "epoch": 0.02, "learning_rate": 2.9783158655583665e-05, "loss": 85.9557, "step": 43 }, { "compression_loss": 84.15126037597656, "epoch": 0.02, "learning_rate": 2.9777737621973256e-05, "loss": 84.7466, "step": 44 }, { "compression_loss": 85.33430480957031, "epoch": 0.02, "learning_rate": 2.9772316588362848e-05, "loss": 85.847, "step": 45 }, { "compression_loss": 85.37297821044922, "epoch": 0.02, "learning_rate": 2.976689555475244e-05, "loss": 85.9185, "step": 46 }, { "compression_loss": 84.7735595703125, "epoch": 0.02, "learning_rate": 2.976147452114203e-05, "loss": 85.4191, "step": 47 }, { "compression_loss": 87.0035171508789, "epoch": 0.02, "learning_rate": 2.9756053487531625e-05, "loss": 87.7724, "step": 48 }, { "compression_loss": 84.44734191894531, "epoch": 0.02, "learning_rate": 2.9750632453921216e-05, "loss": 84.9495, "step": 49 }, { "compression_loss": 84.46060180664062, "epoch": 0.02, "learning_rate": 2.9745211420310807e-05, "loss": 85.0173, "step": 50 }, { "compression_loss": 85.94522094726562, "epoch": 0.02, "learning_rate": 2.97397903867004e-05, "loss": 86.6114, "step": 51 }, { "compression_loss": 83.65790557861328, "epoch": 0.02, "learning_rate": 2.973436935308999e-05, "loss": 84.0334, "step": 52 }, { "compression_loss": 83.9991226196289, "epoch": 0.02, "learning_rate": 2.972894831947958e-05, "loss": 84.4697, "step": 53 }, { "compression_loss": 84.89405059814453, "epoch": 0.02, "learning_rate": 2.9723527285869173e-05, "loss": 85.8699, "step": 54 }, { "compression_loss": 83.4554443359375, "epoch": 0.02, "learning_rate": 2.9718106252258764e-05, "loss": 84.0317, "step": 55 }, { "compression_loss": 83.94524383544922, "epoch": 0.02, "learning_rate": 2.971268521864836e-05, "loss": 84.4982, "step": 56 }, { "compression_loss": 84.80667114257812, "epoch": 0.02, "learning_rate": 2.970726418503795e-05, "loss": 85.1245, "step": 57 }, { "compression_loss": 83.76476287841797, "epoch": 0.02, "learning_rate": 2.970184315142754e-05, "loss": 84.3971, "step": 58 }, { "compression_loss": 85.11968231201172, "epoch": 0.02, "learning_rate": 2.9696422117817132e-05, "loss": 85.9125, "step": 59 }, { "compression_loss": 86.35231018066406, "epoch": 0.02, "learning_rate": 2.9691001084206724e-05, "loss": 86.8509, "step": 60 }, { "compression_loss": 85.25308227539062, "epoch": 0.02, "learning_rate": 2.9685580050596315e-05, "loss": 85.6895, "step": 61 }, { "compression_loss": 84.11752319335938, "epoch": 0.02, "learning_rate": 2.9680159016985906e-05, "loss": 84.5853, "step": 62 }, { "compression_loss": 84.76602172851562, "epoch": 0.02, "learning_rate": 2.9674737983375497e-05, "loss": 85.1739, "step": 63 }, { "compression_loss": 84.66615295410156, "epoch": 0.02, "learning_rate": 2.9669316949765092e-05, "loss": 85.2174, "step": 64 }, { "compression_loss": 82.84949493408203, "epoch": 0.02, "learning_rate": 2.9663895916154683e-05, "loss": 83.3929, "step": 65 }, { "compression_loss": 85.27127838134766, "epoch": 0.02, "learning_rate": 2.9658474882544275e-05, "loss": 85.8926, "step": 66 }, { "compression_loss": 83.6759262084961, "epoch": 0.02, "learning_rate": 2.9653053848933866e-05, "loss": 84.1068, "step": 67 }, { "compression_loss": 85.19695281982422, "epoch": 0.02, "learning_rate": 2.9647632815323457e-05, "loss": 85.8803, "step": 68 }, { "compression_loss": 85.49566650390625, "epoch": 0.02, "learning_rate": 2.9642211781713045e-05, "loss": 86.24, "step": 69 }, { "compression_loss": 83.94522094726562, "epoch": 0.03, "learning_rate": 2.9636790748102636e-05, "loss": 84.542, "step": 70 }, { "compression_loss": 87.18675231933594, "epoch": 0.03, "learning_rate": 2.9631369714492228e-05, "loss": 88.2827, "step": 71 }, { "compression_loss": 84.9115219116211, "epoch": 0.03, "learning_rate": 2.9625948680881822e-05, "loss": 85.5927, "step": 72 }, { "compression_loss": 85.96586608886719, "epoch": 0.03, "learning_rate": 2.9620527647271414e-05, "loss": 86.6112, "step": 73 }, { "compression_loss": 85.32550048828125, "epoch": 0.03, "learning_rate": 2.9615106613661005e-05, "loss": 85.7618, "step": 74 }, { "compression_loss": 85.0137710571289, "epoch": 0.03, "learning_rate": 2.9609685580050596e-05, "loss": 85.6486, "step": 75 }, { "compression_loss": 84.52037048339844, "epoch": 0.03, "learning_rate": 2.9604264546440187e-05, "loss": 84.9587, "step": 76 }, { "compression_loss": 83.85589599609375, "epoch": 0.03, "learning_rate": 2.959884351282978e-05, "loss": 84.3376, "step": 77 }, { "compression_loss": 85.57453918457031, "epoch": 0.03, "learning_rate": 2.959342247921937e-05, "loss": 86.0414, "step": 78 }, { "compression_loss": 83.15218353271484, "epoch": 0.03, "learning_rate": 2.958800144560896e-05, "loss": 83.7526, "step": 79 }, { "compression_loss": 86.13349914550781, "epoch": 0.03, "learning_rate": 2.9582580411998556e-05, "loss": 86.8314, "step": 80 }, { "compression_loss": 84.47675323486328, "epoch": 0.03, "learning_rate": 2.9577159378388147e-05, "loss": 85.0033, "step": 81 }, { "compression_loss": 84.72871398925781, "epoch": 0.03, "learning_rate": 2.957173834477774e-05, "loss": 85.213, "step": 82 }, { "compression_loss": 82.2294921875, "epoch": 0.03, "learning_rate": 2.956631731116733e-05, "loss": 82.8447, "step": 83 }, { "compression_loss": 86.16065979003906, "epoch": 0.03, "learning_rate": 2.956089627755692e-05, "loss": 87.0405, "step": 84 }, { "compression_loss": 85.44457244873047, "epoch": 0.03, "learning_rate": 2.9555475243946512e-05, "loss": 86.0783, "step": 85 }, { "compression_loss": 83.39029693603516, "epoch": 0.03, "learning_rate": 2.9550054210336103e-05, "loss": 84.1099, "step": 86 }, { "compression_loss": 85.02418518066406, "epoch": 0.03, "learning_rate": 2.9544633176725695e-05, "loss": 85.3356, "step": 87 }, { "compression_loss": 84.53054809570312, "epoch": 0.03, "learning_rate": 2.953921214311529e-05, "loss": 85.2522, "step": 88 }, { "compression_loss": 84.78958129882812, "epoch": 0.03, "learning_rate": 2.953379110950488e-05, "loss": 85.7768, "step": 89 }, { "compression_loss": 85.17430114746094, "epoch": 0.03, "learning_rate": 2.9528370075894472e-05, "loss": 85.923, "step": 90 }, { "compression_loss": 84.39291381835938, "epoch": 0.03, "learning_rate": 2.9522949042284063e-05, "loss": 85.2816, "step": 91 }, { "compression_loss": 83.58943939208984, "epoch": 0.03, "learning_rate": 2.9517528008673654e-05, "loss": 83.9656, "step": 92 }, { "compression_loss": 85.44347381591797, "epoch": 0.03, "learning_rate": 2.9512106975063246e-05, "loss": 86.1798, "step": 93 }, { "compression_loss": 84.12031555175781, "epoch": 0.03, "learning_rate": 2.9506685941452837e-05, "loss": 84.8126, "step": 94 }, { "compression_loss": 84.60367584228516, "epoch": 0.03, "learning_rate": 2.9501264907842428e-05, "loss": 85.1846, "step": 95 }, { "compression_loss": 84.87464904785156, "epoch": 0.03, "learning_rate": 2.9495843874232023e-05, "loss": 85.4138, "step": 96 }, { "compression_loss": 86.79476928710938, "epoch": 0.04, "learning_rate": 2.9490422840621614e-05, "loss": 87.507, "step": 97 }, { "compression_loss": 83.62125396728516, "epoch": 0.04, "learning_rate": 2.9485001807011205e-05, "loss": 84.1315, "step": 98 }, { "compression_loss": 84.47203826904297, "epoch": 0.04, "learning_rate": 2.9479580773400797e-05, "loss": 85.2119, "step": 99 }, { "compression_loss": 86.0842056274414, "epoch": 0.04, "learning_rate": 2.9474159739790388e-05, "loss": 86.8229, "step": 100 }, { "compression_loss": 84.89802551269531, "epoch": 0.04, "learning_rate": 2.946873870617998e-05, "loss": 85.5711, "step": 101 }, { "compression_loss": 85.60102081298828, "epoch": 0.04, "learning_rate": 2.946331767256957e-05, "loss": 86.3495, "step": 102 }, { "compression_loss": 83.3472671508789, "epoch": 0.04, "learning_rate": 2.9457896638959162e-05, "loss": 83.6085, "step": 103 }, { "compression_loss": 82.88160705566406, "epoch": 0.04, "learning_rate": 2.9452475605348756e-05, "loss": 83.3797, "step": 104 }, { "compression_loss": 85.625, "epoch": 0.04, "learning_rate": 2.9447054571738348e-05, "loss": 86.4479, "step": 105 }, { "compression_loss": 86.70901489257812, "epoch": 0.04, "learning_rate": 2.944163353812794e-05, "loss": 87.1377, "step": 106 }, { "compression_loss": 84.61742401123047, "epoch": 0.04, "learning_rate": 2.943621250451753e-05, "loss": 84.8116, "step": 107 }, { "compression_loss": 83.58950805664062, "epoch": 0.04, "learning_rate": 2.943079147090712e-05, "loss": 83.9844, "step": 108 }, { "compression_loss": 83.7735824584961, "epoch": 0.04, "learning_rate": 2.9425370437296713e-05, "loss": 84.3502, "step": 109 }, { "compression_loss": 85.18844604492188, "epoch": 0.04, "learning_rate": 2.9419949403686304e-05, "loss": 85.9195, "step": 110 }, { "compression_loss": 86.57881164550781, "epoch": 0.04, "learning_rate": 2.9414528370075895e-05, "loss": 87.131, "step": 111 }, { "compression_loss": 85.70211029052734, "epoch": 0.04, "learning_rate": 2.9409107336465487e-05, "loss": 86.6903, "step": 112 }, { "compression_loss": 86.21756744384766, "epoch": 0.04, "learning_rate": 2.9403686302855078e-05, "loss": 86.9937, "step": 113 }, { "compression_loss": 83.72859954833984, "epoch": 0.04, "learning_rate": 2.939826526924467e-05, "loss": 84.2656, "step": 114 }, { "compression_loss": 86.08450317382812, "epoch": 0.04, "learning_rate": 2.939284423563426e-05, "loss": 86.5159, "step": 115 }, { "compression_loss": 84.20480346679688, "epoch": 0.04, "learning_rate": 2.9387423202023852e-05, "loss": 84.5841, "step": 116 }, { "compression_loss": 84.19602966308594, "epoch": 0.04, "learning_rate": 2.9382002168413443e-05, "loss": 84.7418, "step": 117 }, { "compression_loss": 84.48530578613281, "epoch": 0.04, "learning_rate": 2.9376581134803034e-05, "loss": 84.8689, "step": 118 }, { "compression_loss": 85.86943054199219, "epoch": 0.04, "learning_rate": 2.9371160101192626e-05, "loss": 86.9907, "step": 119 }, { "compression_loss": 84.52864074707031, "epoch": 0.04, "learning_rate": 2.936573906758222e-05, "loss": 85.1449, "step": 120 }, { "compression_loss": 84.59535217285156, "epoch": 0.04, "learning_rate": 2.936031803397181e-05, "loss": 85.2032, "step": 121 }, { "compression_loss": 85.05545043945312, "epoch": 0.04, "learning_rate": 2.9354897000361403e-05, "loss": 85.4639, "step": 122 }, { "compression_loss": 85.22352600097656, "epoch": 0.04, "learning_rate": 2.9349475966750994e-05, "loss": 85.6643, "step": 123 }, { "compression_loss": 84.78773498535156, "epoch": 0.04, "learning_rate": 2.9344054933140585e-05, "loss": 85.3309, "step": 124 }, { "compression_loss": 83.1006088256836, "epoch": 0.05, "learning_rate": 2.9338633899530177e-05, "loss": 83.4961, "step": 125 }, { "compression_loss": 83.17539978027344, "epoch": 0.05, "learning_rate": 2.9333212865919768e-05, "loss": 83.5442, "step": 126 }, { "compression_loss": 84.48330688476562, "epoch": 0.05, "learning_rate": 2.932779183230936e-05, "loss": 85.1311, "step": 127 }, { "compression_loss": 83.53936767578125, "epoch": 0.05, "learning_rate": 2.9322370798698954e-05, "loss": 84.6159, "step": 128 }, { "compression_loss": 83.38616943359375, "epoch": 0.05, "learning_rate": 2.9316949765088545e-05, "loss": 83.8492, "step": 129 }, { "compression_loss": 85.45303344726562, "epoch": 0.05, "learning_rate": 2.9311528731478136e-05, "loss": 86.0353, "step": 130 }, { "compression_loss": 85.07221221923828, "epoch": 0.05, "learning_rate": 2.9306107697867728e-05, "loss": 85.5691, "step": 131 }, { "compression_loss": 83.29330444335938, "epoch": 0.05, "learning_rate": 2.930068666425732e-05, "loss": 83.7647, "step": 132 }, { "compression_loss": 85.14433288574219, "epoch": 0.05, "learning_rate": 2.929526563064691e-05, "loss": 85.8328, "step": 133 }, { "compression_loss": 85.50108337402344, "epoch": 0.05, "learning_rate": 2.92898445970365e-05, "loss": 86.5059, "step": 134 }, { "compression_loss": 84.95072937011719, "epoch": 0.05, "learning_rate": 2.9284423563426093e-05, "loss": 85.7958, "step": 135 }, { "compression_loss": 85.05303955078125, "epoch": 0.05, "learning_rate": 2.9279002529815687e-05, "loss": 85.6171, "step": 136 }, { "compression_loss": 84.54081726074219, "epoch": 0.05, "learning_rate": 2.927358149620528e-05, "loss": 85.0202, "step": 137 }, { "compression_loss": 85.31010437011719, "epoch": 0.05, "learning_rate": 2.926816046259487e-05, "loss": 85.8345, "step": 138 }, { "compression_loss": 83.70062255859375, "epoch": 0.05, "learning_rate": 2.926273942898446e-05, "loss": 84.0965, "step": 139 }, { "compression_loss": 83.44458770751953, "epoch": 0.05, "learning_rate": 2.9257318395374052e-05, "loss": 84.1352, "step": 140 }, { "compression_loss": 83.22015380859375, "epoch": 0.05, "learning_rate": 2.9251897361763644e-05, "loss": 83.6216, "step": 141 }, { "compression_loss": 85.96746063232422, "epoch": 0.05, "learning_rate": 2.9246476328153235e-05, "loss": 86.596, "step": 142 }, { "compression_loss": 85.13638305664062, "epoch": 0.05, "learning_rate": 2.9241055294542826e-05, "loss": 85.6781, "step": 143 }, { "compression_loss": 83.8939208984375, "epoch": 0.05, "learning_rate": 2.923563426093242e-05, "loss": 84.519, "step": 144 }, { "compression_loss": 85.68746185302734, "epoch": 0.05, "learning_rate": 2.9230213227322012e-05, "loss": 86.5479, "step": 145 }, { "compression_loss": 84.2289810180664, "epoch": 0.05, "learning_rate": 2.9224792193711603e-05, "loss": 84.6301, "step": 146 }, { "compression_loss": 86.06626892089844, "epoch": 0.05, "learning_rate": 2.9219371160101195e-05, "loss": 86.9928, "step": 147 }, { "compression_loss": 84.20088195800781, "epoch": 0.05, "learning_rate": 2.9213950126490786e-05, "loss": 84.7624, "step": 148 }, { "compression_loss": 84.51805114746094, "epoch": 0.05, "learning_rate": 2.9208529092880377e-05, "loss": 84.8062, "step": 149 }, { "compression_loss": 85.79377746582031, "epoch": 0.05, "learning_rate": 2.920310805926997e-05, "loss": 86.3509, "step": 150 }, { "compression_loss": 85.4161148071289, "epoch": 0.05, "learning_rate": 2.919768702565956e-05, "loss": 86.1045, "step": 151 }, { "compression_loss": 85.09808349609375, "epoch": 0.05, "learning_rate": 2.9192265992049154e-05, "loss": 85.4793, "step": 152 }, { "compression_loss": 83.70523834228516, "epoch": 0.06, "learning_rate": 2.9186844958438746e-05, "loss": 84.3336, "step": 153 }, { "compression_loss": 84.34274291992188, "epoch": 0.06, "learning_rate": 2.9181423924828337e-05, "loss": 85.0162, "step": 154 }, { "compression_loss": 83.56495666503906, "epoch": 0.06, "learning_rate": 2.9176002891217928e-05, "loss": 84.2007, "step": 155 }, { "compression_loss": 86.31279754638672, "epoch": 0.06, "learning_rate": 2.9170581857607516e-05, "loss": 86.9149, "step": 156 }, { "compression_loss": 84.04824829101562, "epoch": 0.06, "learning_rate": 2.9165160823997107e-05, "loss": 84.6308, "step": 157 }, { "compression_loss": 84.08952331542969, "epoch": 0.06, "learning_rate": 2.91597397903867e-05, "loss": 84.5983, "step": 158 }, { "compression_loss": 83.07487487792969, "epoch": 0.06, "learning_rate": 2.915431875677629e-05, "loss": 83.6972, "step": 159 }, { "compression_loss": 83.62516784667969, "epoch": 0.06, "learning_rate": 2.9148897723165885e-05, "loss": 83.9664, "step": 160 }, { "compression_loss": 84.33580017089844, "epoch": 0.06, "learning_rate": 2.9143476689555476e-05, "loss": 84.8215, "step": 161 }, { "compression_loss": 84.27178955078125, "epoch": 0.06, "learning_rate": 2.9138055655945067e-05, "loss": 84.7865, "step": 162 }, { "compression_loss": 83.02823638916016, "epoch": 0.06, "learning_rate": 2.913263462233466e-05, "loss": 83.4993, "step": 163 }, { "compression_loss": 84.03942108154297, "epoch": 0.06, "learning_rate": 2.912721358872425e-05, "loss": 84.668, "step": 164 }, { "compression_loss": 83.82827758789062, "epoch": 0.06, "learning_rate": 2.912179255511384e-05, "loss": 84.2201, "step": 165 }, { "compression_loss": 87.76260375976562, "epoch": 0.06, "learning_rate": 2.9116371521503432e-05, "loss": 88.3837, "step": 166 }, { "compression_loss": 83.46829986572266, "epoch": 0.06, "learning_rate": 2.9110950487893024e-05, "loss": 84.0109, "step": 167 }, { "compression_loss": 84.85670471191406, "epoch": 0.06, "learning_rate": 2.9105529454282618e-05, "loss": 85.5336, "step": 168 }, { "compression_loss": 84.51240539550781, "epoch": 0.06, "learning_rate": 2.910010842067221e-05, "loss": 85.1639, "step": 169 }, { "compression_loss": 85.50804138183594, "epoch": 0.06, "learning_rate": 2.90946873870618e-05, "loss": 86.1512, "step": 170 }, { "compression_loss": 84.58477783203125, "epoch": 0.06, "learning_rate": 2.9089266353451392e-05, "loss": 85.2809, "step": 171 }, { "compression_loss": 85.73362731933594, "epoch": 0.06, "learning_rate": 2.9083845319840983e-05, "loss": 86.4267, "step": 172 }, { "compression_loss": 85.43461608886719, "epoch": 0.06, "learning_rate": 2.9078424286230575e-05, "loss": 85.8003, "step": 173 }, { "compression_loss": 82.18844604492188, "epoch": 0.06, "learning_rate": 2.9073003252620166e-05, "loss": 82.6418, "step": 174 }, { "compression_loss": 84.17156982421875, "epoch": 0.06, "learning_rate": 2.9067582219009757e-05, "loss": 84.5227, "step": 175 }, { "compression_loss": 84.1295166015625, "epoch": 0.06, "learning_rate": 2.9062161185399352e-05, "loss": 84.8891, "step": 176 }, { "compression_loss": 86.37615966796875, "epoch": 0.06, "learning_rate": 2.9056740151788943e-05, "loss": 86.8795, "step": 177 }, { "compression_loss": 84.37887573242188, "epoch": 0.06, "learning_rate": 2.9051319118178534e-05, "loss": 84.6237, "step": 178 }, { "compression_loss": 83.73757934570312, "epoch": 0.06, "learning_rate": 2.9045898084568126e-05, "loss": 84.1828, "step": 179 }, { "compression_loss": 83.5699462890625, "epoch": 0.07, "learning_rate": 2.9040477050957717e-05, "loss": 84.1953, "step": 180 }, { "compression_loss": 85.08660125732422, "epoch": 0.07, "learning_rate": 2.9035056017347308e-05, "loss": 86.1471, "step": 181 }, { "compression_loss": 85.75422668457031, "epoch": 0.07, "learning_rate": 2.90296349837369e-05, "loss": 86.2038, "step": 182 }, { "compression_loss": 85.4156494140625, "epoch": 0.07, "learning_rate": 2.902421395012649e-05, "loss": 85.8195, "step": 183 }, { "compression_loss": 86.55537414550781, "epoch": 0.07, "learning_rate": 2.9018792916516085e-05, "loss": 87.0031, "step": 184 }, { "compression_loss": 85.26417541503906, "epoch": 0.07, "learning_rate": 2.9013371882905677e-05, "loss": 85.9882, "step": 185 }, { "compression_loss": 84.35462188720703, "epoch": 0.07, "learning_rate": 2.9007950849295268e-05, "loss": 84.9287, "step": 186 }, { "compression_loss": 86.52105712890625, "epoch": 0.07, "learning_rate": 2.900252981568486e-05, "loss": 87.0804, "step": 187 }, { "compression_loss": 85.04695129394531, "epoch": 0.07, "learning_rate": 2.899710878207445e-05, "loss": 85.8689, "step": 188 }, { "compression_loss": 83.93319702148438, "epoch": 0.07, "learning_rate": 2.8991687748464042e-05, "loss": 84.4777, "step": 189 }, { "compression_loss": 82.79999542236328, "epoch": 0.07, "learning_rate": 2.8986266714853633e-05, "loss": 83.1094, "step": 190 }, { "compression_loss": 84.12474060058594, "epoch": 0.07, "learning_rate": 2.8980845681243224e-05, "loss": 84.8439, "step": 191 }, { "compression_loss": 86.47573852539062, "epoch": 0.07, "learning_rate": 2.8975424647632816e-05, "loss": 86.9656, "step": 192 }, { "compression_loss": 84.4737777709961, "epoch": 0.07, "learning_rate": 2.897000361402241e-05, "loss": 85.0348, "step": 193 }, { "compression_loss": 84.20248413085938, "epoch": 0.07, "learning_rate": 2.8964582580412e-05, "loss": 84.755, "step": 194 }, { "compression_loss": 83.65550994873047, "epoch": 0.07, "learning_rate": 2.8959161546801593e-05, "loss": 83.9976, "step": 195 }, { "compression_loss": 85.86560821533203, "epoch": 0.07, "learning_rate": 2.8953740513191184e-05, "loss": 86.5534, "step": 196 }, { "compression_loss": 83.85475158691406, "epoch": 0.07, "learning_rate": 2.8948319479580775e-05, "loss": 84.5141, "step": 197 }, { "compression_loss": 82.648681640625, "epoch": 0.07, "learning_rate": 2.8942898445970367e-05, "loss": 82.9361, "step": 198 }, { "compression_loss": 84.45886993408203, "epoch": 0.07, "learning_rate": 2.8937477412359958e-05, "loss": 84.9045, "step": 199 }, { "compression_loss": 83.63796997070312, "epoch": 0.07, "learning_rate": 2.8932056378749546e-05, "loss": 84.1837, "step": 200 }, { "compression_loss": 83.4482421875, "epoch": 0.07, "learning_rate": 2.892663534513914e-05, "loss": 84.0039, "step": 201 }, { "compression_loss": 84.66983795166016, "epoch": 0.07, "learning_rate": 2.892121431152873e-05, "loss": 85.1851, "step": 202 }, { "compression_loss": 84.73308563232422, "epoch": 0.07, "learning_rate": 2.8915793277918323e-05, "loss": 85.2974, "step": 203 }, { "compression_loss": 85.366455078125, "epoch": 0.07, "learning_rate": 2.8910372244307914e-05, "loss": 85.7245, "step": 204 }, { "compression_loss": 85.08921813964844, "epoch": 0.07, "learning_rate": 2.8904951210697505e-05, "loss": 85.5165, "step": 205 }, { "compression_loss": 84.31353759765625, "epoch": 0.07, "learning_rate": 2.8899530177087097e-05, "loss": 84.9178, "step": 206 }, { "compression_loss": 82.2561264038086, "epoch": 0.07, "learning_rate": 2.8894109143476688e-05, "loss": 83.0417, "step": 207 }, { "compression_loss": 84.0971450805664, "epoch": 0.08, "learning_rate": 2.888868810986628e-05, "loss": 84.5445, "step": 208 }, { "compression_loss": 85.28726196289062, "epoch": 0.08, "learning_rate": 2.8883267076255874e-05, "loss": 86.2695, "step": 209 }, { "compression_loss": 85.73007202148438, "epoch": 0.08, "learning_rate": 2.8877846042645465e-05, "loss": 86.4852, "step": 210 }, { "compression_loss": 85.9957275390625, "epoch": 0.08, "learning_rate": 2.8872425009035056e-05, "loss": 86.6997, "step": 211 }, { "compression_loss": 82.64970397949219, "epoch": 0.08, "learning_rate": 2.8867003975424648e-05, "loss": 83.26, "step": 212 }, { "compression_loss": 82.20022583007812, "epoch": 0.08, "learning_rate": 2.886158294181424e-05, "loss": 82.8769, "step": 213 }, { "compression_loss": 85.34929656982422, "epoch": 0.08, "learning_rate": 2.885616190820383e-05, "loss": 85.9546, "step": 214 }, { "compression_loss": 84.69729614257812, "epoch": 0.08, "learning_rate": 2.885074087459342e-05, "loss": 85.3149, "step": 215 }, { "compression_loss": 84.72030639648438, "epoch": 0.08, "learning_rate": 2.8845319840983013e-05, "loss": 85.2795, "step": 216 }, { "compression_loss": 83.61612701416016, "epoch": 0.08, "learning_rate": 2.8839898807372607e-05, "loss": 83.9423, "step": 217 }, { "compression_loss": 86.02745819091797, "epoch": 0.08, "learning_rate": 2.88344777737622e-05, "loss": 86.4119, "step": 218 }, { "compression_loss": 87.23128509521484, "epoch": 0.08, "learning_rate": 2.882905674015179e-05, "loss": 88.1582, "step": 219 }, { "compression_loss": 84.91947937011719, "epoch": 0.08, "learning_rate": 2.882363570654138e-05, "loss": 85.2169, "step": 220 }, { "compression_loss": 83.93670654296875, "epoch": 0.08, "learning_rate": 2.8818214672930973e-05, "loss": 84.3182, "step": 221 }, { "compression_loss": 83.17330932617188, "epoch": 0.08, "learning_rate": 2.8812793639320564e-05, "loss": 83.8349, "step": 222 }, { "compression_loss": 85.38731384277344, "epoch": 0.08, "learning_rate": 2.8807372605710155e-05, "loss": 85.8986, "step": 223 }, { "compression_loss": 83.8514175415039, "epoch": 0.08, "learning_rate": 2.8807372605710155e-05, "loss": 84.273, "step": 224 }, { "compression_loss": 83.87045288085938, "epoch": 0.08, "learning_rate": 2.8801951572099746e-05, "loss": 84.3607, "step": 225 }, { "compression_loss": 84.16893005371094, "epoch": 0.08, "learning_rate": 2.879653053848934e-05, "loss": 84.819, "step": 226 }, { "compression_loss": 85.31173706054688, "epoch": 0.08, "learning_rate": 2.8791109504878932e-05, "loss": 85.785, "step": 227 }, { "compression_loss": 84.22294616699219, "epoch": 0.08, "learning_rate": 2.8785688471268524e-05, "loss": 84.8245, "step": 228 }, { "compression_loss": 85.95883178710938, "epoch": 0.08, "learning_rate": 2.8780267437658115e-05, "loss": 86.6257, "step": 229 }, { "compression_loss": 83.68135833740234, "epoch": 0.08, "learning_rate": 2.8774846404047706e-05, "loss": 84.2593, "step": 230 }, { "compression_loss": 84.52743530273438, "epoch": 0.08, "learning_rate": 2.8769425370437297e-05, "loss": 85.0463, "step": 231 }, { "compression_loss": 84.22059631347656, "epoch": 0.08, "learning_rate": 2.876400433682689e-05, "loss": 84.9759, "step": 232 }, { "compression_loss": 85.41593170166016, "epoch": 0.08, "learning_rate": 2.875858330321648e-05, "loss": 86.4931, "step": 233 }, { "compression_loss": 84.1436538696289, "epoch": 0.08, "learning_rate": 2.8753162269606075e-05, "loss": 84.8359, "step": 234 }, { "compression_loss": 83.04212951660156, "epoch": 0.08, "learning_rate": 2.8747741235995666e-05, "loss": 83.3933, "step": 235 }, { "compression_loss": 83.84089660644531, "epoch": 0.09, "learning_rate": 2.8742320202385257e-05, "loss": 84.2058, "step": 236 }, { "compression_loss": 83.6302490234375, "epoch": 0.09, "learning_rate": 2.873689916877485e-05, "loss": 84.2638, "step": 237 }, { "compression_loss": 83.47061157226562, "epoch": 0.09, "learning_rate": 2.873147813516444e-05, "loss": 84.0138, "step": 238 }, { "compression_loss": 84.7635498046875, "epoch": 0.09, "learning_rate": 2.872605710155403e-05, "loss": 85.2573, "step": 239 }, { "compression_loss": 83.37515258789062, "epoch": 0.09, "learning_rate": 2.8720636067943622e-05, "loss": 83.6158, "step": 240 }, { "compression_loss": 85.3198471069336, "epoch": 0.09, "learning_rate": 2.8715215034333214e-05, "loss": 85.8801, "step": 241 }, { "compression_loss": 85.80567932128906, "epoch": 0.09, "learning_rate": 2.8709794000722808e-05, "loss": 86.6086, "step": 242 }, { "compression_loss": 84.88502502441406, "epoch": 0.09, "learning_rate": 2.87043729671124e-05, "loss": 85.4839, "step": 243 }, { "compression_loss": 83.12512969970703, "epoch": 0.09, "learning_rate": 2.8698951933501987e-05, "loss": 83.4953, "step": 244 }, { "compression_loss": 84.80097961425781, "epoch": 0.09, "learning_rate": 2.869353089989158e-05, "loss": 85.5853, "step": 245 }, { "compression_loss": 84.12393188476562, "epoch": 0.09, "learning_rate": 2.868810986628117e-05, "loss": 84.5505, "step": 246 }, { "compression_loss": 83.5642318725586, "epoch": 0.09, "learning_rate": 2.868268883267076e-05, "loss": 84.2331, "step": 247 }, { "compression_loss": 84.11042022705078, "epoch": 0.09, "learning_rate": 2.8677267799060352e-05, "loss": 84.5657, "step": 248 }, { "compression_loss": 85.09321594238281, "epoch": 0.09, "learning_rate": 2.8671846765449944e-05, "loss": 85.7746, "step": 249 }, { "compression_loss": 84.64033508300781, "epoch": 0.09, "learning_rate": 2.866642573183954e-05, "loss": 85.5257, "step": 250 }, { "epoch": 0.09, "eval_exact_match": 84.88174077578051, "eval_f1": 91.5708068492618, "step": 250 }, { "compression_loss": 83.31948852539062, "epoch": 0.09, "learning_rate": 2.866100469822913e-05, "loss": 83.8193, "step": 251 }, { "compression_loss": 85.98564147949219, "epoch": 0.09, "learning_rate": 2.865558366461872e-05, "loss": 86.8092, "step": 252 }, { "compression_loss": 85.61344146728516, "epoch": 0.09, "learning_rate": 2.8650162631008312e-05, "loss": 86.5297, "step": 253 }, { "compression_loss": 83.49954986572266, "epoch": 0.09, "learning_rate": 2.8644741597397903e-05, "loss": 84.0607, "step": 254 }, { "compression_loss": 84.95011138916016, "epoch": 0.09, "learning_rate": 2.8639320563787495e-05, "loss": 85.4117, "step": 255 }, { "compression_loss": 83.6414794921875, "epoch": 0.09, "learning_rate": 2.8633899530177086e-05, "loss": 84.484, "step": 256 }, { "compression_loss": 85.12683868408203, "epoch": 0.09, "learning_rate": 2.8628478496566677e-05, "loss": 85.5077, "step": 257 }, { "compression_loss": 84.56124114990234, "epoch": 0.09, "learning_rate": 2.8623057462956272e-05, "loss": 85.1298, "step": 258 }, { "compression_loss": 84.70410919189453, "epoch": 0.09, "learning_rate": 2.8617636429345863e-05, "loss": 85.1437, "step": 259 }, { "compression_loss": 85.19192504882812, "epoch": 0.09, "learning_rate": 2.8612215395735454e-05, "loss": 85.9201, "step": 260 }, { "compression_loss": 86.75040435791016, "epoch": 0.09, "learning_rate": 2.8606794362125046e-05, "loss": 87.5057, "step": 261 }, { "compression_loss": 86.11820220947266, "epoch": 0.09, "learning_rate": 2.8601373328514637e-05, "loss": 86.6832, "step": 262 }, { "compression_loss": 84.60743713378906, "epoch": 0.1, "learning_rate": 2.8595952294904228e-05, "loss": 85.2883, "step": 263 }, { "compression_loss": 85.50480651855469, "epoch": 0.1, "learning_rate": 2.859053126129382e-05, "loss": 86.1779, "step": 264 }, { "compression_loss": 86.8486099243164, "epoch": 0.1, "learning_rate": 2.858511022768341e-05, "loss": 87.611, "step": 265 }, { "compression_loss": 83.46945190429688, "epoch": 0.1, "learning_rate": 2.8579689194073005e-05, "loss": 83.82, "step": 266 }, { "compression_loss": 84.58091735839844, "epoch": 0.1, "learning_rate": 2.8574268160462597e-05, "loss": 85.3435, "step": 267 }, { "compression_loss": 83.07101440429688, "epoch": 0.1, "learning_rate": 2.8568847126852188e-05, "loss": 83.7384, "step": 268 }, { "compression_loss": 84.6493148803711, "epoch": 0.1, "learning_rate": 2.856342609324178e-05, "loss": 85.25, "step": 269 }, { "compression_loss": 86.26058197021484, "epoch": 0.1, "learning_rate": 2.855800505963137e-05, "loss": 87.097, "step": 270 }, { "compression_loss": 85.03691101074219, "epoch": 0.1, "learning_rate": 2.8552584026020962e-05, "loss": 85.848, "step": 271 }, { "compression_loss": 85.76789093017578, "epoch": 0.1, "learning_rate": 2.8547162992410553e-05, "loss": 86.6555, "step": 272 }, { "compression_loss": 86.9668960571289, "epoch": 0.1, "learning_rate": 2.8541741958800144e-05, "loss": 87.4609, "step": 273 }, { "compression_loss": 84.21022033691406, "epoch": 0.1, "learning_rate": 2.853632092518974e-05, "loss": 84.9613, "step": 274 }, { "compression_loss": 87.28606414794922, "epoch": 0.1, "learning_rate": 2.853089989157933e-05, "loss": 87.9798, "step": 275 }, { "compression_loss": 84.21675109863281, "epoch": 0.1, "learning_rate": 2.852547885796892e-05, "loss": 84.8398, "step": 276 }, { "compression_loss": 84.26268005371094, "epoch": 0.1, "learning_rate": 2.8520057824358513e-05, "loss": 84.5359, "step": 277 }, { "compression_loss": 84.12464904785156, "epoch": 0.1, "learning_rate": 2.8514636790748104e-05, "loss": 84.5967, "step": 278 }, { "compression_loss": 82.88410949707031, "epoch": 0.1, "learning_rate": 2.8509215757137695e-05, "loss": 83.4667, "step": 279 }, { "compression_loss": 82.26457214355469, "epoch": 0.1, "learning_rate": 2.8503794723527287e-05, "loss": 82.7495, "step": 280 }, { "compression_loss": 83.13746643066406, "epoch": 0.1, "learning_rate": 2.8498373689916878e-05, "loss": 83.5181, "step": 281 }, { "compression_loss": 83.57550048828125, "epoch": 0.1, "learning_rate": 2.8492952656306473e-05, "loss": 83.8179, "step": 282 }, { "compression_loss": 84.67514038085938, "epoch": 0.1, "learning_rate": 2.8487531622696064e-05, "loss": 85.2486, "step": 283 }, { "compression_loss": 83.74687957763672, "epoch": 0.1, "learning_rate": 2.8482110589085655e-05, "loss": 84.3098, "step": 284 }, { "compression_loss": 81.7544937133789, "epoch": 0.1, "learning_rate": 2.8476689555475246e-05, "loss": 82.0451, "step": 285 }, { "compression_loss": 84.91880798339844, "epoch": 0.1, "learning_rate": 2.8471268521864838e-05, "loss": 85.6298, "step": 286 }, { "compression_loss": 85.25128936767578, "epoch": 0.1, "learning_rate": 2.846584748825443e-05, "loss": 85.7442, "step": 287 }, { "compression_loss": 84.39530944824219, "epoch": 0.1, "learning_rate": 2.8460426454644017e-05, "loss": 85.0364, "step": 288 }, { "compression_loss": 83.81135559082031, "epoch": 0.1, "learning_rate": 2.8455005421033608e-05, "loss": 84.0853, "step": 289 }, { "compression_loss": 85.3443603515625, "epoch": 0.1, "learning_rate": 2.8449584387423203e-05, "loss": 86.131, "step": 290 }, { "compression_loss": 85.57710266113281, "epoch": 0.11, "learning_rate": 2.8444163353812794e-05, "loss": 85.9309, "step": 291 }, { "compression_loss": 85.01205444335938, "epoch": 0.11, "learning_rate": 2.8438742320202385e-05, "loss": 85.6251, "step": 292 }, { "compression_loss": 83.52633666992188, "epoch": 0.11, "learning_rate": 2.8433321286591977e-05, "loss": 83.9065, "step": 293 }, { "compression_loss": 82.90347290039062, "epoch": 0.11, "learning_rate": 2.8427900252981568e-05, "loss": 83.3837, "step": 294 }, { "compression_loss": 85.62637329101562, "epoch": 0.11, "learning_rate": 2.842247921937116e-05, "loss": 86.2471, "step": 295 }, { "compression_loss": 83.55464172363281, "epoch": 0.11, "learning_rate": 2.841705818576075e-05, "loss": 83.9977, "step": 296 }, { "compression_loss": 85.49636840820312, "epoch": 0.11, "learning_rate": 2.841163715215034e-05, "loss": 86.0586, "step": 297 }, { "compression_loss": 84.72897338867188, "epoch": 0.11, "learning_rate": 2.8406216118539936e-05, "loss": 85.1983, "step": 298 }, { "compression_loss": 86.55899047851562, "epoch": 0.11, "learning_rate": 2.8400795084929528e-05, "loss": 87.2101, "step": 299 }, { "compression_loss": 85.38566589355469, "epoch": 0.11, "learning_rate": 2.839537405131912e-05, "loss": 86.1023, "step": 300 }, { "compression_loss": 84.99784851074219, "epoch": 0.11, "learning_rate": 2.838995301770871e-05, "loss": 85.4145, "step": 301 }, { "compression_loss": 85.00364685058594, "epoch": 0.11, "learning_rate": 2.83845319840983e-05, "loss": 85.7617, "step": 302 }, { "compression_loss": 84.55653381347656, "epoch": 0.11, "learning_rate": 2.8379110950487893e-05, "loss": 85.3098, "step": 303 }, { "compression_loss": 84.50532531738281, "epoch": 0.11, "learning_rate": 2.8373689916877484e-05, "loss": 85.0471, "step": 304 }, { "compression_loss": 84.37663269042969, "epoch": 0.11, "learning_rate": 2.8368268883267075e-05, "loss": 84.8985, "step": 305 }, { "compression_loss": 83.7489242553711, "epoch": 0.11, "learning_rate": 2.836284784965667e-05, "loss": 84.2225, "step": 306 }, { "compression_loss": 84.07115936279297, "epoch": 0.11, "learning_rate": 2.835742681604626e-05, "loss": 84.4293, "step": 307 }, { "compression_loss": 83.94087982177734, "epoch": 0.11, "learning_rate": 2.8352005782435852e-05, "loss": 84.5368, "step": 308 }, { "compression_loss": 84.23505401611328, "epoch": 0.11, "learning_rate": 2.8346584748825444e-05, "loss": 84.7503, "step": 309 }, { "compression_loss": 84.97795104980469, "epoch": 0.11, "learning_rate": 2.8341163715215035e-05, "loss": 85.9472, "step": 310 }, { "compression_loss": 83.16683959960938, "epoch": 0.11, "learning_rate": 2.8335742681604626e-05, "loss": 83.6325, "step": 311 }, { "compression_loss": 84.75828552246094, "epoch": 0.11, "learning_rate": 2.8330321647994218e-05, "loss": 85.3262, "step": 312 }, { "compression_loss": 84.3406982421875, "epoch": 0.11, "learning_rate": 2.832490061438381e-05, "loss": 84.6572, "step": 313 }, { "compression_loss": 84.17280578613281, "epoch": 0.11, "learning_rate": 2.8319479580773403e-05, "loss": 85.0238, "step": 314 }, { "compression_loss": 85.8260498046875, "epoch": 0.11, "learning_rate": 2.8314058547162995e-05, "loss": 86.2271, "step": 315 }, { "compression_loss": 84.95227813720703, "epoch": 0.11, "learning_rate": 2.8308637513552586e-05, "loss": 85.5704, "step": 316 }, { "compression_loss": 84.26910400390625, "epoch": 0.11, "learning_rate": 2.8303216479942177e-05, "loss": 84.8412, "step": 317 }, { "compression_loss": 84.0335922241211, "epoch": 0.11, "learning_rate": 2.829779544633177e-05, "loss": 84.5621, "step": 318 }, { "compression_loss": 83.34211730957031, "epoch": 0.12, "learning_rate": 2.829237441272136e-05, "loss": 83.7265, "step": 319 }, { "compression_loss": 84.85005187988281, "epoch": 0.12, "learning_rate": 2.828695337911095e-05, "loss": 85.4373, "step": 320 }, { "compression_loss": 85.72159576416016, "epoch": 0.12, "learning_rate": 2.8281532345500542e-05, "loss": 86.226, "step": 321 }, { "compression_loss": 85.24807739257812, "epoch": 0.12, "learning_rate": 2.8276111311890137e-05, "loss": 85.9074, "step": 322 }, { "compression_loss": 83.19929504394531, "epoch": 0.12, "learning_rate": 2.8270690278279728e-05, "loss": 83.6927, "step": 323 }, { "compression_loss": 83.62713623046875, "epoch": 0.12, "learning_rate": 2.826526924466932e-05, "loss": 84.1929, "step": 324 }, { "compression_loss": 84.07491302490234, "epoch": 0.12, "learning_rate": 2.825984821105891e-05, "loss": 84.5025, "step": 325 }, { "compression_loss": 82.826416015625, "epoch": 0.12, "learning_rate": 2.8254427177448502e-05, "loss": 83.2365, "step": 326 }, { "compression_loss": 84.59883117675781, "epoch": 0.12, "learning_rate": 2.8249006143838093e-05, "loss": 85.0881, "step": 327 }, { "compression_loss": 83.85711669921875, "epoch": 0.12, "learning_rate": 2.8243585110227685e-05, "loss": 84.7751, "step": 328 }, { "compression_loss": 84.49691772460938, "epoch": 0.12, "learning_rate": 2.8238164076617276e-05, "loss": 85.0935, "step": 329 }, { "compression_loss": 84.08187103271484, "epoch": 0.12, "learning_rate": 2.823274304300687e-05, "loss": 84.7472, "step": 330 }, { "compression_loss": 83.679931640625, "epoch": 0.12, "learning_rate": 2.822732200939646e-05, "loss": 83.9958, "step": 331 }, { "compression_loss": 82.90784454345703, "epoch": 0.12, "learning_rate": 2.822190097578605e-05, "loss": 83.4156, "step": 332 }, { "compression_loss": 84.24812316894531, "epoch": 0.12, "learning_rate": 2.821647994217564e-05, "loss": 85.0323, "step": 333 }, { "compression_loss": 84.5469970703125, "epoch": 0.12, "learning_rate": 2.8211058908565232e-05, "loss": 85.1947, "step": 334 }, { "compression_loss": 84.42156982421875, "epoch": 0.12, "learning_rate": 2.8205637874954824e-05, "loss": 85.0767, "step": 335 }, { "compression_loss": 84.2613525390625, "epoch": 0.12, "learning_rate": 2.8200216841344415e-05, "loss": 85.0473, "step": 336 }, { "compression_loss": 82.29255676269531, "epoch": 0.12, "learning_rate": 2.8194795807734006e-05, "loss": 82.8693, "step": 337 }, { "compression_loss": 84.73880004882812, "epoch": 0.12, "learning_rate": 2.81893747741236e-05, "loss": 85.0191, "step": 338 }, { "compression_loss": 87.56228637695312, "epoch": 0.12, "learning_rate": 2.8183953740513192e-05, "loss": 88.2117, "step": 339 }, { "compression_loss": 84.78945922851562, "epoch": 0.12, "learning_rate": 2.8178532706902783e-05, "loss": 85.691, "step": 340 }, { "compression_loss": 86.42613983154297, "epoch": 0.12, "learning_rate": 2.8173111673292375e-05, "loss": 87.4344, "step": 341 }, { "compression_loss": 84.55069732666016, "epoch": 0.12, "learning_rate": 2.8167690639681966e-05, "loss": 85.2132, "step": 342 }, { "compression_loss": 83.99800872802734, "epoch": 0.12, "learning_rate": 2.8162269606071557e-05, "loss": 84.3331, "step": 343 }, { "compression_loss": 84.83384704589844, "epoch": 0.12, "learning_rate": 2.815684857246115e-05, "loss": 85.3407, "step": 344 }, { "compression_loss": 85.49198913574219, "epoch": 0.12, "learning_rate": 2.815142753885074e-05, "loss": 86.259, "step": 345 }, { "compression_loss": 84.30152893066406, "epoch": 0.13, "learning_rate": 2.8146006505240334e-05, "loss": 85.1316, "step": 346 }, { "compression_loss": 82.81570434570312, "epoch": 0.13, "learning_rate": 2.8140585471629926e-05, "loss": 83.313, "step": 347 }, { "compression_loss": 82.58336639404297, "epoch": 0.13, "learning_rate": 2.8135164438019517e-05, "loss": 83.1191, "step": 348 }, { "compression_loss": 83.27227783203125, "epoch": 0.13, "learning_rate": 2.8129743404409108e-05, "loss": 83.763, "step": 349 }, { "compression_loss": 84.67218017578125, "epoch": 0.13, "learning_rate": 2.81243223707987e-05, "loss": 85.1254, "step": 350 }, { "compression_loss": 83.6842041015625, "epoch": 0.13, "learning_rate": 2.811890133718829e-05, "loss": 84.2209, "step": 351 }, { "compression_loss": 84.22510528564453, "epoch": 0.13, "learning_rate": 2.8113480303577882e-05, "loss": 84.7259, "step": 352 }, { "compression_loss": 84.50341033935547, "epoch": 0.13, "learning_rate": 2.8108059269967473e-05, "loss": 85.1747, "step": 353 }, { "compression_loss": 84.85893249511719, "epoch": 0.13, "learning_rate": 2.8102638236357068e-05, "loss": 85.5263, "step": 354 }, { "compression_loss": 86.04774475097656, "epoch": 0.13, "learning_rate": 2.809721720274666e-05, "loss": 86.5379, "step": 355 }, { "compression_loss": 84.92120361328125, "epoch": 0.13, "learning_rate": 2.809179616913625e-05, "loss": 85.3589, "step": 356 }, { "compression_loss": 84.98701477050781, "epoch": 0.13, "learning_rate": 2.8086375135525842e-05, "loss": 85.971, "step": 357 }, { "compression_loss": 84.60720825195312, "epoch": 0.13, "learning_rate": 2.8080954101915433e-05, "loss": 85.6357, "step": 358 }, { "compression_loss": 85.13993835449219, "epoch": 0.13, "learning_rate": 2.8075533068305024e-05, "loss": 85.6935, "step": 359 }, { "compression_loss": 84.2608871459961, "epoch": 0.13, "learning_rate": 2.8070112034694616e-05, "loss": 84.7687, "step": 360 }, { "compression_loss": 84.86787414550781, "epoch": 0.13, "learning_rate": 2.8064691001084207e-05, "loss": 85.6009, "step": 361 }, { "compression_loss": 84.41200256347656, "epoch": 0.13, "learning_rate": 2.80592699674738e-05, "loss": 84.9352, "step": 362 }, { "compression_loss": 82.78460693359375, "epoch": 0.13, "learning_rate": 2.8053848933863393e-05, "loss": 83.5202, "step": 363 }, { "compression_loss": 84.7044677734375, "epoch": 0.13, "learning_rate": 2.8048427900252984e-05, "loss": 85.2268, "step": 364 }, { "compression_loss": 83.74159240722656, "epoch": 0.13, "learning_rate": 2.8043006866642575e-05, "loss": 84.1137, "step": 365 }, { "compression_loss": 85.11663818359375, "epoch": 0.13, "learning_rate": 2.8037585833032167e-05, "loss": 85.5134, "step": 366 }, { "compression_loss": 84.53040313720703, "epoch": 0.13, "learning_rate": 2.8032164799421758e-05, "loss": 85.1756, "step": 367 }, { "compression_loss": 84.1705093383789, "epoch": 0.13, "learning_rate": 2.802674376581135e-05, "loss": 84.8304, "step": 368 }, { "compression_loss": 85.02725219726562, "epoch": 0.13, "learning_rate": 2.802132273220094e-05, "loss": 85.6572, "step": 369 }, { "compression_loss": 83.56877136230469, "epoch": 0.13, "learning_rate": 2.8015901698590535e-05, "loss": 84.0059, "step": 370 }, { "compression_loss": 86.46388244628906, "epoch": 0.13, "learning_rate": 2.8010480664980126e-05, "loss": 87.5818, "step": 371 }, { "compression_loss": 85.06462860107422, "epoch": 0.13, "learning_rate": 2.8005059631369718e-05, "loss": 86.1009, "step": 372 }, { "compression_loss": 83.66384887695312, "epoch": 0.13, "learning_rate": 2.799963859775931e-05, "loss": 84.1646, "step": 373 }, { "compression_loss": 84.38356018066406, "epoch": 0.14, "learning_rate": 2.79942175641489e-05, "loss": 84.8471, "step": 374 }, { "compression_loss": 83.08638000488281, "epoch": 0.14, "learning_rate": 2.7988796530538488e-05, "loss": 83.4856, "step": 375 }, { "compression_loss": 84.62490844726562, "epoch": 0.14, "learning_rate": 2.798337549692808e-05, "loss": 85.023, "step": 376 }, { "compression_loss": 85.49107360839844, "epoch": 0.14, "learning_rate": 2.797795446331767e-05, "loss": 86.4457, "step": 377 }, { "compression_loss": 84.84919738769531, "epoch": 0.14, "learning_rate": 2.7972533429707262e-05, "loss": 85.64, "step": 378 }, { "compression_loss": 84.76888275146484, "epoch": 0.14, "learning_rate": 2.7967112396096856e-05, "loss": 85.8023, "step": 379 }, { "compression_loss": 85.3145751953125, "epoch": 0.14, "learning_rate": 2.7961691362486448e-05, "loss": 85.9466, "step": 380 }, { "compression_loss": 85.07283020019531, "epoch": 0.14, "learning_rate": 2.795627032887604e-05, "loss": 85.5837, "step": 381 }, { "compression_loss": 84.96563720703125, "epoch": 0.14, "learning_rate": 2.795084929526563e-05, "loss": 85.7463, "step": 382 }, { "compression_loss": 83.45459747314453, "epoch": 0.14, "learning_rate": 2.794542826165522e-05, "loss": 83.7705, "step": 383 }, { "compression_loss": 84.5423583984375, "epoch": 0.14, "learning_rate": 2.7940007228044813e-05, "loss": 85.2533, "step": 384 }, { "compression_loss": 84.04263305664062, "epoch": 0.14, "learning_rate": 2.7934586194434404e-05, "loss": 84.4861, "step": 385 }, { "compression_loss": 82.8313217163086, "epoch": 0.14, "learning_rate": 2.7929165160823995e-05, "loss": 83.3027, "step": 386 }, { "compression_loss": 83.43119812011719, "epoch": 0.14, "learning_rate": 2.792374412721359e-05, "loss": 84.1757, "step": 387 }, { "compression_loss": 84.67650604248047, "epoch": 0.14, "learning_rate": 2.791832309360318e-05, "loss": 85.311, "step": 388 }, { "compression_loss": 83.91355895996094, "epoch": 0.14, "learning_rate": 2.7912902059992773e-05, "loss": 84.3521, "step": 389 }, { "compression_loss": 84.1539306640625, "epoch": 0.14, "learning_rate": 2.7907481026382364e-05, "loss": 84.7358, "step": 390 }, { "compression_loss": 83.70205688476562, "epoch": 0.14, "learning_rate": 2.7902059992771955e-05, "loss": 84.2649, "step": 391 }, { "compression_loss": 85.3228759765625, "epoch": 0.14, "learning_rate": 2.7896638959161546e-05, "loss": 86.1357, "step": 392 }, { "compression_loss": 85.61737823486328, "epoch": 0.14, "learning_rate": 2.7891217925551138e-05, "loss": 86.0519, "step": 393 }, { "compression_loss": 84.36670684814453, "epoch": 0.14, "learning_rate": 2.788579689194073e-05, "loss": 85.1422, "step": 394 }, { "compression_loss": 84.51896667480469, "epoch": 0.14, "learning_rate": 2.7880375858330324e-05, "loss": 84.9841, "step": 395 }, { "compression_loss": 84.82728576660156, "epoch": 0.14, "learning_rate": 2.7874954824719915e-05, "loss": 85.4557, "step": 396 }, { "compression_loss": 84.74739074707031, "epoch": 0.14, "learning_rate": 2.7869533791109506e-05, "loss": 85.1954, "step": 397 }, { "compression_loss": 85.05335998535156, "epoch": 0.14, "learning_rate": 2.7864112757499097e-05, "loss": 85.8335, "step": 398 }, { "compression_loss": 85.10723876953125, "epoch": 0.14, "learning_rate": 2.785869172388869e-05, "loss": 85.5813, "step": 399 }, { "compression_loss": 85.0053482055664, "epoch": 0.14, "learning_rate": 2.785327069027828e-05, "loss": 85.7789, "step": 400 }, { "compression_loss": 84.19612121582031, "epoch": 0.14, "learning_rate": 2.784784965666787e-05, "loss": 84.8796, "step": 401 }, { "compression_loss": 85.0941162109375, "epoch": 0.15, "learning_rate": 2.7842428623057462e-05, "loss": 85.9739, "step": 402 }, { "compression_loss": 85.22056579589844, "epoch": 0.15, "learning_rate": 2.7837007589447057e-05, "loss": 85.7906, "step": 403 }, { "compression_loss": 85.36778259277344, "epoch": 0.15, "learning_rate": 2.783158655583665e-05, "loss": 85.8937, "step": 404 }, { "compression_loss": 85.18685150146484, "epoch": 0.15, "learning_rate": 2.782616552222624e-05, "loss": 85.9516, "step": 405 }, { "compression_loss": 85.03721618652344, "epoch": 0.15, "learning_rate": 2.782074448861583e-05, "loss": 85.4845, "step": 406 }, { "compression_loss": 82.08303833007812, "epoch": 0.15, "learning_rate": 2.7815323455005422e-05, "loss": 82.5607, "step": 407 }, { "compression_loss": 84.82384490966797, "epoch": 0.15, "learning_rate": 2.7809902421395014e-05, "loss": 85.6117, "step": 408 }, { "compression_loss": 83.6595230102539, "epoch": 0.15, "learning_rate": 2.7804481387784605e-05, "loss": 84.027, "step": 409 }, { "compression_loss": 85.08820343017578, "epoch": 0.15, "learning_rate": 2.7799060354174196e-05, "loss": 85.6546, "step": 410 }, { "compression_loss": 84.58966827392578, "epoch": 0.15, "learning_rate": 2.779363932056379e-05, "loss": 85.4456, "step": 411 }, { "compression_loss": 84.33181762695312, "epoch": 0.15, "learning_rate": 2.7788218286953382e-05, "loss": 84.7151, "step": 412 }, { "compression_loss": 84.177734375, "epoch": 0.15, "learning_rate": 2.7782797253342973e-05, "loss": 84.7179, "step": 413 }, { "compression_loss": 83.44587707519531, "epoch": 0.15, "learning_rate": 2.7777376219732565e-05, "loss": 83.831, "step": 414 }, { "compression_loss": 84.39089965820312, "epoch": 0.15, "learning_rate": 2.7771955186122156e-05, "loss": 85.1462, "step": 415 }, { "compression_loss": 83.85928344726562, "epoch": 0.15, "learning_rate": 2.7766534152511747e-05, "loss": 84.3409, "step": 416 }, { "compression_loss": 85.09506225585938, "epoch": 0.15, "learning_rate": 2.776111311890134e-05, "loss": 85.5133, "step": 417 }, { "compression_loss": 83.85865020751953, "epoch": 0.15, "learning_rate": 2.775569208529093e-05, "loss": 84.1859, "step": 418 }, { "compression_loss": 83.40725708007812, "epoch": 0.15, "learning_rate": 2.775027105168052e-05, "loss": 84.0185, "step": 419 }, { "compression_loss": 84.5224609375, "epoch": 0.15, "learning_rate": 2.7744850018070112e-05, "loss": 85.2118, "step": 420 }, { "compression_loss": 85.75654602050781, "epoch": 0.15, "learning_rate": 2.7739428984459703e-05, "loss": 86.646, "step": 421 }, { "compression_loss": 84.95301055908203, "epoch": 0.15, "learning_rate": 2.7734007950849295e-05, "loss": 85.5105, "step": 422 }, { "compression_loss": 83.64949035644531, "epoch": 0.15, "learning_rate": 2.7728586917238886e-05, "loss": 84.0385, "step": 423 }, { "compression_loss": 83.5283432006836, "epoch": 0.15, "learning_rate": 2.7723165883628477e-05, "loss": 83.9072, "step": 424 }, { "compression_loss": 85.62428283691406, "epoch": 0.15, "learning_rate": 2.771774485001807e-05, "loss": 86.4335, "step": 425 }, { "compression_loss": 85.31475067138672, "epoch": 0.15, "learning_rate": 2.771232381640766e-05, "loss": 85.9117, "step": 426 }, { "compression_loss": 84.24671936035156, "epoch": 0.15, "learning_rate": 2.7706902782797254e-05, "loss": 84.7702, "step": 427 }, { "compression_loss": 84.24510192871094, "epoch": 0.15, "learning_rate": 2.7701481749186846e-05, "loss": 84.8701, "step": 428 }, { "compression_loss": 85.16385650634766, "epoch": 0.16, "learning_rate": 2.7696060715576437e-05, "loss": 85.9041, "step": 429 }, { "compression_loss": 83.88288879394531, "epoch": 0.16, "learning_rate": 2.7690639681966028e-05, "loss": 84.4581, "step": 430 }, { "compression_loss": 84.4958267211914, "epoch": 0.16, "learning_rate": 2.768521864835562e-05, "loss": 85.0397, "step": 431 }, { "compression_loss": 85.82415008544922, "epoch": 0.16, "learning_rate": 2.767979761474521e-05, "loss": 86.6095, "step": 432 }, { "compression_loss": 83.03767395019531, "epoch": 0.16, "learning_rate": 2.7674376581134802e-05, "loss": 83.65, "step": 433 }, { "compression_loss": 83.32955932617188, "epoch": 0.16, "learning_rate": 2.7668955547524393e-05, "loss": 83.5781, "step": 434 }, { "compression_loss": 83.44166564941406, "epoch": 0.16, "learning_rate": 2.7663534513913988e-05, "loss": 83.9231, "step": 435 }, { "compression_loss": 84.4182357788086, "epoch": 0.16, "learning_rate": 2.765811348030358e-05, "loss": 84.6745, "step": 436 }, { "compression_loss": 86.00543975830078, "epoch": 0.16, "learning_rate": 2.765269244669317e-05, "loss": 86.6605, "step": 437 }, { "compression_loss": 83.80541229248047, "epoch": 0.16, "learning_rate": 2.7647271413082762e-05, "loss": 84.3366, "step": 438 }, { "compression_loss": 87.53416442871094, "epoch": 0.16, "learning_rate": 2.7641850379472353e-05, "loss": 88.358, "step": 439 }, { "compression_loss": 84.59054565429688, "epoch": 0.16, "learning_rate": 2.7636429345861944e-05, "loss": 85.475, "step": 440 }, { "compression_loss": 85.02266693115234, "epoch": 0.16, "learning_rate": 2.7631008312251536e-05, "loss": 85.9271, "step": 441 }, { "compression_loss": 82.70005798339844, "epoch": 0.16, "learning_rate": 2.7625587278641127e-05, "loss": 83.1163, "step": 442 }, { "compression_loss": 83.44100952148438, "epoch": 0.16, "learning_rate": 2.762016624503072e-05, "loss": 84.2354, "step": 443 }, { "compression_loss": 85.48541259765625, "epoch": 0.16, "learning_rate": 2.7614745211420313e-05, "loss": 85.8845, "step": 444 }, { "compression_loss": 85.2613296508789, "epoch": 0.16, "learning_rate": 2.7609324177809904e-05, "loss": 86.1576, "step": 445 }, { "compression_loss": 84.44761657714844, "epoch": 0.16, "learning_rate": 2.7603903144199495e-05, "loss": 85.0659, "step": 446 }, { "compression_loss": 83.56965637207031, "epoch": 0.16, "learning_rate": 2.7598482110589087e-05, "loss": 84.2816, "step": 447 }, { "compression_loss": 85.47257995605469, "epoch": 0.16, "learning_rate": 2.7593061076978678e-05, "loss": 86.3765, "step": 448 }, { "compression_loss": 82.71307373046875, "epoch": 0.16, "learning_rate": 2.758764004336827e-05, "loss": 83.1717, "step": 449 }, { "compression_loss": 85.25990295410156, "epoch": 0.16, "learning_rate": 2.758221900975786e-05, "loss": 85.5117, "step": 450 }, { "compression_loss": 85.62578582763672, "epoch": 0.16, "learning_rate": 2.7576797976147455e-05, "loss": 86.3133, "step": 451 }, { "compression_loss": 86.20317077636719, "epoch": 0.16, "learning_rate": 2.7571376942537046e-05, "loss": 86.5913, "step": 452 }, { "compression_loss": 84.9404296875, "epoch": 0.16, "learning_rate": 2.7565955908926638e-05, "loss": 85.7504, "step": 453 }, { "compression_loss": 84.59127807617188, "epoch": 0.16, "learning_rate": 2.756053487531623e-05, "loss": 85.0281, "step": 454 }, { "compression_loss": 84.67822265625, "epoch": 0.16, "learning_rate": 2.755511384170582e-05, "loss": 85.5244, "step": 455 }, { "compression_loss": 84.62399291992188, "epoch": 0.16, "learning_rate": 2.754969280809541e-05, "loss": 85.4819, "step": 456 }, { "compression_loss": 84.78485870361328, "epoch": 0.17, "learning_rate": 2.7544271774485003e-05, "loss": 85.1786, "step": 457 }, { "compression_loss": 85.6932373046875, "epoch": 0.17, "learning_rate": 2.7538850740874594e-05, "loss": 86.3755, "step": 458 }, { "compression_loss": 84.36193084716797, "epoch": 0.17, "learning_rate": 2.753342970726419e-05, "loss": 84.9018, "step": 459 }, { "compression_loss": 83.19586181640625, "epoch": 0.17, "learning_rate": 2.752800867365378e-05, "loss": 83.6738, "step": 460 }, { "compression_loss": 86.06985473632812, "epoch": 0.17, "learning_rate": 2.752258764004337e-05, "loss": 86.8258, "step": 461 }, { "compression_loss": 85.30824279785156, "epoch": 0.17, "learning_rate": 2.751716660643296e-05, "loss": 85.8563, "step": 462 }, { "compression_loss": 85.08732604980469, "epoch": 0.17, "learning_rate": 2.751174557282255e-05, "loss": 85.6331, "step": 463 }, { "compression_loss": 84.57818603515625, "epoch": 0.17, "learning_rate": 2.750632453921214e-05, "loss": 84.9398, "step": 464 }, { "compression_loss": 84.74296569824219, "epoch": 0.17, "learning_rate": 2.7500903505601733e-05, "loss": 85.4838, "step": 465 }, { "compression_loss": 84.15586853027344, "epoch": 0.17, "learning_rate": 2.7495482471991324e-05, "loss": 84.8597, "step": 466 }, { "compression_loss": 85.18586730957031, "epoch": 0.17, "learning_rate": 2.749006143838092e-05, "loss": 85.7884, "step": 467 }, { "compression_loss": 84.52879333496094, "epoch": 0.17, "learning_rate": 2.748464040477051e-05, "loss": 85.0512, "step": 468 }, { "compression_loss": 84.99238586425781, "epoch": 0.17, "learning_rate": 2.74792193711601e-05, "loss": 85.6001, "step": 469 }, { "compression_loss": 84.21357727050781, "epoch": 0.17, "learning_rate": 2.7473798337549693e-05, "loss": 84.7654, "step": 470 }, { "compression_loss": 84.71051788330078, "epoch": 0.17, "learning_rate": 2.7468377303939284e-05, "loss": 85.3188, "step": 471 }, { "compression_loss": 82.78368377685547, "epoch": 0.17, "learning_rate": 2.7462956270328875e-05, "loss": 83.2914, "step": 472 }, { "compression_loss": 84.674560546875, "epoch": 0.17, "learning_rate": 2.7457535236718467e-05, "loss": 85.2202, "step": 473 }, { "compression_loss": 84.3699951171875, "epoch": 0.17, "learning_rate": 2.7452114203108058e-05, "loss": 84.9094, "step": 474 }, { "compression_loss": 84.44137573242188, "epoch": 0.17, "learning_rate": 2.7446693169497652e-05, "loss": 85.3394, "step": 475 }, { "compression_loss": 84.02284240722656, "epoch": 0.17, "learning_rate": 2.7441272135887244e-05, "loss": 84.6896, "step": 476 }, { "compression_loss": 82.76435852050781, "epoch": 0.17, "learning_rate": 2.7435851102276835e-05, "loss": 83.3085, "step": 477 }, { "compression_loss": 87.01901245117188, "epoch": 0.17, "learning_rate": 2.7430430068666426e-05, "loss": 87.6368, "step": 478 }, { "compression_loss": 84.71241760253906, "epoch": 0.17, "learning_rate": 2.7425009035056018e-05, "loss": 85.2415, "step": 479 }, { "compression_loss": 83.83924102783203, "epoch": 0.17, "learning_rate": 2.741958800144561e-05, "loss": 84.1631, "step": 480 }, { "compression_loss": 85.75215911865234, "epoch": 0.17, "learning_rate": 2.74141669678352e-05, "loss": 86.423, "step": 481 }, { "compression_loss": 84.54859161376953, "epoch": 0.17, "learning_rate": 2.740874593422479e-05, "loss": 85.0913, "step": 482 }, { "compression_loss": 84.56538391113281, "epoch": 0.17, "learning_rate": 2.7403324900614386e-05, "loss": 85.167, "step": 483 }, { "compression_loss": 83.82896423339844, "epoch": 0.17, "learning_rate": 2.7397903867003977e-05, "loss": 84.5368, "step": 484 }, { "compression_loss": 85.7573471069336, "epoch": 0.18, "learning_rate": 2.739248283339357e-05, "loss": 86.5162, "step": 485 }, { "compression_loss": 86.42817687988281, "epoch": 0.18, "learning_rate": 2.738706179978316e-05, "loss": 87.0208, "step": 486 }, { "compression_loss": 85.25578308105469, "epoch": 0.18, "learning_rate": 2.738164076617275e-05, "loss": 85.9316, "step": 487 }, { "compression_loss": 84.5115966796875, "epoch": 0.18, "learning_rate": 2.7376219732562342e-05, "loss": 84.973, "step": 488 }, { "compression_loss": 87.6007080078125, "epoch": 0.18, "learning_rate": 2.7370798698951934e-05, "loss": 88.1982, "step": 489 }, { "compression_loss": 87.4936294555664, "epoch": 0.18, "learning_rate": 2.7365377665341525e-05, "loss": 88.2095, "step": 490 }, { "compression_loss": 84.20169067382812, "epoch": 0.18, "learning_rate": 2.735995663173112e-05, "loss": 84.9769, "step": 491 }, { "compression_loss": 85.6832275390625, "epoch": 0.18, "learning_rate": 2.735453559812071e-05, "loss": 86.2472, "step": 492 }, { "compression_loss": 84.46385955810547, "epoch": 0.18, "learning_rate": 2.7349114564510302e-05, "loss": 85.0831, "step": 493 }, { "compression_loss": 84.92608642578125, "epoch": 0.18, "learning_rate": 2.7343693530899893e-05, "loss": 85.5384, "step": 494 }, { "compression_loss": 84.07304382324219, "epoch": 0.18, "learning_rate": 2.7338272497289485e-05, "loss": 84.4826, "step": 495 }, { "compression_loss": 86.9497299194336, "epoch": 0.18, "learning_rate": 2.7332851463679076e-05, "loss": 87.8702, "step": 496 }, { "compression_loss": 84.06285095214844, "epoch": 0.18, "learning_rate": 2.7327430430068667e-05, "loss": 84.719, "step": 497 }, { "compression_loss": 85.79522705078125, "epoch": 0.18, "learning_rate": 2.732200939645826e-05, "loss": 86.206, "step": 498 }, { "compression_loss": 83.08586120605469, "epoch": 0.18, "learning_rate": 2.7316588362847853e-05, "loss": 83.6409, "step": 499 }, { "compression_loss": 83.40800476074219, "epoch": 0.18, "learning_rate": 2.7311167329237444e-05, "loss": 83.7052, "step": 500 }, { "epoch": 0.18, "eval_exact_match": 85.23178807947019, "eval_f1": 91.90115545003138, "step": 500 }, { "compression_loss": 83.5028305053711, "epoch": 0.18, "learning_rate": 2.7305746295627036e-05, "loss": 83.9793, "step": 501 }, { "compression_loss": 84.42515563964844, "epoch": 0.18, "learning_rate": 2.7300325262016627e-05, "loss": 85.1745, "step": 502 }, { "compression_loss": 85.45045471191406, "epoch": 0.18, "learning_rate": 2.7294904228406218e-05, "loss": 86.3126, "step": 503 }, { "compression_loss": 84.3325424194336, "epoch": 0.18, "learning_rate": 2.728948319479581e-05, "loss": 84.9511, "step": 504 }, { "compression_loss": 85.80323028564453, "epoch": 0.18, "learning_rate": 2.72840621611854e-05, "loss": 86.699, "step": 505 }, { "compression_loss": 85.15342712402344, "epoch": 0.18, "learning_rate": 2.727864112757499e-05, "loss": 85.6118, "step": 506 }, { "compression_loss": 82.70216369628906, "epoch": 0.18, "learning_rate": 2.7273220093964583e-05, "loss": 83.097, "step": 507 }, { "compression_loss": 86.38706970214844, "epoch": 0.18, "learning_rate": 2.7267799060354175e-05, "loss": 86.9648, "step": 508 }, { "compression_loss": 85.7657699584961, "epoch": 0.18, "learning_rate": 2.7262378026743766e-05, "loss": 86.6093, "step": 509 }, { "compression_loss": 84.98487854003906, "epoch": 0.18, "learning_rate": 2.7256956993133357e-05, "loss": 85.5847, "step": 510 }, { "compression_loss": 85.17729187011719, "epoch": 0.18, "learning_rate": 2.725153595952295e-05, "loss": 86.0241, "step": 511 }, { "compression_loss": 83.93744659423828, "epoch": 0.19, "learning_rate": 2.724611492591254e-05, "loss": 84.39, "step": 512 }, { "compression_loss": 84.51119995117188, "epoch": 0.19, "learning_rate": 2.724069389230213e-05, "loss": 85.0424, "step": 513 }, { "compression_loss": 83.31319427490234, "epoch": 0.19, "learning_rate": 2.7235272858691722e-05, "loss": 83.785, "step": 514 }, { "compression_loss": 87.15953063964844, "epoch": 0.19, "learning_rate": 2.7229851825081317e-05, "loss": 87.6063, "step": 515 }, { "compression_loss": 83.37252044677734, "epoch": 0.19, "learning_rate": 2.7224430791470908e-05, "loss": 83.7586, "step": 516 }, { "compression_loss": 84.82600402832031, "epoch": 0.19, "learning_rate": 2.72190097578605e-05, "loss": 85.4667, "step": 517 }, { "compression_loss": 85.36139678955078, "epoch": 0.19, "learning_rate": 2.721358872425009e-05, "loss": 86.1638, "step": 518 }, { "compression_loss": 83.14576721191406, "epoch": 0.19, "learning_rate": 2.7208167690639682e-05, "loss": 83.5516, "step": 519 }, { "compression_loss": 82.82257080078125, "epoch": 0.19, "learning_rate": 2.7202746657029273e-05, "loss": 84.0161, "step": 520 }, { "compression_loss": 83.54072570800781, "epoch": 0.19, "learning_rate": 2.7197325623418864e-05, "loss": 84.3874, "step": 521 }, { "compression_loss": 85.03089904785156, "epoch": 0.19, "learning_rate": 2.7191904589808456e-05, "loss": 85.6818, "step": 522 }, { "compression_loss": 84.02704620361328, "epoch": 0.19, "learning_rate": 2.718648355619805e-05, "loss": 84.3989, "step": 523 }, { "compression_loss": 84.38008117675781, "epoch": 0.19, "learning_rate": 2.7181062522587642e-05, "loss": 84.9413, "step": 524 }, { "compression_loss": 83.91609191894531, "epoch": 0.19, "learning_rate": 2.7175641488977233e-05, "loss": 84.2208, "step": 525 }, { "compression_loss": 84.738037109375, "epoch": 0.19, "learning_rate": 2.7170220455366824e-05, "loss": 85.4285, "step": 526 }, { "compression_loss": 83.9432601928711, "epoch": 0.19, "learning_rate": 2.7164799421756416e-05, "loss": 84.5846, "step": 527 }, { "compression_loss": 83.744873046875, "epoch": 0.19, "learning_rate": 2.7159378388146007e-05, "loss": 84.3872, "step": 528 }, { "compression_loss": 85.63014221191406, "epoch": 0.19, "learning_rate": 2.7153957354535598e-05, "loss": 86.2471, "step": 529 }, { "compression_loss": 83.69508361816406, "epoch": 0.19, "learning_rate": 2.714853632092519e-05, "loss": 84.1866, "step": 530 }, { "compression_loss": 83.882080078125, "epoch": 0.19, "learning_rate": 2.7143115287314784e-05, "loss": 84.3375, "step": 531 }, { "compression_loss": 83.30180358886719, "epoch": 0.19, "learning_rate": 2.7137694253704375e-05, "loss": 83.6382, "step": 532 }, { "compression_loss": 83.57830810546875, "epoch": 0.19, "learning_rate": 2.7132273220093967e-05, "loss": 84.3079, "step": 533 }, { "compression_loss": 85.38492584228516, "epoch": 0.19, "learning_rate": 2.7126852186483558e-05, "loss": 85.7804, "step": 534 }, { "compression_loss": 83.14556884765625, "epoch": 0.19, "learning_rate": 2.712143115287315e-05, "loss": 83.5578, "step": 535 }, { "compression_loss": 83.54669189453125, "epoch": 0.19, "learning_rate": 2.711601011926274e-05, "loss": 84.0877, "step": 536 }, { "compression_loss": 83.95381164550781, "epoch": 0.19, "learning_rate": 2.711058908565233e-05, "loss": 84.6181, "step": 537 }, { "compression_loss": 82.86602783203125, "epoch": 0.19, "learning_rate": 2.7105168052041923e-05, "loss": 83.3867, "step": 538 }, { "compression_loss": 84.19087982177734, "epoch": 0.19, "learning_rate": 2.7099747018431518e-05, "loss": 84.8682, "step": 539 }, { "compression_loss": 83.76351165771484, "epoch": 0.2, "learning_rate": 2.709432598482111e-05, "loss": 84.4369, "step": 540 }, { "compression_loss": 84.5472640991211, "epoch": 0.2, "learning_rate": 2.70889049512107e-05, "loss": 85.2444, "step": 541 }, { "compression_loss": 85.62640380859375, "epoch": 0.2, "learning_rate": 2.708348391760029e-05, "loss": 86.7256, "step": 542 }, { "compression_loss": 84.95481872558594, "epoch": 0.2, "learning_rate": 2.7078062883989883e-05, "loss": 85.3758, "step": 543 }, { "compression_loss": 83.33840942382812, "epoch": 0.2, "learning_rate": 2.7072641850379474e-05, "loss": 84.0249, "step": 544 }, { "compression_loss": 84.85759735107422, "epoch": 0.2, "learning_rate": 2.7067220816769065e-05, "loss": 85.2396, "step": 545 }, { "compression_loss": 83.45191192626953, "epoch": 0.2, "learning_rate": 2.7061799783158656e-05, "loss": 83.7699, "step": 546 }, { "compression_loss": 87.4626235961914, "epoch": 0.2, "learning_rate": 2.705637874954825e-05, "loss": 88.304, "step": 547 }, { "compression_loss": 83.5230941772461, "epoch": 0.2, "learning_rate": 2.7050957715937842e-05, "loss": 84.0269, "step": 548 }, { "compression_loss": 84.02068328857422, "epoch": 0.2, "learning_rate": 2.704553668232743e-05, "loss": 84.5525, "step": 549 }, { "compression_loss": 83.90840148925781, "epoch": 0.2, "learning_rate": 2.704011564871702e-05, "loss": 84.5336, "step": 550 }, { "compression_loss": 85.09613037109375, "epoch": 0.2, "learning_rate": 2.7034694615106613e-05, "loss": 85.7203, "step": 551 }, { "compression_loss": 84.09402465820312, "epoch": 0.2, "learning_rate": 2.7029273581496204e-05, "loss": 84.6697, "step": 552 }, { "compression_loss": 84.71495819091797, "epoch": 0.2, "learning_rate": 2.7023852547885795e-05, "loss": 85.2116, "step": 553 }, { "compression_loss": 85.71395874023438, "epoch": 0.2, "learning_rate": 2.7018431514275387e-05, "loss": 86.2336, "step": 554 }, { "compression_loss": 84.18038940429688, "epoch": 0.2, "learning_rate": 2.701301048066498e-05, "loss": 84.4991, "step": 555 }, { "compression_loss": 83.69888305664062, "epoch": 0.2, "learning_rate": 2.7007589447054573e-05, "loss": 84.3282, "step": 556 }, { "compression_loss": 84.62600708007812, "epoch": 0.2, "learning_rate": 2.7002168413444164e-05, "loss": 85.3772, "step": 557 }, { "compression_loss": 85.11152648925781, "epoch": 0.2, "learning_rate": 2.6996747379833755e-05, "loss": 85.9898, "step": 558 }, { "compression_loss": 86.62965393066406, "epoch": 0.2, "learning_rate": 2.6991326346223346e-05, "loss": 87.3503, "step": 559 }, { "compression_loss": 84.8294677734375, "epoch": 0.2, "learning_rate": 2.6985905312612938e-05, "loss": 85.664, "step": 560 }, { "compression_loss": 84.40382385253906, "epoch": 0.2, "learning_rate": 2.698048427900253e-05, "loss": 85.2583, "step": 561 }, { "compression_loss": 85.64547729492188, "epoch": 0.2, "learning_rate": 2.697506324539212e-05, "loss": 86.2665, "step": 562 }, { "compression_loss": 86.16989135742188, "epoch": 0.2, "learning_rate": 2.696964221178171e-05, "loss": 86.9579, "step": 563 }, { "compression_loss": 85.35030364990234, "epoch": 0.2, "learning_rate": 2.6964221178171306e-05, "loss": 85.8612, "step": 564 }, { "compression_loss": 85.29400634765625, "epoch": 0.2, "learning_rate": 2.6958800144560897e-05, "loss": 85.8394, "step": 565 }, { "compression_loss": 84.5017318725586, "epoch": 0.2, "learning_rate": 2.695337911095049e-05, "loss": 85.4259, "step": 566 }, { "compression_loss": 87.07488250732422, "epoch": 0.2, "learning_rate": 2.694795807734008e-05, "loss": 87.9406, "step": 567 }, { "compression_loss": 84.55655670166016, "epoch": 0.21, "learning_rate": 2.694253704372967e-05, "loss": 85.5666, "step": 568 }, { "compression_loss": 84.44180297851562, "epoch": 0.21, "learning_rate": 2.6937116010119262e-05, "loss": 85.191, "step": 569 }, { "compression_loss": 86.03453063964844, "epoch": 0.21, "learning_rate": 2.6931694976508854e-05, "loss": 86.6658, "step": 570 }, { "compression_loss": 85.08280181884766, "epoch": 0.21, "learning_rate": 2.6926273942898445e-05, "loss": 85.8107, "step": 571 }, { "compression_loss": 85.71024322509766, "epoch": 0.21, "learning_rate": 2.692085290928804e-05, "loss": 86.3597, "step": 572 }, { "compression_loss": 81.8438720703125, "epoch": 0.21, "learning_rate": 2.691543187567763e-05, "loss": 82.1211, "step": 573 }, { "compression_loss": 83.76292419433594, "epoch": 0.21, "learning_rate": 2.6910010842067222e-05, "loss": 84.398, "step": 574 }, { "compression_loss": 84.54730224609375, "epoch": 0.21, "learning_rate": 2.6904589808456814e-05, "loss": 85.1204, "step": 575 }, { "compression_loss": 85.99649047851562, "epoch": 0.21, "learning_rate": 2.6899168774846405e-05, "loss": 86.8168, "step": 576 }, { "compression_loss": 84.01283264160156, "epoch": 0.21, "learning_rate": 2.6893747741235996e-05, "loss": 84.5689, "step": 577 }, { "compression_loss": 83.19287872314453, "epoch": 0.21, "learning_rate": 2.6888326707625587e-05, "loss": 83.8263, "step": 578 }, { "compression_loss": 85.89366149902344, "epoch": 0.21, "learning_rate": 2.688290567401518e-05, "loss": 86.4662, "step": 579 }, { "compression_loss": 85.9215316772461, "epoch": 0.21, "learning_rate": 2.6877484640404773e-05, "loss": 86.3643, "step": 580 }, { "compression_loss": 85.8050537109375, "epoch": 0.21, "learning_rate": 2.6872063606794365e-05, "loss": 86.6402, "step": 581 }, { "compression_loss": 85.27418518066406, "epoch": 0.21, "learning_rate": 2.6866642573183956e-05, "loss": 85.7863, "step": 582 }, { "compression_loss": 84.62838745117188, "epoch": 0.21, "learning_rate": 2.6861221539573547e-05, "loss": 85.0683, "step": 583 }, { "compression_loss": 84.85310363769531, "epoch": 0.21, "learning_rate": 2.685580050596314e-05, "loss": 85.1334, "step": 584 }, { "compression_loss": 85.23345947265625, "epoch": 0.21, "learning_rate": 2.685037947235273e-05, "loss": 85.7028, "step": 585 }, { "compression_loss": 84.3292465209961, "epoch": 0.21, "learning_rate": 2.684495843874232e-05, "loss": 84.7264, "step": 586 }, { "compression_loss": 85.13322448730469, "epoch": 0.21, "learning_rate": 2.6839537405131912e-05, "loss": 85.8912, "step": 587 }, { "compression_loss": 85.05770874023438, "epoch": 0.21, "learning_rate": 2.6834116371521507e-05, "loss": 85.8402, "step": 588 }, { "compression_loss": 85.58414459228516, "epoch": 0.21, "learning_rate": 2.6828695337911098e-05, "loss": 86.0406, "step": 589 }, { "compression_loss": 84.98319244384766, "epoch": 0.21, "learning_rate": 2.682327430430069e-05, "loss": 85.4273, "step": 590 }, { "compression_loss": 83.89288330078125, "epoch": 0.21, "learning_rate": 2.681785327069028e-05, "loss": 84.5835, "step": 591 }, { "compression_loss": 84.23556518554688, "epoch": 0.21, "learning_rate": 2.6812432237079872e-05, "loss": 84.716, "step": 592 }, { "compression_loss": 85.10810852050781, "epoch": 0.21, "learning_rate": 2.680701120346946e-05, "loss": 86.0025, "step": 593 }, { "compression_loss": 84.68386840820312, "epoch": 0.21, "learning_rate": 2.680159016985905e-05, "loss": 85.3498, "step": 594 }, { "compression_loss": 84.50961303710938, "epoch": 0.22, "learning_rate": 2.6796169136248642e-05, "loss": 85.2257, "step": 595 }, { "compression_loss": 85.7369613647461, "epoch": 0.22, "learning_rate": 2.6790748102638237e-05, "loss": 86.3061, "step": 596 }, { "compression_loss": 82.30980682373047, "epoch": 0.22, "learning_rate": 2.6785327069027828e-05, "loss": 82.9558, "step": 597 }, { "compression_loss": 83.92352294921875, "epoch": 0.22, "learning_rate": 2.677990603541742e-05, "loss": 85.1298, "step": 598 }, { "compression_loss": 83.79855346679688, "epoch": 0.22, "learning_rate": 2.677448500180701e-05, "loss": 84.6915, "step": 599 }, { "compression_loss": 84.5593490600586, "epoch": 0.22, "learning_rate": 2.6769063968196602e-05, "loss": 84.9418, "step": 600 }, { "compression_loss": 84.74146270751953, "epoch": 0.22, "learning_rate": 2.6763642934586193e-05, "loss": 85.315, "step": 601 }, { "compression_loss": 83.797119140625, "epoch": 0.22, "learning_rate": 2.6758221900975785e-05, "loss": 84.363, "step": 602 }, { "compression_loss": 82.20193481445312, "epoch": 0.22, "learning_rate": 2.6752800867365376e-05, "loss": 82.7038, "step": 603 }, { "compression_loss": 83.82501220703125, "epoch": 0.22, "learning_rate": 2.674737983375497e-05, "loss": 84.3392, "step": 604 }, { "compression_loss": 85.17945098876953, "epoch": 0.22, "learning_rate": 2.6741958800144562e-05, "loss": 85.7907, "step": 605 }, { "compression_loss": 85.05610656738281, "epoch": 0.22, "learning_rate": 2.6736537766534153e-05, "loss": 85.7266, "step": 606 }, { "compression_loss": 84.80133056640625, "epoch": 0.22, "learning_rate": 2.6731116732923744e-05, "loss": 85.7252, "step": 607 }, { "compression_loss": 82.87449645996094, "epoch": 0.22, "learning_rate": 2.6725695699313336e-05, "loss": 83.7658, "step": 608 }, { "compression_loss": 83.46087646484375, "epoch": 0.22, "learning_rate": 2.6720274665702927e-05, "loss": 84.0778, "step": 609 }, { "compression_loss": 84.21000671386719, "epoch": 0.22, "learning_rate": 2.6714853632092518e-05, "loss": 84.7092, "step": 610 }, { "compression_loss": 84.76103210449219, "epoch": 0.22, "learning_rate": 2.670943259848211e-05, "loss": 85.2818, "step": 611 }, { "compression_loss": 83.63018798828125, "epoch": 0.22, "learning_rate": 2.6704011564871704e-05, "loss": 83.9992, "step": 612 }, { "compression_loss": 84.11361694335938, "epoch": 0.22, "learning_rate": 2.6698590531261295e-05, "loss": 84.6375, "step": 613 }, { "compression_loss": 86.19444274902344, "epoch": 0.22, "learning_rate": 2.6693169497650887e-05, "loss": 86.7923, "step": 614 }, { "compression_loss": 84.00220489501953, "epoch": 0.22, "learning_rate": 2.6687748464040478e-05, "loss": 84.5857, "step": 615 }, { "compression_loss": 84.4896469116211, "epoch": 0.22, "learning_rate": 2.668232743043007e-05, "loss": 84.8898, "step": 616 }, { "compression_loss": 82.92973327636719, "epoch": 0.22, "learning_rate": 2.667690639681966e-05, "loss": 83.3916, "step": 617 }, { "compression_loss": 84.02830505371094, "epoch": 0.22, "learning_rate": 2.6671485363209252e-05, "loss": 84.7145, "step": 618 }, { "compression_loss": 85.06642150878906, "epoch": 0.22, "learning_rate": 2.6666064329598843e-05, "loss": 85.3962, "step": 619 }, { "compression_loss": 84.0866470336914, "epoch": 0.22, "learning_rate": 2.6660643295988438e-05, "loss": 84.6011, "step": 620 }, { "compression_loss": 84.03392028808594, "epoch": 0.22, "learning_rate": 2.665522226237803e-05, "loss": 84.4436, "step": 621 }, { "compression_loss": 84.70397186279297, "epoch": 0.22, "learning_rate": 2.664980122876762e-05, "loss": 85.1435, "step": 622 }, { "compression_loss": 83.04107666015625, "epoch": 0.23, "learning_rate": 2.664438019515721e-05, "loss": 83.7206, "step": 623 }, { "compression_loss": 86.71467590332031, "epoch": 0.23, "learning_rate": 2.6638959161546803e-05, "loss": 87.4145, "step": 624 }, { "compression_loss": 83.62972259521484, "epoch": 0.23, "learning_rate": 2.6633538127936394e-05, "loss": 84.0661, "step": 625 }, { "compression_loss": 83.89585876464844, "epoch": 0.23, "learning_rate": 2.6628117094325985e-05, "loss": 84.4831, "step": 626 }, { "compression_loss": 84.44256591796875, "epoch": 0.23, "learning_rate": 2.6622696060715577e-05, "loss": 85.2219, "step": 627 }, { "compression_loss": 83.9019775390625, "epoch": 0.23, "learning_rate": 2.661727502710517e-05, "loss": 84.4711, "step": 628 }, { "compression_loss": 84.996826171875, "epoch": 0.23, "learning_rate": 2.6611853993494763e-05, "loss": 85.2059, "step": 629 }, { "compression_loss": 84.00587463378906, "epoch": 0.23, "learning_rate": 2.6606432959884354e-05, "loss": 84.4076, "step": 630 }, { "compression_loss": 85.65440368652344, "epoch": 0.23, "learning_rate": 2.6601011926273945e-05, "loss": 86.502, "step": 631 }, { "compression_loss": 85.27218627929688, "epoch": 0.23, "learning_rate": 2.6595590892663536e-05, "loss": 85.7402, "step": 632 }, { "compression_loss": 84.62650299072266, "epoch": 0.23, "learning_rate": 2.6590169859053128e-05, "loss": 85.1898, "step": 633 }, { "compression_loss": 83.43378448486328, "epoch": 0.23, "learning_rate": 2.658474882544272e-05, "loss": 83.6889, "step": 634 }, { "compression_loss": 83.2182846069336, "epoch": 0.23, "learning_rate": 2.657932779183231e-05, "loss": 83.5839, "step": 635 }, { "compression_loss": 85.65663146972656, "epoch": 0.23, "learning_rate": 2.65739067582219e-05, "loss": 86.3202, "step": 636 }, { "compression_loss": 84.82257843017578, "epoch": 0.23, "learning_rate": 2.6568485724611493e-05, "loss": 85.3935, "step": 637 }, { "compression_loss": 86.66615295410156, "epoch": 0.23, "learning_rate": 2.6563064691001084e-05, "loss": 87.6857, "step": 638 }, { "compression_loss": 84.85358428955078, "epoch": 0.23, "learning_rate": 2.6557643657390675e-05, "loss": 85.2976, "step": 639 }, { "compression_loss": 85.02157592773438, "epoch": 0.23, "learning_rate": 2.6552222623780267e-05, "loss": 85.5708, "step": 640 }, { "compression_loss": 81.90934753417969, "epoch": 0.23, "learning_rate": 2.6546801590169858e-05, "loss": 82.4875, "step": 641 }, { "compression_loss": 83.7437744140625, "epoch": 0.23, "learning_rate": 2.654138055655945e-05, "loss": 84.1595, "step": 642 }, { "compression_loss": 84.27732849121094, "epoch": 0.23, "learning_rate": 2.653595952294904e-05, "loss": 85.0387, "step": 643 }, { "compression_loss": 84.64398956298828, "epoch": 0.23, "learning_rate": 2.6530538489338635e-05, "loss": 85.439, "step": 644 }, { "compression_loss": 82.91926574707031, "epoch": 0.23, "learning_rate": 2.6525117455728226e-05, "loss": 83.2887, "step": 645 }, { "compression_loss": 85.2219009399414, "epoch": 0.23, "learning_rate": 2.6519696422117818e-05, "loss": 85.6686, "step": 646 }, { "compression_loss": 85.22357177734375, "epoch": 0.23, "learning_rate": 2.651427538850741e-05, "loss": 85.8079, "step": 647 }, { "compression_loss": 84.47567749023438, "epoch": 0.23, "learning_rate": 2.6508854354897e-05, "loss": 85.0316, "step": 648 }, { "compression_loss": 83.8251953125, "epoch": 0.23, "learning_rate": 2.650343332128659e-05, "loss": 84.2074, "step": 649 }, { "compression_loss": 84.56525421142578, "epoch": 0.23, "learning_rate": 2.6498012287676183e-05, "loss": 85.0985, "step": 650 }, { "compression_loss": 83.83744812011719, "epoch": 0.24, "learning_rate": 2.6492591254065774e-05, "loss": 84.7446, "step": 651 }, { "compression_loss": 83.09003448486328, "epoch": 0.24, "learning_rate": 2.648717022045537e-05, "loss": 83.4728, "step": 652 }, { "compression_loss": 84.51747131347656, "epoch": 0.24, "learning_rate": 2.648174918684496e-05, "loss": 85.0319, "step": 653 }, { "compression_loss": 85.80245971679688, "epoch": 0.24, "learning_rate": 2.647632815323455e-05, "loss": 86.3588, "step": 654 }, { "compression_loss": 85.80987548828125, "epoch": 0.24, "learning_rate": 2.6470907119624142e-05, "loss": 86.4152, "step": 655 }, { "compression_loss": 83.9101791381836, "epoch": 0.24, "learning_rate": 2.6465486086013734e-05, "loss": 84.367, "step": 656 }, { "compression_loss": 83.78659057617188, "epoch": 0.24, "learning_rate": 2.6460065052403325e-05, "loss": 84.392, "step": 657 }, { "compression_loss": 84.40013885498047, "epoch": 0.24, "learning_rate": 2.6454644018792916e-05, "loss": 85.144, "step": 658 }, { "compression_loss": 86.27102661132812, "epoch": 0.24, "learning_rate": 2.6449222985182507e-05, "loss": 86.8973, "step": 659 }, { "compression_loss": 84.51985168457031, "epoch": 0.24, "learning_rate": 2.6443801951572102e-05, "loss": 85.0849, "step": 660 }, { "compression_loss": 85.02110290527344, "epoch": 0.24, "learning_rate": 2.6438380917961693e-05, "loss": 85.3978, "step": 661 }, { "compression_loss": 83.92158508300781, "epoch": 0.24, "learning_rate": 2.6432959884351285e-05, "loss": 84.4219, "step": 662 }, { "compression_loss": 84.67571258544922, "epoch": 0.24, "learning_rate": 2.6427538850740876e-05, "loss": 85.2487, "step": 663 }, { "compression_loss": 84.7196273803711, "epoch": 0.24, "learning_rate": 2.6422117817130467e-05, "loss": 85.6721, "step": 664 }, { "compression_loss": 85.76388549804688, "epoch": 0.24, "learning_rate": 2.641669678352006e-05, "loss": 86.4973, "step": 665 }, { "compression_loss": 84.46864318847656, "epoch": 0.24, "learning_rate": 2.641127574990965e-05, "loss": 84.9972, "step": 666 }, { "compression_loss": 83.43309783935547, "epoch": 0.24, "learning_rate": 2.640585471629924e-05, "loss": 83.8153, "step": 667 }, { "compression_loss": 83.20772552490234, "epoch": 0.24, "learning_rate": 2.6400433682688836e-05, "loss": 83.7107, "step": 668 }, { "compression_loss": 85.43841552734375, "epoch": 0.24, "learning_rate": 2.6395012649078427e-05, "loss": 85.8455, "step": 669 }, { "compression_loss": 85.8267593383789, "epoch": 0.24, "learning_rate": 2.6389591615468018e-05, "loss": 86.3354, "step": 670 }, { "compression_loss": 86.06452941894531, "epoch": 0.24, "learning_rate": 2.638417058185761e-05, "loss": 86.952, "step": 671 }, { "compression_loss": 84.59473419189453, "epoch": 0.24, "learning_rate": 2.63787495482472e-05, "loss": 85.2175, "step": 672 }, { "compression_loss": 83.57872772216797, "epoch": 0.24, "learning_rate": 2.6373328514636792e-05, "loss": 84.1073, "step": 673 }, { "compression_loss": 83.20772552490234, "epoch": 0.24, "learning_rate": 2.6367907481026383e-05, "loss": 83.8618, "step": 674 }, { "compression_loss": 82.05908203125, "epoch": 0.24, "learning_rate": 2.6362486447415975e-05, "loss": 82.3442, "step": 675 }, { "compression_loss": 83.22503662109375, "epoch": 0.24, "learning_rate": 2.635706541380557e-05, "loss": 83.3985, "step": 676 }, { "compression_loss": 84.5078125, "epoch": 0.24, "learning_rate": 2.635164438019516e-05, "loss": 85.0265, "step": 677 }, { "compression_loss": 84.18687438964844, "epoch": 0.25, "learning_rate": 2.6346223346584752e-05, "loss": 84.8394, "step": 678 }, { "compression_loss": 85.16847229003906, "epoch": 0.25, "learning_rate": 2.6340802312974343e-05, "loss": 86.4117, "step": 679 }, { "compression_loss": 86.1584243774414, "epoch": 0.25, "learning_rate": 2.633538127936393e-05, "loss": 86.8275, "step": 680 }, { "compression_loss": 84.94137573242188, "epoch": 0.25, "learning_rate": 2.6329960245753522e-05, "loss": 85.6778, "step": 681 }, { "compression_loss": 86.99870300292969, "epoch": 0.25, "learning_rate": 2.6324539212143113e-05, "loss": 87.6771, "step": 682 }, { "compression_loss": 83.50433349609375, "epoch": 0.25, "learning_rate": 2.6319118178532705e-05, "loss": 84.1301, "step": 683 }, { "compression_loss": 82.93773651123047, "epoch": 0.25, "learning_rate": 2.63136971449223e-05, "loss": 83.4483, "step": 684 }, { "compression_loss": 85.95423889160156, "epoch": 0.25, "learning_rate": 2.630827611131189e-05, "loss": 86.5062, "step": 685 }, { "compression_loss": 85.44578552246094, "epoch": 0.25, "learning_rate": 2.6302855077701482e-05, "loss": 85.9807, "step": 686 }, { "compression_loss": 84.06092834472656, "epoch": 0.25, "learning_rate": 2.6297434044091073e-05, "loss": 84.7256, "step": 687 }, { "compression_loss": 83.0500259399414, "epoch": 0.25, "learning_rate": 2.6292013010480664e-05, "loss": 83.5825, "step": 688 }, { "compression_loss": 83.9212646484375, "epoch": 0.25, "learning_rate": 2.6286591976870256e-05, "loss": 84.3623, "step": 689 }, { "compression_loss": 83.69830322265625, "epoch": 0.25, "learning_rate": 2.6281170943259847e-05, "loss": 84.1324, "step": 690 }, { "compression_loss": 85.41788482666016, "epoch": 0.25, "learning_rate": 2.627574990964944e-05, "loss": 85.9307, "step": 691 }, { "compression_loss": 86.06968688964844, "epoch": 0.25, "learning_rate": 2.6270328876039033e-05, "loss": 86.7076, "step": 692 }, { "compression_loss": 86.6199951171875, "epoch": 0.25, "learning_rate": 2.6264907842428624e-05, "loss": 87.0788, "step": 693 }, { "compression_loss": 85.21733856201172, "epoch": 0.25, "learning_rate": 2.6259486808818216e-05, "loss": 85.8283, "step": 694 }, { "compression_loss": 85.46843719482422, "epoch": 0.25, "learning_rate": 2.6254065775207807e-05, "loss": 86.2852, "step": 695 }, { "compression_loss": 86.49384307861328, "epoch": 0.25, "learning_rate": 2.6248644741597398e-05, "loss": 87.2337, "step": 696 }, { "compression_loss": 84.6192855834961, "epoch": 0.25, "learning_rate": 2.624322370798699e-05, "loss": 85.1286, "step": 697 }, { "compression_loss": 83.41229248046875, "epoch": 0.25, "learning_rate": 2.623780267437658e-05, "loss": 83.9369, "step": 698 }, { "compression_loss": 83.8067398071289, "epoch": 0.25, "learning_rate": 2.6232381640766172e-05, "loss": 84.2495, "step": 699 }, { "compression_loss": 82.7241439819336, "epoch": 0.25, "learning_rate": 2.6226960607155767e-05, "loss": 83.1825, "step": 700 }, { "compression_loss": 83.22814178466797, "epoch": 0.25, "learning_rate": 2.6221539573545358e-05, "loss": 83.7172, "step": 701 }, { "compression_loss": 83.88792419433594, "epoch": 0.25, "learning_rate": 2.621611853993495e-05, "loss": 84.5183, "step": 702 }, { "compression_loss": 85.35458374023438, "epoch": 0.25, "learning_rate": 2.621069750632454e-05, "loss": 85.8164, "step": 703 }, { "compression_loss": 83.01829528808594, "epoch": 0.25, "learning_rate": 2.620527647271413e-05, "loss": 83.5957, "step": 704 }, { "compression_loss": 84.52545166015625, "epoch": 0.25, "learning_rate": 2.6199855439103723e-05, "loss": 85.12, "step": 705 }, { "compression_loss": 84.790771484375, "epoch": 0.26, "learning_rate": 2.6194434405493314e-05, "loss": 85.2602, "step": 706 }, { "compression_loss": 82.43743896484375, "epoch": 0.26, "learning_rate": 2.6189013371882905e-05, "loss": 83.1979, "step": 707 }, { "compression_loss": 86.06666564941406, "epoch": 0.26, "learning_rate": 2.61835923382725e-05, "loss": 86.9648, "step": 708 }, { "compression_loss": 84.76333618164062, "epoch": 0.26, "learning_rate": 2.617817130466209e-05, "loss": 85.3937, "step": 709 }, { "compression_loss": 85.86711120605469, "epoch": 0.26, "learning_rate": 2.6172750271051683e-05, "loss": 86.2752, "step": 710 }, { "compression_loss": 86.0439453125, "epoch": 0.26, "learning_rate": 2.6167329237441274e-05, "loss": 87.0003, "step": 711 }, { "compression_loss": 84.66104125976562, "epoch": 0.26, "learning_rate": 2.6161908203830865e-05, "loss": 85.3537, "step": 712 }, { "compression_loss": 85.02753448486328, "epoch": 0.26, "learning_rate": 2.6156487170220456e-05, "loss": 85.3274, "step": 713 }, { "compression_loss": 84.70980072021484, "epoch": 0.26, "learning_rate": 2.6151066136610048e-05, "loss": 85.268, "step": 714 }, { "compression_loss": 86.0601806640625, "epoch": 0.26, "learning_rate": 2.614564510299964e-05, "loss": 86.7679, "step": 715 }, { "compression_loss": 84.53764343261719, "epoch": 0.26, "learning_rate": 2.6140224069389234e-05, "loss": 85.2071, "step": 716 }, { "compression_loss": 84.82659149169922, "epoch": 0.26, "learning_rate": 2.6134803035778825e-05, "loss": 86.0801, "step": 717 }, { "compression_loss": 84.58898162841797, "epoch": 0.26, "learning_rate": 2.6129382002168416e-05, "loss": 84.9804, "step": 718 }, { "compression_loss": 83.1011962890625, "epoch": 0.26, "learning_rate": 2.6123960968558007e-05, "loss": 83.6989, "step": 719 }, { "compression_loss": 85.41062927246094, "epoch": 0.26, "learning_rate": 2.61185399349476e-05, "loss": 85.9962, "step": 720 }, { "compression_loss": 84.54407501220703, "epoch": 0.26, "learning_rate": 2.611311890133719e-05, "loss": 85.2446, "step": 721 }, { "compression_loss": 85.51377868652344, "epoch": 0.26, "learning_rate": 2.610769786772678e-05, "loss": 85.9435, "step": 722 }, { "compression_loss": 86.65360260009766, "epoch": 0.26, "learning_rate": 2.6102276834116373e-05, "loss": 87.4296, "step": 723 }, { "compression_loss": 84.52252960205078, "epoch": 0.26, "learning_rate": 2.6096855800505964e-05, "loss": 85.0747, "step": 724 }, { "compression_loss": 84.1920166015625, "epoch": 0.26, "learning_rate": 2.6091434766895555e-05, "loss": 84.619, "step": 725 }, { "compression_loss": 85.84333801269531, "epoch": 0.26, "learning_rate": 2.6086013733285146e-05, "loss": 86.6483, "step": 726 }, { "compression_loss": 83.14297485351562, "epoch": 0.26, "learning_rate": 2.6080592699674738e-05, "loss": 83.5007, "step": 727 }, { "compression_loss": 83.5848159790039, "epoch": 0.26, "learning_rate": 2.607517166606433e-05, "loss": 83.9649, "step": 728 }, { "compression_loss": 82.91450500488281, "epoch": 0.26, "learning_rate": 2.606975063245392e-05, "loss": 83.3936, "step": 729 }, { "compression_loss": 84.5454330444336, "epoch": 0.26, "learning_rate": 2.606432959884351e-05, "loss": 85.446, "step": 730 }, { "compression_loss": 85.63691711425781, "epoch": 0.26, "learning_rate": 2.6058908565233103e-05, "loss": 86.2234, "step": 731 }, { "compression_loss": 84.20425415039062, "epoch": 0.26, "learning_rate": 2.6053487531622697e-05, "loss": 84.6972, "step": 732 }, { "compression_loss": 83.61749267578125, "epoch": 0.26, "learning_rate": 2.604806649801229e-05, "loss": 83.8587, "step": 733 }, { "compression_loss": 84.61354064941406, "epoch": 0.27, "learning_rate": 2.604264546440188e-05, "loss": 85.1808, "step": 734 }, { "compression_loss": 85.30830383300781, "epoch": 0.27, "learning_rate": 2.603722443079147e-05, "loss": 85.8072, "step": 735 }, { "compression_loss": 83.66436767578125, "epoch": 0.27, "learning_rate": 2.6031803397181062e-05, "loss": 84.24, "step": 736 }, { "compression_loss": 85.43125915527344, "epoch": 0.27, "learning_rate": 2.6026382363570654e-05, "loss": 86.3786, "step": 737 }, { "compression_loss": 86.37230682373047, "epoch": 0.27, "learning_rate": 2.6020961329960245e-05, "loss": 86.9143, "step": 738 }, { "compression_loss": 83.69573211669922, "epoch": 0.27, "learning_rate": 2.6015540296349836e-05, "loss": 84.2259, "step": 739 }, { "compression_loss": 84.09561157226562, "epoch": 0.27, "learning_rate": 2.601011926273943e-05, "loss": 84.5689, "step": 740 }, { "compression_loss": 86.12159729003906, "epoch": 0.27, "learning_rate": 2.6004698229129022e-05, "loss": 86.8681, "step": 741 }, { "compression_loss": 84.26750946044922, "epoch": 0.27, "learning_rate": 2.5999277195518614e-05, "loss": 84.7339, "step": 742 }, { "compression_loss": 82.85518646240234, "epoch": 0.27, "learning_rate": 2.5993856161908205e-05, "loss": 83.2544, "step": 743 }, { "compression_loss": 85.04598999023438, "epoch": 0.27, "learning_rate": 2.5988435128297796e-05, "loss": 85.7733, "step": 744 }, { "compression_loss": 85.30673217773438, "epoch": 0.27, "learning_rate": 2.5983014094687387e-05, "loss": 85.802, "step": 745 }, { "compression_loss": 85.58810424804688, "epoch": 0.27, "learning_rate": 2.597759306107698e-05, "loss": 86.2084, "step": 746 }, { "compression_loss": 85.15409088134766, "epoch": 0.27, "learning_rate": 2.597217202746657e-05, "loss": 85.9782, "step": 747 }, { "compression_loss": 86.00509643554688, "epoch": 0.27, "learning_rate": 2.596675099385616e-05, "loss": 87.0044, "step": 748 }, { "compression_loss": 83.40724182128906, "epoch": 0.27, "learning_rate": 2.5961329960245756e-05, "loss": 83.7149, "step": 749 }, { "compression_loss": 85.51018524169922, "epoch": 0.27, "learning_rate": 2.5955908926635347e-05, "loss": 86.14, "step": 750 }, { "epoch": 0.27, "eval_exact_match": 85.71428571428571, "eval_f1": 92.20352143297343, "step": 750 }, { "compression_loss": 85.58514404296875, "epoch": 0.27, "learning_rate": 2.595048789302494e-05, "loss": 86.5804, "step": 751 }, { "compression_loss": 85.03996276855469, "epoch": 0.27, "learning_rate": 2.594506685941453e-05, "loss": 85.6322, "step": 752 }, { "compression_loss": 84.8355941772461, "epoch": 0.27, "learning_rate": 2.593964582580412e-05, "loss": 85.7291, "step": 753 }, { "compression_loss": 84.82850646972656, "epoch": 0.27, "learning_rate": 2.5934224792193712e-05, "loss": 85.1921, "step": 754 }, { "compression_loss": 85.5968017578125, "epoch": 0.27, "learning_rate": 2.5928803758583303e-05, "loss": 86.1793, "step": 755 }, { "compression_loss": 84.86681365966797, "epoch": 0.27, "learning_rate": 2.5923382724972895e-05, "loss": 85.3309, "step": 756 }, { "compression_loss": 84.12045288085938, "epoch": 0.27, "learning_rate": 2.591796169136249e-05, "loss": 84.6562, "step": 757 }, { "compression_loss": 85.4756088256836, "epoch": 0.27, "learning_rate": 2.591254065775208e-05, "loss": 85.9747, "step": 758 }, { "compression_loss": 84.73968505859375, "epoch": 0.27, "learning_rate": 2.5907119624141672e-05, "loss": 85.3682, "step": 759 }, { "compression_loss": 83.87834930419922, "epoch": 0.27, "learning_rate": 2.5901698590531263e-05, "loss": 84.2553, "step": 760 }, { "compression_loss": 84.83709716796875, "epoch": 0.28, "learning_rate": 2.5896277556920854e-05, "loss": 85.5662, "step": 761 }, { "compression_loss": 85.50460815429688, "epoch": 0.28, "learning_rate": 2.5890856523310446e-05, "loss": 86.2004, "step": 762 }, { "compression_loss": 85.95812225341797, "epoch": 0.28, "learning_rate": 2.5885435489700037e-05, "loss": 86.4587, "step": 763 }, { "compression_loss": 86.56501770019531, "epoch": 0.28, "learning_rate": 2.5880014456089628e-05, "loss": 87.1597, "step": 764 }, { "compression_loss": 83.30992126464844, "epoch": 0.28, "learning_rate": 2.5874593422479223e-05, "loss": 83.6731, "step": 765 }, { "compression_loss": 85.817138671875, "epoch": 0.28, "learning_rate": 2.5869172388868814e-05, "loss": 86.1987, "step": 766 }, { "compression_loss": 85.68914794921875, "epoch": 0.28, "learning_rate": 2.5863751355258402e-05, "loss": 86.1087, "step": 767 }, { "compression_loss": 85.77345275878906, "epoch": 0.28, "learning_rate": 2.5858330321647993e-05, "loss": 86.9234, "step": 768 }, { "compression_loss": 84.91036987304688, "epoch": 0.28, "learning_rate": 2.5852909288037585e-05, "loss": 85.607, "step": 769 }, { "compression_loss": 82.59917449951172, "epoch": 0.28, "learning_rate": 2.5847488254427176e-05, "loss": 83.2773, "step": 770 }, { "compression_loss": 84.7386245727539, "epoch": 0.28, "learning_rate": 2.5842067220816767e-05, "loss": 85.2841, "step": 771 }, { "compression_loss": 82.64828491210938, "epoch": 0.28, "learning_rate": 2.583664618720636e-05, "loss": 83.0098, "step": 772 }, { "compression_loss": 86.09741973876953, "epoch": 0.28, "learning_rate": 2.5831225153595953e-05, "loss": 86.9368, "step": 773 }, { "compression_loss": 82.39185333251953, "epoch": 0.28, "learning_rate": 2.5825804119985544e-05, "loss": 83.1849, "step": 774 }, { "compression_loss": 85.52769470214844, "epoch": 0.28, "learning_rate": 2.5820383086375136e-05, "loss": 85.9332, "step": 775 }, { "compression_loss": 84.21986389160156, "epoch": 0.28, "learning_rate": 2.5814962052764727e-05, "loss": 84.7964, "step": 776 }, { "compression_loss": 83.5674057006836, "epoch": 0.28, "learning_rate": 2.5809541019154318e-05, "loss": 84.2096, "step": 777 }, { "compression_loss": 85.33134460449219, "epoch": 0.28, "learning_rate": 2.580411998554391e-05, "loss": 85.8558, "step": 778 }, { "compression_loss": 84.93482208251953, "epoch": 0.28, "learning_rate": 2.57986989519335e-05, "loss": 85.5827, "step": 779 }, { "compression_loss": 85.6207275390625, "epoch": 0.28, "learning_rate": 2.5793277918323092e-05, "loss": 86.2959, "step": 780 }, { "compression_loss": 86.81668090820312, "epoch": 0.28, "learning_rate": 2.5787856884712687e-05, "loss": 88.0972, "step": 781 }, { "compression_loss": 84.9217300415039, "epoch": 0.28, "learning_rate": 2.5782435851102278e-05, "loss": 85.3725, "step": 782 }, { "compression_loss": 85.15275573730469, "epoch": 0.28, "learning_rate": 2.577701481749187e-05, "loss": 85.7749, "step": 783 }, { "compression_loss": 83.5855712890625, "epoch": 0.28, "learning_rate": 2.577159378388146e-05, "loss": 83.9548, "step": 784 }, { "compression_loss": 84.40412139892578, "epoch": 0.28, "learning_rate": 2.5766172750271052e-05, "loss": 84.7889, "step": 785 }, { "compression_loss": 82.60457611083984, "epoch": 0.28, "learning_rate": 2.5760751716660643e-05, "loss": 83.1935, "step": 786 }, { "compression_loss": 85.69320678710938, "epoch": 0.28, "learning_rate": 2.5755330683050234e-05, "loss": 86.6147, "step": 787 }, { "compression_loss": 85.75263977050781, "epoch": 0.28, "learning_rate": 2.5749909649439826e-05, "loss": 86.9446, "step": 788 }, { "compression_loss": 84.61886596679688, "epoch": 0.29, "learning_rate": 2.574448861582942e-05, "loss": 85.277, "step": 789 }, { "compression_loss": 84.86312866210938, "epoch": 0.29, "learning_rate": 2.573906758221901e-05, "loss": 85.5784, "step": 790 }, { "compression_loss": 86.13746643066406, "epoch": 0.29, "learning_rate": 2.5733646548608603e-05, "loss": 86.8407, "step": 791 }, { "compression_loss": 84.39503479003906, "epoch": 0.29, "learning_rate": 2.5728225514998194e-05, "loss": 84.9747, "step": 792 }, { "compression_loss": 83.84508514404297, "epoch": 0.29, "learning_rate": 2.5722804481387785e-05, "loss": 84.3959, "step": 793 }, { "compression_loss": 84.89189910888672, "epoch": 0.29, "learning_rate": 2.5717383447777377e-05, "loss": 85.6165, "step": 794 }, { "compression_loss": 85.09202575683594, "epoch": 0.29, "learning_rate": 2.5711962414166968e-05, "loss": 85.729, "step": 795 }, { "compression_loss": 82.08912658691406, "epoch": 0.29, "learning_rate": 2.570654138055656e-05, "loss": 82.8813, "step": 796 }, { "compression_loss": 83.87565612792969, "epoch": 0.29, "learning_rate": 2.5701120346946154e-05, "loss": 84.3556, "step": 797 }, { "compression_loss": 84.31647491455078, "epoch": 0.29, "learning_rate": 2.5695699313335745e-05, "loss": 84.8381, "step": 798 }, { "compression_loss": 85.51068115234375, "epoch": 0.29, "learning_rate": 2.5690278279725336e-05, "loss": 86.0678, "step": 799 }, { "compression_loss": 86.28826904296875, "epoch": 0.29, "learning_rate": 2.5684857246114928e-05, "loss": 86.675, "step": 800 }, { "compression_loss": 84.6733169555664, "epoch": 0.29, "learning_rate": 2.567943621250452e-05, "loss": 85.0884, "step": 801 }, { "compression_loss": 82.17766571044922, "epoch": 0.29, "learning_rate": 2.567401517889411e-05, "loss": 82.5785, "step": 802 }, { "compression_loss": 82.61650085449219, "epoch": 0.29, "learning_rate": 2.56685941452837e-05, "loss": 82.9358, "step": 803 }, { "compression_loss": 84.59719848632812, "epoch": 0.29, "learning_rate": 2.5663173111673293e-05, "loss": 85.542, "step": 804 }, { "compression_loss": 83.08328247070312, "epoch": 0.29, "learning_rate": 2.5657752078062887e-05, "loss": 83.5695, "step": 805 }, { "compression_loss": 84.48274993896484, "epoch": 0.29, "learning_rate": 2.565233104445248e-05, "loss": 85.3386, "step": 806 }, { "compression_loss": 84.93070220947266, "epoch": 0.29, "learning_rate": 2.564691001084207e-05, "loss": 85.3489, "step": 807 }, { "compression_loss": 84.42250061035156, "epoch": 0.29, "learning_rate": 2.564148897723166e-05, "loss": 84.8924, "step": 808 }, { "compression_loss": 85.05558776855469, "epoch": 0.29, "learning_rate": 2.5636067943621252e-05, "loss": 85.6861, "step": 809 }, { "compression_loss": 85.50775146484375, "epoch": 0.29, "learning_rate": 2.5630646910010844e-05, "loss": 86.5078, "step": 810 }, { "compression_loss": 84.93310546875, "epoch": 0.29, "learning_rate": 2.562522587640043e-05, "loss": 85.7616, "step": 811 }, { "compression_loss": 83.89894104003906, "epoch": 0.29, "learning_rate": 2.5619804842790023e-05, "loss": 84.3558, "step": 812 }, { "compression_loss": 83.25638580322266, "epoch": 0.29, "learning_rate": 2.5614383809179618e-05, "loss": 83.7746, "step": 813 }, { "compression_loss": 85.4486083984375, "epoch": 0.29, "learning_rate": 2.560896277556921e-05, "loss": 86.0432, "step": 814 }, { "compression_loss": 85.15155029296875, "epoch": 0.29, "learning_rate": 2.56035417419588e-05, "loss": 85.6934, "step": 815 }, { "compression_loss": 86.49371337890625, "epoch": 0.29, "learning_rate": 2.559812070834839e-05, "loss": 87.05, "step": 816 }, { "compression_loss": 84.76737976074219, "epoch": 0.3, "learning_rate": 2.5592699674737983e-05, "loss": 85.1746, "step": 817 }, { "compression_loss": 84.00660705566406, "epoch": 0.3, "learning_rate": 2.5587278641127574e-05, "loss": 84.2881, "step": 818 }, { "compression_loss": 84.91462707519531, "epoch": 0.3, "learning_rate": 2.5581857607517165e-05, "loss": 85.6226, "step": 819 }, { "compression_loss": 84.88776397705078, "epoch": 0.3, "learning_rate": 2.5576436573906756e-05, "loss": 85.5846, "step": 820 }, { "compression_loss": 83.82675170898438, "epoch": 0.3, "learning_rate": 2.557101554029635e-05, "loss": 84.6525, "step": 821 }, { "compression_loss": 83.29803466796875, "epoch": 0.3, "learning_rate": 2.5565594506685942e-05, "loss": 84.1562, "step": 822 }, { "compression_loss": 85.4826889038086, "epoch": 0.3, "learning_rate": 2.5560173473075534e-05, "loss": 86.1046, "step": 823 }, { "compression_loss": 83.44166564941406, "epoch": 0.3, "learning_rate": 2.5554752439465125e-05, "loss": 84.2467, "step": 824 }, { "compression_loss": 82.66215515136719, "epoch": 0.3, "learning_rate": 2.5549331405854716e-05, "loss": 83.2008, "step": 825 }, { "compression_loss": 84.98217010498047, "epoch": 0.3, "learning_rate": 2.5543910372244307e-05, "loss": 85.8063, "step": 826 }, { "compression_loss": 84.55343627929688, "epoch": 0.3, "learning_rate": 2.55384893386339e-05, "loss": 85.1928, "step": 827 }, { "compression_loss": 83.99134826660156, "epoch": 0.3, "learning_rate": 2.553306830502349e-05, "loss": 84.4822, "step": 828 }, { "compression_loss": 84.44384765625, "epoch": 0.3, "learning_rate": 2.5527647271413085e-05, "loss": 85.3235, "step": 829 }, { "compression_loss": 82.3162841796875, "epoch": 0.3, "learning_rate": 2.5522226237802676e-05, "loss": 82.7588, "step": 830 }, { "compression_loss": 85.1368179321289, "epoch": 0.3, "learning_rate": 2.5516805204192267e-05, "loss": 85.5512, "step": 831 }, { "compression_loss": 85.27079772949219, "epoch": 0.3, "learning_rate": 2.551138417058186e-05, "loss": 85.7183, "step": 832 }, { "compression_loss": 83.54054260253906, "epoch": 0.3, "learning_rate": 2.550596313697145e-05, "loss": 83.9777, "step": 833 }, { "compression_loss": 84.98909759521484, "epoch": 0.3, "learning_rate": 2.550054210336104e-05, "loss": 85.5617, "step": 834 }, { "compression_loss": 84.57638549804688, "epoch": 0.3, "learning_rate": 2.5495121069750632e-05, "loss": 84.9917, "step": 835 }, { "compression_loss": 86.16323852539062, "epoch": 0.3, "learning_rate": 2.5489700036140224e-05, "loss": 86.9618, "step": 836 }, { "compression_loss": 83.28775024414062, "epoch": 0.3, "learning_rate": 2.5484279002529818e-05, "loss": 83.9753, "step": 837 }, { "compression_loss": 83.58818054199219, "epoch": 0.3, "learning_rate": 2.547885796891941e-05, "loss": 84.1172, "step": 838 }, { "compression_loss": 86.58973693847656, "epoch": 0.3, "learning_rate": 2.5473436935309e-05, "loss": 87.2877, "step": 839 }, { "compression_loss": 83.58847045898438, "epoch": 0.3, "learning_rate": 2.5468015901698592e-05, "loss": 83.9716, "step": 840 }, { "compression_loss": 85.07415008544922, "epoch": 0.3, "learning_rate": 2.5462594868088183e-05, "loss": 85.9936, "step": 841 }, { "compression_loss": 84.28292846679688, "epoch": 0.3, "learning_rate": 2.5457173834477775e-05, "loss": 84.7472, "step": 842 }, { "compression_loss": 84.78514099121094, "epoch": 0.3, "learning_rate": 2.5451752800867366e-05, "loss": 85.3914, "step": 843 }, { "compression_loss": 86.14940643310547, "epoch": 0.31, "learning_rate": 2.5446331767256957e-05, "loss": 86.5382, "step": 844 }, { "compression_loss": 83.97173309326172, "epoch": 0.31, "learning_rate": 2.5440910733646552e-05, "loss": 84.984, "step": 845 }, { "compression_loss": 82.79801177978516, "epoch": 0.31, "learning_rate": 2.5435489700036143e-05, "loss": 83.2506, "step": 846 }, { "compression_loss": 84.7918930053711, "epoch": 0.31, "learning_rate": 2.5430068666425734e-05, "loss": 85.238, "step": 847 }, { "compression_loss": 84.73894500732422, "epoch": 0.31, "learning_rate": 2.5424647632815326e-05, "loss": 85.5105, "step": 848 }, { "compression_loss": 85.12548828125, "epoch": 0.31, "learning_rate": 2.5419226599204917e-05, "loss": 85.7464, "step": 849 }, { "compression_loss": 84.5750732421875, "epoch": 0.31, "learning_rate": 2.5413805565594508e-05, "loss": 84.8269, "step": 850 }, { "compression_loss": 85.1200180053711, "epoch": 0.31, "learning_rate": 2.54083845319841e-05, "loss": 85.9372, "step": 851 }, { "compression_loss": 84.38545227050781, "epoch": 0.31, "learning_rate": 2.540296349837369e-05, "loss": 84.9179, "step": 852 }, { "compression_loss": 85.18756103515625, "epoch": 0.31, "learning_rate": 2.5397542464763285e-05, "loss": 86.0924, "step": 853 }, { "compression_loss": 85.91270446777344, "epoch": 0.31, "learning_rate": 2.5392121431152873e-05, "loss": 86.4379, "step": 854 }, { "compression_loss": 85.0280990600586, "epoch": 0.31, "learning_rate": 2.5386700397542464e-05, "loss": 85.7848, "step": 855 }, { "compression_loss": 85.4067611694336, "epoch": 0.31, "learning_rate": 2.5381279363932056e-05, "loss": 85.9971, "step": 856 }, { "compression_loss": 85.57905578613281, "epoch": 0.31, "learning_rate": 2.5375858330321647e-05, "loss": 86.2379, "step": 857 }, { "compression_loss": 83.62826538085938, "epoch": 0.31, "learning_rate": 2.537043729671124e-05, "loss": 84.1465, "step": 858 }, { "compression_loss": 84.76446533203125, "epoch": 0.31, "learning_rate": 2.536501626310083e-05, "loss": 85.2119, "step": 859 }, { "compression_loss": 85.62554931640625, "epoch": 0.31, "learning_rate": 2.535959522949042e-05, "loss": 86.0302, "step": 860 }, { "compression_loss": 84.60655975341797, "epoch": 0.31, "learning_rate": 2.5354174195880016e-05, "loss": 85.1782, "step": 861 }, { "compression_loss": 82.26683044433594, "epoch": 0.31, "learning_rate": 2.5348753162269607e-05, "loss": 82.9411, "step": 862 }, { "compression_loss": 83.54515838623047, "epoch": 0.31, "learning_rate": 2.5343332128659198e-05, "loss": 84.3772, "step": 863 }, { "compression_loss": 84.8697509765625, "epoch": 0.31, "learning_rate": 2.533791109504879e-05, "loss": 85.4263, "step": 864 }, { "compression_loss": 83.8523178100586, "epoch": 0.31, "learning_rate": 2.533249006143838e-05, "loss": 84.3855, "step": 865 }, { "compression_loss": 84.142333984375, "epoch": 0.31, "learning_rate": 2.5327069027827972e-05, "loss": 84.7795, "step": 866 }, { "compression_loss": 85.55587768554688, "epoch": 0.31, "learning_rate": 2.5321647994217563e-05, "loss": 86.0871, "step": 867 }, { "compression_loss": 83.67743682861328, "epoch": 0.31, "learning_rate": 2.5316226960607154e-05, "loss": 83.9904, "step": 868 }, { "compression_loss": 85.1765365600586, "epoch": 0.31, "learning_rate": 2.531080592699675e-05, "loss": 85.6874, "step": 869 }, { "compression_loss": 84.99048614501953, "epoch": 0.31, "learning_rate": 2.530538489338634e-05, "loss": 85.6419, "step": 870 }, { "compression_loss": 84.25005340576172, "epoch": 0.31, "learning_rate": 2.529996385977593e-05, "loss": 84.6437, "step": 871 }, { "compression_loss": 84.6412353515625, "epoch": 0.32, "learning_rate": 2.5294542826165523e-05, "loss": 84.9867, "step": 872 }, { "compression_loss": 84.27247619628906, "epoch": 0.32, "learning_rate": 2.5289121792555114e-05, "loss": 85.1424, "step": 873 }, { "compression_loss": 83.73185729980469, "epoch": 0.32, "learning_rate": 2.5283700758944705e-05, "loss": 84.1594, "step": 874 }, { "compression_loss": 84.69617462158203, "epoch": 0.32, "learning_rate": 2.5278279725334297e-05, "loss": 85.4188, "step": 875 }, { "compression_loss": 84.76729583740234, "epoch": 0.32, "learning_rate": 2.5272858691723888e-05, "loss": 85.3913, "step": 876 }, { "compression_loss": 83.3388671875, "epoch": 0.32, "learning_rate": 2.5267437658113483e-05, "loss": 83.9684, "step": 877 }, { "compression_loss": 84.49650573730469, "epoch": 0.32, "learning_rate": 2.5262016624503074e-05, "loss": 85.1693, "step": 878 }, { "compression_loss": 86.7043685913086, "epoch": 0.32, "learning_rate": 2.5256595590892665e-05, "loss": 87.1884, "step": 879 }, { "compression_loss": 83.92679595947266, "epoch": 0.32, "learning_rate": 2.5251174557282256e-05, "loss": 84.4309, "step": 880 }, { "compression_loss": 84.16768646240234, "epoch": 0.32, "learning_rate": 2.5245753523671848e-05, "loss": 84.5168, "step": 881 }, { "compression_loss": 85.12198638916016, "epoch": 0.32, "learning_rate": 2.524033249006144e-05, "loss": 85.4967, "step": 882 }, { "compression_loss": 85.60665130615234, "epoch": 0.32, "learning_rate": 2.523491145645103e-05, "loss": 86.1884, "step": 883 }, { "compression_loss": 84.06935119628906, "epoch": 0.32, "learning_rate": 2.522949042284062e-05, "loss": 84.7728, "step": 884 }, { "compression_loss": 84.01273345947266, "epoch": 0.32, "learning_rate": 2.5224069389230216e-05, "loss": 84.6025, "step": 885 }, { "compression_loss": 84.03848266601562, "epoch": 0.32, "learning_rate": 2.5218648355619807e-05, "loss": 84.3201, "step": 886 }, { "compression_loss": 84.24974822998047, "epoch": 0.32, "learning_rate": 2.52132273220094e-05, "loss": 84.931, "step": 887 }, { "compression_loss": 84.75204467773438, "epoch": 0.32, "learning_rate": 2.520780628839899e-05, "loss": 85.3257, "step": 888 }, { "compression_loss": 85.48555755615234, "epoch": 0.32, "learning_rate": 2.520238525478858e-05, "loss": 86.3948, "step": 889 }, { "compression_loss": 84.9803237915039, "epoch": 0.32, "learning_rate": 2.5196964221178173e-05, "loss": 85.6633, "step": 890 }, { "compression_loss": 84.46197509765625, "epoch": 0.32, "learning_rate": 2.5191543187567764e-05, "loss": 85.0708, "step": 891 }, { "compression_loss": 84.67835998535156, "epoch": 0.32, "learning_rate": 2.5186122153957355e-05, "loss": 85.2908, "step": 892 }, { "compression_loss": 83.90602111816406, "epoch": 0.32, "learning_rate": 2.518070112034695e-05, "loss": 84.3259, "step": 893 }, { "compression_loss": 84.09382629394531, "epoch": 0.32, "learning_rate": 2.517528008673654e-05, "loss": 84.585, "step": 894 }, { "compression_loss": 84.30528259277344, "epoch": 0.32, "learning_rate": 2.5169859053126132e-05, "loss": 84.8925, "step": 895 }, { "compression_loss": 85.58379364013672, "epoch": 0.32, "learning_rate": 2.5164438019515724e-05, "loss": 86.3855, "step": 896 }, { "compression_loss": 84.13833618164062, "epoch": 0.32, "learning_rate": 2.5159016985905315e-05, "loss": 84.6056, "step": 897 }, { "compression_loss": 84.22270965576172, "epoch": 0.32, "learning_rate": 2.5153595952294903e-05, "loss": 84.6324, "step": 898 }, { "compression_loss": 82.92858123779297, "epoch": 0.32, "learning_rate": 2.5148174918684494e-05, "loss": 83.5019, "step": 899 }, { "compression_loss": 83.396240234375, "epoch": 0.33, "learning_rate": 2.5142753885074085e-05, "loss": 83.9197, "step": 900 }, { "compression_loss": 84.55399322509766, "epoch": 0.33, "learning_rate": 2.513733285146368e-05, "loss": 85.2014, "step": 901 }, { "compression_loss": 85.17497253417969, "epoch": 0.33, "learning_rate": 2.513191181785327e-05, "loss": 85.606, "step": 902 }, { "compression_loss": 82.42367553710938, "epoch": 0.33, "learning_rate": 2.5126490784242862e-05, "loss": 82.7872, "step": 903 }, { "compression_loss": 84.00953674316406, "epoch": 0.33, "learning_rate": 2.5121069750632454e-05, "loss": 84.4742, "step": 904 }, { "compression_loss": 83.71182250976562, "epoch": 0.33, "learning_rate": 2.5115648717022045e-05, "loss": 84.4882, "step": 905 }, { "compression_loss": 85.54178619384766, "epoch": 0.33, "learning_rate": 2.5110227683411636e-05, "loss": 86.335, "step": 906 }, { "compression_loss": 84.19461059570312, "epoch": 0.33, "learning_rate": 2.5104806649801228e-05, "loss": 84.5851, "step": 907 }, { "compression_loss": 84.15776062011719, "epoch": 0.33, "learning_rate": 2.509938561619082e-05, "loss": 84.9394, "step": 908 }, { "compression_loss": 85.40973663330078, "epoch": 0.33, "learning_rate": 2.5093964582580414e-05, "loss": 85.8895, "step": 909 }, { "compression_loss": 83.8176040649414, "epoch": 0.33, "learning_rate": 2.5088543548970005e-05, "loss": 84.3743, "step": 910 }, { "compression_loss": 84.77047729492188, "epoch": 0.33, "learning_rate": 2.5083122515359596e-05, "loss": 85.7325, "step": 911 }, { "compression_loss": 83.87710571289062, "epoch": 0.33, "learning_rate": 2.5077701481749187e-05, "loss": 84.5727, "step": 912 }, { "compression_loss": 83.93240356445312, "epoch": 0.33, "learning_rate": 2.507228044813878e-05, "loss": 84.4666, "step": 913 }, { "compression_loss": 84.43946075439453, "epoch": 0.33, "learning_rate": 2.506685941452837e-05, "loss": 84.7254, "step": 914 }, { "compression_loss": 83.87937927246094, "epoch": 0.33, "learning_rate": 2.506143838091796e-05, "loss": 84.1649, "step": 915 }, { "compression_loss": 84.44563293457031, "epoch": 0.33, "learning_rate": 2.5056017347307552e-05, "loss": 84.9938, "step": 916 }, { "compression_loss": 86.11697387695312, "epoch": 0.33, "learning_rate": 2.5050596313697147e-05, "loss": 86.7602, "step": 917 }, { "compression_loss": 84.94989776611328, "epoch": 0.33, "learning_rate": 2.504517528008674e-05, "loss": 85.5873, "step": 918 }, { "compression_loss": 84.70264434814453, "epoch": 0.33, "learning_rate": 2.503975424647633e-05, "loss": 85.2953, "step": 919 }, { "compression_loss": 84.1766128540039, "epoch": 0.33, "learning_rate": 2.503433321286592e-05, "loss": 84.6155, "step": 920 }, { "compression_loss": 83.69920349121094, "epoch": 0.33, "learning_rate": 2.5028912179255512e-05, "loss": 84.167, "step": 921 }, { "compression_loss": 83.14641571044922, "epoch": 0.33, "learning_rate": 2.5023491145645103e-05, "loss": 83.5246, "step": 922 }, { "compression_loss": 83.09089660644531, "epoch": 0.33, "learning_rate": 2.5018070112034695e-05, "loss": 83.6218, "step": 923 }, { "compression_loss": 83.26396942138672, "epoch": 0.33, "learning_rate": 2.5012649078424286e-05, "loss": 83.8837, "step": 924 }, { "compression_loss": 85.9346923828125, "epoch": 0.33, "learning_rate": 2.500722804481388e-05, "loss": 86.8047, "step": 925 }, { "compression_loss": 84.29994201660156, "epoch": 0.33, "learning_rate": 2.5001807011203472e-05, "loss": 84.8228, "step": 926 }, { "compression_loss": 85.42365264892578, "epoch": 0.34, "learning_rate": 2.4996385977593063e-05, "loss": 85.9857, "step": 927 }, { "compression_loss": 86.19358825683594, "epoch": 0.34, "learning_rate": 2.4990964943982654e-05, "loss": 86.6489, "step": 928 }, { "compression_loss": 84.2470703125, "epoch": 0.34, "learning_rate": 2.4985543910372246e-05, "loss": 84.8483, "step": 929 }, { "compression_loss": 84.7751235961914, "epoch": 0.34, "learning_rate": 2.4980122876761837e-05, "loss": 85.2594, "step": 930 }, { "compression_loss": 85.71968078613281, "epoch": 0.34, "learning_rate": 2.4974701843151428e-05, "loss": 86.6767, "step": 931 }, { "compression_loss": 84.3134994506836, "epoch": 0.34, "learning_rate": 2.496928080954102e-05, "loss": 84.8089, "step": 932 }, { "compression_loss": 84.36758422851562, "epoch": 0.34, "learning_rate": 2.496385977593061e-05, "loss": 84.8909, "step": 933 }, { "compression_loss": 83.53785705566406, "epoch": 0.34, "learning_rate": 2.4958438742320205e-05, "loss": 83.9614, "step": 934 }, { "compression_loss": 85.02967834472656, "epoch": 0.34, "learning_rate": 2.4953017708709797e-05, "loss": 85.6057, "step": 935 }, { "compression_loss": 83.75160217285156, "epoch": 0.34, "learning_rate": 2.4947596675099388e-05, "loss": 84.2024, "step": 936 }, { "compression_loss": 85.04425048828125, "epoch": 0.34, "learning_rate": 2.494217564148898e-05, "loss": 85.9373, "step": 937 }, { "compression_loss": 83.83051300048828, "epoch": 0.34, "learning_rate": 2.493675460787857e-05, "loss": 84.6793, "step": 938 }, { "compression_loss": 83.65213012695312, "epoch": 0.34, "learning_rate": 2.4931333574268162e-05, "loss": 84.0314, "step": 939 }, { "compression_loss": 84.5115737915039, "epoch": 0.34, "learning_rate": 2.4925912540657753e-05, "loss": 84.9506, "step": 940 }, { "compression_loss": 81.73818969726562, "epoch": 0.34, "learning_rate": 2.4920491507047344e-05, "loss": 82.1741, "step": 941 }, { "compression_loss": 84.04300689697266, "epoch": 0.34, "learning_rate": 2.4915070473436936e-05, "loss": 84.6938, "step": 942 }, { "compression_loss": 84.4754867553711, "epoch": 0.34, "learning_rate": 2.4909649439826527e-05, "loss": 85.0683, "step": 943 }, { "compression_loss": 85.10247802734375, "epoch": 0.34, "learning_rate": 2.4904228406216118e-05, "loss": 85.6844, "step": 944 }, { "compression_loss": 83.84111785888672, "epoch": 0.34, "learning_rate": 2.489880737260571e-05, "loss": 84.3007, "step": 945 }, { "compression_loss": 84.4124984741211, "epoch": 0.34, "learning_rate": 2.48933863389953e-05, "loss": 85.018, "step": 946 }, { "compression_loss": 84.44734191894531, "epoch": 0.34, "learning_rate": 2.4887965305384892e-05, "loss": 84.903, "step": 947 }, { "compression_loss": 85.24586486816406, "epoch": 0.34, "learning_rate": 2.4882544271774483e-05, "loss": 85.8951, "step": 948 }, { "compression_loss": 83.1393814086914, "epoch": 0.34, "learning_rate": 2.4877123238164075e-05, "loss": 83.6052, "step": 949 }, { "compression_loss": 83.67294311523438, "epoch": 0.34, "learning_rate": 2.487170220455367e-05, "loss": 84.4767, "step": 950 }, { "compression_loss": 85.71334075927734, "epoch": 0.34, "learning_rate": 2.486628117094326e-05, "loss": 86.4087, "step": 951 }, { "compression_loss": 85.02029418945312, "epoch": 0.34, "learning_rate": 2.4860860137332852e-05, "loss": 85.7032, "step": 952 }, { "compression_loss": 84.93768310546875, "epoch": 0.34, "learning_rate": 2.4855439103722443e-05, "loss": 85.4703, "step": 953 }, { "compression_loss": 84.4019775390625, "epoch": 0.34, "learning_rate": 2.4850018070112034e-05, "loss": 84.6892, "step": 954 }, { "compression_loss": 85.13780212402344, "epoch": 0.35, "learning_rate": 2.4844597036501626e-05, "loss": 85.5733, "step": 955 }, { "compression_loss": 84.29534912109375, "epoch": 0.35, "learning_rate": 2.4839176002891217e-05, "loss": 84.9331, "step": 956 }, { "compression_loss": 85.84622192382812, "epoch": 0.35, "learning_rate": 2.4833754969280808e-05, "loss": 86.3007, "step": 957 }, { "compression_loss": 85.21907043457031, "epoch": 0.35, "learning_rate": 2.4828333935670403e-05, "loss": 85.9105, "step": 958 }, { "compression_loss": 84.91978454589844, "epoch": 0.35, "learning_rate": 2.4822912902059994e-05, "loss": 85.6626, "step": 959 }, { "compression_loss": 85.50533294677734, "epoch": 0.35, "learning_rate": 2.4817491868449585e-05, "loss": 86.3111, "step": 960 }, { "compression_loss": 83.47358703613281, "epoch": 0.35, "learning_rate": 2.4812070834839177e-05, "loss": 84.2214, "step": 961 }, { "compression_loss": 83.90904998779297, "epoch": 0.35, "learning_rate": 2.4806649801228768e-05, "loss": 84.5363, "step": 962 }, { "compression_loss": 86.62151336669922, "epoch": 0.35, "learning_rate": 2.480122876761836e-05, "loss": 86.9768, "step": 963 }, { "compression_loss": 83.89350891113281, "epoch": 0.35, "learning_rate": 2.479580773400795e-05, "loss": 84.9002, "step": 964 }, { "compression_loss": 85.24290466308594, "epoch": 0.35, "learning_rate": 2.479038670039754e-05, "loss": 85.8674, "step": 965 }, { "compression_loss": 84.18820190429688, "epoch": 0.35, "learning_rate": 2.4784965666787136e-05, "loss": 84.6957, "step": 966 }, { "compression_loss": 83.14897918701172, "epoch": 0.35, "learning_rate": 2.4779544633176728e-05, "loss": 83.6285, "step": 967 }, { "compression_loss": 84.13102722167969, "epoch": 0.35, "learning_rate": 2.477412359956632e-05, "loss": 84.562, "step": 968 }, { "compression_loss": 85.18853759765625, "epoch": 0.35, "learning_rate": 2.476870256595591e-05, "loss": 85.5765, "step": 969 }, { "compression_loss": 85.92871856689453, "epoch": 0.35, "learning_rate": 2.47632815323455e-05, "loss": 86.7296, "step": 970 }, { "compression_loss": 83.18180847167969, "epoch": 0.35, "learning_rate": 2.4757860498735093e-05, "loss": 83.5739, "step": 971 }, { "compression_loss": 84.54383850097656, "epoch": 0.35, "learning_rate": 2.4752439465124684e-05, "loss": 85.399, "step": 972 }, { "compression_loss": 85.76522827148438, "epoch": 0.35, "learning_rate": 2.4747018431514275e-05, "loss": 86.5421, "step": 973 }, { "compression_loss": 83.01105499267578, "epoch": 0.35, "learning_rate": 2.474159739790387e-05, "loss": 83.2556, "step": 974 }, { "compression_loss": 83.60668182373047, "epoch": 0.35, "learning_rate": 2.473617636429346e-05, "loss": 84.1919, "step": 975 }, { "compression_loss": 86.41700744628906, "epoch": 0.35, "learning_rate": 2.4730755330683052e-05, "loss": 86.762, "step": 976 }, { "compression_loss": 85.73381042480469, "epoch": 0.35, "learning_rate": 2.4725334297072644e-05, "loss": 86.384, "step": 977 }, { "compression_loss": 84.56001281738281, "epoch": 0.35, "learning_rate": 2.4719913263462235e-05, "loss": 85.2999, "step": 978 }, { "compression_loss": 86.37754821777344, "epoch": 0.35, "learning_rate": 2.4714492229851826e-05, "loss": 86.7567, "step": 979 }, { "compression_loss": 83.37914276123047, "epoch": 0.35, "learning_rate": 2.4709071196241418e-05, "loss": 84.1458, "step": 980 }, { "compression_loss": 83.66268920898438, "epoch": 0.35, "learning_rate": 2.470365016263101e-05, "loss": 84.4309, "step": 981 }, { "compression_loss": 84.05776977539062, "epoch": 0.35, "learning_rate": 2.4698229129020603e-05, "loss": 84.7011, "step": 982 }, { "compression_loss": 85.12825012207031, "epoch": 0.36, "learning_rate": 2.4692808095410195e-05, "loss": 85.5371, "step": 983 }, { "compression_loss": 84.76219177246094, "epoch": 0.36, "learning_rate": 2.4687387061799786e-05, "loss": 85.4723, "step": 984 }, { "compression_loss": 82.43470764160156, "epoch": 0.36, "learning_rate": 2.4681966028189374e-05, "loss": 82.7711, "step": 985 }, { "compression_loss": 85.94712829589844, "epoch": 0.36, "learning_rate": 2.4676544994578965e-05, "loss": 86.6189, "step": 986 }, { "compression_loss": 83.81987762451172, "epoch": 0.36, "learning_rate": 2.4671123960968556e-05, "loss": 84.684, "step": 987 }, { "compression_loss": 82.70208740234375, "epoch": 0.36, "learning_rate": 2.4665702927358148e-05, "loss": 83.2036, "step": 988 }, { "compression_loss": 83.15705871582031, "epoch": 0.36, "learning_rate": 2.466028189374774e-05, "loss": 83.5022, "step": 989 }, { "compression_loss": 85.65031433105469, "epoch": 0.36, "learning_rate": 2.4654860860137334e-05, "loss": 86.1975, "step": 990 }, { "compression_loss": 84.91020202636719, "epoch": 0.36, "learning_rate": 2.4649439826526925e-05, "loss": 85.431, "step": 991 }, { "compression_loss": 83.90269470214844, "epoch": 0.36, "learning_rate": 2.4644018792916516e-05, "loss": 84.3102, "step": 992 }, { "compression_loss": 85.45319366455078, "epoch": 0.36, "learning_rate": 2.4638597759306107e-05, "loss": 85.8222, "step": 993 }, { "compression_loss": 85.20757293701172, "epoch": 0.36, "learning_rate": 2.46331767256957e-05, "loss": 85.7662, "step": 994 }, { "compression_loss": 84.00325012207031, "epoch": 0.36, "learning_rate": 2.462775569208529e-05, "loss": 84.4823, "step": 995 }, { "compression_loss": 86.526123046875, "epoch": 0.36, "learning_rate": 2.462233465847488e-05, "loss": 87.4973, "step": 996 }, { "compression_loss": 84.31915283203125, "epoch": 0.36, "learning_rate": 2.4616913624864473e-05, "loss": 84.8709, "step": 997 }, { "compression_loss": 85.92611694335938, "epoch": 0.36, "learning_rate": 2.4611492591254067e-05, "loss": 86.5458, "step": 998 }, { "compression_loss": 83.46392822265625, "epoch": 0.36, "learning_rate": 2.460607155764366e-05, "loss": 84.0, "step": 999 }, { "compression_loss": 84.9072265625, "epoch": 0.36, "learning_rate": 2.460065052403325e-05, "loss": 85.5256, "step": 1000 }, { "epoch": 0.36, "eval_exact_match": 85.44938505203406, "eval_f1": 92.13597516813063, "step": 1000 }, { "compression_loss": 82.56554412841797, "epoch": 0.36, "learning_rate": 2.459522949042284e-05, "loss": 83.4504, "step": 1001 }, { "compression_loss": 84.620849609375, "epoch": 0.36, "learning_rate": 2.4589808456812432e-05, "loss": 85.4732, "step": 1002 }, { "compression_loss": 83.03596496582031, "epoch": 0.36, "learning_rate": 2.4584387423202024e-05, "loss": 83.4376, "step": 1003 }, { "compression_loss": 83.36070251464844, "epoch": 0.36, "learning_rate": 2.4578966389591615e-05, "loss": 83.7616, "step": 1004 }, { "compression_loss": 83.09944152832031, "epoch": 0.36, "learning_rate": 2.4573545355981206e-05, "loss": 83.5175, "step": 1005 }, { "compression_loss": 86.04373168945312, "epoch": 0.36, "learning_rate": 2.45681243223708e-05, "loss": 86.5167, "step": 1006 }, { "compression_loss": 85.50700378417969, "epoch": 0.36, "learning_rate": 2.4562703288760392e-05, "loss": 86.2621, "step": 1007 }, { "compression_loss": 84.89134216308594, "epoch": 0.36, "learning_rate": 2.4557282255149983e-05, "loss": 85.3701, "step": 1008 }, { "compression_loss": 83.21021270751953, "epoch": 0.36, "learning_rate": 2.4551861221539575e-05, "loss": 83.5779, "step": 1009 }, { "compression_loss": 84.375244140625, "epoch": 0.37, "learning_rate": 2.4546440187929166e-05, "loss": 84.8761, "step": 1010 }, { "compression_loss": 83.42822265625, "epoch": 0.37, "learning_rate": 2.4541019154318757e-05, "loss": 84.0898, "step": 1011 }, { "compression_loss": 84.24783325195312, "epoch": 0.37, "learning_rate": 2.453559812070835e-05, "loss": 85.0362, "step": 1012 }, { "compression_loss": 82.80480194091797, "epoch": 0.37, "learning_rate": 2.453017708709794e-05, "loss": 83.5586, "step": 1013 }, { "compression_loss": 85.11318969726562, "epoch": 0.37, "learning_rate": 2.4524756053487534e-05, "loss": 85.3615, "step": 1014 }, { "compression_loss": 83.32952880859375, "epoch": 0.37, "learning_rate": 2.4519335019877126e-05, "loss": 83.8388, "step": 1015 }, { "compression_loss": 83.19050598144531, "epoch": 0.37, "learning_rate": 2.4513913986266717e-05, "loss": 83.7345, "step": 1016 }, { "compression_loss": 84.77371215820312, "epoch": 0.37, "learning_rate": 2.4508492952656308e-05, "loss": 85.5614, "step": 1017 }, { "compression_loss": 85.55835723876953, "epoch": 0.37, "learning_rate": 2.45030719190459e-05, "loss": 86.1553, "step": 1018 }, { "compression_loss": 83.74728393554688, "epoch": 0.37, "learning_rate": 2.449765088543549e-05, "loss": 84.316, "step": 1019 }, { "compression_loss": 83.01455688476562, "epoch": 0.37, "learning_rate": 2.4492229851825082e-05, "loss": 83.5338, "step": 1020 }, { "compression_loss": 84.33680725097656, "epoch": 0.37, "learning_rate": 2.4486808818214673e-05, "loss": 85.0655, "step": 1021 }, { "compression_loss": 86.58787536621094, "epoch": 0.37, "learning_rate": 2.4481387784604268e-05, "loss": 86.9589, "step": 1022 }, { "compression_loss": 85.56205749511719, "epoch": 0.37, "learning_rate": 2.447596675099386e-05, "loss": 86.3862, "step": 1023 }, { "compression_loss": 84.34419250488281, "epoch": 0.37, "learning_rate": 2.447054571738345e-05, "loss": 85.1101, "step": 1024 }, { "compression_loss": 84.57621765136719, "epoch": 0.37, "learning_rate": 2.4465124683773042e-05, "loss": 85.1347, "step": 1025 }, { "compression_loss": 85.90812683105469, "epoch": 0.37, "learning_rate": 2.4459703650162633e-05, "loss": 86.7833, "step": 1026 }, { "compression_loss": 85.52029418945312, "epoch": 0.37, "learning_rate": 2.4454282616552224e-05, "loss": 86.1765, "step": 1027 }, { "compression_loss": 83.89515686035156, "epoch": 0.37, "learning_rate": 2.4448861582941816e-05, "loss": 84.2073, "step": 1028 }, { "compression_loss": 84.15425109863281, "epoch": 0.37, "learning_rate": 2.4443440549331403e-05, "loss": 84.772, "step": 1029 }, { "compression_loss": 85.30426025390625, "epoch": 0.37, "learning_rate": 2.4438019515720998e-05, "loss": 85.8873, "step": 1030 }, { "compression_loss": 85.30856323242188, "epoch": 0.37, "learning_rate": 2.443259848211059e-05, "loss": 85.833, "step": 1031 }, { "compression_loss": 84.3972396850586, "epoch": 0.37, "learning_rate": 2.442717744850018e-05, "loss": 85.0497, "step": 1032 }, { "compression_loss": 83.55692291259766, "epoch": 0.37, "learning_rate": 2.4421756414889772e-05, "loss": 84.0312, "step": 1033 }, { "compression_loss": 84.916259765625, "epoch": 0.37, "learning_rate": 2.4416335381279363e-05, "loss": 85.4356, "step": 1034 }, { "compression_loss": 83.02468872070312, "epoch": 0.37, "learning_rate": 2.4410914347668954e-05, "loss": 83.5046, "step": 1035 }, { "compression_loss": 84.88188171386719, "epoch": 0.37, "learning_rate": 2.4405493314058546e-05, "loss": 85.7192, "step": 1036 }, { "compression_loss": 86.28982543945312, "epoch": 0.37, "learning_rate": 2.4400072280448137e-05, "loss": 87.0165, "step": 1037 }, { "compression_loss": 85.39530944824219, "epoch": 0.38, "learning_rate": 2.439465124683773e-05, "loss": 86.108, "step": 1038 }, { "compression_loss": 84.0799789428711, "epoch": 0.38, "learning_rate": 2.4389230213227323e-05, "loss": 84.5824, "step": 1039 }, { "compression_loss": 83.69551849365234, "epoch": 0.38, "learning_rate": 2.4383809179616914e-05, "loss": 84.2878, "step": 1040 }, { "compression_loss": 84.44645690917969, "epoch": 0.38, "learning_rate": 2.4378388146006505e-05, "loss": 85.1776, "step": 1041 }, { "compression_loss": 84.50399780273438, "epoch": 0.38, "learning_rate": 2.4372967112396097e-05, "loss": 84.9923, "step": 1042 }, { "compression_loss": 85.3138427734375, "epoch": 0.38, "learning_rate": 2.4367546078785688e-05, "loss": 85.7238, "step": 1043 }, { "compression_loss": 85.00729370117188, "epoch": 0.38, "learning_rate": 2.436212504517528e-05, "loss": 85.6944, "step": 1044 }, { "compression_loss": 84.91807556152344, "epoch": 0.38, "learning_rate": 2.435670401156487e-05, "loss": 85.8801, "step": 1045 }, { "compression_loss": 83.57588195800781, "epoch": 0.38, "learning_rate": 2.4351282977954465e-05, "loss": 83.9721, "step": 1046 }, { "compression_loss": 82.85032653808594, "epoch": 0.38, "learning_rate": 2.4345861944344056e-05, "loss": 83.6245, "step": 1047 }, { "compression_loss": 85.14318084716797, "epoch": 0.38, "learning_rate": 2.4340440910733648e-05, "loss": 85.7397, "step": 1048 }, { "compression_loss": 85.72210693359375, "epoch": 0.38, "learning_rate": 2.433501987712324e-05, "loss": 86.4479, "step": 1049 }, { "compression_loss": 83.71017456054688, "epoch": 0.38, "learning_rate": 2.432959884351283e-05, "loss": 84.3499, "step": 1050 }, { "compression_loss": 85.482177734375, "epoch": 0.38, "learning_rate": 2.432417780990242e-05, "loss": 86.3331, "step": 1051 }, { "compression_loss": 84.38421630859375, "epoch": 0.38, "learning_rate": 2.4318756776292013e-05, "loss": 84.8224, "step": 1052 }, { "compression_loss": 84.03350830078125, "epoch": 0.38, "learning_rate": 2.4313335742681604e-05, "loss": 84.66, "step": 1053 }, { "compression_loss": 85.19165802001953, "epoch": 0.38, "learning_rate": 2.43079147090712e-05, "loss": 85.8071, "step": 1054 }, { "compression_loss": 86.55905151367188, "epoch": 0.38, "learning_rate": 2.430249367546079e-05, "loss": 86.975, "step": 1055 }, { "compression_loss": 82.17628479003906, "epoch": 0.38, "learning_rate": 2.429707264185038e-05, "loss": 82.7642, "step": 1056 }, { "compression_loss": 84.34781646728516, "epoch": 0.38, "learning_rate": 2.4291651608239973e-05, "loss": 84.9679, "step": 1057 }, { "compression_loss": 85.47159576416016, "epoch": 0.38, "learning_rate": 2.4286230574629564e-05, "loss": 86.2797, "step": 1058 }, { "compression_loss": 83.63294982910156, "epoch": 0.38, "learning_rate": 2.4280809541019155e-05, "loss": 84.0505, "step": 1059 }, { "compression_loss": 83.73222351074219, "epoch": 0.38, "learning_rate": 2.4275388507408746e-05, "loss": 84.4734, "step": 1060 }, { "compression_loss": 85.70764923095703, "epoch": 0.38, "learning_rate": 2.4269967473798338e-05, "loss": 86.1202, "step": 1061 }, { "compression_loss": 85.9281997680664, "epoch": 0.38, "learning_rate": 2.4264546440187932e-05, "loss": 86.5888, "step": 1062 }, { "compression_loss": 84.77808380126953, "epoch": 0.38, "learning_rate": 2.4259125406577524e-05, "loss": 85.3802, "step": 1063 }, { "compression_loss": 84.75013732910156, "epoch": 0.38, "learning_rate": 2.4253704372967115e-05, "loss": 85.5101, "step": 1064 }, { "compression_loss": 84.8255844116211, "epoch": 0.38, "learning_rate": 2.4248283339356706e-05, "loss": 85.2945, "step": 1065 }, { "compression_loss": 84.21197509765625, "epoch": 0.39, "learning_rate": 2.4242862305746297e-05, "loss": 84.584, "step": 1066 }, { "compression_loss": 84.43329620361328, "epoch": 0.39, "learning_rate": 2.423744127213589e-05, "loss": 84.7583, "step": 1067 }, { "compression_loss": 83.54495239257812, "epoch": 0.39, "learning_rate": 2.423202023852548e-05, "loss": 84.4097, "step": 1068 }, { "compression_loss": 83.14190673828125, "epoch": 0.39, "learning_rate": 2.422659920491507e-05, "loss": 83.5565, "step": 1069 }, { "compression_loss": 82.4168472290039, "epoch": 0.39, "learning_rate": 2.4221178171304666e-05, "loss": 82.9024, "step": 1070 }, { "compression_loss": 83.58914947509766, "epoch": 0.39, "learning_rate": 2.4215757137694257e-05, "loss": 84.1151, "step": 1071 }, { "compression_loss": 86.35258483886719, "epoch": 0.39, "learning_rate": 2.4210336104083845e-05, "loss": 87.3485, "step": 1072 }, { "compression_loss": 85.3751220703125, "epoch": 0.39, "learning_rate": 2.4204915070473436e-05, "loss": 86.4353, "step": 1073 }, { "compression_loss": 84.82194519042969, "epoch": 0.39, "learning_rate": 2.4199494036863028e-05, "loss": 85.4114, "step": 1074 }, { "compression_loss": 85.03106689453125, "epoch": 0.39, "learning_rate": 2.419407300325262e-05, "loss": 85.7324, "step": 1075 }, { "compression_loss": 85.19378662109375, "epoch": 0.39, "learning_rate": 2.418865196964221e-05, "loss": 85.7478, "step": 1076 }, { "compression_loss": 82.15807342529297, "epoch": 0.39, "learning_rate": 2.41832309360318e-05, "loss": 82.7108, "step": 1077 }, { "compression_loss": 84.89813232421875, "epoch": 0.39, "learning_rate": 2.4177809902421396e-05, "loss": 85.549, "step": 1078 }, { "compression_loss": 84.24378967285156, "epoch": 0.39, "learning_rate": 2.4172388868810987e-05, "loss": 84.548, "step": 1079 }, { "compression_loss": 84.36459350585938, "epoch": 0.39, "learning_rate": 2.416696783520058e-05, "loss": 84.8607, "step": 1080 }, { "compression_loss": 85.21354675292969, "epoch": 0.39, "learning_rate": 2.416154680159017e-05, "loss": 85.8682, "step": 1081 }, { "compression_loss": 83.93722534179688, "epoch": 0.39, "learning_rate": 2.415612576797976e-05, "loss": 84.7936, "step": 1082 }, { "compression_loss": 82.91729736328125, "epoch": 0.39, "learning_rate": 2.4150704734369352e-05, "loss": 83.7853, "step": 1083 }, { "compression_loss": 84.14550018310547, "epoch": 0.39, "learning_rate": 2.4145283700758944e-05, "loss": 85.0423, "step": 1084 }, { "compression_loss": 85.38776397705078, "epoch": 0.39, "learning_rate": 2.4139862667148535e-05, "loss": 86.1793, "step": 1085 }, { "compression_loss": 86.20372772216797, "epoch": 0.39, "learning_rate": 2.413444163353813e-05, "loss": 86.7467, "step": 1086 }, { "compression_loss": 85.33506774902344, "epoch": 0.39, "learning_rate": 2.412902059992772e-05, "loss": 85.5226, "step": 1087 }, { "compression_loss": 85.81690979003906, "epoch": 0.39, "learning_rate": 2.4123599566317312e-05, "loss": 86.5391, "step": 1088 }, { "compression_loss": 85.61870574951172, "epoch": 0.39, "learning_rate": 2.4118178532706903e-05, "loss": 86.2584, "step": 1089 }, { "compression_loss": 86.49847412109375, "epoch": 0.39, "learning_rate": 2.4112757499096495e-05, "loss": 87.1185, "step": 1090 }, { "compression_loss": 83.25227355957031, "epoch": 0.39, "learning_rate": 2.4107336465486086e-05, "loss": 83.7246, "step": 1091 }, { "compression_loss": 84.50752258300781, "epoch": 0.39, "learning_rate": 2.4101915431875677e-05, "loss": 85.1347, "step": 1092 }, { "compression_loss": 83.95006561279297, "epoch": 0.4, "learning_rate": 2.409649439826527e-05, "loss": 84.4715, "step": 1093 }, { "compression_loss": 84.25433349609375, "epoch": 0.4, "learning_rate": 2.4091073364654863e-05, "loss": 85.1522, "step": 1094 }, { "compression_loss": 84.457763671875, "epoch": 0.4, "learning_rate": 2.4085652331044454e-05, "loss": 85.0362, "step": 1095 }, { "compression_loss": 84.05923461914062, "epoch": 0.4, "learning_rate": 2.4080231297434046e-05, "loss": 85.0961, "step": 1096 }, { "compression_loss": 85.8155288696289, "epoch": 0.4, "learning_rate": 2.4074810263823637e-05, "loss": 86.5348, "step": 1097 }, { "compression_loss": 83.67774963378906, "epoch": 0.4, "learning_rate": 2.4069389230213228e-05, "loss": 84.2622, "step": 1098 }, { "compression_loss": 84.8853759765625, "epoch": 0.4, "learning_rate": 2.406396819660282e-05, "loss": 85.2059, "step": 1099 }, { "compression_loss": 83.23188781738281, "epoch": 0.4, "learning_rate": 2.405854716299241e-05, "loss": 83.6939, "step": 1100 }, { "compression_loss": 84.55037689208984, "epoch": 0.4, "learning_rate": 2.4053126129382002e-05, "loss": 85.0672, "step": 1101 }, { "compression_loss": 84.52494049072266, "epoch": 0.4, "learning_rate": 2.4047705095771597e-05, "loss": 85.115, "step": 1102 }, { "compression_loss": 84.90206909179688, "epoch": 0.4, "learning_rate": 2.4042284062161188e-05, "loss": 85.2906, "step": 1103 }, { "compression_loss": 83.11637878417969, "epoch": 0.4, "learning_rate": 2.403686302855078e-05, "loss": 83.3486, "step": 1104 }, { "compression_loss": 85.66752624511719, "epoch": 0.4, "learning_rate": 2.403144199494037e-05, "loss": 86.1991, "step": 1105 }, { "compression_loss": 86.06024932861328, "epoch": 0.4, "learning_rate": 2.4026020961329962e-05, "loss": 86.8108, "step": 1106 }, { "compression_loss": 84.6148452758789, "epoch": 0.4, "learning_rate": 2.4020599927719553e-05, "loss": 85.0977, "step": 1107 }, { "compression_loss": 82.58467102050781, "epoch": 0.4, "learning_rate": 2.4015178894109144e-05, "loss": 83.1507, "step": 1108 }, { "compression_loss": 84.7341079711914, "epoch": 0.4, "learning_rate": 2.4009757860498736e-05, "loss": 85.2142, "step": 1109 }, { "compression_loss": 85.13760375976562, "epoch": 0.4, "learning_rate": 2.400433682688833e-05, "loss": 85.6725, "step": 1110 }, { "compression_loss": 84.4765625, "epoch": 0.4, "learning_rate": 2.399891579327792e-05, "loss": 84.8263, "step": 1111 }, { "compression_loss": 83.66708374023438, "epoch": 0.4, "learning_rate": 2.3993494759667513e-05, "loss": 84.0507, "step": 1112 }, { "compression_loss": 84.2161636352539, "epoch": 0.4, "learning_rate": 2.3988073726057104e-05, "loss": 84.9759, "step": 1113 }, { "compression_loss": 84.99919128417969, "epoch": 0.4, "learning_rate": 2.3982652692446695e-05, "loss": 85.6736, "step": 1114 }, { "compression_loss": 86.01378631591797, "epoch": 0.4, "learning_rate": 2.3977231658836287e-05, "loss": 86.5985, "step": 1115 }, { "compression_loss": 86.5175552368164, "epoch": 0.4, "learning_rate": 2.3971810625225875e-05, "loss": 87.1822, "step": 1116 }, { "compression_loss": 86.16950988769531, "epoch": 0.4, "learning_rate": 2.3966389591615466e-05, "loss": 86.5545, "step": 1117 }, { "compression_loss": 84.18606567382812, "epoch": 0.4, "learning_rate": 2.3960968558005057e-05, "loss": 84.8543, "step": 1118 }, { "compression_loss": 83.52253723144531, "epoch": 0.4, "learning_rate": 2.3955547524394652e-05, "loss": 84.196, "step": 1119 }, { "compression_loss": 84.44020080566406, "epoch": 0.4, "learning_rate": 2.3950126490784243e-05, "loss": 85.2341, "step": 1120 }, { "compression_loss": 83.9029541015625, "epoch": 0.41, "learning_rate": 2.3944705457173834e-05, "loss": 84.451, "step": 1121 }, { "compression_loss": 83.44732666015625, "epoch": 0.41, "learning_rate": 2.3939284423563426e-05, "loss": 83.9018, "step": 1122 }, { "compression_loss": 85.6992416381836, "epoch": 0.41, "learning_rate": 2.3933863389953017e-05, "loss": 86.3586, "step": 1123 }, { "compression_loss": 84.9192886352539, "epoch": 0.41, "learning_rate": 2.3928442356342608e-05, "loss": 85.8553, "step": 1124 }, { "compression_loss": 83.92528533935547, "epoch": 0.41, "learning_rate": 2.39230213227322e-05, "loss": 84.6566, "step": 1125 }, { "compression_loss": 85.18891143798828, "epoch": 0.41, "learning_rate": 2.391760028912179e-05, "loss": 86.008, "step": 1126 }, { "compression_loss": 84.26443481445312, "epoch": 0.41, "learning_rate": 2.3912179255511385e-05, "loss": 84.9735, "step": 1127 }, { "compression_loss": 84.19216918945312, "epoch": 0.41, "learning_rate": 2.3906758221900977e-05, "loss": 85.029, "step": 1128 }, { "compression_loss": 84.59747314453125, "epoch": 0.41, "learning_rate": 2.3901337188290568e-05, "loss": 85.5958, "step": 1129 }, { "compression_loss": 82.85922241210938, "epoch": 0.41, "learning_rate": 2.389591615468016e-05, "loss": 83.575, "step": 1130 }, { "compression_loss": 85.66470336914062, "epoch": 0.41, "learning_rate": 2.389049512106975e-05, "loss": 86.2449, "step": 1131 }, { "compression_loss": 86.01346588134766, "epoch": 0.41, "learning_rate": 2.388507408745934e-05, "loss": 86.7543, "step": 1132 }, { "compression_loss": 83.77510070800781, "epoch": 0.41, "learning_rate": 2.3879653053848933e-05, "loss": 84.432, "step": 1133 }, { "compression_loss": 83.20730590820312, "epoch": 0.41, "learning_rate": 2.3874232020238524e-05, "loss": 83.7889, "step": 1134 }, { "compression_loss": 83.9296875, "epoch": 0.41, "learning_rate": 2.386881098662812e-05, "loss": 84.5255, "step": 1135 }, { "compression_loss": 85.21353149414062, "epoch": 0.41, "learning_rate": 2.386338995301771e-05, "loss": 85.655, "step": 1136 }, { "compression_loss": 83.36035919189453, "epoch": 0.41, "learning_rate": 2.38579689194073e-05, "loss": 84.1144, "step": 1137 }, { "compression_loss": 87.28843688964844, "epoch": 0.41, "learning_rate": 2.3852547885796893e-05, "loss": 87.7656, "step": 1138 }, { "compression_loss": 83.9898452758789, "epoch": 0.41, "learning_rate": 2.3847126852186484e-05, "loss": 84.3952, "step": 1139 }, { "compression_loss": 85.04411315917969, "epoch": 0.41, "learning_rate": 2.3841705818576075e-05, "loss": 85.8395, "step": 1140 }, { "compression_loss": 82.87483978271484, "epoch": 0.41, "learning_rate": 2.3836284784965667e-05, "loss": 83.5656, "step": 1141 }, { "compression_loss": 84.54560852050781, "epoch": 0.41, "learning_rate": 2.3830863751355258e-05, "loss": 84.9121, "step": 1142 }, { "compression_loss": 84.10966491699219, "epoch": 0.41, "learning_rate": 2.3825442717744852e-05, "loss": 84.4364, "step": 1143 }, { "compression_loss": 85.41942596435547, "epoch": 0.41, "learning_rate": 2.3820021684134444e-05, "loss": 86.1794, "step": 1144 }, { "compression_loss": 85.00371551513672, "epoch": 0.41, "learning_rate": 2.3814600650524035e-05, "loss": 85.6339, "step": 1145 }, { "compression_loss": 85.43299102783203, "epoch": 0.41, "learning_rate": 2.3809179616913626e-05, "loss": 86.1856, "step": 1146 }, { "compression_loss": 85.05221557617188, "epoch": 0.41, "learning_rate": 2.3803758583303218e-05, "loss": 85.6512, "step": 1147 }, { "compression_loss": 84.00667572021484, "epoch": 0.41, "learning_rate": 2.379833754969281e-05, "loss": 84.4156, "step": 1148 }, { "compression_loss": 84.66850280761719, "epoch": 0.42, "learning_rate": 2.37929165160824e-05, "loss": 85.0001, "step": 1149 }, { "compression_loss": 82.41879272460938, "epoch": 0.42, "learning_rate": 2.378749548247199e-05, "loss": 82.7201, "step": 1150 }, { "compression_loss": 84.94659423828125, "epoch": 0.42, "learning_rate": 2.3782074448861586e-05, "loss": 85.5515, "step": 1151 }, { "compression_loss": 85.379150390625, "epoch": 0.42, "learning_rate": 2.3776653415251177e-05, "loss": 86.0204, "step": 1152 }, { "compression_loss": 85.08391571044922, "epoch": 0.42, "learning_rate": 2.377123238164077e-05, "loss": 85.7726, "step": 1153 }, { "compression_loss": 83.1212158203125, "epoch": 0.42, "learning_rate": 2.376581134803036e-05, "loss": 83.8323, "step": 1154 }, { "compression_loss": 86.02227783203125, "epoch": 0.42, "learning_rate": 2.376039031441995e-05, "loss": 86.7347, "step": 1155 }, { "compression_loss": 85.03831481933594, "epoch": 0.42, "learning_rate": 2.3754969280809542e-05, "loss": 85.8202, "step": 1156 }, { "compression_loss": 83.68540954589844, "epoch": 0.42, "learning_rate": 2.3749548247199134e-05, "loss": 84.1236, "step": 1157 }, { "compression_loss": 83.81538391113281, "epoch": 0.42, "learning_rate": 2.3744127213588725e-05, "loss": 84.0905, "step": 1158 }, { "compression_loss": 85.60152435302734, "epoch": 0.42, "learning_rate": 2.3738706179978316e-05, "loss": 86.3227, "step": 1159 }, { "compression_loss": 84.33792114257812, "epoch": 0.42, "learning_rate": 2.3733285146367907e-05, "loss": 85.0826, "step": 1160 }, { "compression_loss": 85.91139221191406, "epoch": 0.42, "learning_rate": 2.37278641127575e-05, "loss": 86.5373, "step": 1161 }, { "compression_loss": 85.26598358154297, "epoch": 0.42, "learning_rate": 2.372244307914709e-05, "loss": 85.8304, "step": 1162 }, { "compression_loss": 83.69670104980469, "epoch": 0.42, "learning_rate": 2.371702204553668e-05, "loss": 84.0284, "step": 1163 }, { "compression_loss": 86.4041748046875, "epoch": 0.42, "learning_rate": 2.3711601011926273e-05, "loss": 87.086, "step": 1164 }, { "compression_loss": 85.16148376464844, "epoch": 0.42, "learning_rate": 2.3706179978315864e-05, "loss": 85.9328, "step": 1165 }, { "compression_loss": 84.93984985351562, "epoch": 0.42, "learning_rate": 2.3700758944705455e-05, "loss": 85.6204, "step": 1166 }, { "compression_loss": 82.98091888427734, "epoch": 0.42, "learning_rate": 2.369533791109505e-05, "loss": 83.4474, "step": 1167 }, { "compression_loss": 85.29712677001953, "epoch": 0.42, "learning_rate": 2.368991687748464e-05, "loss": 85.868, "step": 1168 }, { "compression_loss": 83.69659423828125, "epoch": 0.42, "learning_rate": 2.3684495843874232e-05, "loss": 84.6351, "step": 1169 }, { "compression_loss": 83.30363464355469, "epoch": 0.42, "learning_rate": 2.3679074810263824e-05, "loss": 83.5812, "step": 1170 }, { "compression_loss": 83.12786102294922, "epoch": 0.42, "learning_rate": 2.3673653776653415e-05, "loss": 83.3957, "step": 1171 }, { "compression_loss": 82.65850830078125, "epoch": 0.42, "learning_rate": 2.3668232743043006e-05, "loss": 83.1368, "step": 1172 }, { "compression_loss": 85.73980712890625, "epoch": 0.42, "learning_rate": 2.3662811709432597e-05, "loss": 87.0035, "step": 1173 }, { "compression_loss": 84.06488037109375, "epoch": 0.42, "learning_rate": 2.365739067582219e-05, "loss": 84.5321, "step": 1174 }, { "compression_loss": 86.23471069335938, "epoch": 0.42, "learning_rate": 2.3651969642211783e-05, "loss": 86.9551, "step": 1175 }, { "compression_loss": 83.92575073242188, "epoch": 0.43, "learning_rate": 2.3646548608601375e-05, "loss": 84.4896, "step": 1176 }, { "compression_loss": 83.92453002929688, "epoch": 0.43, "learning_rate": 2.3641127574990966e-05, "loss": 84.6059, "step": 1177 }, { "compression_loss": 84.66252136230469, "epoch": 0.43, "learning_rate": 2.3635706541380557e-05, "loss": 85.0888, "step": 1178 }, { "compression_loss": 83.33633422851562, "epoch": 0.43, "learning_rate": 2.363028550777015e-05, "loss": 83.7232, "step": 1179 }, { "compression_loss": 85.46729278564453, "epoch": 0.43, "learning_rate": 2.362486447415974e-05, "loss": 86.1979, "step": 1180 }, { "compression_loss": 83.7985610961914, "epoch": 0.43, "learning_rate": 2.361944344054933e-05, "loss": 84.3026, "step": 1181 }, { "compression_loss": 83.93997192382812, "epoch": 0.43, "learning_rate": 2.3614022406938922e-05, "loss": 84.3424, "step": 1182 }, { "compression_loss": 85.48312377929688, "epoch": 0.43, "learning_rate": 2.3608601373328517e-05, "loss": 85.8954, "step": 1183 }, { "compression_loss": 86.76251220703125, "epoch": 0.43, "learning_rate": 2.3603180339718108e-05, "loss": 87.5621, "step": 1184 }, { "compression_loss": 84.6446762084961, "epoch": 0.43, "learning_rate": 2.35977593061077e-05, "loss": 85.4951, "step": 1185 }, { "compression_loss": 84.63136291503906, "epoch": 0.43, "learning_rate": 2.359233827249729e-05, "loss": 85.2111, "step": 1186 }, { "compression_loss": 84.04906463623047, "epoch": 0.43, "learning_rate": 2.3586917238886882e-05, "loss": 84.3723, "step": 1187 }, { "compression_loss": 86.10072326660156, "epoch": 0.43, "learning_rate": 2.3581496205276473e-05, "loss": 86.7295, "step": 1188 }, { "compression_loss": 84.45893859863281, "epoch": 0.43, "learning_rate": 2.3576075171666064e-05, "loss": 85.0886, "step": 1189 }, { "compression_loss": 83.85890197753906, "epoch": 0.43, "learning_rate": 2.3570654138055656e-05, "loss": 84.4008, "step": 1190 }, { "compression_loss": 85.19632720947266, "epoch": 0.43, "learning_rate": 2.356523310444525e-05, "loss": 85.8931, "step": 1191 }, { "compression_loss": 84.9166259765625, "epoch": 0.43, "learning_rate": 2.3559812070834842e-05, "loss": 85.6906, "step": 1192 }, { "compression_loss": 83.84281921386719, "epoch": 0.43, "learning_rate": 2.3554391037224433e-05, "loss": 84.7031, "step": 1193 }, { "compression_loss": 83.7643814086914, "epoch": 0.43, "learning_rate": 2.3548970003614024e-05, "loss": 84.5211, "step": 1194 }, { "compression_loss": 83.37014770507812, "epoch": 0.43, "learning_rate": 2.3543548970003616e-05, "loss": 83.7593, "step": 1195 }, { "compression_loss": 82.2281265258789, "epoch": 0.43, "learning_rate": 2.3538127936393207e-05, "loss": 83.2378, "step": 1196 }, { "compression_loss": 83.51246643066406, "epoch": 0.43, "learning_rate": 2.3532706902782798e-05, "loss": 83.8129, "step": 1197 }, { "compression_loss": 83.26365661621094, "epoch": 0.43, "learning_rate": 2.352728586917239e-05, "loss": 83.8488, "step": 1198 }, { "compression_loss": 85.20063018798828, "epoch": 0.43, "learning_rate": 2.3521864835561984e-05, "loss": 85.8667, "step": 1199 }, { "compression_loss": 84.182861328125, "epoch": 0.43, "learning_rate": 2.3516443801951575e-05, "loss": 84.701, "step": 1200 }, { "compression_loss": 84.57157897949219, "epoch": 0.43, "learning_rate": 2.3511022768341167e-05, "loss": 84.9652, "step": 1201 }, { "compression_loss": 85.37025451660156, "epoch": 0.43, "learning_rate": 2.3505601734730758e-05, "loss": 86.2694, "step": 1202 }, { "compression_loss": 83.67613983154297, "epoch": 0.43, "learning_rate": 2.3500180701120346e-05, "loss": 84.4507, "step": 1203 }, { "compression_loss": 83.64921569824219, "epoch": 0.44, "learning_rate": 2.3494759667509937e-05, "loss": 84.274, "step": 1204 }, { "compression_loss": 84.64368438720703, "epoch": 0.44, "learning_rate": 2.3489338633899528e-05, "loss": 85.1693, "step": 1205 }, { "compression_loss": 84.8961181640625, "epoch": 0.44, "learning_rate": 2.348391760028912e-05, "loss": 85.4532, "step": 1206 }, { "compression_loss": 82.80238342285156, "epoch": 0.44, "learning_rate": 2.3478496566678714e-05, "loss": 83.5005, "step": 1207 }, { "compression_loss": 83.03089141845703, "epoch": 0.44, "learning_rate": 2.3473075533068305e-05, "loss": 83.5553, "step": 1208 }, { "compression_loss": 85.17620849609375, "epoch": 0.44, "learning_rate": 2.3467654499457897e-05, "loss": 85.7066, "step": 1209 }, { "compression_loss": 84.70262145996094, "epoch": 0.44, "learning_rate": 2.3462233465847488e-05, "loss": 85.1596, "step": 1210 }, { "compression_loss": 84.5599365234375, "epoch": 0.44, "learning_rate": 2.345681243223708e-05, "loss": 85.1018, "step": 1211 }, { "compression_loss": 85.27401733398438, "epoch": 0.44, "learning_rate": 2.345139139862667e-05, "loss": 85.9208, "step": 1212 }, { "compression_loss": 83.47859191894531, "epoch": 0.44, "learning_rate": 2.3445970365016262e-05, "loss": 84.4236, "step": 1213 }, { "compression_loss": 85.39884185791016, "epoch": 0.44, "learning_rate": 2.3440549331405853e-05, "loss": 85.7429, "step": 1214 }, { "compression_loss": 85.05274963378906, "epoch": 0.44, "learning_rate": 2.3435128297795448e-05, "loss": 85.6899, "step": 1215 }, { "compression_loss": 85.1895523071289, "epoch": 0.44, "learning_rate": 2.342970726418504e-05, "loss": 85.6518, "step": 1216 }, { "compression_loss": 82.89452362060547, "epoch": 0.44, "learning_rate": 2.342428623057463e-05, "loss": 83.5155, "step": 1217 }, { "compression_loss": 83.04869842529297, "epoch": 0.44, "learning_rate": 2.341886519696422e-05, "loss": 83.5604, "step": 1218 }, { "compression_loss": 83.5170669555664, "epoch": 0.44, "learning_rate": 2.3413444163353813e-05, "loss": 83.998, "step": 1219 }, { "compression_loss": 84.608154296875, "epoch": 0.44, "learning_rate": 2.3408023129743404e-05, "loss": 84.9281, "step": 1220 }, { "compression_loss": 86.00399017333984, "epoch": 0.44, "learning_rate": 2.3402602096132995e-05, "loss": 86.5926, "step": 1221 }, { "compression_loss": 84.27548217773438, "epoch": 0.44, "learning_rate": 2.3397181062522587e-05, "loss": 85.0242, "step": 1222 }, { "compression_loss": 83.66191101074219, "epoch": 0.44, "learning_rate": 2.339176002891218e-05, "loss": 84.1075, "step": 1223 }, { "compression_loss": 85.07029724121094, "epoch": 0.44, "learning_rate": 2.3386338995301773e-05, "loss": 85.8445, "step": 1224 }, { "compression_loss": 84.6280517578125, "epoch": 0.44, "learning_rate": 2.3380917961691364e-05, "loss": 85.6826, "step": 1225 }, { "compression_loss": 84.22452545166016, "epoch": 0.44, "learning_rate": 2.3375496928080955e-05, "loss": 85.0904, "step": 1226 }, { "compression_loss": 84.24761962890625, "epoch": 0.44, "learning_rate": 2.3370075894470546e-05, "loss": 84.914, "step": 1227 }, { "compression_loss": 84.90269470214844, "epoch": 0.44, "learning_rate": 2.3364654860860138e-05, "loss": 85.5027, "step": 1228 }, { "compression_loss": 84.51275634765625, "epoch": 0.44, "learning_rate": 2.335923382724973e-05, "loss": 84.9808, "step": 1229 }, { "compression_loss": 85.86766052246094, "epoch": 0.44, "learning_rate": 2.335381279363932e-05, "loss": 86.5804, "step": 1230 }, { "compression_loss": 82.83000946044922, "epoch": 0.44, "learning_rate": 2.3348391760028915e-05, "loss": 83.2, "step": 1231 }, { "compression_loss": 83.29898071289062, "epoch": 0.45, "learning_rate": 2.3342970726418506e-05, "loss": 84.051, "step": 1232 }, { "compression_loss": 84.6854248046875, "epoch": 0.45, "learning_rate": 2.3337549692808097e-05, "loss": 85.3883, "step": 1233 }, { "compression_loss": 84.61407470703125, "epoch": 0.45, "learning_rate": 2.333212865919769e-05, "loss": 85.3091, "step": 1234 }, { "compression_loss": 87.587890625, "epoch": 0.45, "learning_rate": 2.332670762558728e-05, "loss": 88.302, "step": 1235 }, { "compression_loss": 83.74329376220703, "epoch": 0.45, "learning_rate": 2.332128659197687e-05, "loss": 84.1588, "step": 1236 }, { "compression_loss": 84.71065521240234, "epoch": 0.45, "learning_rate": 2.3315865558366462e-05, "loss": 85.1351, "step": 1237 }, { "compression_loss": 84.6362075805664, "epoch": 0.45, "learning_rate": 2.3310444524756054e-05, "loss": 85.1986, "step": 1238 }, { "compression_loss": 85.21900177001953, "epoch": 0.45, "learning_rate": 2.330502349114565e-05, "loss": 86.0296, "step": 1239 }, { "compression_loss": 83.21783447265625, "epoch": 0.45, "learning_rate": 2.329960245753524e-05, "loss": 83.5223, "step": 1240 }, { "compression_loss": 84.48690032958984, "epoch": 0.45, "learning_rate": 2.329418142392483e-05, "loss": 85.0764, "step": 1241 }, { "compression_loss": 84.56859588623047, "epoch": 0.45, "learning_rate": 2.3288760390314422e-05, "loss": 85.1997, "step": 1242 }, { "compression_loss": 86.55284118652344, "epoch": 0.45, "learning_rate": 2.3283339356704014e-05, "loss": 87.1316, "step": 1243 }, { "compression_loss": 84.12330627441406, "epoch": 0.45, "learning_rate": 2.3277918323093605e-05, "loss": 84.8185, "step": 1244 }, { "compression_loss": 84.12094116210938, "epoch": 0.45, "learning_rate": 2.3272497289483196e-05, "loss": 84.5923, "step": 1245 }, { "compression_loss": 84.51530456542969, "epoch": 0.45, "learning_rate": 2.3267076255872787e-05, "loss": 85.0974, "step": 1246 }, { "compression_loss": 85.06292724609375, "epoch": 0.45, "learning_rate": 2.326165522226238e-05, "loss": 85.5913, "step": 1247 }, { "compression_loss": 83.12071228027344, "epoch": 0.45, "learning_rate": 2.325623418865197e-05, "loss": 83.621, "step": 1248 }, { "compression_loss": 85.53739166259766, "epoch": 0.45, "learning_rate": 2.325081315504156e-05, "loss": 86.3097, "step": 1249 }, { "compression_loss": 83.31149291992188, "epoch": 0.45, "learning_rate": 2.3245392121431152e-05, "loss": 83.8146, "step": 1250 }, { "epoch": 0.45, "eval_exact_match": 86.07379375591296, "eval_f1": 92.43443849171884, "step": 1250 }, { "compression_loss": 84.83229064941406, "epoch": 0.45, "learning_rate": 2.3239971087820744e-05, "loss": 85.3198, "step": 1251 }, { "compression_loss": 84.47201538085938, "epoch": 0.45, "learning_rate": 2.3234550054210335e-05, "loss": 84.9752, "step": 1252 }, { "compression_loss": 85.0939712524414, "epoch": 0.45, "learning_rate": 2.3229129020599926e-05, "loss": 85.8908, "step": 1253 }, { "compression_loss": 83.04492950439453, "epoch": 0.45, "learning_rate": 2.3223707986989517e-05, "loss": 83.4764, "step": 1254 }, { "compression_loss": 86.39665985107422, "epoch": 0.45, "learning_rate": 2.3218286953379112e-05, "loss": 87.0242, "step": 1255 }, { "compression_loss": 82.58690643310547, "epoch": 0.45, "learning_rate": 2.3212865919768703e-05, "loss": 83.268, "step": 1256 }, { "compression_loss": 84.52592468261719, "epoch": 0.45, "learning_rate": 2.3207444886158295e-05, "loss": 85.1322, "step": 1257 }, { "compression_loss": 83.20216369628906, "epoch": 0.45, "learning_rate": 2.3202023852547886e-05, "loss": 83.5324, "step": 1258 }, { "compression_loss": 84.67340087890625, "epoch": 0.46, "learning_rate": 2.3196602818937477e-05, "loss": 85.2801, "step": 1259 }, { "compression_loss": 84.05228424072266, "epoch": 0.46, "learning_rate": 2.319118178532707e-05, "loss": 84.4085, "step": 1260 }, { "compression_loss": 85.239990234375, "epoch": 0.46, "learning_rate": 2.318576075171666e-05, "loss": 85.9846, "step": 1261 }, { "compression_loss": 83.45968627929688, "epoch": 0.46, "learning_rate": 2.318033971810625e-05, "loss": 83.9991, "step": 1262 }, { "compression_loss": 83.79684448242188, "epoch": 0.46, "learning_rate": 2.3174918684495846e-05, "loss": 84.2205, "step": 1263 }, { "compression_loss": 84.95025634765625, "epoch": 0.46, "learning_rate": 2.3169497650885437e-05, "loss": 85.5362, "step": 1264 }, { "compression_loss": 85.99284362792969, "epoch": 0.46, "learning_rate": 2.3164076617275028e-05, "loss": 86.6463, "step": 1265 }, { "compression_loss": 85.16456604003906, "epoch": 0.46, "learning_rate": 2.315865558366462e-05, "loss": 86.0233, "step": 1266 }, { "compression_loss": 86.6165771484375, "epoch": 0.46, "learning_rate": 2.315323455005421e-05, "loss": 87.4237, "step": 1267 }, { "compression_loss": 84.8016357421875, "epoch": 0.46, "learning_rate": 2.3147813516443802e-05, "loss": 85.3565, "step": 1268 }, { "compression_loss": 84.49391174316406, "epoch": 0.46, "learning_rate": 2.3142392482833393e-05, "loss": 85.1726, "step": 1269 }, { "compression_loss": 83.71129608154297, "epoch": 0.46, "learning_rate": 2.3136971449222985e-05, "loss": 84.0262, "step": 1270 }, { "compression_loss": 82.82438659667969, "epoch": 0.46, "learning_rate": 2.313155041561258e-05, "loss": 83.3449, "step": 1271 }, { "compression_loss": 84.45337677001953, "epoch": 0.46, "learning_rate": 2.312612938200217e-05, "loss": 84.9257, "step": 1272 }, { "compression_loss": 86.01373291015625, "epoch": 0.46, "learning_rate": 2.3120708348391762e-05, "loss": 86.4217, "step": 1273 }, { "compression_loss": 84.1395263671875, "epoch": 0.46, "learning_rate": 2.3115287314781353e-05, "loss": 84.4252, "step": 1274 }, { "compression_loss": 84.27407836914062, "epoch": 0.46, "learning_rate": 2.3109866281170944e-05, "loss": 84.7987, "step": 1275 }, { "compression_loss": 86.56597900390625, "epoch": 0.46, "learning_rate": 2.3104445247560536e-05, "loss": 87.3138, "step": 1276 }, { "compression_loss": 82.45576477050781, "epoch": 0.46, "learning_rate": 2.3099024213950127e-05, "loss": 83.0441, "step": 1277 }, { "compression_loss": 84.01739501953125, "epoch": 0.46, "learning_rate": 2.3093603180339718e-05, "loss": 84.8621, "step": 1278 }, { "compression_loss": 83.86383819580078, "epoch": 0.46, "learning_rate": 2.3088182146729313e-05, "loss": 84.4242, "step": 1279 }, { "compression_loss": 83.95391082763672, "epoch": 0.46, "learning_rate": 2.3082761113118904e-05, "loss": 84.5186, "step": 1280 }, { "compression_loss": 84.83026123046875, "epoch": 0.46, "learning_rate": 2.3077340079508495e-05, "loss": 85.4202, "step": 1281 }, { "compression_loss": 84.13655090332031, "epoch": 0.46, "learning_rate": 2.3071919045898087e-05, "loss": 84.5702, "step": 1282 }, { "compression_loss": 83.85963439941406, "epoch": 0.46, "learning_rate": 2.3066498012287678e-05, "loss": 84.5998, "step": 1283 }, { "compression_loss": 85.59855651855469, "epoch": 0.46, "learning_rate": 2.306107697867727e-05, "loss": 86.1307, "step": 1284 }, { "compression_loss": 84.67771911621094, "epoch": 0.46, "learning_rate": 2.305565594506686e-05, "loss": 85.0954, "step": 1285 }, { "compression_loss": 86.02322387695312, "epoch": 0.46, "learning_rate": 2.3050234911456452e-05, "loss": 86.823, "step": 1286 }, { "compression_loss": 83.91072082519531, "epoch": 0.47, "learning_rate": 2.3044813877846046e-05, "loss": 84.5309, "step": 1287 }, { "compression_loss": 86.65304565429688, "epoch": 0.47, "learning_rate": 2.3039392844235638e-05, "loss": 87.1295, "step": 1288 }, { "compression_loss": 83.67817687988281, "epoch": 0.47, "learning_rate": 2.303397181062523e-05, "loss": 84.1133, "step": 1289 }, { "compression_loss": 83.99473571777344, "epoch": 0.47, "learning_rate": 2.3028550777014817e-05, "loss": 84.3919, "step": 1290 }, { "compression_loss": 83.50120544433594, "epoch": 0.47, "learning_rate": 2.3023129743404408e-05, "loss": 83.7226, "step": 1291 }, { "compression_loss": 83.80603790283203, "epoch": 0.47, "learning_rate": 2.3017708709794e-05, "loss": 84.4514, "step": 1292 }, { "compression_loss": 84.59921264648438, "epoch": 0.47, "learning_rate": 2.301228767618359e-05, "loss": 85.4272, "step": 1293 }, { "compression_loss": 84.57135009765625, "epoch": 0.47, "learning_rate": 2.3006866642573182e-05, "loss": 84.8672, "step": 1294 }, { "compression_loss": 84.5486068725586, "epoch": 0.47, "learning_rate": 2.3001445608962777e-05, "loss": 85.0141, "step": 1295 }, { "compression_loss": 85.00639343261719, "epoch": 0.47, "learning_rate": 2.2996024575352368e-05, "loss": 85.5357, "step": 1296 }, { "compression_loss": 83.88211059570312, "epoch": 0.47, "learning_rate": 2.299060354174196e-05, "loss": 84.2515, "step": 1297 }, { "compression_loss": 83.92893981933594, "epoch": 0.47, "learning_rate": 2.298518250813155e-05, "loss": 84.3559, "step": 1298 }, { "compression_loss": 85.13175201416016, "epoch": 0.47, "learning_rate": 2.297976147452114e-05, "loss": 85.6889, "step": 1299 }, { "compression_loss": 84.7450180053711, "epoch": 0.47, "learning_rate": 2.2974340440910733e-05, "loss": 85.5623, "step": 1300 }, { "compression_loss": 84.19590759277344, "epoch": 0.47, "learning_rate": 2.2968919407300324e-05, "loss": 84.6047, "step": 1301 }, { "compression_loss": 85.61615753173828, "epoch": 0.47, "learning_rate": 2.2963498373689915e-05, "loss": 86.2147, "step": 1302 }, { "compression_loss": 85.7218017578125, "epoch": 0.47, "learning_rate": 2.2958077340079507e-05, "loss": 86.4394, "step": 1303 }, { "compression_loss": 84.63580322265625, "epoch": 0.47, "learning_rate": 2.29526563064691e-05, "loss": 85.3283, "step": 1304 }, { "compression_loss": 85.28157043457031, "epoch": 0.47, "learning_rate": 2.2947235272858693e-05, "loss": 85.7459, "step": 1305 }, { "compression_loss": 82.94902801513672, "epoch": 0.47, "learning_rate": 2.2941814239248284e-05, "loss": 83.3983, "step": 1306 }, { "compression_loss": 84.37324523925781, "epoch": 0.47, "learning_rate": 2.2936393205637875e-05, "loss": 84.7558, "step": 1307 }, { "compression_loss": 84.38572692871094, "epoch": 0.47, "learning_rate": 2.2930972172027467e-05, "loss": 85.2766, "step": 1308 }, { "compression_loss": 85.21871185302734, "epoch": 0.47, "learning_rate": 2.2925551138417058e-05, "loss": 85.728, "step": 1309 }, { "compression_loss": 83.9997329711914, "epoch": 0.47, "learning_rate": 2.292013010480665e-05, "loss": 84.6006, "step": 1310 }, { "compression_loss": 84.90603637695312, "epoch": 0.47, "learning_rate": 2.291470907119624e-05, "loss": 85.68, "step": 1311 }, { "compression_loss": 85.63320922851562, "epoch": 0.47, "learning_rate": 2.2909288037585835e-05, "loss": 86.197, "step": 1312 }, { "compression_loss": 83.89335632324219, "epoch": 0.47, "learning_rate": 2.2903867003975426e-05, "loss": 84.2904, "step": 1313 }, { "compression_loss": 83.71379089355469, "epoch": 0.47, "learning_rate": 2.2898445970365018e-05, "loss": 84.3284, "step": 1314 }, { "compression_loss": 83.27613830566406, "epoch": 0.48, "learning_rate": 2.289302493675461e-05, "loss": 83.7462, "step": 1315 }, { "compression_loss": 83.32781982421875, "epoch": 0.48, "learning_rate": 2.28876039031442e-05, "loss": 83.7711, "step": 1316 }, { "compression_loss": 85.496337890625, "epoch": 0.48, "learning_rate": 2.288218286953379e-05, "loss": 86.2423, "step": 1317 }, { "compression_loss": 84.38761901855469, "epoch": 0.48, "learning_rate": 2.2876761835923383e-05, "loss": 84.7943, "step": 1318 }, { "compression_loss": 84.69188690185547, "epoch": 0.48, "learning_rate": 2.2871340802312974e-05, "loss": 85.1741, "step": 1319 }, { "compression_loss": 84.15968322753906, "epoch": 0.48, "learning_rate": 2.286591976870257e-05, "loss": 84.5696, "step": 1320 }, { "compression_loss": 84.338623046875, "epoch": 0.48, "learning_rate": 2.286049873509216e-05, "loss": 84.7222, "step": 1321 }, { "compression_loss": 85.0526351928711, "epoch": 0.48, "learning_rate": 2.285507770148175e-05, "loss": 85.8082, "step": 1322 }, { "compression_loss": 85.18828582763672, "epoch": 0.48, "learning_rate": 2.2849656667871342e-05, "loss": 86.0627, "step": 1323 }, { "compression_loss": 84.83604431152344, "epoch": 0.48, "learning_rate": 2.2844235634260934e-05, "loss": 85.1478, "step": 1324 }, { "compression_loss": 85.47872924804688, "epoch": 0.48, "learning_rate": 2.2838814600650525e-05, "loss": 86.1433, "step": 1325 }, { "compression_loss": 84.89739990234375, "epoch": 0.48, "learning_rate": 2.2833393567040116e-05, "loss": 85.5123, "step": 1326 }, { "compression_loss": 84.80552673339844, "epoch": 0.48, "learning_rate": 2.2827972533429707e-05, "loss": 85.5342, "step": 1327 }, { "compression_loss": 84.47283935546875, "epoch": 0.48, "learning_rate": 2.2822551499819302e-05, "loss": 85.4222, "step": 1328 }, { "compression_loss": 84.66011047363281, "epoch": 0.48, "learning_rate": 2.2817130466208893e-05, "loss": 85.1019, "step": 1329 }, { "compression_loss": 84.67756652832031, "epoch": 0.48, "learning_rate": 2.2811709432598485e-05, "loss": 85.3083, "step": 1330 }, { "compression_loss": 82.86203002929688, "epoch": 0.48, "learning_rate": 2.2806288398988076e-05, "loss": 83.2521, "step": 1331 }, { "compression_loss": 83.857177734375, "epoch": 0.48, "learning_rate": 2.2800867365377667e-05, "loss": 84.3234, "step": 1332 }, { "compression_loss": 84.56466674804688, "epoch": 0.48, "learning_rate": 2.279544633176726e-05, "loss": 85.2038, "step": 1333 }, { "compression_loss": 83.87982940673828, "epoch": 0.48, "learning_rate": 2.2790025298156846e-05, "loss": 84.3743, "step": 1334 }, { "compression_loss": 84.84024047851562, "epoch": 0.48, "learning_rate": 2.2784604264546438e-05, "loss": 85.5766, "step": 1335 }, { "compression_loss": 85.08903503417969, "epoch": 0.48, "learning_rate": 2.2779183230936032e-05, "loss": 85.8383, "step": 1336 }, { "compression_loss": 82.83161926269531, "epoch": 0.48, "learning_rate": 2.2773762197325624e-05, "loss": 83.4074, "step": 1337 }, { "compression_loss": 85.15704345703125, "epoch": 0.48, "learning_rate": 2.2768341163715215e-05, "loss": 85.7827, "step": 1338 }, { "compression_loss": 83.85005187988281, "epoch": 0.48, "learning_rate": 2.2762920130104806e-05, "loss": 84.237, "step": 1339 }, { "compression_loss": 84.51742553710938, "epoch": 0.48, "learning_rate": 2.2757499096494397e-05, "loss": 85.1347, "step": 1340 }, { "compression_loss": 84.78395080566406, "epoch": 0.48, "learning_rate": 2.275207806288399e-05, "loss": 85.5821, "step": 1341 }, { "compression_loss": 86.30696868896484, "epoch": 0.49, "learning_rate": 2.274665702927358e-05, "loss": 87.0848, "step": 1342 }, { "compression_loss": 85.04267120361328, "epoch": 0.49, "learning_rate": 2.274123599566317e-05, "loss": 85.5769, "step": 1343 }, { "compression_loss": 85.65968322753906, "epoch": 0.49, "learning_rate": 2.2735814962052766e-05, "loss": 86.1928, "step": 1344 }, { "compression_loss": 83.72047424316406, "epoch": 0.49, "learning_rate": 2.2730393928442357e-05, "loss": 84.1713, "step": 1345 }, { "compression_loss": 83.74919128417969, "epoch": 0.49, "learning_rate": 2.272497289483195e-05, "loss": 84.4541, "step": 1346 }, { "compression_loss": 84.20841217041016, "epoch": 0.49, "learning_rate": 2.271955186122154e-05, "loss": 84.916, "step": 1347 }, { "compression_loss": 83.68499755859375, "epoch": 0.49, "learning_rate": 2.271413082761113e-05, "loss": 84.3498, "step": 1348 }, { "compression_loss": 85.03475189208984, "epoch": 0.49, "learning_rate": 2.2708709794000722e-05, "loss": 85.7049, "step": 1349 }, { "compression_loss": 83.93406677246094, "epoch": 0.49, "learning_rate": 2.2703288760390313e-05, "loss": 84.5522, "step": 1350 }, { "compression_loss": 83.57637786865234, "epoch": 0.49, "learning_rate": 2.2697867726779905e-05, "loss": 84.1411, "step": 1351 }, { "compression_loss": 83.2172622680664, "epoch": 0.49, "learning_rate": 2.26924466931695e-05, "loss": 83.7036, "step": 1352 }, { "compression_loss": 84.85301208496094, "epoch": 0.49, "learning_rate": 2.268702565955909e-05, "loss": 85.3321, "step": 1353 }, { "compression_loss": 85.08485412597656, "epoch": 0.49, "learning_rate": 2.2681604625948682e-05, "loss": 85.7808, "step": 1354 }, { "compression_loss": 84.00068664550781, "epoch": 0.49, "learning_rate": 2.2676183592338273e-05, "loss": 84.4668, "step": 1355 }, { "compression_loss": 84.50212097167969, "epoch": 0.49, "learning_rate": 2.2670762558727864e-05, "loss": 84.8923, "step": 1356 }, { "compression_loss": 85.02130126953125, "epoch": 0.49, "learning_rate": 2.2665341525117456e-05, "loss": 85.7057, "step": 1357 }, { "compression_loss": 86.82371520996094, "epoch": 0.49, "learning_rate": 2.2659920491507047e-05, "loss": 87.6292, "step": 1358 }, { "compression_loss": 84.5399169921875, "epoch": 0.49, "learning_rate": 2.265449945789664e-05, "loss": 85.0995, "step": 1359 }, { "compression_loss": 85.48749542236328, "epoch": 0.49, "learning_rate": 2.2649078424286233e-05, "loss": 85.9114, "step": 1360 }, { "compression_loss": 84.22306823730469, "epoch": 0.49, "learning_rate": 2.2643657390675824e-05, "loss": 84.5039, "step": 1361 }, { "compression_loss": 84.3080062866211, "epoch": 0.49, "learning_rate": 2.2638236357065416e-05, "loss": 84.7996, "step": 1362 }, { "compression_loss": 84.78108215332031, "epoch": 0.49, "learning_rate": 2.2632815323455007e-05, "loss": 85.3996, "step": 1363 }, { "compression_loss": 85.15040588378906, "epoch": 0.49, "learning_rate": 2.2627394289844598e-05, "loss": 85.5231, "step": 1364 }, { "compression_loss": 84.6453857421875, "epoch": 0.49, "learning_rate": 2.262197325623419e-05, "loss": 85.1444, "step": 1365 }, { "compression_loss": 84.48884582519531, "epoch": 0.49, "learning_rate": 2.261655222262378e-05, "loss": 84.8835, "step": 1366 }, { "compression_loss": 83.8958740234375, "epoch": 0.49, "learning_rate": 2.2611131189013372e-05, "loss": 84.6173, "step": 1367 }, { "compression_loss": 83.74427795410156, "epoch": 0.49, "learning_rate": 2.2605710155402967e-05, "loss": 84.1237, "step": 1368 }, { "compression_loss": 83.20219421386719, "epoch": 0.49, "learning_rate": 2.2600289121792558e-05, "loss": 83.7153, "step": 1369 }, { "compression_loss": 84.82501983642578, "epoch": 0.5, "learning_rate": 2.259486808818215e-05, "loss": 85.367, "step": 1370 }, { "compression_loss": 84.80718994140625, "epoch": 0.5, "learning_rate": 2.258944705457174e-05, "loss": 85.5409, "step": 1371 }, { "compression_loss": 86.28742980957031, "epoch": 0.5, "learning_rate": 2.258402602096133e-05, "loss": 86.991, "step": 1372 }, { "compression_loss": 83.41084289550781, "epoch": 0.5, "learning_rate": 2.2578604987350923e-05, "loss": 84.1636, "step": 1373 }, { "compression_loss": 84.93870544433594, "epoch": 0.5, "learning_rate": 2.2573183953740514e-05, "loss": 85.2975, "step": 1374 }, { "compression_loss": 84.55706787109375, "epoch": 0.5, "learning_rate": 2.2567762920130105e-05, "loss": 85.1319, "step": 1375 }, { "compression_loss": 85.4754638671875, "epoch": 0.5, "learning_rate": 2.25623418865197e-05, "loss": 86.0124, "step": 1376 }, { "compression_loss": 83.77777099609375, "epoch": 0.5, "learning_rate": 2.2556920852909288e-05, "loss": 84.2335, "step": 1377 }, { "compression_loss": 83.56021118164062, "epoch": 0.5, "learning_rate": 2.255149981929888e-05, "loss": 84.2219, "step": 1378 }, { "compression_loss": 82.66796112060547, "epoch": 0.5, "learning_rate": 2.254607878568847e-05, "loss": 83.1481, "step": 1379 }, { "compression_loss": 85.48907470703125, "epoch": 0.5, "learning_rate": 2.2540657752078062e-05, "loss": 86.103, "step": 1380 }, { "compression_loss": 85.49102020263672, "epoch": 0.5, "learning_rate": 2.2535236718467653e-05, "loss": 85.9542, "step": 1381 }, { "compression_loss": 85.54490661621094, "epoch": 0.5, "learning_rate": 2.2529815684857244e-05, "loss": 86.1032, "step": 1382 }, { "compression_loss": 83.13223266601562, "epoch": 0.5, "learning_rate": 2.2524394651246836e-05, "loss": 83.9532, "step": 1383 }, { "compression_loss": 84.03330993652344, "epoch": 0.5, "learning_rate": 2.251897361763643e-05, "loss": 84.8782, "step": 1384 }, { "compression_loss": 84.8471908569336, "epoch": 0.5, "learning_rate": 2.251355258402602e-05, "loss": 85.4135, "step": 1385 }, { "compression_loss": 84.44173431396484, "epoch": 0.5, "learning_rate": 2.2508131550415613e-05, "loss": 84.9519, "step": 1386 }, { "compression_loss": 83.66978454589844, "epoch": 0.5, "learning_rate": 2.2502710516805204e-05, "loss": 84.193, "step": 1387 }, { "compression_loss": 83.04035949707031, "epoch": 0.5, "learning_rate": 2.2497289483194795e-05, "loss": 83.4119, "step": 1388 }, { "compression_loss": 84.55000305175781, "epoch": 0.5, "learning_rate": 2.2491868449584387e-05, "loss": 85.1563, "step": 1389 }, { "compression_loss": 84.08618927001953, "epoch": 0.5, "learning_rate": 2.2486447415973978e-05, "loss": 84.4456, "step": 1390 }, { "compression_loss": 84.48600769042969, "epoch": 0.5, "learning_rate": 2.248102638236357e-05, "loss": 84.8463, "step": 1391 }, { "compression_loss": 85.51531982421875, "epoch": 0.5, "learning_rate": 2.2475605348753164e-05, "loss": 86.1323, "step": 1392 }, { "compression_loss": 85.86405944824219, "epoch": 0.5, "learning_rate": 2.2470184315142755e-05, "loss": 86.3576, "step": 1393 }, { "compression_loss": 83.31752014160156, "epoch": 0.5, "learning_rate": 2.2464763281532346e-05, "loss": 83.6205, "step": 1394 }, { "compression_loss": 84.68412780761719, "epoch": 0.5, "learning_rate": 2.2459342247921938e-05, "loss": 85.1243, "step": 1395 }, { "compression_loss": 85.67514038085938, "epoch": 0.5, "learning_rate": 2.245392121431153e-05, "loss": 86.353, "step": 1396 }, { "compression_loss": 86.33607482910156, "epoch": 0.5, "learning_rate": 2.244850018070112e-05, "loss": 87.4883, "step": 1397 }, { "compression_loss": 85.4250259399414, "epoch": 0.51, "learning_rate": 2.244307914709071e-05, "loss": 86.1721, "step": 1398 }, { "compression_loss": 84.20252990722656, "epoch": 0.51, "learning_rate": 2.2437658113480303e-05, "loss": 84.8381, "step": 1399 }, { "compression_loss": 82.85645294189453, "epoch": 0.51, "learning_rate": 2.2432237079869897e-05, "loss": 83.313, "step": 1400 }, { "compression_loss": 84.00463104248047, "epoch": 0.51, "learning_rate": 2.242681604625949e-05, "loss": 84.3845, "step": 1401 }, { "compression_loss": 84.03643798828125, "epoch": 0.51, "learning_rate": 2.242139501264908e-05, "loss": 84.7157, "step": 1402 }, { "compression_loss": 85.54888916015625, "epoch": 0.51, "learning_rate": 2.241597397903867e-05, "loss": 86.1967, "step": 1403 }, { "compression_loss": 84.36759948730469, "epoch": 0.51, "learning_rate": 2.2410552945428262e-05, "loss": 84.9853, "step": 1404 }, { "compression_loss": 84.57864379882812, "epoch": 0.51, "learning_rate": 2.2405131911817854e-05, "loss": 85.1685, "step": 1405 }, { "compression_loss": 85.32688903808594, "epoch": 0.51, "learning_rate": 2.2399710878207445e-05, "loss": 85.6415, "step": 1406 }, { "compression_loss": 83.056396484375, "epoch": 0.51, "learning_rate": 2.2394289844597036e-05, "loss": 83.5003, "step": 1407 }, { "compression_loss": 85.130615234375, "epoch": 0.51, "learning_rate": 2.238886881098663e-05, "loss": 85.541, "step": 1408 }, { "compression_loss": 85.80003356933594, "epoch": 0.51, "learning_rate": 2.2383447777376222e-05, "loss": 86.4447, "step": 1409 }, { "compression_loss": 84.23219299316406, "epoch": 0.51, "learning_rate": 2.2378026743765814e-05, "loss": 84.9248, "step": 1410 }, { "compression_loss": 86.108642578125, "epoch": 0.51, "learning_rate": 2.2372605710155405e-05, "loss": 87.0654, "step": 1411 }, { "compression_loss": 85.76203155517578, "epoch": 0.51, "learning_rate": 2.2367184676544996e-05, "loss": 86.4188, "step": 1412 }, { "compression_loss": 85.33552551269531, "epoch": 0.51, "learning_rate": 2.2361763642934587e-05, "loss": 85.9295, "step": 1413 }, { "compression_loss": 84.23692321777344, "epoch": 0.51, "learning_rate": 2.235634260932418e-05, "loss": 84.894, "step": 1414 }, { "compression_loss": 85.55612182617188, "epoch": 0.51, "learning_rate": 2.235092157571377e-05, "loss": 86.0128, "step": 1415 }, { "compression_loss": 83.56597900390625, "epoch": 0.51, "learning_rate": 2.2345500542103365e-05, "loss": 83.9175, "step": 1416 }, { "compression_loss": 85.00337219238281, "epoch": 0.51, "learning_rate": 2.2340079508492956e-05, "loss": 85.5449, "step": 1417 }, { "compression_loss": 83.90182495117188, "epoch": 0.51, "learning_rate": 2.2334658474882547e-05, "loss": 84.1768, "step": 1418 }, { "compression_loss": 81.9633560180664, "epoch": 0.51, "learning_rate": 2.232923744127214e-05, "loss": 82.514, "step": 1419 }, { "compression_loss": 84.23060607910156, "epoch": 0.51, "learning_rate": 2.232381640766173e-05, "loss": 84.8523, "step": 1420 }, { "compression_loss": 84.5440673828125, "epoch": 0.51, "learning_rate": 2.2318395374051317e-05, "loss": 84.9776, "step": 1421 }, { "compression_loss": 82.73797607421875, "epoch": 0.51, "learning_rate": 2.231297434044091e-05, "loss": 83.1303, "step": 1422 }, { "compression_loss": 85.13508605957031, "epoch": 0.51, "learning_rate": 2.23075533068305e-05, "loss": 85.7191, "step": 1423 }, { "compression_loss": 84.68423461914062, "epoch": 0.51, "learning_rate": 2.2302132273220095e-05, "loss": 85.3176, "step": 1424 }, { "compression_loss": 83.24640655517578, "epoch": 0.51, "learning_rate": 2.2296711239609686e-05, "loss": 83.7644, "step": 1425 }, { "compression_loss": 84.38481140136719, "epoch": 0.52, "learning_rate": 2.2291290205999277e-05, "loss": 85.064, "step": 1426 }, { "compression_loss": 85.37747955322266, "epoch": 0.52, "learning_rate": 2.228586917238887e-05, "loss": 86.0334, "step": 1427 }, { "compression_loss": 84.77606201171875, "epoch": 0.52, "learning_rate": 2.228044813877846e-05, "loss": 85.3706, "step": 1428 }, { "compression_loss": 84.8058090209961, "epoch": 0.52, "learning_rate": 2.227502710516805e-05, "loss": 85.3214, "step": 1429 }, { "compression_loss": 84.9219970703125, "epoch": 0.52, "learning_rate": 2.2269606071557642e-05, "loss": 85.4817, "step": 1430 }, { "compression_loss": 81.71160888671875, "epoch": 0.52, "learning_rate": 2.2264185037947234e-05, "loss": 82.2549, "step": 1431 }, { "compression_loss": 84.65650939941406, "epoch": 0.52, "learning_rate": 2.2258764004336828e-05, "loss": 85.3658, "step": 1432 }, { "compression_loss": 85.1113052368164, "epoch": 0.52, "learning_rate": 2.225334297072642e-05, "loss": 85.6842, "step": 1433 }, { "compression_loss": 84.94718933105469, "epoch": 0.52, "learning_rate": 2.224792193711601e-05, "loss": 85.4498, "step": 1434 }, { "compression_loss": 85.28156280517578, "epoch": 0.52, "learning_rate": 2.2242500903505602e-05, "loss": 85.7914, "step": 1435 }, { "compression_loss": 85.09717559814453, "epoch": 0.52, "learning_rate": 2.2237079869895193e-05, "loss": 86.0401, "step": 1436 }, { "compression_loss": 84.63835906982422, "epoch": 0.52, "learning_rate": 2.2231658836284785e-05, "loss": 85.1723, "step": 1437 }, { "compression_loss": 84.8756103515625, "epoch": 0.52, "learning_rate": 2.2226237802674376e-05, "loss": 85.6589, "step": 1438 }, { "compression_loss": 83.61784362792969, "epoch": 0.52, "learning_rate": 2.2220816769063967e-05, "loss": 84.0942, "step": 1439 }, { "compression_loss": 85.67369842529297, "epoch": 0.52, "learning_rate": 2.2215395735453562e-05, "loss": 86.3556, "step": 1440 }, { "compression_loss": 84.44281768798828, "epoch": 0.52, "learning_rate": 2.2209974701843153e-05, "loss": 84.9615, "step": 1441 }, { "compression_loss": 84.42283630371094, "epoch": 0.52, "learning_rate": 2.2204553668232744e-05, "loss": 84.8412, "step": 1442 }, { "compression_loss": 84.89297485351562, "epoch": 0.52, "learning_rate": 2.2199132634622336e-05, "loss": 85.8604, "step": 1443 }, { "compression_loss": 84.40336608886719, "epoch": 0.52, "learning_rate": 2.2193711601011927e-05, "loss": 84.8951, "step": 1444 }, { "compression_loss": 84.08271789550781, "epoch": 0.52, "learning_rate": 2.2188290567401518e-05, "loss": 84.6857, "step": 1445 }, { "compression_loss": 83.64823913574219, "epoch": 0.52, "learning_rate": 2.218286953379111e-05, "loss": 84.0098, "step": 1446 }, { "compression_loss": 84.61482238769531, "epoch": 0.52, "learning_rate": 2.21774485001807e-05, "loss": 85.0527, "step": 1447 }, { "compression_loss": 83.69274139404297, "epoch": 0.52, "learning_rate": 2.2172027466570295e-05, "loss": 84.0794, "step": 1448 }, { "compression_loss": 84.36904907226562, "epoch": 0.52, "learning_rate": 2.2166606432959887e-05, "loss": 84.7923, "step": 1449 }, { "compression_loss": 84.32910919189453, "epoch": 0.52, "learning_rate": 2.2161185399349478e-05, "loss": 84.7484, "step": 1450 }, { "compression_loss": 83.78350830078125, "epoch": 0.52, "learning_rate": 2.215576436573907e-05, "loss": 84.2308, "step": 1451 }, { "compression_loss": 84.09471893310547, "epoch": 0.52, "learning_rate": 2.215034333212866e-05, "loss": 84.8025, "step": 1452 }, { "compression_loss": 84.389404296875, "epoch": 0.53, "learning_rate": 2.2144922298518252e-05, "loss": 85.0179, "step": 1453 }, { "compression_loss": 84.75181579589844, "epoch": 0.53, "learning_rate": 2.2139501264907843e-05, "loss": 85.2917, "step": 1454 }, { "compression_loss": 85.24054718017578, "epoch": 0.53, "learning_rate": 2.2134080231297434e-05, "loss": 86.0117, "step": 1455 }, { "compression_loss": 83.38407135009766, "epoch": 0.53, "learning_rate": 2.212865919768703e-05, "loss": 83.6947, "step": 1456 }, { "compression_loss": 85.09048461914062, "epoch": 0.53, "learning_rate": 2.212323816407662e-05, "loss": 85.7448, "step": 1457 }, { "compression_loss": 84.45732116699219, "epoch": 0.53, "learning_rate": 2.211781713046621e-05, "loss": 84.9301, "step": 1458 }, { "compression_loss": 84.33790588378906, "epoch": 0.53, "learning_rate": 2.2112396096855803e-05, "loss": 84.9207, "step": 1459 }, { "compression_loss": 85.06973266601562, "epoch": 0.53, "learning_rate": 2.2106975063245394e-05, "loss": 85.5788, "step": 1460 }, { "compression_loss": 85.99250030517578, "epoch": 0.53, "learning_rate": 2.2101554029634985e-05, "loss": 86.9301, "step": 1461 }, { "compression_loss": 84.10099029541016, "epoch": 0.53, "learning_rate": 2.2096132996024577e-05, "loss": 84.8901, "step": 1462 }, { "compression_loss": 83.95773315429688, "epoch": 0.53, "learning_rate": 2.2090711962414168e-05, "loss": 84.4222, "step": 1463 }, { "compression_loss": 85.61275482177734, "epoch": 0.53, "learning_rate": 2.208529092880376e-05, "loss": 86.418, "step": 1464 }, { "compression_loss": 84.66797637939453, "epoch": 0.53, "learning_rate": 2.207986989519335e-05, "loss": 85.3163, "step": 1465 }, { "compression_loss": 84.25137329101562, "epoch": 0.53, "learning_rate": 2.207444886158294e-05, "loss": 84.9297, "step": 1466 }, { "compression_loss": 84.49406433105469, "epoch": 0.53, "learning_rate": 2.2069027827972533e-05, "loss": 84.8866, "step": 1467 }, { "compression_loss": 83.5956802368164, "epoch": 0.53, "learning_rate": 2.2063606794362124e-05, "loss": 84.0631, "step": 1468 }, { "compression_loss": 85.04178619384766, "epoch": 0.53, "learning_rate": 2.2058185760751715e-05, "loss": 85.4523, "step": 1469 }, { "compression_loss": 84.48574829101562, "epoch": 0.53, "learning_rate": 2.2052764727141307e-05, "loss": 85.2008, "step": 1470 }, { "compression_loss": 85.89198303222656, "epoch": 0.53, "learning_rate": 2.2047343693530898e-05, "loss": 86.3285, "step": 1471 }, { "compression_loss": 82.43681335449219, "epoch": 0.53, "learning_rate": 2.2041922659920493e-05, "loss": 82.8244, "step": 1472 }, { "compression_loss": 84.12095642089844, "epoch": 0.53, "learning_rate": 2.2036501626310084e-05, "loss": 84.6087, "step": 1473 }, { "compression_loss": 84.24569702148438, "epoch": 0.53, "learning_rate": 2.2031080592699675e-05, "loss": 84.7864, "step": 1474 }, { "compression_loss": 82.99015045166016, "epoch": 0.53, "learning_rate": 2.2025659559089267e-05, "loss": 83.6988, "step": 1475 }, { "compression_loss": 83.31392669677734, "epoch": 0.53, "learning_rate": 2.2020238525478858e-05, "loss": 83.9061, "step": 1476 }, { "compression_loss": 83.9373779296875, "epoch": 0.53, "learning_rate": 2.201481749186845e-05, "loss": 84.3075, "step": 1477 }, { "compression_loss": 85.15977478027344, "epoch": 0.53, "learning_rate": 2.200939645825804e-05, "loss": 85.6846, "step": 1478 }, { "compression_loss": 85.06224060058594, "epoch": 0.53, "learning_rate": 2.200397542464763e-05, "loss": 86.1875, "step": 1479 }, { "compression_loss": 83.88160705566406, "epoch": 0.53, "learning_rate": 2.1998554391037223e-05, "loss": 84.5249, "step": 1480 }, { "compression_loss": 83.84994506835938, "epoch": 0.54, "learning_rate": 2.1993133357426818e-05, "loss": 84.4589, "step": 1481 }, { "compression_loss": 84.74229431152344, "epoch": 0.54, "learning_rate": 2.198771232381641e-05, "loss": 85.3175, "step": 1482 }, { "compression_loss": 86.03727722167969, "epoch": 0.54, "learning_rate": 2.1982291290206e-05, "loss": 86.3981, "step": 1483 }, { "compression_loss": 84.11876678466797, "epoch": 0.54, "learning_rate": 2.197687025659559e-05, "loss": 84.5925, "step": 1484 }, { "compression_loss": 82.82362365722656, "epoch": 0.54, "learning_rate": 2.1971449222985183e-05, "loss": 83.2458, "step": 1485 }, { "compression_loss": 84.48145294189453, "epoch": 0.54, "learning_rate": 2.1966028189374774e-05, "loss": 85.0446, "step": 1486 }, { "compression_loss": 86.41845703125, "epoch": 0.54, "learning_rate": 2.1960607155764365e-05, "loss": 87.0801, "step": 1487 }, { "compression_loss": 83.75334930419922, "epoch": 0.54, "learning_rate": 2.1955186122153956e-05, "loss": 84.2057, "step": 1488 }, { "compression_loss": 83.16517639160156, "epoch": 0.54, "learning_rate": 2.194976508854355e-05, "loss": 83.9889, "step": 1489 }, { "compression_loss": 82.36616516113281, "epoch": 0.54, "learning_rate": 2.1944344054933142e-05, "loss": 82.938, "step": 1490 }, { "compression_loss": 83.93672180175781, "epoch": 0.54, "learning_rate": 2.1938923021322734e-05, "loss": 84.6578, "step": 1491 }, { "compression_loss": 85.35514831542969, "epoch": 0.54, "learning_rate": 2.1933501987712325e-05, "loss": 86.6104, "step": 1492 }, { "compression_loss": 84.1763916015625, "epoch": 0.54, "learning_rate": 2.1928080954101916e-05, "loss": 85.0332, "step": 1493 }, { "compression_loss": 84.73175811767578, "epoch": 0.54, "learning_rate": 2.1922659920491507e-05, "loss": 85.0931, "step": 1494 }, { "compression_loss": 83.71873474121094, "epoch": 0.54, "learning_rate": 2.19172388868811e-05, "loss": 84.0075, "step": 1495 }, { "compression_loss": 83.13230895996094, "epoch": 0.54, "learning_rate": 2.191181785327069e-05, "loss": 83.5065, "step": 1496 }, { "compression_loss": 85.92241668701172, "epoch": 0.54, "learning_rate": 2.1906396819660285e-05, "loss": 86.6123, "step": 1497 }, { "compression_loss": 84.3239517211914, "epoch": 0.54, "learning_rate": 2.1900975786049876e-05, "loss": 84.9202, "step": 1498 }, { "compression_loss": 85.108154296875, "epoch": 0.54, "learning_rate": 2.1895554752439467e-05, "loss": 85.7107, "step": 1499 }, { "compression_loss": 83.60191345214844, "epoch": 0.54, "learning_rate": 2.189013371882906e-05, "loss": 84.352, "step": 1500 }, { "epoch": 0.54, "eval_exact_match": 85.44938505203406, "eval_f1": 92.14154215356514, "step": 1500 }, { "compression_loss": 84.11557006835938, "epoch": 0.54, "learning_rate": 2.188471268521865e-05, "loss": 84.638, "step": 1501 }, { "compression_loss": 83.41522216796875, "epoch": 0.54, "learning_rate": 2.187929165160824e-05, "loss": 83.9385, "step": 1502 }, { "compression_loss": 85.46888732910156, "epoch": 0.54, "learning_rate": 2.1873870617997832e-05, "loss": 85.9678, "step": 1503 }, { "compression_loss": 84.57022094726562, "epoch": 0.54, "learning_rate": 2.1868449584387424e-05, "loss": 85.1853, "step": 1504 }, { "compression_loss": 85.20394897460938, "epoch": 0.54, "learning_rate": 2.1863028550777018e-05, "loss": 85.807, "step": 1505 }, { "compression_loss": 85.3830795288086, "epoch": 0.54, "learning_rate": 2.185760751716661e-05, "loss": 86.0483, "step": 1506 }, { "compression_loss": 84.09417724609375, "epoch": 0.54, "learning_rate": 2.18521864835562e-05, "loss": 84.5651, "step": 1507 }, { "compression_loss": 83.1566162109375, "epoch": 0.54, "learning_rate": 2.184676544994579e-05, "loss": 83.5347, "step": 1508 }, { "compression_loss": 86.80380249023438, "epoch": 0.55, "learning_rate": 2.184134441633538e-05, "loss": 87.6324, "step": 1509 }, { "compression_loss": 82.43743896484375, "epoch": 0.55, "learning_rate": 2.183592338272497e-05, "loss": 82.8911, "step": 1510 }, { "compression_loss": 84.43374633789062, "epoch": 0.55, "learning_rate": 2.1830502349114562e-05, "loss": 84.8587, "step": 1511 }, { "compression_loss": 84.25140380859375, "epoch": 0.55, "learning_rate": 2.1825081315504154e-05, "loss": 84.8022, "step": 1512 }, { "compression_loss": 83.42913818359375, "epoch": 0.55, "learning_rate": 2.181966028189375e-05, "loss": 83.7344, "step": 1513 }, { "compression_loss": 83.64912414550781, "epoch": 0.55, "learning_rate": 2.181423924828334e-05, "loss": 83.9918, "step": 1514 }, { "compression_loss": 84.77371215820312, "epoch": 0.55, "learning_rate": 2.180881821467293e-05, "loss": 85.4167, "step": 1515 }, { "compression_loss": 84.38031768798828, "epoch": 0.55, "learning_rate": 2.1803397181062522e-05, "loss": 84.9435, "step": 1516 }, { "compression_loss": 84.984619140625, "epoch": 0.55, "learning_rate": 2.1797976147452113e-05, "loss": 85.6408, "step": 1517 }, { "compression_loss": 84.38133239746094, "epoch": 0.55, "learning_rate": 2.1792555113841705e-05, "loss": 84.9131, "step": 1518 }, { "compression_loss": 84.491455078125, "epoch": 0.55, "learning_rate": 2.1787134080231296e-05, "loss": 85.2691, "step": 1519 }, { "compression_loss": 84.46028900146484, "epoch": 0.55, "learning_rate": 2.1781713046620887e-05, "loss": 84.7458, "step": 1520 }, { "compression_loss": 83.34248352050781, "epoch": 0.55, "learning_rate": 2.1776292013010482e-05, "loss": 83.9855, "step": 1521 }, { "compression_loss": 83.65667724609375, "epoch": 0.55, "learning_rate": 2.1770870979400073e-05, "loss": 84.0796, "step": 1522 }, { "compression_loss": 85.62893676757812, "epoch": 0.55, "learning_rate": 2.1765449945789664e-05, "loss": 86.1213, "step": 1523 }, { "compression_loss": 84.62873840332031, "epoch": 0.55, "learning_rate": 2.1760028912179256e-05, "loss": 85.0901, "step": 1524 }, { "compression_loss": 83.16661071777344, "epoch": 0.55, "learning_rate": 2.1754607878568847e-05, "loss": 83.6107, "step": 1525 }, { "compression_loss": 84.66204833984375, "epoch": 0.55, "learning_rate": 2.174918684495844e-05, "loss": 85.101, "step": 1526 }, { "compression_loss": 85.82911682128906, "epoch": 0.55, "learning_rate": 2.174376581134803e-05, "loss": 86.452, "step": 1527 }, { "compression_loss": 82.50830078125, "epoch": 0.55, "learning_rate": 2.173834477773762e-05, "loss": 82.8505, "step": 1528 }, { "compression_loss": 85.78214263916016, "epoch": 0.55, "learning_rate": 2.1732923744127216e-05, "loss": 86.2777, "step": 1529 }, { "compression_loss": 85.13957977294922, "epoch": 0.55, "learning_rate": 2.1727502710516807e-05, "loss": 85.8319, "step": 1530 }, { "compression_loss": 84.47309875488281, "epoch": 0.55, "learning_rate": 2.1722081676906398e-05, "loss": 84.7955, "step": 1531 }, { "compression_loss": 84.41647338867188, "epoch": 0.55, "learning_rate": 2.171666064329599e-05, "loss": 84.8259, "step": 1532 }, { "compression_loss": 85.13618469238281, "epoch": 0.55, "learning_rate": 2.171123960968558e-05, "loss": 85.8553, "step": 1533 }, { "compression_loss": 85.79964447021484, "epoch": 0.55, "learning_rate": 2.1705818576075172e-05, "loss": 86.5246, "step": 1534 }, { "compression_loss": 83.348876953125, "epoch": 0.55, "learning_rate": 2.1700397542464763e-05, "loss": 83.8748, "step": 1535 }, { "compression_loss": 86.7050552368164, "epoch": 0.56, "learning_rate": 2.1694976508854354e-05, "loss": 87.4972, "step": 1536 }, { "compression_loss": 82.90019989013672, "epoch": 0.56, "learning_rate": 2.168955547524395e-05, "loss": 83.4342, "step": 1537 }, { "compression_loss": 84.02151489257812, "epoch": 0.56, "learning_rate": 2.168413444163354e-05, "loss": 84.4164, "step": 1538 }, { "compression_loss": 84.1946029663086, "epoch": 0.56, "learning_rate": 2.167871340802313e-05, "loss": 84.8834, "step": 1539 }, { "compression_loss": 85.58187866210938, "epoch": 0.56, "learning_rate": 2.1673292374412723e-05, "loss": 85.937, "step": 1540 }, { "compression_loss": 83.742919921875, "epoch": 0.56, "learning_rate": 2.1667871340802314e-05, "loss": 84.2484, "step": 1541 }, { "compression_loss": 82.88249206542969, "epoch": 0.56, "learning_rate": 2.1662450307191905e-05, "loss": 83.4928, "step": 1542 }, { "compression_loss": 84.17747497558594, "epoch": 0.56, "learning_rate": 2.1657029273581497e-05, "loss": 84.5637, "step": 1543 }, { "compression_loss": 85.97671508789062, "epoch": 0.56, "learning_rate": 2.1651608239971088e-05, "loss": 86.7862, "step": 1544 }, { "compression_loss": 84.738037109375, "epoch": 0.56, "learning_rate": 2.1646187206360683e-05, "loss": 85.2007, "step": 1545 }, { "compression_loss": 84.22909545898438, "epoch": 0.56, "learning_rate": 2.1640766172750274e-05, "loss": 84.9832, "step": 1546 }, { "compression_loss": 84.41941833496094, "epoch": 0.56, "learning_rate": 2.1635345139139865e-05, "loss": 85.0265, "step": 1547 }, { "compression_loss": 83.66214752197266, "epoch": 0.56, "learning_rate": 2.1629924105529456e-05, "loss": 84.4587, "step": 1548 }, { "compression_loss": 83.12887573242188, "epoch": 0.56, "learning_rate": 2.1624503071919048e-05, "loss": 83.6841, "step": 1549 }, { "compression_loss": 83.51850891113281, "epoch": 0.56, "learning_rate": 2.161908203830864e-05, "loss": 84.1164, "step": 1550 }, { "compression_loss": 85.34403991699219, "epoch": 0.56, "learning_rate": 2.161366100469823e-05, "loss": 85.851, "step": 1551 }, { "compression_loss": 85.40586853027344, "epoch": 0.56, "learning_rate": 2.1608239971087818e-05, "loss": 85.7753, "step": 1552 }, { "compression_loss": 84.23680114746094, "epoch": 0.56, "learning_rate": 2.1602818937477413e-05, "loss": 85.0439, "step": 1553 }, { "compression_loss": 85.66679382324219, "epoch": 0.56, "learning_rate": 2.1597397903867004e-05, "loss": 86.361, "step": 1554 }, { "compression_loss": 82.68067169189453, "epoch": 0.56, "learning_rate": 2.1591976870256595e-05, "loss": 83.1051, "step": 1555 }, { "compression_loss": 85.53544616699219, "epoch": 0.56, "learning_rate": 2.1586555836646187e-05, "loss": 85.8822, "step": 1556 }, { "compression_loss": 85.9629135131836, "epoch": 0.56, "learning_rate": 2.1581134803035778e-05, "loss": 86.6998, "step": 1557 }, { "compression_loss": 84.03036499023438, "epoch": 0.56, "learning_rate": 2.157571376942537e-05, "loss": 84.2913, "step": 1558 }, { "compression_loss": 85.4150390625, "epoch": 0.56, "learning_rate": 2.157029273581496e-05, "loss": 85.9529, "step": 1559 }, { "compression_loss": 83.92435455322266, "epoch": 0.56, "learning_rate": 2.1564871702204552e-05, "loss": 84.2475, "step": 1560 }, { "compression_loss": 83.38005065917969, "epoch": 0.56, "learning_rate": 2.1559450668594146e-05, "loss": 83.8711, "step": 1561 }, { "compression_loss": 85.27296447753906, "epoch": 0.56, "learning_rate": 2.1554029634983738e-05, "loss": 86.0283, "step": 1562 }, { "compression_loss": 83.04206085205078, "epoch": 0.56, "learning_rate": 2.154860860137333e-05, "loss": 83.8304, "step": 1563 }, { "compression_loss": 83.81806945800781, "epoch": 0.57, "learning_rate": 2.154318756776292e-05, "loss": 84.1781, "step": 1564 }, { "compression_loss": 84.02775573730469, "epoch": 0.57, "learning_rate": 2.153776653415251e-05, "loss": 84.4812, "step": 1565 }, { "compression_loss": 85.02687072753906, "epoch": 0.57, "learning_rate": 2.1532345500542103e-05, "loss": 85.9649, "step": 1566 }, { "compression_loss": 84.2643051147461, "epoch": 0.57, "learning_rate": 2.1526924466931694e-05, "loss": 84.824, "step": 1567 }, { "compression_loss": 83.64141845703125, "epoch": 0.57, "learning_rate": 2.1521503433321285e-05, "loss": 84.2903, "step": 1568 }, { "compression_loss": 83.40257263183594, "epoch": 0.57, "learning_rate": 2.151608239971088e-05, "loss": 84.0868, "step": 1569 }, { "compression_loss": 84.22950744628906, "epoch": 0.57, "learning_rate": 2.151066136610047e-05, "loss": 84.5719, "step": 1570 }, { "compression_loss": 84.40605926513672, "epoch": 0.57, "learning_rate": 2.1505240332490062e-05, "loss": 84.93, "step": 1571 }, { "compression_loss": 83.60055541992188, "epoch": 0.57, "learning_rate": 2.1499819298879654e-05, "loss": 83.9655, "step": 1572 }, { "compression_loss": 85.23939514160156, "epoch": 0.57, "learning_rate": 2.1494398265269245e-05, "loss": 85.9826, "step": 1573 }, { "compression_loss": 85.87540435791016, "epoch": 0.57, "learning_rate": 2.1488977231658836e-05, "loss": 86.4913, "step": 1574 }, { "compression_loss": 82.80998229980469, "epoch": 0.57, "learning_rate": 2.1483556198048428e-05, "loss": 83.171, "step": 1575 }, { "compression_loss": 82.73551177978516, "epoch": 0.57, "learning_rate": 2.147813516443802e-05, "loss": 83.0775, "step": 1576 }, { "compression_loss": 86.55859375, "epoch": 0.57, "learning_rate": 2.1472714130827614e-05, "loss": 87.3165, "step": 1577 }, { "compression_loss": 85.59724426269531, "epoch": 0.57, "learning_rate": 2.1467293097217205e-05, "loss": 86.5891, "step": 1578 }, { "compression_loss": 83.46650695800781, "epoch": 0.57, "learning_rate": 2.1461872063606796e-05, "loss": 84.5042, "step": 1579 }, { "compression_loss": 84.2576904296875, "epoch": 0.57, "learning_rate": 2.1456451029996387e-05, "loss": 84.5829, "step": 1580 }, { "compression_loss": 84.21062469482422, "epoch": 0.57, "learning_rate": 2.145102999638598e-05, "loss": 84.9124, "step": 1581 }, { "compression_loss": 85.7171630859375, "epoch": 0.57, "learning_rate": 2.144560896277557e-05, "loss": 86.2276, "step": 1582 }, { "compression_loss": 84.4788589477539, "epoch": 0.57, "learning_rate": 2.144018792916516e-05, "loss": 84.8539, "step": 1583 }, { "compression_loss": 84.72443389892578, "epoch": 0.57, "learning_rate": 2.1434766895554752e-05, "loss": 85.4019, "step": 1584 }, { "compression_loss": 85.45748901367188, "epoch": 0.57, "learning_rate": 2.1429345861944347e-05, "loss": 86.1887, "step": 1585 }, { "compression_loss": 84.76008605957031, "epoch": 0.57, "learning_rate": 2.142392482833394e-05, "loss": 85.3408, "step": 1586 }, { "compression_loss": 85.33415222167969, "epoch": 0.57, "learning_rate": 2.141850379472353e-05, "loss": 85.7297, "step": 1587 }, { "compression_loss": 84.62443542480469, "epoch": 0.57, "learning_rate": 2.141308276111312e-05, "loss": 84.947, "step": 1588 }, { "compression_loss": 83.67733764648438, "epoch": 0.57, "learning_rate": 2.1407661727502712e-05, "loss": 84.0907, "step": 1589 }, { "compression_loss": 84.17938232421875, "epoch": 0.57, "learning_rate": 2.1402240693892303e-05, "loss": 84.6473, "step": 1590 }, { "compression_loss": 82.66220092773438, "epoch": 0.57, "learning_rate": 2.1396819660281895e-05, "loss": 83.1382, "step": 1591 }, { "compression_loss": 84.56261444091797, "epoch": 0.58, "learning_rate": 2.1391398626671486e-05, "loss": 85.3206, "step": 1592 }, { "compression_loss": 84.62220764160156, "epoch": 0.58, "learning_rate": 2.138597759306108e-05, "loss": 85.47, "step": 1593 }, { "compression_loss": 85.591796875, "epoch": 0.58, "learning_rate": 2.1380556559450672e-05, "loss": 86.1431, "step": 1594 }, { "compression_loss": 84.28289794921875, "epoch": 0.58, "learning_rate": 2.137513552584026e-05, "loss": 84.834, "step": 1595 }, { "compression_loss": 82.30577850341797, "epoch": 0.58, "learning_rate": 2.136971449222985e-05, "loss": 82.7927, "step": 1596 }, { "compression_loss": 86.29557800292969, "epoch": 0.58, "learning_rate": 2.1364293458619442e-05, "loss": 86.9257, "step": 1597 }, { "compression_loss": 84.94123840332031, "epoch": 0.58, "learning_rate": 2.1358872425009034e-05, "loss": 85.4853, "step": 1598 }, { "compression_loss": 82.97615814208984, "epoch": 0.58, "learning_rate": 2.1353451391398625e-05, "loss": 83.3734, "step": 1599 }, { "compression_loss": 83.1455078125, "epoch": 0.58, "learning_rate": 2.1348030357788216e-05, "loss": 83.5875, "step": 1600 }, { "compression_loss": 83.84127807617188, "epoch": 0.58, "learning_rate": 2.134260932417781e-05, "loss": 84.1334, "step": 1601 }, { "compression_loss": 84.48343658447266, "epoch": 0.58, "learning_rate": 2.1337188290567402e-05, "loss": 85.2525, "step": 1602 }, { "compression_loss": 85.56975555419922, "epoch": 0.58, "learning_rate": 2.1331767256956993e-05, "loss": 86.3299, "step": 1603 }, { "compression_loss": 83.96492004394531, "epoch": 0.58, "learning_rate": 2.1326346223346585e-05, "loss": 84.3488, "step": 1604 }, { "compression_loss": 87.36273193359375, "epoch": 0.58, "learning_rate": 2.1320925189736176e-05, "loss": 88.0457, "step": 1605 }, { "compression_loss": 84.13773345947266, "epoch": 0.58, "learning_rate": 2.1315504156125767e-05, "loss": 84.5151, "step": 1606 }, { "compression_loss": 83.42868041992188, "epoch": 0.58, "learning_rate": 2.131008312251536e-05, "loss": 83.8682, "step": 1607 }, { "compression_loss": 84.1367416381836, "epoch": 0.58, "learning_rate": 2.130466208890495e-05, "loss": 84.569, "step": 1608 }, { "compression_loss": 84.38580322265625, "epoch": 0.58, "learning_rate": 2.1299241055294544e-05, "loss": 84.8249, "step": 1609 }, { "compression_loss": 83.425048828125, "epoch": 0.58, "learning_rate": 2.1293820021684136e-05, "loss": 84.0503, "step": 1610 }, { "compression_loss": 83.06198120117188, "epoch": 0.58, "learning_rate": 2.1288398988073727e-05, "loss": 83.6222, "step": 1611 }, { "compression_loss": 85.33294677734375, "epoch": 0.58, "learning_rate": 2.1282977954463318e-05, "loss": 86.0244, "step": 1612 }, { "compression_loss": 83.46206665039062, "epoch": 0.58, "learning_rate": 2.127755692085291e-05, "loss": 83.9776, "step": 1613 }, { "compression_loss": 85.65990447998047, "epoch": 0.58, "learning_rate": 2.12721358872425e-05, "loss": 86.4834, "step": 1614 }, { "compression_loss": 84.331787109375, "epoch": 0.58, "learning_rate": 2.1266714853632092e-05, "loss": 84.8791, "step": 1615 }, { "compression_loss": 85.55797576904297, "epoch": 0.58, "learning_rate": 2.1261293820021683e-05, "loss": 86.0914, "step": 1616 }, { "compression_loss": 82.9869613647461, "epoch": 0.58, "learning_rate": 2.1255872786411278e-05, "loss": 83.5796, "step": 1617 }, { "compression_loss": 84.48580169677734, "epoch": 0.58, "learning_rate": 2.125045175280087e-05, "loss": 85.0046, "step": 1618 }, { "compression_loss": 84.54061889648438, "epoch": 0.59, "learning_rate": 2.124503071919046e-05, "loss": 85.0275, "step": 1619 }, { "compression_loss": 84.433837890625, "epoch": 0.59, "learning_rate": 2.1239609685580052e-05, "loss": 85.1145, "step": 1620 }, { "compression_loss": 85.60121154785156, "epoch": 0.59, "learning_rate": 2.1234188651969643e-05, "loss": 86.3845, "step": 1621 }, { "compression_loss": 84.10930633544922, "epoch": 0.59, "learning_rate": 2.1228767618359234e-05, "loss": 84.7674, "step": 1622 }, { "compression_loss": 83.7098388671875, "epoch": 0.59, "learning_rate": 2.1223346584748826e-05, "loss": 84.1485, "step": 1623 }, { "compression_loss": 83.00119018554688, "epoch": 0.59, "learning_rate": 2.1217925551138417e-05, "loss": 83.5987, "step": 1624 }, { "compression_loss": 84.73841094970703, "epoch": 0.59, "learning_rate": 2.121250451752801e-05, "loss": 85.3099, "step": 1625 }, { "compression_loss": 84.72058868408203, "epoch": 0.59, "learning_rate": 2.1207083483917603e-05, "loss": 85.6016, "step": 1626 }, { "compression_loss": 85.51434326171875, "epoch": 0.59, "learning_rate": 2.1201662450307194e-05, "loss": 86.4341, "step": 1627 }, { "compression_loss": 83.18025970458984, "epoch": 0.59, "learning_rate": 2.1196241416696785e-05, "loss": 83.6569, "step": 1628 }, { "compression_loss": 84.09783172607422, "epoch": 0.59, "learning_rate": 2.1190820383086377e-05, "loss": 84.5623, "step": 1629 }, { "compression_loss": 84.16141510009766, "epoch": 0.59, "learning_rate": 2.1185399349475968e-05, "loss": 84.6034, "step": 1630 }, { "compression_loss": 83.76407623291016, "epoch": 0.59, "learning_rate": 2.117997831586556e-05, "loss": 84.4321, "step": 1631 }, { "compression_loss": 83.23515319824219, "epoch": 0.59, "learning_rate": 2.117455728225515e-05, "loss": 83.727, "step": 1632 }, { "compression_loss": 83.17233276367188, "epoch": 0.59, "learning_rate": 2.1169136248644745e-05, "loss": 83.8446, "step": 1633 }, { "compression_loss": 85.41705322265625, "epoch": 0.59, "learning_rate": 2.1163715215034336e-05, "loss": 85.9413, "step": 1634 }, { "compression_loss": 85.0262680053711, "epoch": 0.59, "learning_rate": 2.1158294181423928e-05, "loss": 85.5607, "step": 1635 }, { "compression_loss": 84.49384307861328, "epoch": 0.59, "learning_rate": 2.115287314781352e-05, "loss": 85.1161, "step": 1636 }, { "compression_loss": 86.05371856689453, "epoch": 0.59, "learning_rate": 2.114745211420311e-05, "loss": 86.551, "step": 1637 }, { "compression_loss": 85.34385681152344, "epoch": 0.59, "learning_rate": 2.11420310805927e-05, "loss": 85.8516, "step": 1638 }, { "compression_loss": 83.4861831665039, "epoch": 0.59, "learning_rate": 2.113661004698229e-05, "loss": 83.7704, "step": 1639 }, { "compression_loss": 85.00462341308594, "epoch": 0.59, "learning_rate": 2.113118901337188e-05, "loss": 85.7461, "step": 1640 }, { "compression_loss": 85.13162231445312, "epoch": 0.59, "learning_rate": 2.1125767979761475e-05, "loss": 85.6994, "step": 1641 }, { "compression_loss": 84.05963897705078, "epoch": 0.59, "learning_rate": 2.1120346946151067e-05, "loss": 84.6387, "step": 1642 }, { "compression_loss": 83.49342346191406, "epoch": 0.59, "learning_rate": 2.1114925912540658e-05, "loss": 83.8998, "step": 1643 }, { "compression_loss": 83.85709381103516, "epoch": 0.59, "learning_rate": 2.110950487893025e-05, "loss": 84.3844, "step": 1644 }, { "compression_loss": 84.96685791015625, "epoch": 0.59, "learning_rate": 2.110408384531984e-05, "loss": 85.5035, "step": 1645 }, { "compression_loss": 84.49556732177734, "epoch": 0.59, "learning_rate": 2.109866281170943e-05, "loss": 84.8496, "step": 1646 }, { "compression_loss": 85.09786987304688, "epoch": 0.6, "learning_rate": 2.1093241778099023e-05, "loss": 85.5146, "step": 1647 }, { "compression_loss": 83.62306213378906, "epoch": 0.6, "learning_rate": 2.1087820744488614e-05, "loss": 84.0198, "step": 1648 }, { "compression_loss": 84.26306915283203, "epoch": 0.6, "learning_rate": 2.108239971087821e-05, "loss": 84.8776, "step": 1649 }, { "compression_loss": 83.10173034667969, "epoch": 0.6, "learning_rate": 2.10769786772678e-05, "loss": 83.6033, "step": 1650 }, { "compression_loss": 83.26126098632812, "epoch": 0.6, "learning_rate": 2.107155764365739e-05, "loss": 83.657, "step": 1651 }, { "compression_loss": 85.86083984375, "epoch": 0.6, "learning_rate": 2.1066136610046983e-05, "loss": 86.4628, "step": 1652 }, { "compression_loss": 81.92225646972656, "epoch": 0.6, "learning_rate": 2.1060715576436574e-05, "loss": 82.4812, "step": 1653 }, { "compression_loss": 84.02020263671875, "epoch": 0.6, "learning_rate": 2.1055294542826165e-05, "loss": 84.8035, "step": 1654 }, { "compression_loss": 85.56396484375, "epoch": 0.6, "learning_rate": 2.1049873509215756e-05, "loss": 86.2436, "step": 1655 }, { "compression_loss": 84.83577728271484, "epoch": 0.6, "learning_rate": 2.1044452475605348e-05, "loss": 85.3248, "step": 1656 }, { "compression_loss": 85.11773681640625, "epoch": 0.6, "learning_rate": 2.1039031441994942e-05, "loss": 85.922, "step": 1657 }, { "compression_loss": 84.6094970703125, "epoch": 0.6, "learning_rate": 2.1033610408384534e-05, "loss": 84.9777, "step": 1658 }, { "compression_loss": 85.3273696899414, "epoch": 0.6, "learning_rate": 2.1028189374774125e-05, "loss": 85.9515, "step": 1659 }, { "compression_loss": 83.84405517578125, "epoch": 0.6, "learning_rate": 2.1022768341163716e-05, "loss": 84.3903, "step": 1660 }, { "compression_loss": 84.76411437988281, "epoch": 0.6, "learning_rate": 2.1017347307553307e-05, "loss": 85.2067, "step": 1661 }, { "compression_loss": 85.09480285644531, "epoch": 0.6, "learning_rate": 2.10119262739429e-05, "loss": 85.6824, "step": 1662 }, { "compression_loss": 84.75674438476562, "epoch": 0.6, "learning_rate": 2.100650524033249e-05, "loss": 85.384, "step": 1663 }, { "compression_loss": 83.98624420166016, "epoch": 0.6, "learning_rate": 2.100108420672208e-05, "loss": 84.4148, "step": 1664 }, { "compression_loss": 84.63844299316406, "epoch": 0.6, "learning_rate": 2.0995663173111673e-05, "loss": 85.2414, "step": 1665 }, { "compression_loss": 84.20159149169922, "epoch": 0.6, "learning_rate": 2.0990242139501267e-05, "loss": 84.5777, "step": 1666 }, { "compression_loss": 82.86372375488281, "epoch": 0.6, "learning_rate": 2.098482110589086e-05, "loss": 83.3361, "step": 1667 }, { "compression_loss": 85.46376037597656, "epoch": 0.6, "learning_rate": 2.097940007228045e-05, "loss": 85.9392, "step": 1668 }, { "compression_loss": 83.37671661376953, "epoch": 0.6, "learning_rate": 2.097397903867004e-05, "loss": 84.0471, "step": 1669 }, { "compression_loss": 84.86619567871094, "epoch": 0.6, "learning_rate": 2.0968558005059632e-05, "loss": 85.7772, "step": 1670 }, { "compression_loss": 85.12931823730469, "epoch": 0.6, "learning_rate": 2.0963136971449224e-05, "loss": 85.9345, "step": 1671 }, { "compression_loss": 85.310546875, "epoch": 0.6, "learning_rate": 2.0957715937838815e-05, "loss": 85.7076, "step": 1672 }, { "compression_loss": 85.34276580810547, "epoch": 0.6, "learning_rate": 2.0952294904228406e-05, "loss": 86.1509, "step": 1673 }, { "compression_loss": 83.5549545288086, "epoch": 0.6, "learning_rate": 2.0946873870618e-05, "loss": 84.0818, "step": 1674 }, { "compression_loss": 82.98208618164062, "epoch": 0.61, "learning_rate": 2.0941452837007592e-05, "loss": 83.3834, "step": 1675 }, { "compression_loss": 84.3784408569336, "epoch": 0.61, "learning_rate": 2.0936031803397183e-05, "loss": 84.9532, "step": 1676 }, { "compression_loss": 85.66159057617188, "epoch": 0.61, "learning_rate": 2.0930610769786775e-05, "loss": 85.8665, "step": 1677 }, { "compression_loss": 84.24844360351562, "epoch": 0.61, "learning_rate": 2.0925189736176366e-05, "loss": 84.7896, "step": 1678 }, { "compression_loss": 84.163330078125, "epoch": 0.61, "learning_rate": 2.0919768702565957e-05, "loss": 84.709, "step": 1679 }, { "compression_loss": 82.66996765136719, "epoch": 0.61, "learning_rate": 2.091434766895555e-05, "loss": 83.2341, "step": 1680 }, { "compression_loss": 84.26110076904297, "epoch": 0.61, "learning_rate": 2.090892663534514e-05, "loss": 84.971, "step": 1681 }, { "compression_loss": 84.347900390625, "epoch": 0.61, "learning_rate": 2.090350560173473e-05, "loss": 84.8584, "step": 1682 }, { "compression_loss": 84.28080749511719, "epoch": 0.61, "learning_rate": 2.0898084568124322e-05, "loss": 84.8189, "step": 1683 }, { "compression_loss": 85.71723175048828, "epoch": 0.61, "learning_rate": 2.0892663534513913e-05, "loss": 86.148, "step": 1684 }, { "compression_loss": 84.79695129394531, "epoch": 0.61, "learning_rate": 2.0887242500903505e-05, "loss": 85.2997, "step": 1685 }, { "compression_loss": 83.9103775024414, "epoch": 0.61, "learning_rate": 2.0881821467293096e-05, "loss": 84.2854, "step": 1686 }, { "compression_loss": 84.01321411132812, "epoch": 0.61, "learning_rate": 2.0876400433682687e-05, "loss": 84.6163, "step": 1687 }, { "compression_loss": 82.62459564208984, "epoch": 0.61, "learning_rate": 2.087097940007228e-05, "loss": 83.1094, "step": 1688 }, { "compression_loss": 84.9227294921875, "epoch": 0.61, "learning_rate": 2.086555836646187e-05, "loss": 85.2539, "step": 1689 }, { "compression_loss": 84.10169982910156, "epoch": 0.61, "learning_rate": 2.0860137332851464e-05, "loss": 84.3841, "step": 1690 }, { "compression_loss": 84.61906433105469, "epoch": 0.61, "learning_rate": 2.0854716299241056e-05, "loss": 85.2536, "step": 1691 }, { "compression_loss": 83.55055236816406, "epoch": 0.61, "learning_rate": 2.0849295265630647e-05, "loss": 84.0458, "step": 1692 }, { "compression_loss": 84.57471466064453, "epoch": 0.61, "learning_rate": 2.084387423202024e-05, "loss": 85.0567, "step": 1693 }, { "compression_loss": 85.57903289794922, "epoch": 0.61, "learning_rate": 2.083845319840983e-05, "loss": 86.2319, "step": 1694 }, { "compression_loss": 84.38053894042969, "epoch": 0.61, "learning_rate": 2.083303216479942e-05, "loss": 84.9907, "step": 1695 }, { "compression_loss": 83.95960235595703, "epoch": 0.61, "learning_rate": 2.0827611131189012e-05, "loss": 84.8583, "step": 1696 }, { "compression_loss": 85.0682373046875, "epoch": 0.61, "learning_rate": 2.0822190097578603e-05, "loss": 85.4746, "step": 1697 }, { "compression_loss": 84.20225524902344, "epoch": 0.61, "learning_rate": 2.0816769063968198e-05, "loss": 84.6468, "step": 1698 }, { "compression_loss": 83.14335632324219, "epoch": 0.61, "learning_rate": 2.081134803035779e-05, "loss": 83.4105, "step": 1699 }, { "compression_loss": 85.92225646972656, "epoch": 0.61, "learning_rate": 2.080592699674738e-05, "loss": 86.5301, "step": 1700 }, { "compression_loss": 85.78196716308594, "epoch": 0.61, "learning_rate": 2.0800505963136972e-05, "loss": 86.488, "step": 1701 }, { "compression_loss": 84.44926452636719, "epoch": 0.62, "learning_rate": 2.0795084929526563e-05, "loss": 85.0149, "step": 1702 }, { "compression_loss": 83.64439392089844, "epoch": 0.62, "learning_rate": 2.0789663895916154e-05, "loss": 83.9537, "step": 1703 }, { "compression_loss": 85.70069122314453, "epoch": 0.62, "learning_rate": 2.0784242862305746e-05, "loss": 86.2606, "step": 1704 }, { "compression_loss": 84.30343627929688, "epoch": 0.62, "learning_rate": 2.0778821828695337e-05, "loss": 84.7906, "step": 1705 }, { "compression_loss": 83.77322387695312, "epoch": 0.62, "learning_rate": 2.077340079508493e-05, "loss": 84.2239, "step": 1706 }, { "compression_loss": 83.7769546508789, "epoch": 0.62, "learning_rate": 2.0767979761474523e-05, "loss": 84.5482, "step": 1707 }, { "compression_loss": 85.1205825805664, "epoch": 0.62, "learning_rate": 2.0762558727864114e-05, "loss": 85.5827, "step": 1708 }, { "compression_loss": 84.6402359008789, "epoch": 0.62, "learning_rate": 2.0757137694253705e-05, "loss": 85.3668, "step": 1709 }, { "compression_loss": 83.18498992919922, "epoch": 0.62, "learning_rate": 2.0751716660643297e-05, "loss": 83.5432, "step": 1710 }, { "compression_loss": 82.88655090332031, "epoch": 0.62, "learning_rate": 2.0746295627032888e-05, "loss": 83.4051, "step": 1711 }, { "compression_loss": 85.381591796875, "epoch": 0.62, "learning_rate": 2.074087459342248e-05, "loss": 85.8479, "step": 1712 }, { "compression_loss": 83.21268463134766, "epoch": 0.62, "learning_rate": 2.073545355981207e-05, "loss": 83.8568, "step": 1713 }, { "compression_loss": 84.28121948242188, "epoch": 0.62, "learning_rate": 2.0730032526201665e-05, "loss": 84.9571, "step": 1714 }, { "compression_loss": 86.08332824707031, "epoch": 0.62, "learning_rate": 2.0724611492591256e-05, "loss": 86.6254, "step": 1715 }, { "compression_loss": 84.53546142578125, "epoch": 0.62, "learning_rate": 2.0719190458980848e-05, "loss": 84.9565, "step": 1716 }, { "compression_loss": 84.90792846679688, "epoch": 0.62, "learning_rate": 2.071376942537044e-05, "loss": 85.5737, "step": 1717 }, { "compression_loss": 85.4339828491211, "epoch": 0.62, "learning_rate": 2.070834839176003e-05, "loss": 86.2086, "step": 1718 }, { "compression_loss": 83.4041519165039, "epoch": 0.62, "learning_rate": 2.070292735814962e-05, "loss": 83.7785, "step": 1719 }, { "compression_loss": 83.94746398925781, "epoch": 0.62, "learning_rate": 2.0697506324539213e-05, "loss": 84.4747, "step": 1720 }, { "compression_loss": 84.91436004638672, "epoch": 0.62, "learning_rate": 2.0692085290928804e-05, "loss": 85.4377, "step": 1721 }, { "compression_loss": 85.03204345703125, "epoch": 0.62, "learning_rate": 2.06866642573184e-05, "loss": 85.7184, "step": 1722 }, { "compression_loss": 83.75485229492188, "epoch": 0.62, "learning_rate": 2.068124322370799e-05, "loss": 84.1508, "step": 1723 }, { "compression_loss": 82.97654724121094, "epoch": 0.62, "learning_rate": 2.067582219009758e-05, "loss": 83.3529, "step": 1724 }, { "compression_loss": 84.08763122558594, "epoch": 0.62, "learning_rate": 2.0670401156487173e-05, "loss": 84.6883, "step": 1725 }, { "compression_loss": 84.34471130371094, "epoch": 0.62, "learning_rate": 2.066498012287676e-05, "loss": 84.6561, "step": 1726 }, { "compression_loss": 84.07438659667969, "epoch": 0.62, "learning_rate": 2.0659559089266352e-05, "loss": 84.428, "step": 1727 }, { "compression_loss": 85.03861999511719, "epoch": 0.62, "learning_rate": 2.0654138055655943e-05, "loss": 85.4086, "step": 1728 }, { "compression_loss": 85.00863647460938, "epoch": 0.62, "learning_rate": 2.0648717022045534e-05, "loss": 85.3704, "step": 1729 }, { "compression_loss": 84.62789916992188, "epoch": 0.63, "learning_rate": 2.064329598843513e-05, "loss": 85.3471, "step": 1730 }, { "compression_loss": 84.83045959472656, "epoch": 0.63, "learning_rate": 2.063787495482472e-05, "loss": 85.2701, "step": 1731 }, { "compression_loss": 84.17194366455078, "epoch": 0.63, "learning_rate": 2.063245392121431e-05, "loss": 84.8861, "step": 1732 }, { "compression_loss": 84.08442687988281, "epoch": 0.63, "learning_rate": 2.0627032887603903e-05, "loss": 84.6964, "step": 1733 }, { "compression_loss": 84.17875671386719, "epoch": 0.63, "learning_rate": 2.0621611853993494e-05, "loss": 84.5543, "step": 1734 }, { "compression_loss": 83.82427978515625, "epoch": 0.63, "learning_rate": 2.0616190820383085e-05, "loss": 84.6311, "step": 1735 }, { "compression_loss": 85.41879272460938, "epoch": 0.63, "learning_rate": 2.0610769786772677e-05, "loss": 85.9595, "step": 1736 }, { "compression_loss": 83.99642944335938, "epoch": 0.63, "learning_rate": 2.0605348753162268e-05, "loss": 84.3505, "step": 1737 }, { "compression_loss": 82.59516143798828, "epoch": 0.63, "learning_rate": 2.0599927719551862e-05, "loss": 82.9884, "step": 1738 }, { "compression_loss": 84.28915405273438, "epoch": 0.63, "learning_rate": 2.0594506685941454e-05, "loss": 84.8718, "step": 1739 }, { "compression_loss": 85.78889465332031, "epoch": 0.63, "learning_rate": 2.0589085652331045e-05, "loss": 86.5511, "step": 1740 }, { "compression_loss": 86.79498291015625, "epoch": 0.63, "learning_rate": 2.0583664618720636e-05, "loss": 87.3536, "step": 1741 }, { "compression_loss": 84.69413757324219, "epoch": 0.63, "learning_rate": 2.0578243585110228e-05, "loss": 85.1359, "step": 1742 }, { "compression_loss": 83.32907104492188, "epoch": 0.63, "learning_rate": 2.057282255149982e-05, "loss": 83.8108, "step": 1743 }, { "compression_loss": 84.02296447753906, "epoch": 0.63, "learning_rate": 2.056740151788941e-05, "loss": 84.5077, "step": 1744 }, { "compression_loss": 84.3870849609375, "epoch": 0.63, "learning_rate": 2.0561980484279e-05, "loss": 84.7445, "step": 1745 }, { "compression_loss": 83.37236022949219, "epoch": 0.63, "learning_rate": 2.0556559450668596e-05, "loss": 83.6677, "step": 1746 }, { "compression_loss": 85.97270965576172, "epoch": 0.63, "learning_rate": 2.0551138417058187e-05, "loss": 86.5903, "step": 1747 }, { "compression_loss": 85.7466812133789, "epoch": 0.63, "learning_rate": 2.054571738344778e-05, "loss": 86.6629, "step": 1748 }, { "compression_loss": 85.53579711914062, "epoch": 0.63, "learning_rate": 2.054029634983737e-05, "loss": 86.0262, "step": 1749 }, { "compression_loss": 84.10038757324219, "epoch": 0.63, "learning_rate": 2.053487531622696e-05, "loss": 84.7715, "step": 1750 }, { "epoch": 0.63, "eval_exact_match": 85.76158940397352, "eval_f1": 92.3576407280219, "step": 1750 }, { "compression_loss": 84.12628936767578, "epoch": 0.63, "learning_rate": 2.0529454282616552e-05, "loss": 84.813, "step": 1751 }, { "compression_loss": 85.60906219482422, "epoch": 0.63, "learning_rate": 2.0524033249006144e-05, "loss": 86.3327, "step": 1752 }, { "compression_loss": 83.09773254394531, "epoch": 0.63, "learning_rate": 2.0518612215395735e-05, "loss": 83.6307, "step": 1753 }, { "compression_loss": 84.81788635253906, "epoch": 0.63, "learning_rate": 2.051319118178533e-05, "loss": 85.6208, "step": 1754 }, { "compression_loss": 87.19355010986328, "epoch": 0.63, "learning_rate": 2.050777014817492e-05, "loss": 88.1528, "step": 1755 }, { "compression_loss": 84.06321716308594, "epoch": 0.63, "learning_rate": 2.0502349114564512e-05, "loss": 84.3882, "step": 1756 }, { "compression_loss": 84.8026123046875, "epoch": 0.63, "learning_rate": 2.0496928080954103e-05, "loss": 85.4821, "step": 1757 }, { "compression_loss": 83.92207336425781, "epoch": 0.64, "learning_rate": 2.0491507047343695e-05, "loss": 84.5824, "step": 1758 }, { "compression_loss": 82.97296142578125, "epoch": 0.64, "learning_rate": 2.0486086013733286e-05, "loss": 83.212, "step": 1759 }, { "compression_loss": 84.7027587890625, "epoch": 0.64, "learning_rate": 2.0480664980122877e-05, "loss": 85.4804, "step": 1760 }, { "compression_loss": 85.1297378540039, "epoch": 0.64, "learning_rate": 2.047524394651247e-05, "loss": 85.4072, "step": 1761 }, { "compression_loss": 85.01946258544922, "epoch": 0.64, "learning_rate": 2.0469822912902063e-05, "loss": 85.4081, "step": 1762 }, { "compression_loss": 87.04853820800781, "epoch": 0.64, "learning_rate": 2.0464401879291654e-05, "loss": 87.3944, "step": 1763 }, { "compression_loss": 84.71018981933594, "epoch": 0.64, "learning_rate": 2.0458980845681246e-05, "loss": 85.0879, "step": 1764 }, { "compression_loss": 85.46414184570312, "epoch": 0.64, "learning_rate": 2.0453559812070837e-05, "loss": 85.9136, "step": 1765 }, { "compression_loss": 84.41419982910156, "epoch": 0.64, "learning_rate": 2.0448138778460428e-05, "loss": 84.7555, "step": 1766 }, { "compression_loss": 82.9613265991211, "epoch": 0.64, "learning_rate": 2.044271774485002e-05, "loss": 83.3352, "step": 1767 }, { "compression_loss": 85.87503051757812, "epoch": 0.64, "learning_rate": 2.043729671123961e-05, "loss": 86.6452, "step": 1768 }, { "compression_loss": 83.88626861572266, "epoch": 0.64, "learning_rate": 2.0431875677629202e-05, "loss": 84.6432, "step": 1769 }, { "compression_loss": 84.13851165771484, "epoch": 0.64, "learning_rate": 2.0426454644018793e-05, "loss": 84.8536, "step": 1770 }, { "compression_loss": 83.68356323242188, "epoch": 0.64, "learning_rate": 2.0421033610408385e-05, "loss": 84.1563, "step": 1771 }, { "compression_loss": 84.60453796386719, "epoch": 0.64, "learning_rate": 2.0415612576797976e-05, "loss": 85.0469, "step": 1772 }, { "compression_loss": 84.19390106201172, "epoch": 0.64, "learning_rate": 2.0410191543187567e-05, "loss": 84.6329, "step": 1773 }, { "compression_loss": 84.81317138671875, "epoch": 0.64, "learning_rate": 2.040477050957716e-05, "loss": 85.2953, "step": 1774 }, { "compression_loss": 83.33395385742188, "epoch": 0.64, "learning_rate": 2.039934947596675e-05, "loss": 83.8863, "step": 1775 }, { "compression_loss": 83.58299255371094, "epoch": 0.64, "learning_rate": 2.039392844235634e-05, "loss": 84.3248, "step": 1776 }, { "compression_loss": 83.36347198486328, "epoch": 0.64, "learning_rate": 2.0388507408745932e-05, "loss": 83.8222, "step": 1777 }, { "compression_loss": 82.97241973876953, "epoch": 0.64, "learning_rate": 2.0383086375135527e-05, "loss": 83.3544, "step": 1778 }, { "compression_loss": 83.61502075195312, "epoch": 0.64, "learning_rate": 2.0377665341525118e-05, "loss": 84.2983, "step": 1779 }, { "compression_loss": 84.15110778808594, "epoch": 0.64, "learning_rate": 2.037224430791471e-05, "loss": 84.5033, "step": 1780 }, { "compression_loss": 85.10730743408203, "epoch": 0.64, "learning_rate": 2.03668232743043e-05, "loss": 85.6629, "step": 1781 }, { "compression_loss": 83.68052673339844, "epoch": 0.64, "learning_rate": 2.0361402240693892e-05, "loss": 84.4198, "step": 1782 }, { "compression_loss": 84.91154479980469, "epoch": 0.64, "learning_rate": 2.0355981207083483e-05, "loss": 85.5139, "step": 1783 }, { "compression_loss": 84.36613464355469, "epoch": 0.64, "learning_rate": 2.0350560173473075e-05, "loss": 85.0553, "step": 1784 }, { "compression_loss": 85.07794952392578, "epoch": 0.65, "learning_rate": 2.0345139139862666e-05, "loss": 85.56, "step": 1785 }, { "compression_loss": 84.20357513427734, "epoch": 0.65, "learning_rate": 2.033971810625226e-05, "loss": 84.9319, "step": 1786 }, { "compression_loss": 85.64927673339844, "epoch": 0.65, "learning_rate": 2.0334297072641852e-05, "loss": 86.3663, "step": 1787 }, { "compression_loss": 84.57513427734375, "epoch": 0.65, "learning_rate": 2.0328876039031443e-05, "loss": 85.2476, "step": 1788 }, { "compression_loss": 86.44544982910156, "epoch": 0.65, "learning_rate": 2.0323455005421034e-05, "loss": 87.1733, "step": 1789 }, { "compression_loss": 84.18739318847656, "epoch": 0.65, "learning_rate": 2.0318033971810626e-05, "loss": 84.7469, "step": 1790 }, { "compression_loss": 83.23033905029297, "epoch": 0.65, "learning_rate": 2.0312612938200217e-05, "loss": 83.5717, "step": 1791 }, { "compression_loss": 84.82223510742188, "epoch": 0.65, "learning_rate": 2.0307191904589808e-05, "loss": 85.7183, "step": 1792 }, { "compression_loss": 83.36814880371094, "epoch": 0.65, "learning_rate": 2.03017708709794e-05, "loss": 83.8574, "step": 1793 }, { "compression_loss": 83.93284606933594, "epoch": 0.65, "learning_rate": 2.0296349837368994e-05, "loss": 84.4597, "step": 1794 }, { "compression_loss": 83.64675903320312, "epoch": 0.65, "learning_rate": 2.0290928803758585e-05, "loss": 84.4608, "step": 1795 }, { "compression_loss": 84.40106201171875, "epoch": 0.65, "learning_rate": 2.0285507770148177e-05, "loss": 85.2309, "step": 1796 }, { "compression_loss": 84.7864990234375, "epoch": 0.65, "learning_rate": 2.0280086736537768e-05, "loss": 85.4174, "step": 1797 }, { "compression_loss": 84.06256103515625, "epoch": 0.65, "learning_rate": 2.027466570292736e-05, "loss": 84.7011, "step": 1798 }, { "compression_loss": 84.598388671875, "epoch": 0.65, "learning_rate": 2.026924466931695e-05, "loss": 85.2873, "step": 1799 }, { "compression_loss": 85.62577819824219, "epoch": 0.65, "learning_rate": 2.026382363570654e-05, "loss": 86.1556, "step": 1800 }, { "compression_loss": 85.2457046508789, "epoch": 0.65, "learning_rate": 2.0258402602096133e-05, "loss": 86.0947, "step": 1801 }, { "compression_loss": 83.83161926269531, "epoch": 0.65, "learning_rate": 2.0252981568485728e-05, "loss": 84.2909, "step": 1802 }, { "compression_loss": 85.2077865600586, "epoch": 0.65, "learning_rate": 2.024756053487532e-05, "loss": 85.626, "step": 1803 }, { "compression_loss": 84.45893859863281, "epoch": 0.65, "learning_rate": 2.024213950126491e-05, "loss": 85.1102, "step": 1804 }, { "compression_loss": 83.63899993896484, "epoch": 0.65, "learning_rate": 2.02367184676545e-05, "loss": 84.1994, "step": 1805 }, { "compression_loss": 85.17565155029297, "epoch": 0.65, "learning_rate": 2.0231297434044093e-05, "loss": 86.1181, "step": 1806 }, { "compression_loss": 84.28865814208984, "epoch": 0.65, "learning_rate": 2.0225876400433684e-05, "loss": 84.8784, "step": 1807 }, { "compression_loss": 86.51679992675781, "epoch": 0.65, "learning_rate": 2.0220455366823275e-05, "loss": 87.1344, "step": 1808 }, { "compression_loss": 85.17156982421875, "epoch": 0.65, "learning_rate": 2.0215034333212867e-05, "loss": 85.6403, "step": 1809 }, { "compression_loss": 86.6005859375, "epoch": 0.65, "learning_rate": 2.020961329960246e-05, "loss": 87.3569, "step": 1810 }, { "compression_loss": 83.8669662475586, "epoch": 0.65, "learning_rate": 2.0204192265992052e-05, "loss": 84.7076, "step": 1811 }, { "compression_loss": 83.9124755859375, "epoch": 0.65, "learning_rate": 2.0198771232381644e-05, "loss": 84.3804, "step": 1812 }, { "compression_loss": 85.1745834350586, "epoch": 0.66, "learning_rate": 2.019335019877123e-05, "loss": 85.7507, "step": 1813 }, { "compression_loss": 85.13076782226562, "epoch": 0.66, "learning_rate": 2.0187929165160823e-05, "loss": 85.7507, "step": 1814 }, { "compression_loss": 84.28453063964844, "epoch": 0.66, "learning_rate": 2.0182508131550414e-05, "loss": 84.6959, "step": 1815 }, { "compression_loss": 85.37049865722656, "epoch": 0.66, "learning_rate": 2.0177087097940005e-05, "loss": 85.8539, "step": 1816 }, { "compression_loss": 85.32650756835938, "epoch": 0.66, "learning_rate": 2.0171666064329597e-05, "loss": 85.9267, "step": 1817 }, { "compression_loss": 83.14389038085938, "epoch": 0.66, "learning_rate": 2.016624503071919e-05, "loss": 83.6436, "step": 1818 }, { "compression_loss": 86.73710632324219, "epoch": 0.66, "learning_rate": 2.0160823997108783e-05, "loss": 87.7086, "step": 1819 }, { "compression_loss": 85.70358276367188, "epoch": 0.66, "learning_rate": 2.0155402963498374e-05, "loss": 86.7652, "step": 1820 }, { "compression_loss": 82.84999084472656, "epoch": 0.66, "learning_rate": 2.0149981929887965e-05, "loss": 83.3401, "step": 1821 }, { "compression_loss": 84.7356948852539, "epoch": 0.66, "learning_rate": 2.0144560896277556e-05, "loss": 85.1867, "step": 1822 }, { "compression_loss": 83.96473693847656, "epoch": 0.66, "learning_rate": 2.0139139862667148e-05, "loss": 84.5022, "step": 1823 }, { "compression_loss": 85.7467041015625, "epoch": 0.66, "learning_rate": 2.013371882905674e-05, "loss": 86.4493, "step": 1824 }, { "compression_loss": 84.30570983886719, "epoch": 0.66, "learning_rate": 2.012829779544633e-05, "loss": 84.9267, "step": 1825 }, { "compression_loss": 85.07154846191406, "epoch": 0.66, "learning_rate": 2.0122876761835925e-05, "loss": 85.4554, "step": 1826 }, { "compression_loss": 83.26900482177734, "epoch": 0.66, "learning_rate": 2.0117455728225516e-05, "loss": 83.926, "step": 1827 }, { "compression_loss": 85.00462341308594, "epoch": 0.66, "learning_rate": 2.0112034694615107e-05, "loss": 85.7499, "step": 1828 }, { "compression_loss": 84.97056579589844, "epoch": 0.66, "learning_rate": 2.01066136610047e-05, "loss": 85.4868, "step": 1829 }, { "compression_loss": 86.24160766601562, "epoch": 0.66, "learning_rate": 2.010119262739429e-05, "loss": 86.9447, "step": 1830 }, { "compression_loss": 85.22576141357422, "epoch": 0.66, "learning_rate": 2.009577159378388e-05, "loss": 86.2852, "step": 1831 }, { "compression_loss": 85.20901489257812, "epoch": 0.66, "learning_rate": 2.0090350560173473e-05, "loss": 85.5906, "step": 1832 }, { "compression_loss": 85.37254333496094, "epoch": 0.66, "learning_rate": 2.0084929526563064e-05, "loss": 86.1099, "step": 1833 }, { "compression_loss": 85.79489135742188, "epoch": 0.66, "learning_rate": 2.007950849295266e-05, "loss": 86.3197, "step": 1834 }, { "compression_loss": 83.11629486083984, "epoch": 0.66, "learning_rate": 2.007408745934225e-05, "loss": 83.4387, "step": 1835 }, { "compression_loss": 84.52052307128906, "epoch": 0.66, "learning_rate": 2.006866642573184e-05, "loss": 85.078, "step": 1836 }, { "compression_loss": 83.7518310546875, "epoch": 0.66, "learning_rate": 2.0063245392121432e-05, "loss": 84.1091, "step": 1837 }, { "compression_loss": 84.32771301269531, "epoch": 0.66, "learning_rate": 2.0057824358511024e-05, "loss": 84.6353, "step": 1838 }, { "compression_loss": 85.44227600097656, "epoch": 0.66, "learning_rate": 2.0052403324900615e-05, "loss": 85.8713, "step": 1839 }, { "compression_loss": 84.62870788574219, "epoch": 0.66, "learning_rate": 2.0046982291290206e-05, "loss": 84.9883, "step": 1840 }, { "compression_loss": 85.72312927246094, "epoch": 0.67, "learning_rate": 2.0041561257679797e-05, "loss": 86.1742, "step": 1841 }, { "compression_loss": 83.77287292480469, "epoch": 0.67, "learning_rate": 2.0036140224069392e-05, "loss": 84.204, "step": 1842 }, { "compression_loss": 84.67402648925781, "epoch": 0.67, "learning_rate": 2.0030719190458983e-05, "loss": 85.428, "step": 1843 }, { "compression_loss": 83.91264343261719, "epoch": 0.67, "learning_rate": 2.0025298156848575e-05, "loss": 84.3962, "step": 1844 }, { "compression_loss": 84.19786071777344, "epoch": 0.67, "learning_rate": 2.0019877123238166e-05, "loss": 84.7801, "step": 1845 }, { "compression_loss": 85.00827026367188, "epoch": 0.67, "learning_rate": 2.0014456089627757e-05, "loss": 85.5126, "step": 1846 }, { "compression_loss": 85.16276550292969, "epoch": 0.67, "learning_rate": 2.000903505601735e-05, "loss": 85.5307, "step": 1847 }, { "compression_loss": 82.33145141601562, "epoch": 0.67, "learning_rate": 2.000361402240694e-05, "loss": 82.8901, "step": 1848 }, { "compression_loss": 83.2293701171875, "epoch": 0.67, "learning_rate": 1.999819298879653e-05, "loss": 83.706, "step": 1849 }, { "compression_loss": 84.96719360351562, "epoch": 0.67, "learning_rate": 1.9992771955186122e-05, "loss": 85.6332, "step": 1850 }, { "compression_loss": 85.24760437011719, "epoch": 0.67, "learning_rate": 1.9987350921575717e-05, "loss": 85.7726, "step": 1851 }, { "compression_loss": 82.74010467529297, "epoch": 0.67, "learning_rate": 1.9981929887965308e-05, "loss": 83.3565, "step": 1852 }, { "compression_loss": 84.22332763671875, "epoch": 0.67, "learning_rate": 1.99765088543549e-05, "loss": 84.6736, "step": 1853 }, { "compression_loss": 84.3412094116211, "epoch": 0.67, "learning_rate": 1.997108782074449e-05, "loss": 84.8709, "step": 1854 }, { "compression_loss": 84.36441802978516, "epoch": 0.67, "learning_rate": 1.9965666787134082e-05, "loss": 84.8827, "step": 1855 }, { "compression_loss": 84.19349670410156, "epoch": 0.67, "learning_rate": 1.9960245753523673e-05, "loss": 84.443, "step": 1856 }, { "compression_loss": 86.06536102294922, "epoch": 0.67, "learning_rate": 1.995482471991326e-05, "loss": 86.9002, "step": 1857 }, { "compression_loss": 83.68692016601562, "epoch": 0.67, "learning_rate": 1.9949403686302852e-05, "loss": 84.0882, "step": 1858 }, { "compression_loss": 83.7032699584961, "epoch": 0.67, "learning_rate": 1.9943982652692447e-05, "loss": 84.423, "step": 1859 }, { "compression_loss": 86.61066436767578, "epoch": 0.67, "learning_rate": 1.993856161908204e-05, "loss": 87.2468, "step": 1860 }, { "compression_loss": 85.43045806884766, "epoch": 0.67, "learning_rate": 1.993314058547163e-05, "loss": 85.889, "step": 1861 }, { "compression_loss": 84.32503509521484, "epoch": 0.67, "learning_rate": 1.992771955186122e-05, "loss": 85.0737, "step": 1862 }, { "compression_loss": 85.39618682861328, "epoch": 0.67, "learning_rate": 1.9922298518250812e-05, "loss": 85.9855, "step": 1863 }, { "compression_loss": 83.82003784179688, "epoch": 0.67, "learning_rate": 1.9916877484640403e-05, "loss": 84.5017, "step": 1864 }, { "compression_loss": 83.35002899169922, "epoch": 0.67, "learning_rate": 1.9911456451029995e-05, "loss": 83.7017, "step": 1865 }, { "compression_loss": 84.08308410644531, "epoch": 0.67, "learning_rate": 1.9906035417419586e-05, "loss": 84.7269, "step": 1866 }, { "compression_loss": 83.62100219726562, "epoch": 0.67, "learning_rate": 1.990061438380918e-05, "loss": 83.9545, "step": 1867 }, { "compression_loss": 84.891845703125, "epoch": 0.68, "learning_rate": 1.9895193350198772e-05, "loss": 85.3748, "step": 1868 }, { "compression_loss": 84.6103515625, "epoch": 0.68, "learning_rate": 1.9889772316588363e-05, "loss": 84.9677, "step": 1869 }, { "compression_loss": 84.35540771484375, "epoch": 0.68, "learning_rate": 1.9884351282977954e-05, "loss": 84.708, "step": 1870 }, { "compression_loss": 84.33538818359375, "epoch": 0.68, "learning_rate": 1.9878930249367546e-05, "loss": 85.1435, "step": 1871 }, { "compression_loss": 83.6181640625, "epoch": 0.68, "learning_rate": 1.9873509215757137e-05, "loss": 84.4371, "step": 1872 }, { "compression_loss": 84.25605773925781, "epoch": 0.68, "learning_rate": 1.9868088182146728e-05, "loss": 84.7344, "step": 1873 }, { "compression_loss": 86.09810638427734, "epoch": 0.68, "learning_rate": 1.986266714853632e-05, "loss": 87.0224, "step": 1874 }, { "compression_loss": 85.3687515258789, "epoch": 0.68, "learning_rate": 1.9857246114925914e-05, "loss": 85.8622, "step": 1875 }, { "compression_loss": 84.180419921875, "epoch": 0.68, "learning_rate": 1.9851825081315505e-05, "loss": 84.5423, "step": 1876 }, { "compression_loss": 86.46019744873047, "epoch": 0.68, "learning_rate": 1.9846404047705097e-05, "loss": 87.1038, "step": 1877 }, { "compression_loss": 84.80641174316406, "epoch": 0.68, "learning_rate": 1.9840983014094688e-05, "loss": 85.7802, "step": 1878 }, { "compression_loss": 84.66524505615234, "epoch": 0.68, "learning_rate": 1.983556198048428e-05, "loss": 85.0587, "step": 1879 }, { "compression_loss": 82.33110809326172, "epoch": 0.68, "learning_rate": 1.983014094687387e-05, "loss": 82.9289, "step": 1880 }, { "compression_loss": 85.60203552246094, "epoch": 0.68, "learning_rate": 1.9824719913263462e-05, "loss": 86.3627, "step": 1881 }, { "compression_loss": 84.93688201904297, "epoch": 0.68, "learning_rate": 1.9819298879653053e-05, "loss": 85.9693, "step": 1882 }, { "compression_loss": 81.4453353881836, "epoch": 0.68, "learning_rate": 1.9813877846042648e-05, "loss": 81.9117, "step": 1883 }, { "compression_loss": 84.48271179199219, "epoch": 0.68, "learning_rate": 1.980845681243224e-05, "loss": 85.2969, "step": 1884 }, { "compression_loss": 85.96417236328125, "epoch": 0.68, "learning_rate": 1.980303577882183e-05, "loss": 86.6201, "step": 1885 }, { "compression_loss": 84.61930847167969, "epoch": 0.68, "learning_rate": 1.979761474521142e-05, "loss": 85.3228, "step": 1886 }, { "compression_loss": 84.67696380615234, "epoch": 0.68, "learning_rate": 1.9792193711601013e-05, "loss": 85.2795, "step": 1887 }, { "compression_loss": 83.60417175292969, "epoch": 0.68, "learning_rate": 1.9786772677990604e-05, "loss": 84.1787, "step": 1888 }, { "compression_loss": 83.74278259277344, "epoch": 0.68, "learning_rate": 1.9781351644380195e-05, "loss": 84.1823, "step": 1889 }, { "compression_loss": 84.21014404296875, "epoch": 0.68, "learning_rate": 1.9775930610769787e-05, "loss": 84.6713, "step": 1890 }, { "compression_loss": 86.05155944824219, "epoch": 0.68, "learning_rate": 1.977050957715938e-05, "loss": 86.6679, "step": 1891 }, { "compression_loss": 83.66445922851562, "epoch": 0.68, "learning_rate": 1.9765088543548973e-05, "loss": 83.959, "step": 1892 }, { "compression_loss": 83.16856384277344, "epoch": 0.68, "learning_rate": 1.9759667509938564e-05, "loss": 83.6021, "step": 1893 }, { "compression_loss": 83.72027587890625, "epoch": 0.68, "learning_rate": 1.9754246476328155e-05, "loss": 84.2533, "step": 1894 }, { "compression_loss": 83.36604309082031, "epoch": 0.68, "learning_rate": 1.9748825442717746e-05, "loss": 83.9622, "step": 1895 }, { "compression_loss": 85.28438568115234, "epoch": 0.69, "learning_rate": 1.9743404409107338e-05, "loss": 85.5737, "step": 1896 }, { "compression_loss": 85.19398498535156, "epoch": 0.69, "learning_rate": 1.973798337549693e-05, "loss": 85.7587, "step": 1897 }, { "compression_loss": 84.42037963867188, "epoch": 0.69, "learning_rate": 1.973256234188652e-05, "loss": 84.808, "step": 1898 }, { "compression_loss": 85.05895233154297, "epoch": 0.69, "learning_rate": 1.9727141308276115e-05, "loss": 85.4495, "step": 1899 }, { "compression_loss": 85.10646057128906, "epoch": 0.69, "learning_rate": 1.9721720274665703e-05, "loss": 85.5555, "step": 1900 }, { "compression_loss": 83.36518859863281, "epoch": 0.69, "learning_rate": 1.9716299241055294e-05, "loss": 83.7718, "step": 1901 }, { "compression_loss": 85.95356750488281, "epoch": 0.69, "learning_rate": 1.9710878207444885e-05, "loss": 86.5686, "step": 1902 }, { "compression_loss": 83.53612518310547, "epoch": 0.69, "learning_rate": 1.9705457173834477e-05, "loss": 84.1311, "step": 1903 }, { "compression_loss": 84.6773681640625, "epoch": 0.69, "learning_rate": 1.9700036140224068e-05, "loss": 85.2438, "step": 1904 }, { "compression_loss": 83.86109924316406, "epoch": 0.69, "learning_rate": 1.969461510661366e-05, "loss": 84.2888, "step": 1905 }, { "compression_loss": 84.64608764648438, "epoch": 0.69, "learning_rate": 1.968919407300325e-05, "loss": 85.1787, "step": 1906 }, { "compression_loss": 86.07058715820312, "epoch": 0.69, "learning_rate": 1.9683773039392845e-05, "loss": 86.659, "step": 1907 }, { "compression_loss": 83.60892486572266, "epoch": 0.69, "learning_rate": 1.9678352005782436e-05, "loss": 84.1626, "step": 1908 }, { "compression_loss": 83.95305633544922, "epoch": 0.69, "learning_rate": 1.9672930972172028e-05, "loss": 84.5006, "step": 1909 }, { "compression_loss": 87.02344512939453, "epoch": 0.69, "learning_rate": 1.966750993856162e-05, "loss": 87.653, "step": 1910 }, { "compression_loss": 82.35194396972656, "epoch": 0.69, "learning_rate": 1.966208890495121e-05, "loss": 82.965, "step": 1911 }, { "compression_loss": 84.68157958984375, "epoch": 0.69, "learning_rate": 1.96566678713408e-05, "loss": 85.0483, "step": 1912 }, { "compression_loss": 83.67158508300781, "epoch": 0.69, "learning_rate": 1.9651246837730393e-05, "loss": 84.1459, "step": 1913 }, { "compression_loss": 84.90357208251953, "epoch": 0.69, "learning_rate": 1.9645825804119984e-05, "loss": 85.6024, "step": 1914 }, { "compression_loss": 83.77345275878906, "epoch": 0.69, "learning_rate": 1.964040477050958e-05, "loss": 84.7967, "step": 1915 }, { "compression_loss": 85.85171508789062, "epoch": 0.69, "learning_rate": 1.963498373689917e-05, "loss": 86.2656, "step": 1916 }, { "compression_loss": 84.3345718383789, "epoch": 0.69, "learning_rate": 1.962956270328876e-05, "loss": 84.9202, "step": 1917 }, { "compression_loss": 84.24330139160156, "epoch": 0.69, "learning_rate": 1.9624141669678352e-05, "loss": 84.6552, "step": 1918 }, { "compression_loss": 86.73200988769531, "epoch": 0.69, "learning_rate": 1.9618720636067944e-05, "loss": 87.0897, "step": 1919 }, { "compression_loss": 83.81442260742188, "epoch": 0.69, "learning_rate": 1.9613299602457535e-05, "loss": 84.163, "step": 1920 }, { "compression_loss": 84.20565795898438, "epoch": 0.69, "learning_rate": 1.9607878568847126e-05, "loss": 84.8534, "step": 1921 }, { "compression_loss": 83.7820816040039, "epoch": 0.69, "learning_rate": 1.9602457535236717e-05, "loss": 84.6576, "step": 1922 }, { "compression_loss": 84.66929626464844, "epoch": 0.69, "learning_rate": 1.9597036501626312e-05, "loss": 85.171, "step": 1923 }, { "compression_loss": 81.96961212158203, "epoch": 0.7, "learning_rate": 1.9591615468015903e-05, "loss": 82.4528, "step": 1924 }, { "compression_loss": 84.38629150390625, "epoch": 0.7, "learning_rate": 1.9586194434405495e-05, "loss": 84.6881, "step": 1925 }, { "compression_loss": 86.01752471923828, "epoch": 0.7, "learning_rate": 1.9580773400795086e-05, "loss": 86.8268, "step": 1926 }, { "compression_loss": 82.70803833007812, "epoch": 0.7, "learning_rate": 1.9575352367184677e-05, "loss": 83.3135, "step": 1927 }, { "compression_loss": 85.2809829711914, "epoch": 0.7, "learning_rate": 1.956993133357427e-05, "loss": 85.7712, "step": 1928 }, { "compression_loss": 82.7823486328125, "epoch": 0.7, "learning_rate": 1.956451029996386e-05, "loss": 83.3944, "step": 1929 }, { "compression_loss": 84.31654357910156, "epoch": 0.7, "learning_rate": 1.955908926635345e-05, "loss": 84.9173, "step": 1930 }, { "compression_loss": 83.08100891113281, "epoch": 0.7, "learning_rate": 1.9553668232743046e-05, "loss": 83.3769, "step": 1931 }, { "compression_loss": 83.0281982421875, "epoch": 0.7, "learning_rate": 1.9548247199132637e-05, "loss": 83.6072, "step": 1932 }, { "compression_loss": 84.46095275878906, "epoch": 0.7, "learning_rate": 1.9542826165522228e-05, "loss": 84.713, "step": 1933 }, { "compression_loss": 85.0530014038086, "epoch": 0.7, "learning_rate": 1.953740513191182e-05, "loss": 85.9177, "step": 1934 }, { "compression_loss": 84.32693481445312, "epoch": 0.7, "learning_rate": 1.953198409830141e-05, "loss": 85.0872, "step": 1935 }, { "compression_loss": 84.89031982421875, "epoch": 0.7, "learning_rate": 1.9526563064691002e-05, "loss": 85.2917, "step": 1936 }, { "compression_loss": 85.359619140625, "epoch": 0.7, "learning_rate": 1.9521142031080593e-05, "loss": 86.0896, "step": 1937 }, { "compression_loss": 85.20724487304688, "epoch": 0.7, "learning_rate": 1.9515720997470185e-05, "loss": 85.5419, "step": 1938 }, { "compression_loss": 84.35941314697266, "epoch": 0.7, "learning_rate": 1.951029996385978e-05, "loss": 84.7935, "step": 1939 }, { "compression_loss": 84.73667907714844, "epoch": 0.7, "learning_rate": 1.950487893024937e-05, "loss": 85.2354, "step": 1940 }, { "compression_loss": 84.11709594726562, "epoch": 0.7, "learning_rate": 1.9499457896638962e-05, "loss": 84.5477, "step": 1941 }, { "compression_loss": 85.27350616455078, "epoch": 0.7, "learning_rate": 1.9494036863028553e-05, "loss": 86.136, "step": 1942 }, { "compression_loss": 85.54182434082031, "epoch": 0.7, "learning_rate": 1.9488615829418144e-05, "loss": 86.366, "step": 1943 }, { "compression_loss": 85.35440063476562, "epoch": 0.7, "learning_rate": 1.9483194795807732e-05, "loss": 85.857, "step": 1944 }, { "compression_loss": 84.44931030273438, "epoch": 0.7, "learning_rate": 1.9477773762197324e-05, "loss": 85.3148, "step": 1945 }, { "compression_loss": 84.28517150878906, "epoch": 0.7, "learning_rate": 1.9472352728586915e-05, "loss": 85.0207, "step": 1946 }, { "compression_loss": 83.41946411132812, "epoch": 0.7, "learning_rate": 1.946693169497651e-05, "loss": 84.2223, "step": 1947 }, { "compression_loss": 84.77737426757812, "epoch": 0.7, "learning_rate": 1.94615106613661e-05, "loss": 85.2332, "step": 1948 }, { "compression_loss": 83.77409362792969, "epoch": 0.7, "learning_rate": 1.9456089627755692e-05, "loss": 84.3132, "step": 1949 }, { "compression_loss": 82.24516296386719, "epoch": 0.7, "learning_rate": 1.9450668594145283e-05, "loss": 83.242, "step": 1950 }, { "compression_loss": 84.1697769165039, "epoch": 0.71, "learning_rate": 1.9445247560534875e-05, "loss": 84.8327, "step": 1951 }, { "compression_loss": 83.71888732910156, "epoch": 0.71, "learning_rate": 1.9439826526924466e-05, "loss": 84.134, "step": 1952 }, { "compression_loss": 84.5479736328125, "epoch": 0.71, "learning_rate": 1.9434405493314057e-05, "loss": 85.1903, "step": 1953 }, { "compression_loss": 84.83930206298828, "epoch": 0.71, "learning_rate": 1.942898445970365e-05, "loss": 85.4027, "step": 1954 }, { "compression_loss": 84.83145904541016, "epoch": 0.71, "learning_rate": 1.9423563426093243e-05, "loss": 85.5232, "step": 1955 }, { "compression_loss": 85.26309204101562, "epoch": 0.71, "learning_rate": 1.9418142392482834e-05, "loss": 85.9092, "step": 1956 }, { "compression_loss": 86.0819091796875, "epoch": 0.71, "learning_rate": 1.9412721358872426e-05, "loss": 86.776, "step": 1957 }, { "compression_loss": 84.74052429199219, "epoch": 0.71, "learning_rate": 1.9407300325262017e-05, "loss": 85.3733, "step": 1958 }, { "compression_loss": 83.00009155273438, "epoch": 0.71, "learning_rate": 1.9401879291651608e-05, "loss": 83.6981, "step": 1959 }, { "compression_loss": 85.76395416259766, "epoch": 0.71, "learning_rate": 1.93964582580412e-05, "loss": 86.2114, "step": 1960 }, { "compression_loss": 85.19013214111328, "epoch": 0.71, "learning_rate": 1.939103722443079e-05, "loss": 85.6369, "step": 1961 }, { "compression_loss": 84.52638244628906, "epoch": 0.71, "learning_rate": 1.9385616190820382e-05, "loss": 84.8048, "step": 1962 }, { "compression_loss": 84.73199462890625, "epoch": 0.71, "learning_rate": 1.9380195157209977e-05, "loss": 85.7797, "step": 1963 }, { "compression_loss": 85.62911224365234, "epoch": 0.71, "learning_rate": 1.9374774123599568e-05, "loss": 86.2604, "step": 1964 }, { "compression_loss": 84.35924530029297, "epoch": 0.71, "learning_rate": 1.936935308998916e-05, "loss": 84.8148, "step": 1965 }, { "compression_loss": 85.51866149902344, "epoch": 0.71, "learning_rate": 1.936393205637875e-05, "loss": 86.2815, "step": 1966 }, { "compression_loss": 83.22669219970703, "epoch": 0.71, "learning_rate": 1.935851102276834e-05, "loss": 83.8374, "step": 1967 }, { "compression_loss": 85.23262023925781, "epoch": 0.71, "learning_rate": 1.9353089989157933e-05, "loss": 85.7461, "step": 1968 }, { "compression_loss": 86.33535766601562, "epoch": 0.71, "learning_rate": 1.9347668955547524e-05, "loss": 87.0125, "step": 1969 }, { "compression_loss": 84.57262420654297, "epoch": 0.71, "learning_rate": 1.9342247921937115e-05, "loss": 85.1497, "step": 1970 }, { "compression_loss": 86.44084167480469, "epoch": 0.71, "learning_rate": 1.933682688832671e-05, "loss": 86.9658, "step": 1971 }, { "compression_loss": 83.92367553710938, "epoch": 0.71, "learning_rate": 1.93314058547163e-05, "loss": 84.4037, "step": 1972 }, { "compression_loss": 84.20463562011719, "epoch": 0.71, "learning_rate": 1.9325984821105893e-05, "loss": 84.7248, "step": 1973 }, { "compression_loss": 84.81463623046875, "epoch": 0.71, "learning_rate": 1.9320563787495484e-05, "loss": 85.169, "step": 1974 }, { "compression_loss": 86.20549011230469, "epoch": 0.71, "learning_rate": 1.9315142753885075e-05, "loss": 87.1757, "step": 1975 }, { "compression_loss": 84.42357635498047, "epoch": 0.71, "learning_rate": 1.9309721720274667e-05, "loss": 84.9105, "step": 1976 }, { "compression_loss": 83.65306091308594, "epoch": 0.71, "learning_rate": 1.9304300686664258e-05, "loss": 84.2062, "step": 1977 }, { "compression_loss": 83.39171600341797, "epoch": 0.71, "learning_rate": 1.929887965305385e-05, "loss": 83.7081, "step": 1978 }, { "compression_loss": 84.54782104492188, "epoch": 0.72, "learning_rate": 1.9293458619443444e-05, "loss": 84.9543, "step": 1979 }, { "compression_loss": 82.86323547363281, "epoch": 0.72, "learning_rate": 1.9288037585833035e-05, "loss": 83.375, "step": 1980 }, { "compression_loss": 84.56954956054688, "epoch": 0.72, "learning_rate": 1.9282616552222626e-05, "loss": 85.1157, "step": 1981 }, { "compression_loss": 84.91114044189453, "epoch": 0.72, "learning_rate": 1.9277195518612218e-05, "loss": 85.2846, "step": 1982 }, { "compression_loss": 85.23185729980469, "epoch": 0.72, "learning_rate": 1.927177448500181e-05, "loss": 86.0435, "step": 1983 }, { "compression_loss": 83.90826416015625, "epoch": 0.72, "learning_rate": 1.92663534513914e-05, "loss": 84.4934, "step": 1984 }, { "compression_loss": 83.85952758789062, "epoch": 0.72, "learning_rate": 1.926093241778099e-05, "loss": 84.4988, "step": 1985 }, { "compression_loss": 84.55894470214844, "epoch": 0.72, "learning_rate": 1.9255511384170583e-05, "loss": 85.2723, "step": 1986 }, { "compression_loss": 83.882568359375, "epoch": 0.72, "learning_rate": 1.9250090350560174e-05, "loss": 84.5322, "step": 1987 }, { "compression_loss": 85.54021453857422, "epoch": 0.72, "learning_rate": 1.9244669316949765e-05, "loss": 86.1717, "step": 1988 }, { "compression_loss": 86.07132720947266, "epoch": 0.72, "learning_rate": 1.9239248283339356e-05, "loss": 86.6721, "step": 1989 }, { "compression_loss": 83.57331848144531, "epoch": 0.72, "learning_rate": 1.9233827249728948e-05, "loss": 84.0395, "step": 1990 }, { "compression_loss": 83.59434509277344, "epoch": 0.72, "learning_rate": 1.922840621611854e-05, "loss": 84.1792, "step": 1991 }, { "compression_loss": 84.1577377319336, "epoch": 0.72, "learning_rate": 1.922298518250813e-05, "loss": 84.8801, "step": 1992 }, { "compression_loss": 84.44110107421875, "epoch": 0.72, "learning_rate": 1.921756414889772e-05, "loss": 85.1687, "step": 1993 }, { "compression_loss": 83.51229095458984, "epoch": 0.72, "learning_rate": 1.9212143115287313e-05, "loss": 84.1944, "step": 1994 }, { "compression_loss": 85.2889633178711, "epoch": 0.72, "learning_rate": 1.9206722081676907e-05, "loss": 85.8139, "step": 1995 }, { "compression_loss": 86.87852478027344, "epoch": 0.72, "learning_rate": 1.92013010480665e-05, "loss": 87.2642, "step": 1996 }, { "compression_loss": 84.26960754394531, "epoch": 0.72, "learning_rate": 1.919588001445609e-05, "loss": 84.9249, "step": 1997 }, { "compression_loss": 84.45311737060547, "epoch": 0.72, "learning_rate": 1.919045898084568e-05, "loss": 85.0677, "step": 1998 }, { "compression_loss": 83.78138732910156, "epoch": 0.72, "learning_rate": 1.9185037947235273e-05, "loss": 84.3277, "step": 1999 }, { "compression_loss": 84.79625701904297, "epoch": 0.72, "learning_rate": 1.9179616913624864e-05, "loss": 85.0947, "step": 2000 }, { "epoch": 0.72, "eval_exact_match": 86.2251655629139, "eval_f1": 92.61769807111047, "step": 2000 }, { "compression_loss": 84.16133117675781, "epoch": 0.72, "learning_rate": 1.9174195880014455e-05, "loss": 84.7837, "step": 2001 }, { "compression_loss": 84.31310272216797, "epoch": 0.72, "learning_rate": 1.9168774846404046e-05, "loss": 84.9415, "step": 2002 }, { "compression_loss": 83.994384765625, "epoch": 0.72, "learning_rate": 1.916335381279364e-05, "loss": 84.4979, "step": 2003 }, { "compression_loss": 82.34806823730469, "epoch": 0.72, "learning_rate": 1.9157932779183232e-05, "loss": 82.891, "step": 2004 }, { "compression_loss": 84.14397430419922, "epoch": 0.72, "learning_rate": 1.9152511745572824e-05, "loss": 84.7095, "step": 2005 }, { "compression_loss": 84.67050170898438, "epoch": 0.72, "learning_rate": 1.9147090711962415e-05, "loss": 85.2611, "step": 2006 }, { "compression_loss": 82.79341888427734, "epoch": 0.73, "learning_rate": 1.9141669678352006e-05, "loss": 83.2405, "step": 2007 }, { "compression_loss": 82.759765625, "epoch": 0.73, "learning_rate": 1.9136248644741597e-05, "loss": 83.2239, "step": 2008 }, { "compression_loss": 84.59903717041016, "epoch": 0.73, "learning_rate": 1.913082761113119e-05, "loss": 85.4714, "step": 2009 }, { "compression_loss": 84.00729370117188, "epoch": 0.73, "learning_rate": 1.912540657752078e-05, "loss": 84.6991, "step": 2010 }, { "compression_loss": 85.71609497070312, "epoch": 0.73, "learning_rate": 1.9119985543910375e-05, "loss": 86.4896, "step": 2011 }, { "compression_loss": 84.20545959472656, "epoch": 0.73, "learning_rate": 1.9114564510299966e-05, "loss": 84.808, "step": 2012 }, { "compression_loss": 85.580322265625, "epoch": 0.73, "learning_rate": 1.9109143476689557e-05, "loss": 86.0533, "step": 2013 }, { "compression_loss": 85.33883666992188, "epoch": 0.73, "learning_rate": 1.910372244307915e-05, "loss": 85.8452, "step": 2014 }, { "compression_loss": 83.93828582763672, "epoch": 0.73, "learning_rate": 1.909830140946874e-05, "loss": 84.4915, "step": 2015 }, { "compression_loss": 85.01702117919922, "epoch": 0.73, "learning_rate": 1.909288037585833e-05, "loss": 85.5022, "step": 2016 }, { "compression_loss": 87.53817749023438, "epoch": 0.73, "learning_rate": 1.9087459342247922e-05, "loss": 88.3178, "step": 2017 }, { "compression_loss": 84.35372924804688, "epoch": 0.73, "learning_rate": 1.9082038308637513e-05, "loss": 84.953, "step": 2018 }, { "compression_loss": 85.15936279296875, "epoch": 0.73, "learning_rate": 1.9076617275027108e-05, "loss": 85.5037, "step": 2019 }, { "compression_loss": 84.9639892578125, "epoch": 0.73, "learning_rate": 1.90711962414167e-05, "loss": 85.3724, "step": 2020 }, { "compression_loss": 84.84734344482422, "epoch": 0.73, "learning_rate": 1.906577520780629e-05, "loss": 85.3612, "step": 2021 }, { "compression_loss": 84.331787109375, "epoch": 0.73, "learning_rate": 1.9060354174195882e-05, "loss": 84.919, "step": 2022 }, { "compression_loss": 86.18891906738281, "epoch": 0.73, "learning_rate": 1.9054933140585473e-05, "loss": 86.8037, "step": 2023 }, { "compression_loss": 83.46307373046875, "epoch": 0.73, "learning_rate": 1.9049512106975064e-05, "loss": 84.13, "step": 2024 }, { "compression_loss": 85.02799224853516, "epoch": 0.73, "learning_rate": 1.9044091073364656e-05, "loss": 85.6852, "step": 2025 }, { "compression_loss": 84.13005828857422, "epoch": 0.73, "learning_rate": 1.9038670039754247e-05, "loss": 85.2455, "step": 2026 }, { "compression_loss": 84.98754119873047, "epoch": 0.73, "learning_rate": 1.9033249006143842e-05, "loss": 85.3337, "step": 2027 }, { "compression_loss": 85.71652221679688, "epoch": 0.73, "learning_rate": 1.9027827972533433e-05, "loss": 86.4781, "step": 2028 }, { "compression_loss": 84.58113861083984, "epoch": 0.73, "learning_rate": 1.9022406938923024e-05, "loss": 85.0207, "step": 2029 }, { "compression_loss": 86.27953338623047, "epoch": 0.73, "learning_rate": 1.9016985905312616e-05, "loss": 87.1625, "step": 2030 }, { "compression_loss": 84.92343139648438, "epoch": 0.73, "learning_rate": 1.9011564871702203e-05, "loss": 85.7133, "step": 2031 }, { "compression_loss": 85.25473022460938, "epoch": 0.73, "learning_rate": 1.9006143838091795e-05, "loss": 85.8774, "step": 2032 }, { "compression_loss": 84.51024627685547, "epoch": 0.73, "learning_rate": 1.9000722804481386e-05, "loss": 84.9507, "step": 2033 }, { "compression_loss": 85.10601806640625, "epoch": 0.74, "learning_rate": 1.8995301770870977e-05, "loss": 85.646, "step": 2034 }, { "compression_loss": 83.59193420410156, "epoch": 0.74, "learning_rate": 1.898988073726057e-05, "loss": 84.3332, "step": 2035 }, { "compression_loss": 84.56084442138672, "epoch": 0.74, "learning_rate": 1.8984459703650163e-05, "loss": 85.2838, "step": 2036 }, { "compression_loss": 84.37460327148438, "epoch": 0.74, "learning_rate": 1.8979038670039754e-05, "loss": 84.9421, "step": 2037 }, { "compression_loss": 85.30799865722656, "epoch": 0.74, "learning_rate": 1.8973617636429346e-05, "loss": 86.0235, "step": 2038 }, { "compression_loss": 85.82328796386719, "epoch": 0.74, "learning_rate": 1.8968196602818937e-05, "loss": 86.4112, "step": 2039 }, { "compression_loss": 84.75348663330078, "epoch": 0.74, "learning_rate": 1.8962775569208528e-05, "loss": 85.1889, "step": 2040 }, { "compression_loss": 84.90089416503906, "epoch": 0.74, "learning_rate": 1.895735453559812e-05, "loss": 85.3148, "step": 2041 }, { "compression_loss": 83.77031707763672, "epoch": 0.74, "learning_rate": 1.895193350198771e-05, "loss": 84.5759, "step": 2042 }, { "compression_loss": 82.48464965820312, "epoch": 0.74, "learning_rate": 1.8946512468377302e-05, "loss": 82.9198, "step": 2043 }, { "compression_loss": 83.09901428222656, "epoch": 0.74, "learning_rate": 1.8941091434766897e-05, "loss": 83.6244, "step": 2044 }, { "compression_loss": 85.6348876953125, "epoch": 0.74, "learning_rate": 1.8935670401156488e-05, "loss": 86.1015, "step": 2045 }, { "compression_loss": 84.87652587890625, "epoch": 0.74, "learning_rate": 1.893024936754608e-05, "loss": 85.437, "step": 2046 }, { "compression_loss": 83.30697631835938, "epoch": 0.74, "learning_rate": 1.892482833393567e-05, "loss": 83.9798, "step": 2047 }, { "compression_loss": 85.6115493774414, "epoch": 0.74, "learning_rate": 1.8919407300325262e-05, "loss": 86.4902, "step": 2048 }, { "compression_loss": 84.12393188476562, "epoch": 0.74, "learning_rate": 1.8913986266714853e-05, "loss": 84.5717, "step": 2049 }, { "compression_loss": 84.46772003173828, "epoch": 0.74, "learning_rate": 1.8908565233104444e-05, "loss": 85.2866, "step": 2050 }, { "compression_loss": 83.20391082763672, "epoch": 0.74, "learning_rate": 1.8903144199494036e-05, "loss": 83.5786, "step": 2051 }, { "compression_loss": 84.97639465332031, "epoch": 0.74, "learning_rate": 1.889772316588363e-05, "loss": 85.7463, "step": 2052 }, { "compression_loss": 85.6247329711914, "epoch": 0.74, "learning_rate": 1.889230213227322e-05, "loss": 86.4106, "step": 2053 }, { "compression_loss": 83.79481506347656, "epoch": 0.74, "learning_rate": 1.8886881098662813e-05, "loss": 84.4184, "step": 2054 }, { "compression_loss": 84.98355102539062, "epoch": 0.74, "learning_rate": 1.8881460065052404e-05, "loss": 85.732, "step": 2055 }, { "compression_loss": 84.02346801757812, "epoch": 0.74, "learning_rate": 1.8876039031441995e-05, "loss": 84.4911, "step": 2056 }, { "compression_loss": 86.24073791503906, "epoch": 0.74, "learning_rate": 1.8870617997831587e-05, "loss": 87.0053, "step": 2057 }, { "compression_loss": 85.52064514160156, "epoch": 0.74, "learning_rate": 1.8865196964221178e-05, "loss": 85.9404, "step": 2058 }, { "compression_loss": 83.26744079589844, "epoch": 0.74, "learning_rate": 1.885977593061077e-05, "loss": 83.7543, "step": 2059 }, { "compression_loss": 83.0379638671875, "epoch": 0.74, "learning_rate": 1.8854354897000364e-05, "loss": 83.3869, "step": 2060 }, { "compression_loss": 84.77791595458984, "epoch": 0.74, "learning_rate": 1.8848933863389955e-05, "loss": 85.4591, "step": 2061 }, { "compression_loss": 85.60137939453125, "epoch": 0.75, "learning_rate": 1.8843512829779546e-05, "loss": 86.433, "step": 2062 }, { "compression_loss": 83.95710754394531, "epoch": 0.75, "learning_rate": 1.8838091796169138e-05, "loss": 84.4422, "step": 2063 }, { "compression_loss": 84.73562622070312, "epoch": 0.75, "learning_rate": 1.883267076255873e-05, "loss": 85.4118, "step": 2064 }, { "compression_loss": 84.35697937011719, "epoch": 0.75, "learning_rate": 1.882724972894832e-05, "loss": 84.8245, "step": 2065 }, { "compression_loss": 86.1748275756836, "epoch": 0.75, "learning_rate": 1.882182869533791e-05, "loss": 86.4783, "step": 2066 }, { "compression_loss": 84.96597290039062, "epoch": 0.75, "learning_rate": 1.8816407661727503e-05, "loss": 85.4457, "step": 2067 }, { "compression_loss": 85.09773254394531, "epoch": 0.75, "learning_rate": 1.8810986628117097e-05, "loss": 85.6026, "step": 2068 }, { "compression_loss": 81.40939331054688, "epoch": 0.75, "learning_rate": 1.880556559450669e-05, "loss": 81.8655, "step": 2069 }, { "compression_loss": 84.84972381591797, "epoch": 0.75, "learning_rate": 1.880014456089628e-05, "loss": 85.8804, "step": 2070 }, { "compression_loss": 84.73274230957031, "epoch": 0.75, "learning_rate": 1.879472352728587e-05, "loss": 85.1403, "step": 2071 }, { "compression_loss": 83.70166015625, "epoch": 0.75, "learning_rate": 1.8789302493675462e-05, "loss": 84.3391, "step": 2072 }, { "compression_loss": 87.34829711914062, "epoch": 0.75, "learning_rate": 1.8783881460065054e-05, "loss": 87.9051, "step": 2073 }, { "compression_loss": 85.58076477050781, "epoch": 0.75, "learning_rate": 1.8778460426454645e-05, "loss": 86.4295, "step": 2074 }, { "compression_loss": 84.93400573730469, "epoch": 0.75, "learning_rate": 1.8773039392844233e-05, "loss": 85.6622, "step": 2075 }, { "compression_loss": 83.39118957519531, "epoch": 0.75, "learning_rate": 1.8767618359233828e-05, "loss": 83.9936, "step": 2076 }, { "compression_loss": 85.98261260986328, "epoch": 0.75, "learning_rate": 1.876219732562342e-05, "loss": 86.8143, "step": 2077 }, { "compression_loss": 84.34473419189453, "epoch": 0.75, "learning_rate": 1.875677629201301e-05, "loss": 84.8245, "step": 2078 }, { "compression_loss": 85.57327270507812, "epoch": 0.75, "learning_rate": 1.87513552584026e-05, "loss": 86.1937, "step": 2079 }, { "compression_loss": 84.99339294433594, "epoch": 0.75, "learning_rate": 1.8745934224792193e-05, "loss": 85.7013, "step": 2080 }, { "compression_loss": 83.9273910522461, "epoch": 0.75, "learning_rate": 1.8740513191181784e-05, "loss": 84.4762, "step": 2081 }, { "compression_loss": 82.94035339355469, "epoch": 0.75, "learning_rate": 1.8735092157571375e-05, "loss": 83.4013, "step": 2082 }, { "compression_loss": 84.09803771972656, "epoch": 0.75, "learning_rate": 1.8729671123960966e-05, "loss": 84.5929, "step": 2083 }, { "compression_loss": 83.4703369140625, "epoch": 0.75, "learning_rate": 1.872425009035056e-05, "loss": 84.1076, "step": 2084 }, { "compression_loss": 83.90863037109375, "epoch": 0.75, "learning_rate": 1.8718829056740152e-05, "loss": 84.4429, "step": 2085 }, { "compression_loss": 83.03179931640625, "epoch": 0.75, "learning_rate": 1.8713408023129744e-05, "loss": 83.5928, "step": 2086 }, { "compression_loss": 83.58963012695312, "epoch": 0.75, "learning_rate": 1.8707986989519335e-05, "loss": 84.0617, "step": 2087 }, { "compression_loss": 83.46900939941406, "epoch": 0.75, "learning_rate": 1.8702565955908926e-05, "loss": 84.3906, "step": 2088 }, { "compression_loss": 86.1820068359375, "epoch": 0.75, "learning_rate": 1.8697144922298517e-05, "loss": 86.8772, "step": 2089 }, { "compression_loss": 85.0175552368164, "epoch": 0.76, "learning_rate": 1.869172388868811e-05, "loss": 85.5785, "step": 2090 }, { "compression_loss": 84.90860748291016, "epoch": 0.76, "learning_rate": 1.86863028550777e-05, "loss": 85.3564, "step": 2091 }, { "compression_loss": 83.91588592529297, "epoch": 0.76, "learning_rate": 1.8680881821467295e-05, "loss": 84.6185, "step": 2092 }, { "compression_loss": 83.42811584472656, "epoch": 0.76, "learning_rate": 1.8675460787856886e-05, "loss": 84.076, "step": 2093 }, { "compression_loss": 84.20663452148438, "epoch": 0.76, "learning_rate": 1.8670039754246477e-05, "loss": 84.7831, "step": 2094 }, { "compression_loss": 85.95207214355469, "epoch": 0.76, "learning_rate": 1.866461872063607e-05, "loss": 86.8466, "step": 2095 }, { "compression_loss": 85.07962036132812, "epoch": 0.76, "learning_rate": 1.865919768702566e-05, "loss": 85.7044, "step": 2096 }, { "compression_loss": 83.47418212890625, "epoch": 0.76, "learning_rate": 1.865377665341525e-05, "loss": 83.909, "step": 2097 }, { "compression_loss": 84.47417449951172, "epoch": 0.76, "learning_rate": 1.8648355619804842e-05, "loss": 85.1197, "step": 2098 }, { "compression_loss": 83.89250183105469, "epoch": 0.76, "learning_rate": 1.8642934586194434e-05, "loss": 84.3705, "step": 2099 }, { "compression_loss": 84.57183074951172, "epoch": 0.76, "learning_rate": 1.8637513552584028e-05, "loss": 85.0966, "step": 2100 }, { "compression_loss": 83.16070556640625, "epoch": 0.76, "learning_rate": 1.863209251897362e-05, "loss": 83.4187, "step": 2101 }, { "compression_loss": 85.68775939941406, "epoch": 0.76, "learning_rate": 1.862667148536321e-05, "loss": 86.1046, "step": 2102 }, { "compression_loss": 84.36760711669922, "epoch": 0.76, "learning_rate": 1.8621250451752802e-05, "loss": 85.0585, "step": 2103 }, { "compression_loss": 82.15170288085938, "epoch": 0.76, "learning_rate": 1.8615829418142393e-05, "loss": 82.4388, "step": 2104 }, { "compression_loss": 84.78736877441406, "epoch": 0.76, "learning_rate": 1.8610408384531985e-05, "loss": 85.3911, "step": 2105 }, { "compression_loss": 84.36750793457031, "epoch": 0.76, "learning_rate": 1.8604987350921576e-05, "loss": 84.9329, "step": 2106 }, { "compression_loss": 86.26722717285156, "epoch": 0.76, "learning_rate": 1.8599566317311167e-05, "loss": 87.2954, "step": 2107 }, { "compression_loss": 84.8011474609375, "epoch": 0.76, "learning_rate": 1.8594145283700762e-05, "loss": 85.4768, "step": 2108 }, { "compression_loss": 82.61885070800781, "epoch": 0.76, "learning_rate": 1.8588724250090353e-05, "loss": 83.0457, "step": 2109 }, { "compression_loss": 83.40200805664062, "epoch": 0.76, "learning_rate": 1.8583303216479944e-05, "loss": 83.8443, "step": 2110 }, { "compression_loss": 84.02953338623047, "epoch": 0.76, "learning_rate": 1.8577882182869536e-05, "loss": 84.3756, "step": 2111 }, { "compression_loss": 86.03936004638672, "epoch": 0.76, "learning_rate": 1.8572461149259127e-05, "loss": 86.8048, "step": 2112 }, { "compression_loss": 85.351806640625, "epoch": 0.76, "learning_rate": 1.8567040115648718e-05, "loss": 86.0569, "step": 2113 }, { "compression_loss": 84.12202453613281, "epoch": 0.76, "learning_rate": 1.856161908203831e-05, "loss": 84.5886, "step": 2114 }, { "compression_loss": 83.30049133300781, "epoch": 0.76, "learning_rate": 1.85561980484279e-05, "loss": 83.7911, "step": 2115 }, { "compression_loss": 85.80644226074219, "epoch": 0.76, "learning_rate": 1.8550777014817495e-05, "loss": 86.3079, "step": 2116 }, { "compression_loss": 83.439208984375, "epoch": 0.77, "learning_rate": 1.8545355981207087e-05, "loss": 83.9915, "step": 2117 }, { "compression_loss": 85.73486328125, "epoch": 0.77, "learning_rate": 1.8539934947596675e-05, "loss": 86.5952, "step": 2118 }, { "compression_loss": 85.57719421386719, "epoch": 0.77, "learning_rate": 1.8534513913986266e-05, "loss": 86.3446, "step": 2119 }, { "compression_loss": 84.37187957763672, "epoch": 0.77, "learning_rate": 1.8529092880375857e-05, "loss": 85.4217, "step": 2120 }, { "compression_loss": 84.29840850830078, "epoch": 0.77, "learning_rate": 1.852367184676545e-05, "loss": 84.7929, "step": 2121 }, { "compression_loss": 83.42018127441406, "epoch": 0.77, "learning_rate": 1.851825081315504e-05, "loss": 83.8565, "step": 2122 }, { "compression_loss": 85.58717346191406, "epoch": 0.77, "learning_rate": 1.851282977954463e-05, "loss": 86.1352, "step": 2123 }, { "compression_loss": 84.29293060302734, "epoch": 0.77, "learning_rate": 1.8507408745934226e-05, "loss": 84.7236, "step": 2124 }, { "compression_loss": 84.07042694091797, "epoch": 0.77, "learning_rate": 1.8501987712323817e-05, "loss": 84.5072, "step": 2125 }, { "compression_loss": 84.33631896972656, "epoch": 0.77, "learning_rate": 1.8496566678713408e-05, "loss": 84.9397, "step": 2126 }, { "compression_loss": 85.27522277832031, "epoch": 0.77, "learning_rate": 1.8491145645103e-05, "loss": 86.049, "step": 2127 }, { "compression_loss": 83.90545654296875, "epoch": 0.77, "learning_rate": 1.848572461149259e-05, "loss": 84.3619, "step": 2128 }, { "compression_loss": 83.95227813720703, "epoch": 0.77, "learning_rate": 1.8480303577882182e-05, "loss": 84.318, "step": 2129 }, { "compression_loss": 84.67566680908203, "epoch": 0.77, "learning_rate": 1.8474882544271773e-05, "loss": 85.4765, "step": 2130 }, { "compression_loss": 86.35749816894531, "epoch": 0.77, "learning_rate": 1.8469461510661364e-05, "loss": 87.1123, "step": 2131 }, { "compression_loss": 85.65604400634766, "epoch": 0.77, "learning_rate": 1.846404047705096e-05, "loss": 86.1596, "step": 2132 }, { "compression_loss": 85.18085479736328, "epoch": 0.77, "learning_rate": 1.845861944344055e-05, "loss": 85.5632, "step": 2133 }, { "compression_loss": 84.70120239257812, "epoch": 0.77, "learning_rate": 1.845319840983014e-05, "loss": 85.1649, "step": 2134 }, { "compression_loss": 82.93447875976562, "epoch": 0.77, "learning_rate": 1.8447777376219733e-05, "loss": 83.4383, "step": 2135 }, { "compression_loss": 85.05753326416016, "epoch": 0.77, "learning_rate": 1.8442356342609324e-05, "loss": 85.5331, "step": 2136 }, { "compression_loss": 84.72875213623047, "epoch": 0.77, "learning_rate": 1.8436935308998915e-05, "loss": 85.1808, "step": 2137 }, { "compression_loss": 84.16373443603516, "epoch": 0.77, "learning_rate": 1.8431514275388507e-05, "loss": 85.0138, "step": 2138 }, { "compression_loss": 84.87238311767578, "epoch": 0.77, "learning_rate": 1.8426093241778098e-05, "loss": 85.3682, "step": 2139 }, { "compression_loss": 84.256591796875, "epoch": 0.77, "learning_rate": 1.8420672208167693e-05, "loss": 84.9473, "step": 2140 }, { "compression_loss": 85.20829772949219, "epoch": 0.77, "learning_rate": 1.8415251174557284e-05, "loss": 85.9599, "step": 2141 }, { "compression_loss": 84.49700164794922, "epoch": 0.77, "learning_rate": 1.8409830140946875e-05, "loss": 84.9213, "step": 2142 }, { "compression_loss": 86.07563781738281, "epoch": 0.77, "learning_rate": 1.8404409107336467e-05, "loss": 86.6139, "step": 2143 }, { "compression_loss": 84.0475845336914, "epoch": 0.77, "learning_rate": 1.8398988073726058e-05, "loss": 84.7647, "step": 2144 }, { "compression_loss": 82.63961029052734, "epoch": 0.78, "learning_rate": 1.839356704011565e-05, "loss": 83.6164, "step": 2145 }, { "compression_loss": 85.20015716552734, "epoch": 0.78, "learning_rate": 1.838814600650524e-05, "loss": 85.5949, "step": 2146 }, { "compression_loss": 86.35896301269531, "epoch": 0.78, "learning_rate": 1.838272497289483e-05, "loss": 86.8952, "step": 2147 }, { "compression_loss": 84.41336822509766, "epoch": 0.78, "learning_rate": 1.8377303939284426e-05, "loss": 84.8031, "step": 2148 }, { "compression_loss": 82.55758666992188, "epoch": 0.78, "learning_rate": 1.8371882905674018e-05, "loss": 83.1118, "step": 2149 }, { "compression_loss": 85.03269958496094, "epoch": 0.78, "learning_rate": 1.836646187206361e-05, "loss": 85.5265, "step": 2150 }, { "compression_loss": 83.78730773925781, "epoch": 0.78, "learning_rate": 1.83610408384532e-05, "loss": 84.2442, "step": 2151 }, { "compression_loss": 85.36634826660156, "epoch": 0.78, "learning_rate": 1.835561980484279e-05, "loss": 86.1056, "step": 2152 }, { "compression_loss": 83.65278625488281, "epoch": 0.78, "learning_rate": 1.8350198771232383e-05, "loss": 84.0964, "step": 2153 }, { "compression_loss": 85.33540344238281, "epoch": 0.78, "learning_rate": 1.8344777737621974e-05, "loss": 86.0296, "step": 2154 }, { "compression_loss": 83.6975326538086, "epoch": 0.78, "learning_rate": 1.8339356704011565e-05, "loss": 84.2365, "step": 2155 }, { "compression_loss": 85.26226806640625, "epoch": 0.78, "learning_rate": 1.833393567040116e-05, "loss": 85.7277, "step": 2156 }, { "compression_loss": 84.146240234375, "epoch": 0.78, "learning_rate": 1.832851463679075e-05, "loss": 84.6838, "step": 2157 }, { "compression_loss": 83.73979949951172, "epoch": 0.78, "learning_rate": 1.8323093603180342e-05, "loss": 84.4584, "step": 2158 }, { "compression_loss": 84.79816436767578, "epoch": 0.78, "learning_rate": 1.8317672569569934e-05, "loss": 85.2496, "step": 2159 }, { "compression_loss": 83.74000549316406, "epoch": 0.78, "learning_rate": 1.8312251535959525e-05, "loss": 84.5083, "step": 2160 }, { "compression_loss": 83.75987243652344, "epoch": 0.78, "learning_rate": 1.8306830502349116e-05, "loss": 84.1434, "step": 2161 }, { "compression_loss": 84.13566589355469, "epoch": 0.78, "learning_rate": 1.8301409468738704e-05, "loss": 84.5663, "step": 2162 }, { "compression_loss": 83.5035171508789, "epoch": 0.78, "learning_rate": 1.8295988435128295e-05, "loss": 83.9019, "step": 2163 }, { "compression_loss": 85.90365600585938, "epoch": 0.78, "learning_rate": 1.829056740151789e-05, "loss": 86.1888, "step": 2164 }, { "compression_loss": 84.7069091796875, "epoch": 0.78, "learning_rate": 1.828514636790748e-05, "loss": 85.4361, "step": 2165 }, { "compression_loss": 84.147216796875, "epoch": 0.78, "learning_rate": 1.8279725334297073e-05, "loss": 84.5638, "step": 2166 }, { "compression_loss": 83.84619140625, "epoch": 0.78, "learning_rate": 1.8274304300686664e-05, "loss": 84.2228, "step": 2167 }, { "compression_loss": 86.53445434570312, "epoch": 0.78, "learning_rate": 1.8268883267076255e-05, "loss": 87.0304, "step": 2168 }, { "compression_loss": 84.43502044677734, "epoch": 0.78, "learning_rate": 1.8263462233465846e-05, "loss": 84.8648, "step": 2169 }, { "compression_loss": 84.2187728881836, "epoch": 0.78, "learning_rate": 1.8258041199855438e-05, "loss": 84.6519, "step": 2170 }, { "compression_loss": 84.71273040771484, "epoch": 0.78, "learning_rate": 1.825262016624503e-05, "loss": 85.447, "step": 2171 }, { "compression_loss": 83.90106201171875, "epoch": 0.78, "learning_rate": 1.8247199132634624e-05, "loss": 84.5473, "step": 2172 }, { "compression_loss": 83.77529907226562, "epoch": 0.79, "learning_rate": 1.8241778099024215e-05, "loss": 84.295, "step": 2173 }, { "compression_loss": 85.17288970947266, "epoch": 0.79, "learning_rate": 1.8236357065413806e-05, "loss": 85.6861, "step": 2174 }, { "compression_loss": 84.68570709228516, "epoch": 0.79, "learning_rate": 1.8230936031803397e-05, "loss": 84.9764, "step": 2175 }, { "compression_loss": 83.58988952636719, "epoch": 0.79, "learning_rate": 1.822551499819299e-05, "loss": 84.0874, "step": 2176 }, { "compression_loss": 82.99443817138672, "epoch": 0.79, "learning_rate": 1.822009396458258e-05, "loss": 83.369, "step": 2177 }, { "compression_loss": 82.9337158203125, "epoch": 0.79, "learning_rate": 1.821467293097217e-05, "loss": 83.3223, "step": 2178 }, { "compression_loss": 83.31470489501953, "epoch": 0.79, "learning_rate": 1.8209251897361762e-05, "loss": 83.909, "step": 2179 }, { "compression_loss": 83.1577377319336, "epoch": 0.79, "learning_rate": 1.8203830863751357e-05, "loss": 83.7326, "step": 2180 }, { "compression_loss": 84.34613037109375, "epoch": 0.79, "learning_rate": 1.819840983014095e-05, "loss": 85.3882, "step": 2181 }, { "compression_loss": 85.08381652832031, "epoch": 0.79, "learning_rate": 1.819298879653054e-05, "loss": 85.5377, "step": 2182 }, { "compression_loss": 84.47865295410156, "epoch": 0.79, "learning_rate": 1.818756776292013e-05, "loss": 85.1659, "step": 2183 }, { "compression_loss": 84.42112731933594, "epoch": 0.79, "learning_rate": 1.8182146729309722e-05, "loss": 84.6342, "step": 2184 }, { "compression_loss": 84.04423522949219, "epoch": 0.79, "learning_rate": 1.8176725695699313e-05, "loss": 84.7114, "step": 2185 }, { "compression_loss": 84.85232543945312, "epoch": 0.79, "learning_rate": 1.8171304662088905e-05, "loss": 85.5407, "step": 2186 }, { "compression_loss": 83.65196990966797, "epoch": 0.79, "learning_rate": 1.8165883628478496e-05, "loss": 84.1009, "step": 2187 }, { "compression_loss": 86.47654724121094, "epoch": 0.79, "learning_rate": 1.816046259486809e-05, "loss": 87.2439, "step": 2188 }, { "compression_loss": 83.47201538085938, "epoch": 0.79, "learning_rate": 1.8155041561257682e-05, "loss": 84.1454, "step": 2189 }, { "compression_loss": 84.0723876953125, "epoch": 0.79, "learning_rate": 1.8149620527647273e-05, "loss": 84.4594, "step": 2190 }, { "compression_loss": 83.72789001464844, "epoch": 0.79, "learning_rate": 1.8144199494036864e-05, "loss": 84.1507, "step": 2191 }, { "compression_loss": 83.02297973632812, "epoch": 0.79, "learning_rate": 1.8138778460426456e-05, "loss": 83.4804, "step": 2192 }, { "compression_loss": 82.44975280761719, "epoch": 0.79, "learning_rate": 1.8133357426816047e-05, "loss": 83.0295, "step": 2193 }, { "compression_loss": 84.7569808959961, "epoch": 0.79, "learning_rate": 1.812793639320564e-05, "loss": 85.135, "step": 2194 }, { "compression_loss": 85.82070922851562, "epoch": 0.79, "learning_rate": 1.812251535959523e-05, "loss": 86.3304, "step": 2195 }, { "compression_loss": 85.80522918701172, "epoch": 0.79, "learning_rate": 1.8117094325984824e-05, "loss": 86.7866, "step": 2196 }, { "compression_loss": 85.07243347167969, "epoch": 0.79, "learning_rate": 1.8111673292374416e-05, "loss": 85.5292, "step": 2197 }, { "compression_loss": 84.98872375488281, "epoch": 0.79, "learning_rate": 1.8106252258764007e-05, "loss": 85.6523, "step": 2198 }, { "compression_loss": 84.447509765625, "epoch": 0.79, "learning_rate": 1.8100831225153598e-05, "loss": 84.9098, "step": 2199 }, { "compression_loss": 84.93170928955078, "epoch": 0.8, "learning_rate": 1.809541019154319e-05, "loss": 85.4276, "step": 2200 }, { "compression_loss": 84.63319396972656, "epoch": 0.8, "learning_rate": 1.808998915793278e-05, "loss": 84.967, "step": 2201 }, { "compression_loss": 83.5347900390625, "epoch": 0.8, "learning_rate": 1.8084568124322372e-05, "loss": 84.1237, "step": 2202 }, { "compression_loss": 85.04907989501953, "epoch": 0.8, "learning_rate": 1.8079147090711963e-05, "loss": 85.668, "step": 2203 }, { "compression_loss": 85.61656951904297, "epoch": 0.8, "learning_rate": 1.8073726057101558e-05, "loss": 86.3741, "step": 2204 }, { "compression_loss": 83.62774658203125, "epoch": 0.8, "learning_rate": 1.8068305023491146e-05, "loss": 83.872, "step": 2205 }, { "compression_loss": 84.86397552490234, "epoch": 0.8, "learning_rate": 1.8062883989880737e-05, "loss": 85.5429, "step": 2206 }, { "compression_loss": 84.07752990722656, "epoch": 0.8, "learning_rate": 1.8057462956270328e-05, "loss": 84.5281, "step": 2207 }, { "compression_loss": 85.12767791748047, "epoch": 0.8, "learning_rate": 1.805204192265992e-05, "loss": 85.9464, "step": 2208 }, { "compression_loss": 84.64253234863281, "epoch": 0.8, "learning_rate": 1.804662088904951e-05, "loss": 84.9829, "step": 2209 }, { "compression_loss": 83.97538757324219, "epoch": 0.8, "learning_rate": 1.8041199855439102e-05, "loss": 84.7216, "step": 2210 }, { "compression_loss": 82.46255493164062, "epoch": 0.8, "learning_rate": 1.8035778821828693e-05, "loss": 83.0063, "step": 2211 }, { "compression_loss": 85.09501647949219, "epoch": 0.8, "learning_rate": 1.8030357788218288e-05, "loss": 85.5998, "step": 2212 }, { "compression_loss": 83.12896728515625, "epoch": 0.8, "learning_rate": 1.802493675460788e-05, "loss": 83.529, "step": 2213 }, { "compression_loss": 85.75579833984375, "epoch": 0.8, "learning_rate": 1.801951572099747e-05, "loss": 86.1189, "step": 2214 }, { "compression_loss": 85.3761978149414, "epoch": 0.8, "learning_rate": 1.8014094687387062e-05, "loss": 85.9905, "step": 2215 }, { "compression_loss": 84.65193176269531, "epoch": 0.8, "learning_rate": 1.8008673653776653e-05, "loss": 85.1701, "step": 2216 }, { "compression_loss": 83.26800537109375, "epoch": 0.8, "learning_rate": 1.8003252620166244e-05, "loss": 83.9906, "step": 2217 }, { "compression_loss": 85.87843322753906, "epoch": 0.8, "learning_rate": 1.7997831586555836e-05, "loss": 86.3502, "step": 2218 }, { "compression_loss": 85.38130187988281, "epoch": 0.8, "learning_rate": 1.7992410552945427e-05, "loss": 85.9768, "step": 2219 }, { "compression_loss": 85.1170425415039, "epoch": 0.8, "learning_rate": 1.7986989519335018e-05, "loss": 85.7479, "step": 2220 }, { "compression_loss": 85.83321380615234, "epoch": 0.8, "learning_rate": 1.7981568485724613e-05, "loss": 86.4936, "step": 2221 }, { "compression_loss": 83.45478057861328, "epoch": 0.8, "learning_rate": 1.7976147452114204e-05, "loss": 83.8185, "step": 2222 }, { "compression_loss": 83.95059204101562, "epoch": 0.8, "learning_rate": 1.7970726418503795e-05, "loss": 84.4101, "step": 2223 }, { "compression_loss": 83.90037536621094, "epoch": 0.8, "learning_rate": 1.7965305384893387e-05, "loss": 84.4763, "step": 2224 }, { "compression_loss": 84.29611206054688, "epoch": 0.8, "learning_rate": 1.7959884351282978e-05, "loss": 84.8615, "step": 2225 }, { "compression_loss": 82.09652709960938, "epoch": 0.8, "learning_rate": 1.795446331767257e-05, "loss": 82.4034, "step": 2226 }, { "compression_loss": 83.2279281616211, "epoch": 0.8, "learning_rate": 1.794904228406216e-05, "loss": 83.522, "step": 2227 }, { "compression_loss": 85.27132415771484, "epoch": 0.81, "learning_rate": 1.7943621250451752e-05, "loss": 85.9283, "step": 2228 }, { "compression_loss": 83.0810317993164, "epoch": 0.81, "learning_rate": 1.7938200216841346e-05, "loss": 83.7987, "step": 2229 }, { "compression_loss": 85.85652923583984, "epoch": 0.81, "learning_rate": 1.7932779183230938e-05, "loss": 86.6009, "step": 2230 }, { "compression_loss": 85.48040771484375, "epoch": 0.81, "learning_rate": 1.792735814962053e-05, "loss": 86.1599, "step": 2231 }, { "compression_loss": 83.11674499511719, "epoch": 0.81, "learning_rate": 1.792193711601012e-05, "loss": 83.9395, "step": 2232 }, { "compression_loss": 85.49388885498047, "epoch": 0.81, "learning_rate": 1.791651608239971e-05, "loss": 86.2695, "step": 2233 }, { "compression_loss": 84.18678283691406, "epoch": 0.81, "learning_rate": 1.7911095048789303e-05, "loss": 84.5656, "step": 2234 }, { "compression_loss": 82.87353515625, "epoch": 0.81, "learning_rate": 1.7905674015178894e-05, "loss": 83.3398, "step": 2235 }, { "compression_loss": 83.76292419433594, "epoch": 0.81, "learning_rate": 1.7900252981568485e-05, "loss": 84.4157, "step": 2236 }, { "compression_loss": 86.12518310546875, "epoch": 0.81, "learning_rate": 1.789483194795808e-05, "loss": 86.6193, "step": 2237 }, { "compression_loss": 85.57655334472656, "epoch": 0.81, "learning_rate": 1.788941091434767e-05, "loss": 85.967, "step": 2238 }, { "compression_loss": 85.64836883544922, "epoch": 0.81, "learning_rate": 1.7883989880737262e-05, "loss": 86.316, "step": 2239 }, { "compression_loss": 84.54397583007812, "epoch": 0.81, "learning_rate": 1.7878568847126854e-05, "loss": 85.1675, "step": 2240 }, { "compression_loss": 84.29027557373047, "epoch": 0.81, "learning_rate": 1.7873147813516445e-05, "loss": 84.9417, "step": 2241 }, { "compression_loss": 82.83541107177734, "epoch": 0.81, "learning_rate": 1.7867726779906036e-05, "loss": 83.6562, "step": 2242 }, { "compression_loss": 81.88864135742188, "epoch": 0.81, "learning_rate": 1.7862305746295628e-05, "loss": 82.449, "step": 2243 }, { "compression_loss": 83.25100708007812, "epoch": 0.81, "learning_rate": 1.785688471268522e-05, "loss": 84.2783, "step": 2244 }, { "compression_loss": 83.20354461669922, "epoch": 0.81, "learning_rate": 1.7851463679074814e-05, "loss": 83.766, "step": 2245 }, { "compression_loss": 83.46464538574219, "epoch": 0.81, "learning_rate": 1.7846042645464405e-05, "loss": 84.136, "step": 2246 }, { "compression_loss": 83.66315460205078, "epoch": 0.81, "learning_rate": 1.7840621611853996e-05, "loss": 84.378, "step": 2247 }, { "compression_loss": 82.85883331298828, "epoch": 0.81, "learning_rate": 1.7835200578243587e-05, "loss": 83.5132, "step": 2248 }, { "compression_loss": 85.64717102050781, "epoch": 0.81, "learning_rate": 1.7829779544633175e-05, "loss": 86.2517, "step": 2249 }, { "compression_loss": 83.97855377197266, "epoch": 0.81, "learning_rate": 1.7824358511022766e-05, "loss": 84.6132, "step": 2250 }, { "epoch": 0.81, "eval_exact_match": 86.04541154210028, "eval_f1": 92.6762139520899, "step": 2250 }, { "compression_loss": 84.05482482910156, "epoch": 0.81, "learning_rate": 1.7818937477412358e-05, "loss": 84.6543, "step": 2251 }, { "compression_loss": 84.12405395507812, "epoch": 0.81, "learning_rate": 1.781351644380195e-05, "loss": 84.9445, "step": 2252 }, { "compression_loss": 85.19157409667969, "epoch": 0.81, "learning_rate": 1.7808095410191544e-05, "loss": 85.7902, "step": 2253 }, { "compression_loss": 84.52313232421875, "epoch": 0.81, "learning_rate": 1.7802674376581135e-05, "loss": 85.3416, "step": 2254 }, { "compression_loss": 84.92998504638672, "epoch": 0.81, "learning_rate": 1.7797253342970726e-05, "loss": 85.5354, "step": 2255 }, { "compression_loss": 84.37930297851562, "epoch": 0.82, "learning_rate": 1.7791832309360317e-05, "loss": 84.8158, "step": 2256 }, { "compression_loss": 84.77955627441406, "epoch": 0.82, "learning_rate": 1.778641127574991e-05, "loss": 85.3482, "step": 2257 }, { "compression_loss": 84.77769470214844, "epoch": 0.82, "learning_rate": 1.77809902421395e-05, "loss": 85.4476, "step": 2258 }, { "compression_loss": 84.96226501464844, "epoch": 0.82, "learning_rate": 1.777556920852909e-05, "loss": 85.3046, "step": 2259 }, { "compression_loss": 84.61144256591797, "epoch": 0.82, "learning_rate": 1.7770148174918683e-05, "loss": 85.111, "step": 2260 }, { "compression_loss": 84.68234252929688, "epoch": 0.82, "learning_rate": 1.7764727141308277e-05, "loss": 85.0992, "step": 2261 }, { "compression_loss": 86.13784790039062, "epoch": 0.82, "learning_rate": 1.775930610769787e-05, "loss": 87.1771, "step": 2262 }, { "compression_loss": 84.00442504882812, "epoch": 0.82, "learning_rate": 1.775388507408746e-05, "loss": 84.3742, "step": 2263 }, { "compression_loss": 84.30745697021484, "epoch": 0.82, "learning_rate": 1.774846404047705e-05, "loss": 84.6338, "step": 2264 }, { "compression_loss": 84.63550567626953, "epoch": 0.82, "learning_rate": 1.7743043006866642e-05, "loss": 85.2849, "step": 2265 }, { "compression_loss": 85.00225830078125, "epoch": 0.82, "learning_rate": 1.7737621973256234e-05, "loss": 85.5576, "step": 2266 }, { "compression_loss": 83.63764953613281, "epoch": 0.82, "learning_rate": 1.7732200939645825e-05, "loss": 83.9792, "step": 2267 }, { "compression_loss": 84.16853332519531, "epoch": 0.82, "learning_rate": 1.7726779906035416e-05, "loss": 84.9911, "step": 2268 }, { "compression_loss": 84.33523559570312, "epoch": 0.82, "learning_rate": 1.772135887242501e-05, "loss": 84.611, "step": 2269 }, { "compression_loss": 87.30604553222656, "epoch": 0.82, "learning_rate": 1.7715937838814602e-05, "loss": 87.846, "step": 2270 }, { "compression_loss": 84.87356567382812, "epoch": 0.82, "learning_rate": 1.7710516805204193e-05, "loss": 85.4552, "step": 2271 }, { "compression_loss": 84.50663757324219, "epoch": 0.82, "learning_rate": 1.7705095771593785e-05, "loss": 85.3484, "step": 2272 }, { "compression_loss": 82.8994140625, "epoch": 0.82, "learning_rate": 1.7699674737983376e-05, "loss": 83.3271, "step": 2273 }, { "compression_loss": 86.26974487304688, "epoch": 0.82, "learning_rate": 1.7694253704372967e-05, "loss": 86.9006, "step": 2274 }, { "compression_loss": 85.06329345703125, "epoch": 0.82, "learning_rate": 1.768883267076256e-05, "loss": 85.4923, "step": 2275 }, { "compression_loss": 82.83486938476562, "epoch": 0.82, "learning_rate": 1.768341163715215e-05, "loss": 83.2922, "step": 2276 }, { "compression_loss": 85.66291809082031, "epoch": 0.82, "learning_rate": 1.7677990603541744e-05, "loss": 86.5051, "step": 2277 }, { "compression_loss": 86.40011596679688, "epoch": 0.82, "learning_rate": 1.7672569569931336e-05, "loss": 86.7693, "step": 2278 }, { "compression_loss": 84.44207000732422, "epoch": 0.82, "learning_rate": 1.7667148536320927e-05, "loss": 85.1243, "step": 2279 }, { "compression_loss": 86.08918762207031, "epoch": 0.82, "learning_rate": 1.7661727502710518e-05, "loss": 86.7675, "step": 2280 }, { "compression_loss": 84.09404754638672, "epoch": 0.82, "learning_rate": 1.765630646910011e-05, "loss": 84.437, "step": 2281 }, { "compression_loss": 82.74324798583984, "epoch": 0.82, "learning_rate": 1.76508854354897e-05, "loss": 83.0794, "step": 2282 }, { "compression_loss": 85.43952178955078, "epoch": 0.83, "learning_rate": 1.7645464401879292e-05, "loss": 85.9798, "step": 2283 }, { "compression_loss": 83.53089141845703, "epoch": 0.83, "learning_rate": 1.7640043368268883e-05, "loss": 83.8699, "step": 2284 }, { "compression_loss": 84.28424835205078, "epoch": 0.83, "learning_rate": 1.7634622334658478e-05, "loss": 84.8653, "step": 2285 }, { "compression_loss": 82.58802032470703, "epoch": 0.83, "learning_rate": 1.762920130104807e-05, "loss": 83.0444, "step": 2286 }, { "compression_loss": 83.5343017578125, "epoch": 0.83, "learning_rate": 1.762378026743766e-05, "loss": 84.0491, "step": 2287 }, { "compression_loss": 82.33271026611328, "epoch": 0.83, "learning_rate": 1.7618359233827252e-05, "loss": 82.7959, "step": 2288 }, { "compression_loss": 85.70759582519531, "epoch": 0.83, "learning_rate": 1.7612938200216843e-05, "loss": 86.3019, "step": 2289 }, { "compression_loss": 84.58889770507812, "epoch": 0.83, "learning_rate": 1.7607517166606434e-05, "loss": 84.947, "step": 2290 }, { "compression_loss": 83.04020690917969, "epoch": 0.83, "learning_rate": 1.7602096132996026e-05, "loss": 83.4201, "step": 2291 }, { "compression_loss": 84.35774993896484, "epoch": 0.83, "learning_rate": 1.7596675099385617e-05, "loss": 84.903, "step": 2292 }, { "compression_loss": 84.45170593261719, "epoch": 0.83, "learning_rate": 1.7591254065775208e-05, "loss": 85.001, "step": 2293 }, { "compression_loss": 84.48912811279297, "epoch": 0.83, "learning_rate": 1.75858330321648e-05, "loss": 84.8751, "step": 2294 }, { "compression_loss": 83.06946563720703, "epoch": 0.83, "learning_rate": 1.758041199855439e-05, "loss": 83.6532, "step": 2295 }, { "compression_loss": 85.09800720214844, "epoch": 0.83, "learning_rate": 1.7574990964943982e-05, "loss": 85.5878, "step": 2296 }, { "compression_loss": 84.66632843017578, "epoch": 0.83, "learning_rate": 1.7569569931333573e-05, "loss": 85.2131, "step": 2297 }, { "compression_loss": 84.24275207519531, "epoch": 0.83, "learning_rate": 1.7564148897723164e-05, "loss": 84.7976, "step": 2298 }, { "compression_loss": 83.47265625, "epoch": 0.83, "learning_rate": 1.7558727864112756e-05, "loss": 83.8708, "step": 2299 }, { "compression_loss": 84.85169219970703, "epoch": 0.83, "learning_rate": 1.7553306830502347e-05, "loss": 85.1009, "step": 2300 }, { "compression_loss": 84.4005126953125, "epoch": 0.83, "learning_rate": 1.754788579689194e-05, "loss": 85.0526, "step": 2301 }, { "compression_loss": 84.8602294921875, "epoch": 0.83, "learning_rate": 1.7542464763281533e-05, "loss": 85.4549, "step": 2302 }, { "compression_loss": 84.67205810546875, "epoch": 0.83, "learning_rate": 1.7537043729671124e-05, "loss": 84.9039, "step": 2303 }, { "compression_loss": 84.64973449707031, "epoch": 0.83, "learning_rate": 1.7531622696060715e-05, "loss": 85.147, "step": 2304 }, { "compression_loss": 84.12599182128906, "epoch": 0.83, "learning_rate": 1.7526201662450307e-05, "loss": 84.5324, "step": 2305 }, { "compression_loss": 85.02347564697266, "epoch": 0.83, "learning_rate": 1.7520780628839898e-05, "loss": 85.5514, "step": 2306 }, { "compression_loss": 84.93448638916016, "epoch": 0.83, "learning_rate": 1.751535959522949e-05, "loss": 85.3647, "step": 2307 }, { "compression_loss": 82.95809936523438, "epoch": 0.83, "learning_rate": 1.750993856161908e-05, "loss": 83.5418, "step": 2308 }, { "compression_loss": 82.06156921386719, "epoch": 0.83, "learning_rate": 1.7504517528008675e-05, "loss": 82.7069, "step": 2309 }, { "compression_loss": 85.05628967285156, "epoch": 0.83, "learning_rate": 1.7499096494398267e-05, "loss": 85.9258, "step": 2310 }, { "compression_loss": 83.31570434570312, "epoch": 0.84, "learning_rate": 1.7493675460787858e-05, "loss": 84.0629, "step": 2311 }, { "compression_loss": 84.20753479003906, "epoch": 0.84, "learning_rate": 1.748825442717745e-05, "loss": 84.5863, "step": 2312 }, { "compression_loss": 85.02690887451172, "epoch": 0.84, "learning_rate": 1.748283339356704e-05, "loss": 85.8131, "step": 2313 }, { "compression_loss": 83.87106323242188, "epoch": 0.84, "learning_rate": 1.747741235995663e-05, "loss": 84.1262, "step": 2314 }, { "compression_loss": 82.20828247070312, "epoch": 0.84, "learning_rate": 1.7471991326346223e-05, "loss": 82.7293, "step": 2315 }, { "compression_loss": 84.54429626464844, "epoch": 0.84, "learning_rate": 1.7466570292735814e-05, "loss": 85.1295, "step": 2316 }, { "compression_loss": 84.6925277709961, "epoch": 0.84, "learning_rate": 1.746114925912541e-05, "loss": 85.2286, "step": 2317 }, { "compression_loss": 83.93551635742188, "epoch": 0.84, "learning_rate": 1.7455728225515e-05, "loss": 84.4543, "step": 2318 }, { "compression_loss": 85.55079650878906, "epoch": 0.84, "learning_rate": 1.745030719190459e-05, "loss": 85.9578, "step": 2319 }, { "compression_loss": 82.50849914550781, "epoch": 0.84, "learning_rate": 1.7444886158294183e-05, "loss": 82.8998, "step": 2320 }, { "compression_loss": 86.57546997070312, "epoch": 0.84, "learning_rate": 1.7439465124683774e-05, "loss": 87.4659, "step": 2321 }, { "compression_loss": 84.51107788085938, "epoch": 0.84, "learning_rate": 1.7434044091073365e-05, "loss": 85.0483, "step": 2322 }, { "compression_loss": 82.39772033691406, "epoch": 0.84, "learning_rate": 1.7428623057462956e-05, "loss": 82.9656, "step": 2323 }, { "compression_loss": 83.99677276611328, "epoch": 0.84, "learning_rate": 1.7423202023852548e-05, "loss": 84.5476, "step": 2324 }, { "compression_loss": 82.79469299316406, "epoch": 0.84, "learning_rate": 1.7417780990242142e-05, "loss": 83.1648, "step": 2325 }, { "compression_loss": 84.39349365234375, "epoch": 0.84, "learning_rate": 1.7412359956631734e-05, "loss": 85.3136, "step": 2326 }, { "compression_loss": 84.36051177978516, "epoch": 0.84, "learning_rate": 1.7406938923021325e-05, "loss": 84.7529, "step": 2327 }, { "compression_loss": 84.07791137695312, "epoch": 0.84, "learning_rate": 1.7401517889410916e-05, "loss": 84.6319, "step": 2328 }, { "compression_loss": 82.53824615478516, "epoch": 0.84, "learning_rate": 1.7396096855800507e-05, "loss": 83.0994, "step": 2329 }, { "compression_loss": 83.22750091552734, "epoch": 0.84, "learning_rate": 1.73906758221901e-05, "loss": 83.9728, "step": 2330 }, { "compression_loss": 83.93545532226562, "epoch": 0.84, "learning_rate": 1.738525478857969e-05, "loss": 84.6601, "step": 2331 }, { "compression_loss": 86.37881469726562, "epoch": 0.84, "learning_rate": 1.737983375496928e-05, "loss": 87.2769, "step": 2332 }, { "compression_loss": 84.88794708251953, "epoch": 0.84, "learning_rate": 1.7374412721358876e-05, "loss": 85.1865, "step": 2333 }, { "compression_loss": 84.49977111816406, "epoch": 0.84, "learning_rate": 1.7368991687748467e-05, "loss": 85.0817, "step": 2334 }, { "compression_loss": 83.47174072265625, "epoch": 0.84, "learning_rate": 1.736357065413806e-05, "loss": 84.3247, "step": 2335 }, { "compression_loss": 83.0064926147461, "epoch": 0.84, "learning_rate": 1.7358149620527646e-05, "loss": 83.4407, "step": 2336 }, { "compression_loss": 84.75996398925781, "epoch": 0.84, "learning_rate": 1.7352728586917238e-05, "loss": 85.1852, "step": 2337 }, { "compression_loss": 84.20669555664062, "epoch": 0.84, "learning_rate": 1.734730755330683e-05, "loss": 84.7526, "step": 2338 }, { "compression_loss": 82.8330078125, "epoch": 0.85, "learning_rate": 1.734188651969642e-05, "loss": 83.1948, "step": 2339 }, { "compression_loss": 84.37242126464844, "epoch": 0.85, "learning_rate": 1.733646548608601e-05, "loss": 85.0381, "step": 2340 }, { "compression_loss": 85.48834228515625, "epoch": 0.85, "learning_rate": 1.7331044452475606e-05, "loss": 86.2506, "step": 2341 }, { "compression_loss": 84.75030517578125, "epoch": 0.85, "learning_rate": 1.7325623418865197e-05, "loss": 85.3866, "step": 2342 }, { "compression_loss": 85.17236328125, "epoch": 0.85, "learning_rate": 1.732020238525479e-05, "loss": 85.9495, "step": 2343 }, { "compression_loss": 85.36505126953125, "epoch": 0.85, "learning_rate": 1.731478135164438e-05, "loss": 86.0769, "step": 2344 }, { "compression_loss": 85.37822723388672, "epoch": 0.85, "learning_rate": 1.730936031803397e-05, "loss": 86.1351, "step": 2345 }, { "compression_loss": 85.52725982666016, "epoch": 0.85, "learning_rate": 1.7303939284423562e-05, "loss": 86.0978, "step": 2346 }, { "compression_loss": 83.34869384765625, "epoch": 0.85, "learning_rate": 1.7298518250813154e-05, "loss": 84.057, "step": 2347 }, { "compression_loss": 83.72027587890625, "epoch": 0.85, "learning_rate": 1.7293097217202745e-05, "loss": 83.9867, "step": 2348 }, { "compression_loss": 84.41604614257812, "epoch": 0.85, "learning_rate": 1.728767618359234e-05, "loss": 84.8549, "step": 2349 }, { "compression_loss": 83.19081115722656, "epoch": 0.85, "learning_rate": 1.728225514998193e-05, "loss": 83.5605, "step": 2350 }, { "compression_loss": 84.60352325439453, "epoch": 0.85, "learning_rate": 1.7276834116371522e-05, "loss": 84.8567, "step": 2351 }, { "compression_loss": 84.49392700195312, "epoch": 0.85, "learning_rate": 1.7271413082761113e-05, "loss": 85.1171, "step": 2352 }, { "compression_loss": 84.84856414794922, "epoch": 0.85, "learning_rate": 1.7265992049150705e-05, "loss": 85.2448, "step": 2353 }, { "compression_loss": 81.82931518554688, "epoch": 0.85, "learning_rate": 1.7260571015540296e-05, "loss": 82.3042, "step": 2354 }, { "compression_loss": 83.72589874267578, "epoch": 0.85, "learning_rate": 1.7255149981929887e-05, "loss": 84.1209, "step": 2355 }, { "compression_loss": 83.29045104980469, "epoch": 0.85, "learning_rate": 1.724972894831948e-05, "loss": 84.029, "step": 2356 }, { "compression_loss": 84.14619445800781, "epoch": 0.85, "learning_rate": 1.7244307914709073e-05, "loss": 84.7144, "step": 2357 }, { "compression_loss": 84.20167541503906, "epoch": 0.85, "learning_rate": 1.7238886881098664e-05, "loss": 84.6844, "step": 2358 }, { "compression_loss": 85.55296325683594, "epoch": 0.85, "learning_rate": 1.7233465847488256e-05, "loss": 86.3792, "step": 2359 }, { "compression_loss": 83.23988342285156, "epoch": 0.85, "learning_rate": 1.7228044813877847e-05, "loss": 83.612, "step": 2360 }, { "compression_loss": 85.00595092773438, "epoch": 0.85, "learning_rate": 1.722262378026744e-05, "loss": 85.4138, "step": 2361 }, { "compression_loss": 83.92724609375, "epoch": 0.85, "learning_rate": 1.721720274665703e-05, "loss": 84.3429, "step": 2362 }, { "compression_loss": 84.07852172851562, "epoch": 0.85, "learning_rate": 1.721178171304662e-05, "loss": 84.5235, "step": 2363 }, { "compression_loss": 84.0244369506836, "epoch": 0.85, "learning_rate": 1.7206360679436212e-05, "loss": 84.3809, "step": 2364 }, { "compression_loss": 85.51638793945312, "epoch": 0.85, "learning_rate": 1.7200939645825807e-05, "loss": 85.9851, "step": 2365 }, { "compression_loss": 85.43421173095703, "epoch": 0.86, "learning_rate": 1.7195518612215398e-05, "loss": 85.8714, "step": 2366 }, { "compression_loss": 82.58808898925781, "epoch": 0.86, "learning_rate": 1.719009757860499e-05, "loss": 83.0361, "step": 2367 }, { "compression_loss": 84.47817993164062, "epoch": 0.86, "learning_rate": 1.718467654499458e-05, "loss": 84.7531, "step": 2368 }, { "compression_loss": 83.75332641601562, "epoch": 0.86, "learning_rate": 1.7179255511384172e-05, "loss": 84.1303, "step": 2369 }, { "compression_loss": 83.72381591796875, "epoch": 0.86, "learning_rate": 1.7173834477773763e-05, "loss": 84.0513, "step": 2370 }, { "compression_loss": 83.4599609375, "epoch": 0.86, "learning_rate": 1.7168413444163354e-05, "loss": 84.0314, "step": 2371 }, { "compression_loss": 83.25949096679688, "epoch": 0.86, "learning_rate": 1.7162992410552946e-05, "loss": 83.6022, "step": 2372 }, { "compression_loss": 85.54035949707031, "epoch": 0.86, "learning_rate": 1.715757137694254e-05, "loss": 86.5284, "step": 2373 }, { "compression_loss": 84.42242431640625, "epoch": 0.86, "learning_rate": 1.715215034333213e-05, "loss": 84.9526, "step": 2374 }, { "compression_loss": 84.49928283691406, "epoch": 0.86, "learning_rate": 1.7146729309721723e-05, "loss": 84.9931, "step": 2375 }, { "compression_loss": 83.15990447998047, "epoch": 0.86, "learning_rate": 1.7141308276111314e-05, "loss": 83.6237, "step": 2376 }, { "compression_loss": 83.6567611694336, "epoch": 0.86, "learning_rate": 1.7135887242500905e-05, "loss": 84.0778, "step": 2377 }, { "compression_loss": 83.57905578613281, "epoch": 0.86, "learning_rate": 1.7130466208890497e-05, "loss": 84.167, "step": 2378 }, { "compression_loss": 83.0923080444336, "epoch": 0.86, "learning_rate": 1.7125045175280088e-05, "loss": 83.7274, "step": 2379 }, { "compression_loss": 85.52828979492188, "epoch": 0.86, "learning_rate": 1.7119624141669676e-05, "loss": 86.186, "step": 2380 }, { "compression_loss": 86.7586669921875, "epoch": 0.86, "learning_rate": 1.711420310805927e-05, "loss": 87.3196, "step": 2381 }, { "compression_loss": 83.46885681152344, "epoch": 0.86, "learning_rate": 1.7108782074448862e-05, "loss": 84.1453, "step": 2382 }, { "compression_loss": 84.8032455444336, "epoch": 0.86, "learning_rate": 1.7103361040838453e-05, "loss": 85.1595, "step": 2383 }, { "compression_loss": 83.74412536621094, "epoch": 0.86, "learning_rate": 1.7097940007228044e-05, "loss": 84.1847, "step": 2384 }, { "compression_loss": 83.86151123046875, "epoch": 0.86, "learning_rate": 1.7092518973617636e-05, "loss": 84.4664, "step": 2385 }, { "compression_loss": 84.337890625, "epoch": 0.86, "learning_rate": 1.7087097940007227e-05, "loss": 84.9066, "step": 2386 }, { "compression_loss": 83.47718048095703, "epoch": 0.86, "learning_rate": 1.7081676906396818e-05, "loss": 83.908, "step": 2387 }, { "compression_loss": 85.02007293701172, "epoch": 0.86, "learning_rate": 1.707625587278641e-05, "loss": 85.6862, "step": 2388 }, { "compression_loss": 84.27936553955078, "epoch": 0.86, "learning_rate": 1.7070834839176004e-05, "loss": 84.801, "step": 2389 }, { "compression_loss": 83.05674743652344, "epoch": 0.86, "learning_rate": 1.7065413805565595e-05, "loss": 83.3618, "step": 2390 }, { "compression_loss": 83.07132720947266, "epoch": 0.86, "learning_rate": 1.7059992771955187e-05, "loss": 83.5509, "step": 2391 }, { "compression_loss": 84.51325988769531, "epoch": 0.86, "learning_rate": 1.7054571738344778e-05, "loss": 85.0149, "step": 2392 }, { "compression_loss": 83.82196044921875, "epoch": 0.86, "learning_rate": 1.704915070473437e-05, "loss": 84.5345, "step": 2393 }, { "compression_loss": 83.70767974853516, "epoch": 0.87, "learning_rate": 1.704372967112396e-05, "loss": 84.1487, "step": 2394 }, { "compression_loss": 85.10963439941406, "epoch": 0.87, "learning_rate": 1.7038308637513552e-05, "loss": 85.8873, "step": 2395 }, { "compression_loss": 84.03378295898438, "epoch": 0.87, "learning_rate": 1.7032887603903143e-05, "loss": 84.5711, "step": 2396 }, { "compression_loss": 85.10565948486328, "epoch": 0.87, "learning_rate": 1.7027466570292738e-05, "loss": 85.5359, "step": 2397 }, { "compression_loss": 85.40866088867188, "epoch": 0.87, "learning_rate": 1.702204553668233e-05, "loss": 85.8381, "step": 2398 }, { "compression_loss": 84.88444519042969, "epoch": 0.87, "learning_rate": 1.701662450307192e-05, "loss": 85.2957, "step": 2399 }, { "compression_loss": 85.87774658203125, "epoch": 0.87, "learning_rate": 1.701120346946151e-05, "loss": 86.6033, "step": 2400 }, { "compression_loss": 85.0479736328125, "epoch": 0.87, "learning_rate": 1.7005782435851103e-05, "loss": 85.524, "step": 2401 }, { "compression_loss": 83.71699523925781, "epoch": 0.87, "learning_rate": 1.7000361402240694e-05, "loss": 84.4141, "step": 2402 }, { "compression_loss": 83.18568420410156, "epoch": 0.87, "learning_rate": 1.6994940368630285e-05, "loss": 83.7112, "step": 2403 }, { "compression_loss": 83.35711669921875, "epoch": 0.87, "learning_rate": 1.6989519335019877e-05, "loss": 84.05, "step": 2404 }, { "compression_loss": 83.23106384277344, "epoch": 0.87, "learning_rate": 1.6984098301409468e-05, "loss": 83.6104, "step": 2405 }, { "compression_loss": 84.28050231933594, "epoch": 0.87, "learning_rate": 1.6978677267799062e-05, "loss": 84.6908, "step": 2406 }, { "compression_loss": 84.3178939819336, "epoch": 0.87, "learning_rate": 1.6973256234188654e-05, "loss": 84.7122, "step": 2407 }, { "compression_loss": 84.39445495605469, "epoch": 0.87, "learning_rate": 1.6967835200578245e-05, "loss": 84.765, "step": 2408 }, { "compression_loss": 85.93628692626953, "epoch": 0.87, "learning_rate": 1.6962414166967836e-05, "loss": 86.6768, "step": 2409 }, { "compression_loss": 86.18328857421875, "epoch": 0.87, "learning_rate": 1.6956993133357428e-05, "loss": 86.6161, "step": 2410 }, { "compression_loss": 85.56941223144531, "epoch": 0.87, "learning_rate": 1.695157209974702e-05, "loss": 86.5274, "step": 2411 }, { "compression_loss": 84.67398834228516, "epoch": 0.87, "learning_rate": 1.694615106613661e-05, "loss": 85.4275, "step": 2412 }, { "compression_loss": 84.24882507324219, "epoch": 0.87, "learning_rate": 1.69407300325262e-05, "loss": 84.7884, "step": 2413 }, { "compression_loss": 84.45199584960938, "epoch": 0.87, "learning_rate": 1.6935308998915796e-05, "loss": 84.9988, "step": 2414 }, { "compression_loss": 83.82763671875, "epoch": 0.87, "learning_rate": 1.6929887965305387e-05, "loss": 84.3595, "step": 2415 }, { "compression_loss": 85.09339904785156, "epoch": 0.87, "learning_rate": 1.692446693169498e-05, "loss": 86.0269, "step": 2416 }, { "compression_loss": 83.5274887084961, "epoch": 0.87, "learning_rate": 1.691904589808457e-05, "loss": 83.9253, "step": 2417 }, { "compression_loss": 82.75946044921875, "epoch": 0.87, "learning_rate": 1.691362486447416e-05, "loss": 83.4858, "step": 2418 }, { "compression_loss": 84.01708984375, "epoch": 0.87, "learning_rate": 1.6908203830863752e-05, "loss": 84.4874, "step": 2419 }, { "compression_loss": 82.77346801757812, "epoch": 0.87, "learning_rate": 1.6902782797253344e-05, "loss": 83.3199, "step": 2420 }, { "compression_loss": 83.78738403320312, "epoch": 0.87, "learning_rate": 1.6897361763642935e-05, "loss": 84.4411, "step": 2421 }, { "compression_loss": 86.26560974121094, "epoch": 0.88, "learning_rate": 1.689194073003253e-05, "loss": 86.9519, "step": 2422 }, { "compression_loss": 84.06353759765625, "epoch": 0.88, "learning_rate": 1.6886519696422117e-05, "loss": 84.4799, "step": 2423 }, { "compression_loss": 85.83209228515625, "epoch": 0.88, "learning_rate": 1.688109866281171e-05, "loss": 86.4565, "step": 2424 }, { "compression_loss": 84.6039047241211, "epoch": 0.88, "learning_rate": 1.68756776292013e-05, "loss": 85.1897, "step": 2425 }, { "compression_loss": 84.78253173828125, "epoch": 0.88, "learning_rate": 1.687025659559089e-05, "loss": 85.5315, "step": 2426 }, { "compression_loss": 84.31488037109375, "epoch": 0.88, "learning_rate": 1.6864835561980483e-05, "loss": 84.8297, "step": 2427 }, { "compression_loss": 84.5914077758789, "epoch": 0.88, "learning_rate": 1.6859414528370074e-05, "loss": 85.489, "step": 2428 }, { "compression_loss": 84.17920684814453, "epoch": 0.88, "learning_rate": 1.6853993494759665e-05, "loss": 84.6405, "step": 2429 }, { "compression_loss": 84.27900695800781, "epoch": 0.88, "learning_rate": 1.684857246114926e-05, "loss": 84.9559, "step": 2430 }, { "compression_loss": 87.4637680053711, "epoch": 0.88, "learning_rate": 1.684315142753885e-05, "loss": 88.4734, "step": 2431 }, { "compression_loss": 86.13726043701172, "epoch": 0.88, "learning_rate": 1.6837730393928442e-05, "loss": 86.8095, "step": 2432 }, { "compression_loss": 84.40376281738281, "epoch": 0.88, "learning_rate": 1.6832309360318034e-05, "loss": 85.0981, "step": 2433 }, { "compression_loss": 85.63941955566406, "epoch": 0.88, "learning_rate": 1.6826888326707625e-05, "loss": 85.9688, "step": 2434 }, { "compression_loss": 85.7405776977539, "epoch": 0.88, "learning_rate": 1.6821467293097216e-05, "loss": 86.5365, "step": 2435 }, { "compression_loss": 84.02584838867188, "epoch": 0.88, "learning_rate": 1.6816046259486807e-05, "loss": 84.5969, "step": 2436 }, { "compression_loss": 85.07403564453125, "epoch": 0.88, "learning_rate": 1.68106252258764e-05, "loss": 85.4829, "step": 2437 }, { "compression_loss": 85.67713165283203, "epoch": 0.88, "learning_rate": 1.6805204192265993e-05, "loss": 86.1343, "step": 2438 }, { "compression_loss": 83.48741149902344, "epoch": 0.88, "learning_rate": 1.6799783158655585e-05, "loss": 84.0018, "step": 2439 }, { "compression_loss": 85.0478515625, "epoch": 0.88, "learning_rate": 1.6794362125045176e-05, "loss": 85.5664, "step": 2440 }, { "compression_loss": 83.84986877441406, "epoch": 0.88, "learning_rate": 1.6788941091434767e-05, "loss": 84.2746, "step": 2441 }, { "compression_loss": 82.89002227783203, "epoch": 0.88, "learning_rate": 1.678352005782436e-05, "loss": 83.207, "step": 2442 }, { "compression_loss": 85.60704040527344, "epoch": 0.88, "learning_rate": 1.677809902421395e-05, "loss": 86.165, "step": 2443 }, { "compression_loss": 83.95288848876953, "epoch": 0.88, "learning_rate": 1.677267799060354e-05, "loss": 84.8403, "step": 2444 }, { "compression_loss": 82.1376953125, "epoch": 0.88, "learning_rate": 1.6767256956993132e-05, "loss": 82.5858, "step": 2445 }, { "compression_loss": 83.43477630615234, "epoch": 0.88, "learning_rate": 1.6761835923382727e-05, "loss": 83.9201, "step": 2446 }, { "compression_loss": 85.67372131347656, "epoch": 0.88, "learning_rate": 1.6756414889772318e-05, "loss": 86.1808, "step": 2447 }, { "compression_loss": 85.51310729980469, "epoch": 0.88, "learning_rate": 1.675099385616191e-05, "loss": 85.9541, "step": 2448 }, { "compression_loss": 85.28701782226562, "epoch": 0.89, "learning_rate": 1.67455728225515e-05, "loss": 85.8504, "step": 2449 }, { "compression_loss": 86.4062728881836, "epoch": 0.89, "learning_rate": 1.6740151788941092e-05, "loss": 86.9621, "step": 2450 }, { "compression_loss": 84.58208465576172, "epoch": 0.89, "learning_rate": 1.6734730755330683e-05, "loss": 85.7457, "step": 2451 }, { "compression_loss": 84.6762466430664, "epoch": 0.89, "learning_rate": 1.6729309721720275e-05, "loss": 85.5308, "step": 2452 }, { "compression_loss": 83.77755737304688, "epoch": 0.89, "learning_rate": 1.6723888688109866e-05, "loss": 84.5005, "step": 2453 }, { "compression_loss": 84.28720092773438, "epoch": 0.89, "learning_rate": 1.671846765449946e-05, "loss": 84.9463, "step": 2454 }, { "compression_loss": 82.84628295898438, "epoch": 0.89, "learning_rate": 1.6713046620889052e-05, "loss": 83.7401, "step": 2455 }, { "compression_loss": 84.56668853759766, "epoch": 0.89, "learning_rate": 1.6707625587278643e-05, "loss": 84.9865, "step": 2456 }, { "compression_loss": 85.2985610961914, "epoch": 0.89, "learning_rate": 1.6702204553668234e-05, "loss": 85.8727, "step": 2457 }, { "compression_loss": 83.84828186035156, "epoch": 0.89, "learning_rate": 1.6696783520057826e-05, "loss": 84.5418, "step": 2458 }, { "compression_loss": 84.19763946533203, "epoch": 0.89, "learning_rate": 1.6691362486447417e-05, "loss": 84.505, "step": 2459 }, { "compression_loss": 85.3404541015625, "epoch": 0.89, "learning_rate": 1.6685941452837008e-05, "loss": 86.1891, "step": 2460 }, { "compression_loss": 86.68822479248047, "epoch": 0.89, "learning_rate": 1.66805204192266e-05, "loss": 87.4573, "step": 2461 }, { "compression_loss": 84.19580078125, "epoch": 0.89, "learning_rate": 1.6675099385616194e-05, "loss": 84.8481, "step": 2462 }, { "compression_loss": 84.88018798828125, "epoch": 0.89, "learning_rate": 1.6669678352005785e-05, "loss": 85.5452, "step": 2463 }, { "compression_loss": 85.30058288574219, "epoch": 0.89, "learning_rate": 1.6664257318395377e-05, "loss": 86.0771, "step": 2464 }, { "compression_loss": 82.57679748535156, "epoch": 0.89, "learning_rate": 1.6658836284784968e-05, "loss": 83.232, "step": 2465 }, { "compression_loss": 84.41148376464844, "epoch": 0.89, "learning_rate": 1.665341525117456e-05, "loss": 84.8229, "step": 2466 }, { "compression_loss": 84.10636138916016, "epoch": 0.89, "learning_rate": 1.6647994217564147e-05, "loss": 84.4423, "step": 2467 }, { "compression_loss": 83.32711791992188, "epoch": 0.89, "learning_rate": 1.6642573183953738e-05, "loss": 84.0227, "step": 2468 }, { "compression_loss": 83.617919921875, "epoch": 0.89, "learning_rate": 1.663715215034333e-05, "loss": 84.2019, "step": 2469 }, { "compression_loss": 85.64545440673828, "epoch": 0.89, "learning_rate": 1.6631731116732924e-05, "loss": 86.2507, "step": 2470 }, { "compression_loss": 83.00070190429688, "epoch": 0.89, "learning_rate": 1.6626310083122515e-05, "loss": 83.573, "step": 2471 }, { "compression_loss": 83.34610748291016, "epoch": 0.89, "learning_rate": 1.6620889049512107e-05, "loss": 84.0212, "step": 2472 }, { "compression_loss": 84.88604736328125, "epoch": 0.89, "learning_rate": 1.6615468015901698e-05, "loss": 85.2395, "step": 2473 }, { "compression_loss": 86.18759155273438, "epoch": 0.89, "learning_rate": 1.661004698229129e-05, "loss": 86.8116, "step": 2474 }, { "compression_loss": 83.44576263427734, "epoch": 0.89, "learning_rate": 1.660462594868088e-05, "loss": 84.0229, "step": 2475 }, { "compression_loss": 85.4007797241211, "epoch": 0.89, "learning_rate": 1.6599204915070472e-05, "loss": 85.9518, "step": 2476 }, { "compression_loss": 84.30384826660156, "epoch": 0.9, "learning_rate": 1.6593783881460063e-05, "loss": 84.979, "step": 2477 }, { "compression_loss": 84.76640319824219, "epoch": 0.9, "learning_rate": 1.6588362847849658e-05, "loss": 85.5269, "step": 2478 }, { "compression_loss": 85.58573150634766, "epoch": 0.9, "learning_rate": 1.658294181423925e-05, "loss": 86.1862, "step": 2479 }, { "compression_loss": 83.563232421875, "epoch": 0.9, "learning_rate": 1.657752078062884e-05, "loss": 84.146, "step": 2480 }, { "compression_loss": 85.23738098144531, "epoch": 0.9, "learning_rate": 1.657209974701843e-05, "loss": 86.1776, "step": 2481 }, { "compression_loss": 84.06541442871094, "epoch": 0.9, "learning_rate": 1.6566678713408023e-05, "loss": 84.6254, "step": 2482 }, { "compression_loss": 85.37602996826172, "epoch": 0.9, "learning_rate": 1.6561257679797614e-05, "loss": 86.3958, "step": 2483 }, { "compression_loss": 86.46434020996094, "epoch": 0.9, "learning_rate": 1.6555836646187205e-05, "loss": 87.0858, "step": 2484 }, { "compression_loss": 82.78111267089844, "epoch": 0.9, "learning_rate": 1.6550415612576797e-05, "loss": 83.4403, "step": 2485 }, { "compression_loss": 84.56702423095703, "epoch": 0.9, "learning_rate": 1.654499457896639e-05, "loss": 85.1797, "step": 2486 }, { "compression_loss": 84.11627197265625, "epoch": 0.9, "learning_rate": 1.6539573545355983e-05, "loss": 84.5362, "step": 2487 }, { "compression_loss": 84.504638671875, "epoch": 0.9, "learning_rate": 1.6534152511745574e-05, "loss": 84.8824, "step": 2488 }, { "compression_loss": 82.73564147949219, "epoch": 0.9, "learning_rate": 1.6528731478135165e-05, "loss": 83.2501, "step": 2489 }, { "compression_loss": 83.37862396240234, "epoch": 0.9, "learning_rate": 1.6523310444524756e-05, "loss": 83.7062, "step": 2490 }, { "compression_loss": 83.67987823486328, "epoch": 0.9, "learning_rate": 1.6517889410914348e-05, "loss": 84.2846, "step": 2491 }, { "compression_loss": 86.94694519042969, "epoch": 0.9, "learning_rate": 1.651246837730394e-05, "loss": 88.0609, "step": 2492 }, { "compression_loss": 85.38864135742188, "epoch": 0.9, "learning_rate": 1.650704734369353e-05, "loss": 85.6577, "step": 2493 }, { "compression_loss": 85.42037963867188, "epoch": 0.9, "learning_rate": 1.6501626310083125e-05, "loss": 86.0001, "step": 2494 }, { "compression_loss": 85.62657165527344, "epoch": 0.9, "learning_rate": 1.6496205276472716e-05, "loss": 86.206, "step": 2495 }, { "compression_loss": 84.87696838378906, "epoch": 0.9, "learning_rate": 1.6490784242862307e-05, "loss": 85.255, "step": 2496 }, { "compression_loss": 83.35794067382812, "epoch": 0.9, "learning_rate": 1.64853632092519e-05, "loss": 83.8489, "step": 2497 }, { "compression_loss": 84.35737609863281, "epoch": 0.9, "learning_rate": 1.647994217564149e-05, "loss": 84.9511, "step": 2498 }, { "compression_loss": 83.7759017944336, "epoch": 0.9, "learning_rate": 1.647452114203108e-05, "loss": 84.1247, "step": 2499 }, { "compression_loss": 82.496337890625, "epoch": 0.9, "learning_rate": 1.6469100108420673e-05, "loss": 83.1906, "step": 2500 }, { "epoch": 0.9, "eval_exact_match": 86.32923368022706, "eval_f1": 92.64119865126868, "step": 2500 }, { "compression_loss": 84.17835998535156, "epoch": 0.9, "learning_rate": 1.6463679074810264e-05, "loss": 84.815, "step": 2501 }, { "compression_loss": 82.87300109863281, "epoch": 0.9, "learning_rate": 1.645825804119986e-05, "loss": 83.2978, "step": 2502 }, { "compression_loss": 84.6961441040039, "epoch": 0.9, "learning_rate": 1.645283700758945e-05, "loss": 85.4109, "step": 2503 }, { "compression_loss": 81.35321044921875, "epoch": 0.9, "learning_rate": 1.644741597397904e-05, "loss": 82.1259, "step": 2504 }, { "compression_loss": 84.02201843261719, "epoch": 0.91, "learning_rate": 1.6441994940368632e-05, "loss": 84.4928, "step": 2505 }, { "compression_loss": 84.46051025390625, "epoch": 0.91, "learning_rate": 1.6436573906758224e-05, "loss": 85.1614, "step": 2506 }, { "compression_loss": 85.46723937988281, "epoch": 0.91, "learning_rate": 1.6431152873147815e-05, "loss": 85.7173, "step": 2507 }, { "compression_loss": 83.20946502685547, "epoch": 0.91, "learning_rate": 1.6425731839537406e-05, "loss": 83.7789, "step": 2508 }, { "compression_loss": 83.52547454833984, "epoch": 0.91, "learning_rate": 1.6420310805926997e-05, "loss": 84.3099, "step": 2509 }, { "compression_loss": 82.58203887939453, "epoch": 0.91, "learning_rate": 1.641488977231659e-05, "loss": 83.0318, "step": 2510 }, { "compression_loss": 84.27296447753906, "epoch": 0.91, "learning_rate": 1.640946873870618e-05, "loss": 84.6784, "step": 2511 }, { "compression_loss": 85.42732238769531, "epoch": 0.91, "learning_rate": 1.640404770509577e-05, "loss": 86.0803, "step": 2512 }, { "compression_loss": 84.07925415039062, "epoch": 0.91, "learning_rate": 1.6398626671485362e-05, "loss": 84.7472, "step": 2513 }, { "compression_loss": 83.75921630859375, "epoch": 0.91, "learning_rate": 1.6393205637874954e-05, "loss": 84.9287, "step": 2514 }, { "compression_loss": 84.53658294677734, "epoch": 0.91, "learning_rate": 1.6387784604264545e-05, "loss": 85.3039, "step": 2515 }, { "compression_loss": 83.20831298828125, "epoch": 0.91, "learning_rate": 1.6382363570654136e-05, "loss": 83.6592, "step": 2516 }, { "compression_loss": 84.32170104980469, "epoch": 0.91, "learning_rate": 1.6376942537043728e-05, "loss": 84.8712, "step": 2517 }, { "compression_loss": 84.19378662109375, "epoch": 0.91, "learning_rate": 1.6371521503433322e-05, "loss": 84.7174, "step": 2518 }, { "compression_loss": 84.20655822753906, "epoch": 0.91, "learning_rate": 1.6366100469822913e-05, "loss": 84.8121, "step": 2519 }, { "compression_loss": 82.82979583740234, "epoch": 0.91, "learning_rate": 1.6360679436212505e-05, "loss": 83.1086, "step": 2520 }, { "compression_loss": 83.28410339355469, "epoch": 0.91, "learning_rate": 1.6355258402602096e-05, "loss": 83.953, "step": 2521 }, { "compression_loss": 86.92627716064453, "epoch": 0.91, "learning_rate": 1.6349837368991687e-05, "loss": 87.6268, "step": 2522 }, { "compression_loss": 83.55538177490234, "epoch": 0.91, "learning_rate": 1.634441633538128e-05, "loss": 84.2508, "step": 2523 }, { "compression_loss": 84.29920959472656, "epoch": 0.91, "learning_rate": 1.633899530177087e-05, "loss": 84.7297, "step": 2524 }, { "compression_loss": 85.73756408691406, "epoch": 0.91, "learning_rate": 1.633357426816046e-05, "loss": 86.4409, "step": 2525 }, { "compression_loss": 82.7918472290039, "epoch": 0.91, "learning_rate": 1.6328153234550056e-05, "loss": 83.3362, "step": 2526 }, { "compression_loss": 86.41104125976562, "epoch": 0.91, "learning_rate": 1.6322732200939647e-05, "loss": 87.3563, "step": 2527 }, { "compression_loss": 85.11347961425781, "epoch": 0.91, "learning_rate": 1.631731116732924e-05, "loss": 85.5338, "step": 2528 }, { "compression_loss": 84.68770599365234, "epoch": 0.91, "learning_rate": 1.631189013371883e-05, "loss": 85.2671, "step": 2529 }, { "compression_loss": 81.87948608398438, "epoch": 0.91, "learning_rate": 1.630646910010842e-05, "loss": 82.3812, "step": 2530 }, { "compression_loss": 84.07829284667969, "epoch": 0.91, "learning_rate": 1.6301048066498012e-05, "loss": 84.6341, "step": 2531 }, { "compression_loss": 85.39639282226562, "epoch": 0.92, "learning_rate": 1.6295627032887603e-05, "loss": 86.0176, "step": 2532 }, { "compression_loss": 84.85330200195312, "epoch": 0.92, "learning_rate": 1.6290205999277195e-05, "loss": 85.4746, "step": 2533 }, { "compression_loss": 85.35958862304688, "epoch": 0.92, "learning_rate": 1.628478496566679e-05, "loss": 85.9698, "step": 2534 }, { "compression_loss": 84.35794067382812, "epoch": 0.92, "learning_rate": 1.627936393205638e-05, "loss": 84.9805, "step": 2535 }, { "compression_loss": 84.33769226074219, "epoch": 0.92, "learning_rate": 1.6273942898445972e-05, "loss": 84.6562, "step": 2536 }, { "compression_loss": 85.57400512695312, "epoch": 0.92, "learning_rate": 1.6268521864835563e-05, "loss": 86.2839, "step": 2537 }, { "compression_loss": 85.7848129272461, "epoch": 0.92, "learning_rate": 1.6263100831225154e-05, "loss": 86.6172, "step": 2538 }, { "compression_loss": 84.659912109375, "epoch": 0.92, "learning_rate": 1.6257679797614746e-05, "loss": 85.3531, "step": 2539 }, { "compression_loss": 84.86151123046875, "epoch": 0.92, "learning_rate": 1.6252258764004337e-05, "loss": 85.3858, "step": 2540 }, { "compression_loss": 85.96923065185547, "epoch": 0.92, "learning_rate": 1.6246837730393928e-05, "loss": 86.5051, "step": 2541 }, { "compression_loss": 86.00213623046875, "epoch": 0.92, "learning_rate": 1.6241416696783523e-05, "loss": 86.6295, "step": 2542 }, { "compression_loss": 80.92813110351562, "epoch": 0.92, "learning_rate": 1.6235995663173114e-05, "loss": 81.5386, "step": 2543 }, { "compression_loss": 84.8662109375, "epoch": 0.92, "learning_rate": 1.6230574629562705e-05, "loss": 85.5116, "step": 2544 }, { "compression_loss": 85.36830139160156, "epoch": 0.92, "learning_rate": 1.6225153595952297e-05, "loss": 85.8952, "step": 2545 }, { "compression_loss": 85.12498474121094, "epoch": 0.92, "learning_rate": 1.6219732562341888e-05, "loss": 85.7481, "step": 2546 }, { "compression_loss": 83.71710205078125, "epoch": 0.92, "learning_rate": 1.621431152873148e-05, "loss": 84.1356, "step": 2547 }, { "compression_loss": 84.489013671875, "epoch": 0.92, "learning_rate": 1.620889049512107e-05, "loss": 85.0432, "step": 2548 }, { "compression_loss": 85.53398132324219, "epoch": 0.92, "learning_rate": 1.6203469461510662e-05, "loss": 85.9499, "step": 2549 }, { "compression_loss": 83.57048797607422, "epoch": 0.92, "learning_rate": 1.6198048427900256e-05, "loss": 84.1351, "step": 2550 }, { "compression_loss": 83.63545227050781, "epoch": 0.92, "learning_rate": 1.6192627394289848e-05, "loss": 84.2646, "step": 2551 }, { "compression_loss": 86.2314682006836, "epoch": 0.92, "learning_rate": 1.618720636067944e-05, "loss": 86.5976, "step": 2552 }, { "compression_loss": 84.01355743408203, "epoch": 0.92, "learning_rate": 1.618178532706903e-05, "loss": 84.4433, "step": 2553 }, { "compression_loss": 82.95928192138672, "epoch": 0.92, "learning_rate": 1.6176364293458618e-05, "loss": 83.5674, "step": 2554 }, { "compression_loss": 81.96687316894531, "epoch": 0.92, "learning_rate": 1.617094325984821e-05, "loss": 82.7165, "step": 2555 }, { "compression_loss": 86.23240661621094, "epoch": 0.92, "learning_rate": 1.61655222262378e-05, "loss": 87.5517, "step": 2556 }, { "compression_loss": 84.41038513183594, "epoch": 0.92, "learning_rate": 1.6160101192627392e-05, "loss": 85.2459, "step": 2557 }, { "compression_loss": 83.94979095458984, "epoch": 0.92, "learning_rate": 1.6154680159016987e-05, "loss": 84.8452, "step": 2558 }, { "compression_loss": 83.27479553222656, "epoch": 0.92, "learning_rate": 1.6149259125406578e-05, "loss": 83.4497, "step": 2559 }, { "compression_loss": 84.6856689453125, "epoch": 0.93, "learning_rate": 1.614383809179617e-05, "loss": 85.2685, "step": 2560 }, { "compression_loss": 86.65743255615234, "epoch": 0.93, "learning_rate": 1.613841705818576e-05, "loss": 87.1238, "step": 2561 }, { "compression_loss": 84.20133209228516, "epoch": 0.93, "learning_rate": 1.6132996024575352e-05, "loss": 84.7494, "step": 2562 }, { "compression_loss": 81.63788604736328, "epoch": 0.93, "learning_rate": 1.6127574990964943e-05, "loss": 82.2267, "step": 2563 }, { "compression_loss": 84.08796691894531, "epoch": 0.93, "learning_rate": 1.6122153957354534e-05, "loss": 84.4949, "step": 2564 }, { "compression_loss": 84.70677947998047, "epoch": 0.93, "learning_rate": 1.6116732923744126e-05, "loss": 85.471, "step": 2565 }, { "compression_loss": 86.77037811279297, "epoch": 0.93, "learning_rate": 1.611131189013372e-05, "loss": 87.5973, "step": 2566 }, { "compression_loss": 83.40731811523438, "epoch": 0.93, "learning_rate": 1.610589085652331e-05, "loss": 83.9966, "step": 2567 }, { "compression_loss": 82.45310974121094, "epoch": 0.93, "learning_rate": 1.6100469822912903e-05, "loss": 83.2548, "step": 2568 }, { "compression_loss": 84.23986053466797, "epoch": 0.93, "learning_rate": 1.6095048789302494e-05, "loss": 84.4754, "step": 2569 }, { "compression_loss": 83.70089721679688, "epoch": 0.93, "learning_rate": 1.6089627755692085e-05, "loss": 84.4231, "step": 2570 }, { "compression_loss": 85.49147033691406, "epoch": 0.93, "learning_rate": 1.6084206722081677e-05, "loss": 86.2834, "step": 2571 }, { "compression_loss": 85.26628112792969, "epoch": 0.93, "learning_rate": 1.6078785688471268e-05, "loss": 85.8343, "step": 2572 }, { "compression_loss": 81.3011245727539, "epoch": 0.93, "learning_rate": 1.607336465486086e-05, "loss": 81.708, "step": 2573 }, { "compression_loss": 83.63426971435547, "epoch": 0.93, "learning_rate": 1.6067943621250454e-05, "loss": 84.5115, "step": 2574 }, { "compression_loss": 83.06608581542969, "epoch": 0.93, "learning_rate": 1.6062522587640045e-05, "loss": 83.4703, "step": 2575 }, { "compression_loss": 85.6130599975586, "epoch": 0.93, "learning_rate": 1.6057101554029636e-05, "loss": 86.2309, "step": 2576 }, { "compression_loss": 85.01814270019531, "epoch": 0.93, "learning_rate": 1.6051680520419228e-05, "loss": 85.4013, "step": 2577 }, { "compression_loss": 85.3349838256836, "epoch": 0.93, "learning_rate": 1.604625948680882e-05, "loss": 86.5061, "step": 2578 }, { "compression_loss": 83.13912963867188, "epoch": 0.93, "learning_rate": 1.604083845319841e-05, "loss": 83.6826, "step": 2579 }, { "compression_loss": 83.38500213623047, "epoch": 0.93, "learning_rate": 1.6035417419588e-05, "loss": 84.0814, "step": 2580 }, { "compression_loss": 82.93614196777344, "epoch": 0.93, "learning_rate": 1.6029996385977593e-05, "loss": 83.2939, "step": 2581 }, { "compression_loss": 86.31053924560547, "epoch": 0.93, "learning_rate": 1.6024575352367187e-05, "loss": 86.7382, "step": 2582 }, { "compression_loss": 85.62596130371094, "epoch": 0.93, "learning_rate": 1.601915431875678e-05, "loss": 86.226, "step": 2583 }, { "compression_loss": 84.90099334716797, "epoch": 0.93, "learning_rate": 1.601373328514637e-05, "loss": 85.6585, "step": 2584 }, { "compression_loss": 82.86402893066406, "epoch": 0.93, "learning_rate": 1.600831225153596e-05, "loss": 83.1452, "step": 2585 }, { "compression_loss": 84.6829605102539, "epoch": 0.93, "learning_rate": 1.6002891217925552e-05, "loss": 85.0368, "step": 2586 }, { "compression_loss": 84.82439422607422, "epoch": 0.93, "learning_rate": 1.5997470184315144e-05, "loss": 85.1192, "step": 2587 }, { "compression_loss": 84.28652954101562, "epoch": 0.94, "learning_rate": 1.5992049150704735e-05, "loss": 84.7515, "step": 2588 }, { "compression_loss": 82.93184661865234, "epoch": 0.94, "learning_rate": 1.5986628117094326e-05, "loss": 83.3735, "step": 2589 }, { "compression_loss": 85.2568588256836, "epoch": 0.94, "learning_rate": 1.5981207083483917e-05, "loss": 85.8383, "step": 2590 }, { "compression_loss": 83.86543273925781, "epoch": 0.94, "learning_rate": 1.5975786049873512e-05, "loss": 84.205, "step": 2591 }, { "compression_loss": 85.92518615722656, "epoch": 0.94, "learning_rate": 1.5970365016263103e-05, "loss": 86.5203, "step": 2592 }, { "compression_loss": 86.76521301269531, "epoch": 0.94, "learning_rate": 1.5964943982652695e-05, "loss": 87.423, "step": 2593 }, { "compression_loss": 83.69552612304688, "epoch": 0.94, "learning_rate": 1.5959522949042286e-05, "loss": 84.3453, "step": 2594 }, { "compression_loss": 85.69941711425781, "epoch": 0.94, "learning_rate": 1.5954101915431877e-05, "loss": 86.1985, "step": 2595 }, { "compression_loss": 83.46837615966797, "epoch": 0.94, "learning_rate": 1.594868088182147e-05, "loss": 84.0938, "step": 2596 }, { "compression_loss": 84.8257064819336, "epoch": 0.94, "learning_rate": 1.5943259848211056e-05, "loss": 85.5209, "step": 2597 }, { "compression_loss": 84.06352233886719, "epoch": 0.94, "learning_rate": 1.5937838814600648e-05, "loss": 84.5958, "step": 2598 }, { "compression_loss": 86.07173156738281, "epoch": 0.94, "learning_rate": 1.5932417780990242e-05, "loss": 86.688, "step": 2599 }, { "compression_loss": 85.40507507324219, "epoch": 0.94, "learning_rate": 1.5926996747379834e-05, "loss": 86.2609, "step": 2600 }, { "compression_loss": 86.59733581542969, "epoch": 0.94, "learning_rate": 1.5921575713769425e-05, "loss": 87.1174, "step": 2601 }, { "compression_loss": 85.85137939453125, "epoch": 0.94, "learning_rate": 1.5916154680159016e-05, "loss": 86.7138, "step": 2602 }, { "compression_loss": 85.55230712890625, "epoch": 0.94, "learning_rate": 1.5910733646548607e-05, "loss": 86.3221, "step": 2603 }, { "compression_loss": 84.91188049316406, "epoch": 0.94, "learning_rate": 1.59053126129382e-05, "loss": 85.3794, "step": 2604 }, { "compression_loss": 84.41908264160156, "epoch": 0.94, "learning_rate": 1.589989157932779e-05, "loss": 84.8967, "step": 2605 }, { "compression_loss": 83.31057739257812, "epoch": 0.94, "learning_rate": 1.589447054571738e-05, "loss": 83.7442, "step": 2606 }, { "compression_loss": 82.81810760498047, "epoch": 0.94, "learning_rate": 1.5889049512106976e-05, "loss": 83.3736, "step": 2607 }, { "compression_loss": 83.61045837402344, "epoch": 0.94, "learning_rate": 1.5883628478496567e-05, "loss": 84.1542, "step": 2608 }, { "compression_loss": 84.74989318847656, "epoch": 0.94, "learning_rate": 1.587820744488616e-05, "loss": 85.2906, "step": 2609 }, { "compression_loss": 85.11576080322266, "epoch": 0.94, "learning_rate": 1.587278641127575e-05, "loss": 85.5825, "step": 2610 }, { "compression_loss": 85.55763244628906, "epoch": 0.94, "learning_rate": 1.586736537766534e-05, "loss": 86.0278, "step": 2611 }, { "compression_loss": 84.83685302734375, "epoch": 0.94, "learning_rate": 1.5861944344054932e-05, "loss": 85.4776, "step": 2612 }, { "compression_loss": 82.08568572998047, "epoch": 0.94, "learning_rate": 1.5856523310444524e-05, "loss": 82.7587, "step": 2613 }, { "compression_loss": 84.63970947265625, "epoch": 0.94, "learning_rate": 1.5851102276834115e-05, "loss": 85.435, "step": 2614 }, { "compression_loss": 85.78587341308594, "epoch": 0.95, "learning_rate": 1.584568124322371e-05, "loss": 86.6082, "step": 2615 }, { "compression_loss": 86.65483856201172, "epoch": 0.95, "learning_rate": 1.58402602096133e-05, "loss": 87.1394, "step": 2616 }, { "compression_loss": 85.54873657226562, "epoch": 0.95, "learning_rate": 1.5834839176002892e-05, "loss": 86.0521, "step": 2617 }, { "compression_loss": 85.26480102539062, "epoch": 0.95, "learning_rate": 1.5829418142392483e-05, "loss": 85.7499, "step": 2618 }, { "compression_loss": 83.9061508178711, "epoch": 0.95, "learning_rate": 1.5823997108782075e-05, "loss": 84.5334, "step": 2619 }, { "compression_loss": 83.86271667480469, "epoch": 0.95, "learning_rate": 1.5818576075171666e-05, "loss": 84.3022, "step": 2620 }, { "compression_loss": 86.19499206542969, "epoch": 0.95, "learning_rate": 1.5813155041561257e-05, "loss": 86.5013, "step": 2621 }, { "compression_loss": 83.94390106201172, "epoch": 0.95, "learning_rate": 1.580773400795085e-05, "loss": 84.326, "step": 2622 }, { "compression_loss": 83.80081176757812, "epoch": 0.95, "learning_rate": 1.5802312974340443e-05, "loss": 84.4863, "step": 2623 }, { "compression_loss": 86.36105346679688, "epoch": 0.95, "learning_rate": 1.5796891940730034e-05, "loss": 86.8384, "step": 2624 }, { "compression_loss": 85.05619812011719, "epoch": 0.95, "learning_rate": 1.5791470907119626e-05, "loss": 85.5714, "step": 2625 }, { "compression_loss": 84.28675842285156, "epoch": 0.95, "learning_rate": 1.5786049873509217e-05, "loss": 84.9066, "step": 2626 }, { "compression_loss": 83.95225524902344, "epoch": 0.95, "learning_rate": 1.5780628839898808e-05, "loss": 84.5434, "step": 2627 }, { "compression_loss": 83.82612609863281, "epoch": 0.95, "learning_rate": 1.57752078062884e-05, "loss": 84.1419, "step": 2628 }, { "compression_loss": 84.19170379638672, "epoch": 0.95, "learning_rate": 1.576978677267799e-05, "loss": 84.8544, "step": 2629 }, { "compression_loss": 84.42359924316406, "epoch": 0.95, "learning_rate": 1.5764365739067582e-05, "loss": 84.8224, "step": 2630 }, { "compression_loss": 84.6458740234375, "epoch": 0.95, "learning_rate": 1.5758944705457177e-05, "loss": 85.1753, "step": 2631 }, { "compression_loss": 85.68301391601562, "epoch": 0.95, "learning_rate": 1.5753523671846768e-05, "loss": 86.325, "step": 2632 }, { "compression_loss": 84.76922607421875, "epoch": 0.95, "learning_rate": 1.574810263823636e-05, "loss": 85.2785, "step": 2633 }, { "compression_loss": 84.76658630371094, "epoch": 0.95, "learning_rate": 1.574268160462595e-05, "loss": 85.232, "step": 2634 }, { "compression_loss": 84.05110168457031, "epoch": 0.95, "learning_rate": 1.573726057101554e-05, "loss": 84.8236, "step": 2635 }, { "compression_loss": 85.70475769042969, "epoch": 0.95, "learning_rate": 1.5731839537405133e-05, "loss": 86.5363, "step": 2636 }, { "compression_loss": 83.9416275024414, "epoch": 0.95, "learning_rate": 1.5726418503794724e-05, "loss": 84.3896, "step": 2637 }, { "compression_loss": 85.26898193359375, "epoch": 0.95, "learning_rate": 1.5720997470184315e-05, "loss": 85.7395, "step": 2638 }, { "compression_loss": 83.54664611816406, "epoch": 0.95, "learning_rate": 1.571557643657391e-05, "loss": 84.2594, "step": 2639 }, { "compression_loss": 84.42684173583984, "epoch": 0.95, "learning_rate": 1.57101554029635e-05, "loss": 85.0828, "step": 2640 }, { "compression_loss": 84.9552001953125, "epoch": 0.95, "learning_rate": 1.570473436935309e-05, "loss": 85.6981, "step": 2641 }, { "compression_loss": 86.24699401855469, "epoch": 0.95, "learning_rate": 1.569931333574268e-05, "loss": 86.6939, "step": 2642 }, { "compression_loss": 85.193115234375, "epoch": 0.96, "learning_rate": 1.5693892302132272e-05, "loss": 85.6818, "step": 2643 }, { "compression_loss": 84.24517822265625, "epoch": 0.96, "learning_rate": 1.5688471268521863e-05, "loss": 85.0419, "step": 2644 }, { "compression_loss": 85.69401550292969, "epoch": 0.96, "learning_rate": 1.5683050234911454e-05, "loss": 86.5902, "step": 2645 }, { "compression_loss": 84.90972900390625, "epoch": 0.96, "learning_rate": 1.5677629201301046e-05, "loss": 85.559, "step": 2646 }, { "compression_loss": 84.72769165039062, "epoch": 0.96, "learning_rate": 1.567220816769064e-05, "loss": 85.2431, "step": 2647 }, { "compression_loss": 85.6619873046875, "epoch": 0.96, "learning_rate": 1.566678713408023e-05, "loss": 86.4673, "step": 2648 }, { "compression_loss": 84.3076171875, "epoch": 0.96, "learning_rate": 1.5661366100469823e-05, "loss": 84.8819, "step": 2649 }, { "compression_loss": 85.27986907958984, "epoch": 0.96, "learning_rate": 1.5655945066859414e-05, "loss": 85.7775, "step": 2650 }, { "compression_loss": 84.579345703125, "epoch": 0.96, "learning_rate": 1.5650524033249005e-05, "loss": 85.0653, "step": 2651 }, { "compression_loss": 86.9261703491211, "epoch": 0.96, "learning_rate": 1.5645102999638597e-05, "loss": 87.8499, "step": 2652 }, { "compression_loss": 84.45365905761719, "epoch": 0.96, "learning_rate": 1.5639681966028188e-05, "loss": 84.7783, "step": 2653 }, { "compression_loss": 84.47492980957031, "epoch": 0.96, "learning_rate": 1.563426093241778e-05, "loss": 84.8939, "step": 2654 }, { "compression_loss": 85.53033447265625, "epoch": 0.96, "learning_rate": 1.5628839898807374e-05, "loss": 85.9526, "step": 2655 }, { "compression_loss": 83.20867919921875, "epoch": 0.96, "learning_rate": 1.5623418865196965e-05, "loss": 83.6116, "step": 2656 }, { "compression_loss": 82.68685913085938, "epoch": 0.96, "learning_rate": 1.5617997831586556e-05, "loss": 83.0232, "step": 2657 }, { "compression_loss": 84.5362777709961, "epoch": 0.96, "learning_rate": 1.5612576797976148e-05, "loss": 85.1295, "step": 2658 }, { "compression_loss": 81.9798583984375, "epoch": 0.96, "learning_rate": 1.560715576436574e-05, "loss": 82.2742, "step": 2659 }, { "compression_loss": 84.55927276611328, "epoch": 0.96, "learning_rate": 1.560173473075533e-05, "loss": 85.4351, "step": 2660 }, { "compression_loss": 84.60536193847656, "epoch": 0.96, "learning_rate": 1.559631369714492e-05, "loss": 85.0066, "step": 2661 }, { "compression_loss": 84.19912719726562, "epoch": 0.96, "learning_rate": 1.5590892663534513e-05, "loss": 84.7609, "step": 2662 }, { "compression_loss": 84.40074920654297, "epoch": 0.96, "learning_rate": 1.5585471629924107e-05, "loss": 85.214, "step": 2663 }, { "compression_loss": 85.58430480957031, "epoch": 0.96, "learning_rate": 1.55800505963137e-05, "loss": 85.9817, "step": 2664 }, { "compression_loss": 83.94121551513672, "epoch": 0.96, "learning_rate": 1.557462956270329e-05, "loss": 84.3364, "step": 2665 }, { "compression_loss": 86.39655303955078, "epoch": 0.96, "learning_rate": 1.556920852909288e-05, "loss": 87.0027, "step": 2666 }, { "compression_loss": 84.35736083984375, "epoch": 0.96, "learning_rate": 1.5563787495482473e-05, "loss": 84.7114, "step": 2667 }, { "compression_loss": 82.826904296875, "epoch": 0.96, "learning_rate": 1.5558366461872064e-05, "loss": 83.2368, "step": 2668 }, { "compression_loss": 83.03487396240234, "epoch": 0.96, "learning_rate": 1.5552945428261655e-05, "loss": 83.5861, "step": 2669 }, { "compression_loss": 84.66354370117188, "epoch": 0.96, "learning_rate": 1.5547524394651246e-05, "loss": 85.626, "step": 2670 }, { "compression_loss": 85.69612121582031, "epoch": 0.97, "learning_rate": 1.554210336104084e-05, "loss": 86.5278, "step": 2671 }, { "compression_loss": 83.3206787109375, "epoch": 0.97, "learning_rate": 1.5536682327430432e-05, "loss": 83.6848, "step": 2672 }, { "compression_loss": 82.05732727050781, "epoch": 0.97, "learning_rate": 1.5531261293820024e-05, "loss": 82.3408, "step": 2673 }, { "compression_loss": 83.71939086914062, "epoch": 0.97, "learning_rate": 1.5525840260209615e-05, "loss": 84.1414, "step": 2674 }, { "compression_loss": 83.44413757324219, "epoch": 0.97, "learning_rate": 1.5520419226599206e-05, "loss": 83.9163, "step": 2675 }, { "compression_loss": 84.13740539550781, "epoch": 0.97, "learning_rate": 1.5514998192988797e-05, "loss": 84.523, "step": 2676 }, { "compression_loss": 82.99435424804688, "epoch": 0.97, "learning_rate": 1.550957715937839e-05, "loss": 83.2981, "step": 2677 }, { "compression_loss": 85.216552734375, "epoch": 0.97, "learning_rate": 1.550415612576798e-05, "loss": 85.6353, "step": 2678 }, { "compression_loss": 85.11637878417969, "epoch": 0.97, "learning_rate": 1.5498735092157575e-05, "loss": 85.7182, "step": 2679 }, { "compression_loss": 86.57014465332031, "epoch": 0.97, "learning_rate": 1.5493314058547166e-05, "loss": 87.6357, "step": 2680 }, { "compression_loss": 83.44576263427734, "epoch": 0.97, "learning_rate": 1.5487893024936757e-05, "loss": 83.8207, "step": 2681 }, { "compression_loss": 87.11689758300781, "epoch": 0.97, "learning_rate": 1.548247199132635e-05, "loss": 87.852, "step": 2682 }, { "compression_loss": 81.67670440673828, "epoch": 0.97, "learning_rate": 1.547705095771594e-05, "loss": 82.0796, "step": 2683 }, { "compression_loss": 86.12721252441406, "epoch": 0.97, "learning_rate": 1.5471629924105528e-05, "loss": 86.9144, "step": 2684 }, { "compression_loss": 84.87325286865234, "epoch": 0.97, "learning_rate": 1.546620889049512e-05, "loss": 85.4516, "step": 2685 }, { "compression_loss": 84.80799865722656, "epoch": 0.97, "learning_rate": 1.546078785688471e-05, "loss": 85.0337, "step": 2686 }, { "compression_loss": 82.45205688476562, "epoch": 0.97, "learning_rate": 1.5455366823274305e-05, "loss": 82.7792, "step": 2687 }, { "compression_loss": 85.69965362548828, "epoch": 0.97, "learning_rate": 1.5449945789663896e-05, "loss": 86.1508, "step": 2688 }, { "compression_loss": 82.99661254882812, "epoch": 0.97, "learning_rate": 1.5444524756053487e-05, "loss": 83.405, "step": 2689 }, { "compression_loss": 84.40739440917969, "epoch": 0.97, "learning_rate": 1.543910372244308e-05, "loss": 84.8328, "step": 2690 }, { "compression_loss": 84.36257934570312, "epoch": 0.97, "learning_rate": 1.543368268883267e-05, "loss": 85.0729, "step": 2691 }, { "compression_loss": 84.36119842529297, "epoch": 0.97, "learning_rate": 1.542826165522226e-05, "loss": 85.03, "step": 2692 }, { "compression_loss": 85.07585144042969, "epoch": 0.97, "learning_rate": 1.5422840621611852e-05, "loss": 85.707, "step": 2693 }, { "compression_loss": 83.45337677001953, "epoch": 0.97, "learning_rate": 1.5417419588001444e-05, "loss": 84.2062, "step": 2694 }, { "compression_loss": 84.54654693603516, "epoch": 0.97, "learning_rate": 1.541199855439104e-05, "loss": 85.0878, "step": 2695 }, { "compression_loss": 84.16435241699219, "epoch": 0.97, "learning_rate": 1.540657752078063e-05, "loss": 84.6027, "step": 2696 }, { "compression_loss": 85.08702087402344, "epoch": 0.97, "learning_rate": 1.540115648717022e-05, "loss": 85.7508, "step": 2697 }, { "compression_loss": 82.67284393310547, "epoch": 0.98, "learning_rate": 1.5395735453559812e-05, "loss": 83.4068, "step": 2698 }, { "compression_loss": 84.23304748535156, "epoch": 0.98, "learning_rate": 1.5390314419949403e-05, "loss": 84.9924, "step": 2699 }, { "compression_loss": 84.28909301757812, "epoch": 0.98, "learning_rate": 1.5384893386338995e-05, "loss": 84.6644, "step": 2700 }, { "compression_loss": 83.11312866210938, "epoch": 0.98, "learning_rate": 1.5379472352728586e-05, "loss": 83.555, "step": 2701 }, { "compression_loss": 85.61949157714844, "epoch": 0.98, "learning_rate": 1.5374051319118177e-05, "loss": 86.2017, "step": 2702 }, { "compression_loss": 84.27027893066406, "epoch": 0.98, "learning_rate": 1.5368630285507772e-05, "loss": 84.7124, "step": 2703 }, { "compression_loss": 84.34115600585938, "epoch": 0.98, "learning_rate": 1.5363209251897363e-05, "loss": 84.841, "step": 2704 }, { "compression_loss": 85.71672058105469, "epoch": 0.98, "learning_rate": 1.5357788218286954e-05, "loss": 86.2613, "step": 2705 }, { "compression_loss": 84.35554504394531, "epoch": 0.98, "learning_rate": 1.5352367184676546e-05, "loss": 84.9832, "step": 2706 }, { "compression_loss": 85.6414566040039, "epoch": 0.98, "learning_rate": 1.5346946151066137e-05, "loss": 85.929, "step": 2707 }, { "compression_loss": 84.40217590332031, "epoch": 0.98, "learning_rate": 1.5341525117455728e-05, "loss": 84.7813, "step": 2708 }, { "compression_loss": 84.49473571777344, "epoch": 0.98, "learning_rate": 1.533610408384532e-05, "loss": 85.1124, "step": 2709 }, { "compression_loss": 82.21752166748047, "epoch": 0.98, "learning_rate": 1.533068305023491e-05, "loss": 82.6622, "step": 2710 }, { "compression_loss": 86.21784973144531, "epoch": 0.98, "learning_rate": 1.5325262016624505e-05, "loss": 86.5789, "step": 2711 }, { "compression_loss": 83.91838073730469, "epoch": 0.98, "learning_rate": 1.5319840983014097e-05, "loss": 84.5934, "step": 2712 }, { "compression_loss": 85.13056945800781, "epoch": 0.98, "learning_rate": 1.5314419949403688e-05, "loss": 85.7943, "step": 2713 }, { "compression_loss": 84.38031005859375, "epoch": 0.98, "learning_rate": 1.530899891579328e-05, "loss": 85.0146, "step": 2714 }, { "compression_loss": 85.37968444824219, "epoch": 0.98, "learning_rate": 1.530357788218287e-05, "loss": 85.9092, "step": 2715 }, { "compression_loss": 84.00294494628906, "epoch": 0.98, "learning_rate": 1.5298156848572462e-05, "loss": 85.2398, "step": 2716 }, { "compression_loss": 85.07756805419922, "epoch": 0.98, "learning_rate": 1.5292735814962053e-05, "loss": 85.6356, "step": 2717 }, { "compression_loss": 83.35316467285156, "epoch": 0.98, "learning_rate": 1.5287314781351644e-05, "loss": 84.6173, "step": 2718 }, { "compression_loss": 83.470703125, "epoch": 0.98, "learning_rate": 1.528189374774124e-05, "loss": 84.2198, "step": 2719 }, { "compression_loss": 85.02030944824219, "epoch": 0.98, "learning_rate": 1.527647271413083e-05, "loss": 85.6304, "step": 2720 }, { "compression_loss": 83.43875122070312, "epoch": 0.98, "learning_rate": 1.527105168052042e-05, "loss": 83.7345, "step": 2721 }, { "compression_loss": 85.14555358886719, "epoch": 0.98, "learning_rate": 1.5265630646910013e-05, "loss": 85.5167, "step": 2722 }, { "compression_loss": 86.23530578613281, "epoch": 0.98, "learning_rate": 1.5260209613299604e-05, "loss": 86.6991, "step": 2723 }, { "compression_loss": 84.37177276611328, "epoch": 0.98, "learning_rate": 1.5254788579689194e-05, "loss": 84.8873, "step": 2724 }, { "compression_loss": 82.00285339355469, "epoch": 0.98, "learning_rate": 1.5249367546078785e-05, "loss": 82.5408, "step": 2725 }, { "compression_loss": 84.28521728515625, "epoch": 0.99, "learning_rate": 1.5243946512468376e-05, "loss": 84.8911, "step": 2726 }, { "compression_loss": 84.44064331054688, "epoch": 0.99, "learning_rate": 1.5238525478857971e-05, "loss": 85.1277, "step": 2727 }, { "compression_loss": 83.8371353149414, "epoch": 0.99, "learning_rate": 1.5233104445247562e-05, "loss": 84.5481, "step": 2728 }, { "compression_loss": 86.1279296875, "epoch": 0.99, "learning_rate": 1.5227683411637153e-05, "loss": 86.7342, "step": 2729 }, { "compression_loss": 85.56608581542969, "epoch": 0.99, "learning_rate": 1.5222262378026745e-05, "loss": 86.1521, "step": 2730 }, { "compression_loss": 84.41986083984375, "epoch": 0.99, "learning_rate": 1.5216841344416336e-05, "loss": 84.8557, "step": 2731 }, { "compression_loss": 85.5318832397461, "epoch": 0.99, "learning_rate": 1.5211420310805927e-05, "loss": 86.0652, "step": 2732 }, { "compression_loss": 84.52315521240234, "epoch": 0.99, "learning_rate": 1.5205999277195518e-05, "loss": 85.2119, "step": 2733 }, { "compression_loss": 83.979248046875, "epoch": 0.99, "learning_rate": 1.520057824358511e-05, "loss": 84.5713, "step": 2734 }, { "compression_loss": 84.15145111083984, "epoch": 0.99, "learning_rate": 1.5195157209974704e-05, "loss": 84.8064, "step": 2735 }, { "compression_loss": 83.29281616210938, "epoch": 0.99, "learning_rate": 1.5189736176364296e-05, "loss": 83.7631, "step": 2736 }, { "compression_loss": 84.55465698242188, "epoch": 0.99, "learning_rate": 1.5184315142753887e-05, "loss": 84.9917, "step": 2737 }, { "compression_loss": 83.81919860839844, "epoch": 0.99, "learning_rate": 1.5178894109143478e-05, "loss": 84.4722, "step": 2738 }, { "compression_loss": 85.29952239990234, "epoch": 0.99, "learning_rate": 1.5173473075533068e-05, "loss": 85.55, "step": 2739 }, { "compression_loss": 84.35635375976562, "epoch": 0.99, "learning_rate": 1.5168052041922659e-05, "loss": 84.6185, "step": 2740 }, { "compression_loss": 85.44465637207031, "epoch": 0.99, "learning_rate": 1.516263100831225e-05, "loss": 85.8547, "step": 2741 }, { "compression_loss": 83.92649841308594, "epoch": 0.99, "learning_rate": 1.5157209974701842e-05, "loss": 84.3091, "step": 2742 }, { "compression_loss": 84.5505599975586, "epoch": 0.99, "learning_rate": 1.5151788941091436e-05, "loss": 85.1331, "step": 2743 }, { "compression_loss": 84.53800964355469, "epoch": 0.99, "learning_rate": 1.5146367907481028e-05, "loss": 85.2718, "step": 2744 }, { "compression_loss": 86.03894805908203, "epoch": 0.99, "learning_rate": 1.5140946873870619e-05, "loss": 86.9691, "step": 2745 }, { "compression_loss": 84.70919036865234, "epoch": 0.99, "learning_rate": 1.513552584026021e-05, "loss": 85.1013, "step": 2746 }, { "compression_loss": 84.6736831665039, "epoch": 0.99, "learning_rate": 1.5130104806649801e-05, "loss": 85.0886, "step": 2747 }, { "compression_loss": 84.56513977050781, "epoch": 0.99, "learning_rate": 1.5124683773039393e-05, "loss": 85.0795, "step": 2748 }, { "compression_loss": 82.4090576171875, "epoch": 0.99, "learning_rate": 1.5119262739428984e-05, "loss": 82.9446, "step": 2749 }, { "compression_loss": 86.54396057128906, "epoch": 0.99, "learning_rate": 1.5113841705818575e-05, "loss": 87.3825, "step": 2750 }, { "epoch": 0.99, "eval_exact_match": 86.85903500473037, "eval_f1": 92.9356430682311, "step": 2750 }, { "compression_loss": 84.88859558105469, "epoch": 0.99, "learning_rate": 1.510842067220817e-05, "loss": 85.4541, "step": 2751 }, { "compression_loss": 87.44498443603516, "epoch": 0.99, "learning_rate": 1.5102999638597761e-05, "loss": 88.3074, "step": 2752 }, { "compression_loss": 84.43220520019531, "epoch": 0.99, "learning_rate": 1.5097578604987352e-05, "loss": 85.1008, "step": 2753 }, { "compression_loss": 84.66212463378906, "epoch": 1.0, "learning_rate": 1.5092157571376944e-05, "loss": 85.0522, "step": 2754 }, { "compression_loss": 84.80653381347656, "epoch": 1.0, "learning_rate": 1.5086736537766535e-05, "loss": 85.554, "step": 2755 }, { "compression_loss": 84.0218276977539, "epoch": 1.0, "learning_rate": 1.5081315504156126e-05, "loss": 84.4654, "step": 2756 }, { "compression_loss": 84.71709442138672, "epoch": 1.0, "learning_rate": 1.5075894470545717e-05, "loss": 85.1446, "step": 2757 }, { "compression_loss": 82.4185791015625, "epoch": 1.0, "learning_rate": 1.5070473436935309e-05, "loss": 82.7648, "step": 2758 }, { "compression_loss": 84.55377197265625, "epoch": 1.0, "learning_rate": 1.5065052403324903e-05, "loss": 85.161, "step": 2759 }, { "compression_loss": 83.76071166992188, "epoch": 1.0, "learning_rate": 1.5059631369714493e-05, "loss": 84.1809, "step": 2760 }, { "compression_loss": 85.12535095214844, "epoch": 1.0, "learning_rate": 1.5054210336104084e-05, "loss": 85.5466, "step": 2761 }, { "compression_loss": 84.63972473144531, "epoch": 1.0, "learning_rate": 1.5048789302493676e-05, "loss": 85.0158, "step": 2762 }, { "compression_loss": 84.51396179199219, "epoch": 1.0, "learning_rate": 1.5043368268883267e-05, "loss": 84.9399, "step": 2763 }, { "compression_loss": 85.67932891845703, "epoch": 1.0, "learning_rate": 1.5037947235272858e-05, "loss": 86.0849, "step": 2764 }, { "compression_loss": 84.31236267089844, "epoch": 1.0, "learning_rate": 1.503252620166245e-05, "loss": 84.8946, "step": 2765 }, { "compression_loss": 84.94093322753906, "epoch": 1.0, "learning_rate": 1.502710516805204e-05, "loss": 85.4911, "step": 2766 }, { "compression_loss": 82.287353515625, "epoch": 1.0, "learning_rate": 1.5021684134441635e-05, "loss": 82.6368, "step": 2767 }, { "compression_loss": 83.85541534423828, "epoch": 1.0, "learning_rate": 1.5016263100831227e-05, "loss": 84.1138, "step": 2768 }, { "compression_loss": 85.96159362792969, "epoch": 1.0, "learning_rate": 1.5010842067220818e-05, "loss": 86.3078, "step": 2769 }, { "compression_loss": 85.88269805908203, "epoch": 1.0, "learning_rate": 1.5005421033610409e-05, "loss": 86.5642, "step": 2770 }, { "compression_loss": 83.29478454589844, "epoch": 1.0, "learning_rate": 1.5e-05, "loss": 83.7907, "step": 2771 }, { "compression_loss": 83.0792465209961, "epoch": 1.0, "learning_rate": 1.4994578966389592e-05, "loss": 83.5285, "step": 2772 }, { "compression_loss": 82.3663558959961, "epoch": 1.0, "learning_rate": 1.4989157932779183e-05, "loss": 82.6084, "step": 2773 }, { "compression_loss": 84.66055297851562, "epoch": 1.0, "learning_rate": 1.4983736899168776e-05, "loss": 85.2662, "step": 2774 }, { "compression_loss": 85.13783264160156, "epoch": 1.0, "learning_rate": 1.4978315865558367e-05, "loss": 85.9488, "step": 2775 }, { "compression_loss": 83.45896911621094, "epoch": 1.0, "learning_rate": 1.4972894831947958e-05, "loss": 83.8596, "step": 2776 }, { "compression_loss": 83.53733825683594, "epoch": 1.0, "learning_rate": 1.496747379833755e-05, "loss": 83.8899, "step": 2777 }, { "compression_loss": 85.24707794189453, "epoch": 1.0, "learning_rate": 1.4962052764727143e-05, "loss": 85.5101, "step": 2778 }, { "compression_loss": 82.89987182617188, "epoch": 1.0, "learning_rate": 1.4956631731116734e-05, "loss": 83.5486, "step": 2779 }, { "compression_loss": 82.8880386352539, "epoch": 1.0, "learning_rate": 1.4951210697506325e-05, "loss": 83.3531, "step": 2780 }, { "compression_loss": 84.65835571289062, "epoch": 1.01, "learning_rate": 1.4945789663895916e-05, "loss": 85.0001, "step": 2781 }, { "compression_loss": 84.20577239990234, "epoch": 1.01, "learning_rate": 1.4940368630285508e-05, "loss": 84.5348, "step": 2782 }, { "compression_loss": 84.04437255859375, "epoch": 1.01, "learning_rate": 1.4934947596675099e-05, "loss": 84.4628, "step": 2783 }, { "compression_loss": 84.15535736083984, "epoch": 1.01, "learning_rate": 1.492952656306469e-05, "loss": 84.7705, "step": 2784 }, { "compression_loss": 84.57269287109375, "epoch": 1.01, "learning_rate": 1.4924105529454282e-05, "loss": 85.0519, "step": 2785 }, { "compression_loss": 85.01579284667969, "epoch": 1.01, "learning_rate": 1.4918684495843875e-05, "loss": 85.6644, "step": 2786 }, { "compression_loss": 84.93170166015625, "epoch": 1.01, "learning_rate": 1.4913263462233466e-05, "loss": 85.2465, "step": 2787 }, { "compression_loss": 84.23555755615234, "epoch": 1.01, "learning_rate": 1.4907842428623057e-05, "loss": 84.5745, "step": 2788 }, { "compression_loss": 85.3051986694336, "epoch": 1.01, "learning_rate": 1.4902421395012648e-05, "loss": 85.8277, "step": 2789 }, { "compression_loss": 83.82972717285156, "epoch": 1.01, "learning_rate": 1.4897000361402241e-05, "loss": 84.2704, "step": 2790 }, { "compression_loss": 84.88539123535156, "epoch": 1.01, "learning_rate": 1.4891579327791833e-05, "loss": 85.4477, "step": 2791 }, { "compression_loss": 83.72135925292969, "epoch": 1.01, "learning_rate": 1.4886158294181424e-05, "loss": 83.9678, "step": 2792 }, { "compression_loss": 84.95622253417969, "epoch": 1.01, "learning_rate": 1.4880737260571015e-05, "loss": 85.3917, "step": 2793 }, { "compression_loss": 85.21983337402344, "epoch": 1.01, "learning_rate": 1.4875316226960608e-05, "loss": 86.4434, "step": 2794 }, { "compression_loss": 85.69450378417969, "epoch": 1.01, "learning_rate": 1.48698951933502e-05, "loss": 86.1808, "step": 2795 }, { "compression_loss": 82.8948745727539, "epoch": 1.01, "learning_rate": 1.486447415973979e-05, "loss": 83.2718, "step": 2796 }, { "compression_loss": 85.26873016357422, "epoch": 1.01, "learning_rate": 1.4859053126129382e-05, "loss": 85.6097, "step": 2797 }, { "compression_loss": 82.94911193847656, "epoch": 1.01, "learning_rate": 1.4853632092518975e-05, "loss": 83.5873, "step": 2798 }, { "compression_loss": 84.74934387207031, "epoch": 1.01, "learning_rate": 1.4848211058908566e-05, "loss": 85.2209, "step": 2799 }, { "compression_loss": 83.90336608886719, "epoch": 1.01, "learning_rate": 1.4842790025298157e-05, "loss": 84.2214, "step": 2800 }, { "compression_loss": 84.7835464477539, "epoch": 1.01, "learning_rate": 1.4837368991687749e-05, "loss": 85.1673, "step": 2801 }, { "compression_loss": 83.15460205078125, "epoch": 1.01, "learning_rate": 1.4831947958077342e-05, "loss": 83.4092, "step": 2802 }, { "compression_loss": 85.38360595703125, "epoch": 1.01, "learning_rate": 1.4826526924466933e-05, "loss": 85.772, "step": 2803 }, { "compression_loss": 84.88453674316406, "epoch": 1.01, "learning_rate": 1.4821105890856523e-05, "loss": 85.4417, "step": 2804 }, { "compression_loss": 83.16944885253906, "epoch": 1.01, "learning_rate": 1.4815684857246114e-05, "loss": 83.5905, "step": 2805 }, { "compression_loss": 84.02969360351562, "epoch": 1.01, "learning_rate": 1.4810263823635707e-05, "loss": 84.8157, "step": 2806 }, { "compression_loss": 83.47278594970703, "epoch": 1.01, "learning_rate": 1.4804842790025298e-05, "loss": 84.0247, "step": 2807 }, { "compression_loss": 84.73721313476562, "epoch": 1.01, "learning_rate": 1.479942175641489e-05, "loss": 85.3469, "step": 2808 }, { "compression_loss": 83.84883880615234, "epoch": 1.02, "learning_rate": 1.479400072280448e-05, "loss": 84.4804, "step": 2809 }, { "compression_loss": 84.01741027832031, "epoch": 1.02, "learning_rate": 1.4788579689194074e-05, "loss": 84.6284, "step": 2810 }, { "compression_loss": 85.21453857421875, "epoch": 1.02, "learning_rate": 1.4783158655583665e-05, "loss": 85.9626, "step": 2811 }, { "compression_loss": 83.67230987548828, "epoch": 1.02, "learning_rate": 1.4777737621973256e-05, "loss": 84.0158, "step": 2812 }, { "compression_loss": 84.1436767578125, "epoch": 1.02, "learning_rate": 1.4772316588362847e-05, "loss": 84.6275, "step": 2813 }, { "compression_loss": 85.24539184570312, "epoch": 1.02, "learning_rate": 1.476689555475244e-05, "loss": 85.9432, "step": 2814 }, { "compression_loss": 83.61759185791016, "epoch": 1.02, "learning_rate": 1.476689555475244e-05, "loss": 84.194, "step": 2815 }, { "compression_loss": 84.505126953125, "epoch": 1.02, "learning_rate": 1.4761474521142032e-05, "loss": 85.0064, "step": 2816 }, { "compression_loss": 84.22279357910156, "epoch": 1.02, "learning_rate": 1.4756053487531623e-05, "loss": 84.9491, "step": 2817 }, { "compression_loss": 83.55370330810547, "epoch": 1.02, "learning_rate": 1.4750632453921214e-05, "loss": 83.9763, "step": 2818 }, { "compression_loss": 83.91438293457031, "epoch": 1.02, "learning_rate": 1.4745211420310807e-05, "loss": 84.2868, "step": 2819 }, { "compression_loss": 83.46112060546875, "epoch": 1.02, "learning_rate": 1.4739790386700398e-05, "loss": 83.8742, "step": 2820 }, { "compression_loss": 84.34194946289062, "epoch": 1.02, "learning_rate": 1.473436935308999e-05, "loss": 84.8905, "step": 2821 }, { "compression_loss": 84.64505004882812, "epoch": 1.02, "learning_rate": 1.4728948319479581e-05, "loss": 85.0303, "step": 2822 }, { "compression_loss": 85.14433288574219, "epoch": 1.02, "learning_rate": 1.4723527285869174e-05, "loss": 85.7286, "step": 2823 }, { "compression_loss": 86.64788818359375, "epoch": 1.02, "learning_rate": 1.4718106252258765e-05, "loss": 87.1721, "step": 2824 }, { "compression_loss": 85.67556762695312, "epoch": 1.02, "learning_rate": 1.4712685218648356e-05, "loss": 86.4934, "step": 2825 }, { "compression_loss": 83.12922668457031, "epoch": 1.02, "learning_rate": 1.4707264185037948e-05, "loss": 83.3625, "step": 2826 }, { "compression_loss": 82.4727783203125, "epoch": 1.02, "learning_rate": 1.4701843151427539e-05, "loss": 82.8738, "step": 2827 }, { "compression_loss": 84.90737915039062, "epoch": 1.02, "learning_rate": 1.469642211781713e-05, "loss": 85.2759, "step": 2828 }, { "compression_loss": 83.87376403808594, "epoch": 1.02, "learning_rate": 1.4691001084206722e-05, "loss": 84.2605, "step": 2829 }, { "compression_loss": 83.84677124023438, "epoch": 1.02, "learning_rate": 1.4685580050596313e-05, "loss": 84.3011, "step": 2830 }, { "compression_loss": 85.17512512207031, "epoch": 1.02, "learning_rate": 1.4680159016985906e-05, "loss": 85.6254, "step": 2831 }, { "compression_loss": 85.88882446289062, "epoch": 1.02, "learning_rate": 1.4674737983375497e-05, "loss": 86.4022, "step": 2832 }, { "compression_loss": 83.63390350341797, "epoch": 1.02, "learning_rate": 1.4669316949765088e-05, "loss": 84.0104, "step": 2833 }, { "compression_loss": 85.27865600585938, "epoch": 1.02, "learning_rate": 1.466389591615468e-05, "loss": 85.762, "step": 2834 }, { "compression_loss": 84.53759765625, "epoch": 1.02, "learning_rate": 1.4658474882544273e-05, "loss": 85.5075, "step": 2835 }, { "compression_loss": 84.24067687988281, "epoch": 1.02, "learning_rate": 1.4653053848933864e-05, "loss": 84.8041, "step": 2836 }, { "compression_loss": 84.60271453857422, "epoch": 1.03, "learning_rate": 1.4647632815323455e-05, "loss": 85.0174, "step": 2837 }, { "compression_loss": 84.07368469238281, "epoch": 1.03, "learning_rate": 1.4642211781713046e-05, "loss": 84.4248, "step": 2838 }, { "compression_loss": 83.70211029052734, "epoch": 1.03, "learning_rate": 1.463679074810264e-05, "loss": 84.4383, "step": 2839 }, { "compression_loss": 83.97952270507812, "epoch": 1.03, "learning_rate": 1.463136971449223e-05, "loss": 84.4011, "step": 2840 }, { "compression_loss": 84.17816162109375, "epoch": 1.03, "learning_rate": 1.4625948680881822e-05, "loss": 84.5472, "step": 2841 }, { "compression_loss": 83.32171630859375, "epoch": 1.03, "learning_rate": 1.4620527647271413e-05, "loss": 83.7382, "step": 2842 }, { "compression_loss": 85.50838470458984, "epoch": 1.03, "learning_rate": 1.4615106613661006e-05, "loss": 86.0308, "step": 2843 }, { "compression_loss": 82.82868957519531, "epoch": 1.03, "learning_rate": 1.4609685580050597e-05, "loss": 83.1089, "step": 2844 }, { "compression_loss": 83.36868286132812, "epoch": 1.03, "learning_rate": 1.4604264546440189e-05, "loss": 83.7752, "step": 2845 }, { "compression_loss": 83.26683044433594, "epoch": 1.03, "learning_rate": 1.459884351282978e-05, "loss": 83.6453, "step": 2846 }, { "compression_loss": 83.8846664428711, "epoch": 1.03, "learning_rate": 1.4593422479219373e-05, "loss": 84.3428, "step": 2847 }, { "compression_loss": 84.95441436767578, "epoch": 1.03, "learning_rate": 1.4588001445608964e-05, "loss": 85.3769, "step": 2848 }, { "compression_loss": 83.50224304199219, "epoch": 1.03, "learning_rate": 1.4582580411998554e-05, "loss": 83.6957, "step": 2849 }, { "compression_loss": 85.87486267089844, "epoch": 1.03, "learning_rate": 1.4577159378388145e-05, "loss": 86.3635, "step": 2850 }, { "compression_loss": 85.1000747680664, "epoch": 1.03, "learning_rate": 1.4571738344777738e-05, "loss": 85.5794, "step": 2851 }, { "compression_loss": 83.74165344238281, "epoch": 1.03, "learning_rate": 1.456631731116733e-05, "loss": 83.9671, "step": 2852 }, { "compression_loss": 85.41387939453125, "epoch": 1.03, "learning_rate": 1.456089627755692e-05, "loss": 85.9563, "step": 2853 }, { "compression_loss": 86.15290832519531, "epoch": 1.03, "learning_rate": 1.4555475243946512e-05, "loss": 86.7303, "step": 2854 }, { "compression_loss": 83.208984375, "epoch": 1.03, "learning_rate": 1.4550054210336105e-05, "loss": 83.6423, "step": 2855 }, { "compression_loss": 85.3936538696289, "epoch": 1.03, "learning_rate": 1.4544633176725696e-05, "loss": 85.9778, "step": 2856 }, { "compression_loss": 83.93507385253906, "epoch": 1.03, "learning_rate": 1.4539212143115287e-05, "loss": 84.4797, "step": 2857 }, { "compression_loss": 82.84221649169922, "epoch": 1.03, "learning_rate": 1.4533791109504879e-05, "loss": 83.3116, "step": 2858 }, { "compression_loss": 86.89672088623047, "epoch": 1.03, "learning_rate": 1.4528370075894472e-05, "loss": 87.527, "step": 2859 }, { "compression_loss": 85.49610137939453, "epoch": 1.03, "learning_rate": 1.4522949042284063e-05, "loss": 85.9444, "step": 2860 }, { "compression_loss": 84.28730773925781, "epoch": 1.03, "learning_rate": 1.4517528008673654e-05, "loss": 84.5527, "step": 2861 }, { "compression_loss": 83.13758850097656, "epoch": 1.03, "learning_rate": 1.4512106975063245e-05, "loss": 83.3764, "step": 2862 }, { "compression_loss": 85.01494598388672, "epoch": 1.03, "learning_rate": 1.4506685941452838e-05, "loss": 85.307, "step": 2863 }, { "compression_loss": 83.88414001464844, "epoch": 1.04, "learning_rate": 1.450126490784243e-05, "loss": 84.3549, "step": 2864 }, { "compression_loss": 85.00325775146484, "epoch": 1.04, "learning_rate": 1.4495843874232021e-05, "loss": 85.4238, "step": 2865 }, { "compression_loss": 84.31089782714844, "epoch": 1.04, "learning_rate": 1.4490422840621612e-05, "loss": 84.8108, "step": 2866 }, { "compression_loss": 84.87890625, "epoch": 1.04, "learning_rate": 1.4485001807011205e-05, "loss": 85.1989, "step": 2867 }, { "compression_loss": 84.22257232666016, "epoch": 1.04, "learning_rate": 1.4479580773400796e-05, "loss": 84.649, "step": 2868 }, { "compression_loss": 83.87982177734375, "epoch": 1.04, "learning_rate": 1.4474159739790388e-05, "loss": 84.4018, "step": 2869 }, { "compression_loss": 83.89234924316406, "epoch": 1.04, "learning_rate": 1.4468738706179979e-05, "loss": 84.6621, "step": 2870 }, { "compression_loss": 83.8968734741211, "epoch": 1.04, "learning_rate": 1.446331767256957e-05, "loss": 84.3344, "step": 2871 }, { "compression_loss": 84.54133605957031, "epoch": 1.04, "learning_rate": 1.4457896638959161e-05, "loss": 84.9303, "step": 2872 }, { "compression_loss": 85.20521545410156, "epoch": 1.04, "learning_rate": 1.4452475605348753e-05, "loss": 85.8492, "step": 2873 }, { "compression_loss": 84.99789428710938, "epoch": 1.04, "learning_rate": 1.4447054571738344e-05, "loss": 85.496, "step": 2874 }, { "compression_loss": 84.66571807861328, "epoch": 1.04, "learning_rate": 1.4441633538127937e-05, "loss": 85.0817, "step": 2875 }, { "compression_loss": 84.75926208496094, "epoch": 1.04, "learning_rate": 1.4436212504517528e-05, "loss": 85.2024, "step": 2876 }, { "compression_loss": 84.3006591796875, "epoch": 1.04, "learning_rate": 1.443079147090712e-05, "loss": 84.9326, "step": 2877 }, { "compression_loss": 85.6314697265625, "epoch": 1.04, "learning_rate": 1.442537043729671e-05, "loss": 86.1502, "step": 2878 }, { "compression_loss": 85.62997436523438, "epoch": 1.04, "learning_rate": 1.4419949403686304e-05, "loss": 85.9154, "step": 2879 }, { "compression_loss": 86.54203033447266, "epoch": 1.04, "learning_rate": 1.4414528370075895e-05, "loss": 86.9557, "step": 2880 }, { "compression_loss": 84.23023986816406, "epoch": 1.04, "learning_rate": 1.4409107336465486e-05, "loss": 84.597, "step": 2881 }, { "compression_loss": 84.2055892944336, "epoch": 1.04, "learning_rate": 1.4403686302855078e-05, "loss": 84.7301, "step": 2882 }, { "compression_loss": 84.96717834472656, "epoch": 1.04, "learning_rate": 1.439826526924467e-05, "loss": 85.2552, "step": 2883 }, { "compression_loss": 83.97416687011719, "epoch": 1.04, "learning_rate": 1.4392844235634262e-05, "loss": 84.6282, "step": 2884 }, { "compression_loss": 83.32499694824219, "epoch": 1.04, "learning_rate": 1.4387423202023853e-05, "loss": 83.7151, "step": 2885 }, { "compression_loss": 85.95823669433594, "epoch": 1.04, "learning_rate": 1.4382002168413444e-05, "loss": 86.3434, "step": 2886 }, { "compression_loss": 83.49003601074219, "epoch": 1.04, "learning_rate": 1.4376581134803037e-05, "loss": 84.1081, "step": 2887 }, { "compression_loss": 84.27690124511719, "epoch": 1.04, "learning_rate": 1.4371160101192629e-05, "loss": 84.6396, "step": 2888 }, { "compression_loss": 86.11274719238281, "epoch": 1.04, "learning_rate": 1.436573906758222e-05, "loss": 86.8547, "step": 2889 }, { "compression_loss": 85.44222259521484, "epoch": 1.04, "learning_rate": 1.4360318033971811e-05, "loss": 86.0502, "step": 2890 }, { "compression_loss": 85.17752075195312, "epoch": 1.04, "learning_rate": 1.4354897000361404e-05, "loss": 85.6664, "step": 2891 }, { "compression_loss": 84.31103515625, "epoch": 1.05, "learning_rate": 1.4349475966750994e-05, "loss": 84.6787, "step": 2892 }, { "compression_loss": 84.43385314941406, "epoch": 1.05, "learning_rate": 1.4344054933140585e-05, "loss": 84.7997, "step": 2893 }, { "compression_loss": 83.79693603515625, "epoch": 1.05, "learning_rate": 1.4338633899530176e-05, "loss": 84.2827, "step": 2894 }, { "compression_loss": 84.23147583007812, "epoch": 1.05, "learning_rate": 1.433321286591977e-05, "loss": 85.0599, "step": 2895 }, { "compression_loss": 84.14328002929688, "epoch": 1.05, "learning_rate": 1.432779183230936e-05, "loss": 84.4505, "step": 2896 }, { "compression_loss": 85.02003479003906, "epoch": 1.05, "learning_rate": 1.4322370798698952e-05, "loss": 85.6239, "step": 2897 }, { "compression_loss": 85.23614501953125, "epoch": 1.05, "learning_rate": 1.4316949765088543e-05, "loss": 85.7215, "step": 2898 }, { "compression_loss": 83.886474609375, "epoch": 1.05, "learning_rate": 1.4311528731478136e-05, "loss": 84.2788, "step": 2899 }, { "compression_loss": 84.06773376464844, "epoch": 1.05, "learning_rate": 1.4306107697867727e-05, "loss": 84.8635, "step": 2900 }, { "compression_loss": 84.96226501464844, "epoch": 1.05, "learning_rate": 1.4300686664257318e-05, "loss": 85.6646, "step": 2901 }, { "compression_loss": 84.49552917480469, "epoch": 1.05, "learning_rate": 1.429526563064691e-05, "loss": 85.0113, "step": 2902 }, { "compression_loss": 84.50006866455078, "epoch": 1.05, "learning_rate": 1.4289844597036503e-05, "loss": 85.0398, "step": 2903 }, { "compression_loss": 84.54183959960938, "epoch": 1.05, "learning_rate": 1.4284423563426094e-05, "loss": 85.1642, "step": 2904 }, { "compression_loss": 84.97988891601562, "epoch": 1.05, "learning_rate": 1.4279002529815685e-05, "loss": 85.3032, "step": 2905 }, { "compression_loss": 83.98231506347656, "epoch": 1.05, "learning_rate": 1.4273581496205277e-05, "loss": 84.2629, "step": 2906 }, { "compression_loss": 84.82935333251953, "epoch": 1.05, "learning_rate": 1.426816046259487e-05, "loss": 85.2728, "step": 2907 }, { "compression_loss": 84.78274536132812, "epoch": 1.05, "learning_rate": 1.426273942898446e-05, "loss": 85.1449, "step": 2908 }, { "compression_loss": 85.37591552734375, "epoch": 1.05, "learning_rate": 1.4257318395374052e-05, "loss": 85.7001, "step": 2909 }, { "compression_loss": 84.1402587890625, "epoch": 1.05, "learning_rate": 1.4251897361763643e-05, "loss": 84.5873, "step": 2910 }, { "compression_loss": 83.3270263671875, "epoch": 1.05, "learning_rate": 1.4246476328153236e-05, "loss": 83.746, "step": 2911 }, { "compression_loss": 84.66554260253906, "epoch": 1.05, "learning_rate": 1.4241055294542828e-05, "loss": 85.2049, "step": 2912 }, { "compression_loss": 84.09649658203125, "epoch": 1.05, "learning_rate": 1.4235634260932419e-05, "loss": 84.6901, "step": 2913 }, { "compression_loss": 84.22727966308594, "epoch": 1.05, "learning_rate": 1.4230213227322008e-05, "loss": 84.7719, "step": 2914 }, { "compression_loss": 83.7751235961914, "epoch": 1.05, "learning_rate": 1.4224792193711601e-05, "loss": 84.2602, "step": 2915 }, { "compression_loss": 84.48756408691406, "epoch": 1.05, "learning_rate": 1.4219371160101193e-05, "loss": 84.9091, "step": 2916 }, { "compression_loss": 84.33837890625, "epoch": 1.05, "learning_rate": 1.4213950126490784e-05, "loss": 84.9757, "step": 2917 }, { "compression_loss": 84.18097686767578, "epoch": 1.05, "learning_rate": 1.4208529092880375e-05, "loss": 84.5297, "step": 2918 }, { "compression_loss": 86.44309997558594, "epoch": 1.05, "learning_rate": 1.4203108059269968e-05, "loss": 87.1108, "step": 2919 }, { "compression_loss": 83.95750427246094, "epoch": 1.06, "learning_rate": 1.419768702565956e-05, "loss": 84.2504, "step": 2920 }, { "compression_loss": 85.33042907714844, "epoch": 1.06, "learning_rate": 1.419226599204915e-05, "loss": 85.731, "step": 2921 }, { "compression_loss": 84.95668029785156, "epoch": 1.06, "learning_rate": 1.4186844958438742e-05, "loss": 85.4142, "step": 2922 }, { "compression_loss": 82.76013946533203, "epoch": 1.06, "learning_rate": 1.4181423924828335e-05, "loss": 83.1508, "step": 2923 }, { "compression_loss": 83.84430694580078, "epoch": 1.06, "learning_rate": 1.4176002891217926e-05, "loss": 84.4379, "step": 2924 }, { "compression_loss": 83.41149139404297, "epoch": 1.06, "learning_rate": 1.4170581857607517e-05, "loss": 83.6829, "step": 2925 }, { "compression_loss": 84.3844223022461, "epoch": 1.06, "learning_rate": 1.4165160823997109e-05, "loss": 84.9935, "step": 2926 }, { "compression_loss": 84.1405029296875, "epoch": 1.06, "learning_rate": 1.4159739790386702e-05, "loss": 84.6629, "step": 2927 }, { "compression_loss": 83.33718872070312, "epoch": 1.06, "learning_rate": 1.4154318756776293e-05, "loss": 83.6949, "step": 2928 }, { "compression_loss": 84.40599822998047, "epoch": 1.06, "learning_rate": 1.4148897723165884e-05, "loss": 84.8985, "step": 2929 }, { "compression_loss": 85.6744613647461, "epoch": 1.06, "learning_rate": 1.4143476689555476e-05, "loss": 86.1637, "step": 2930 }, { "compression_loss": 84.79344177246094, "epoch": 1.06, "learning_rate": 1.4138055655945069e-05, "loss": 85.3571, "step": 2931 }, { "compression_loss": 84.89671325683594, "epoch": 1.06, "learning_rate": 1.413263462233466e-05, "loss": 85.5722, "step": 2932 }, { "compression_loss": 83.30361938476562, "epoch": 1.06, "learning_rate": 1.4127213588724251e-05, "loss": 83.7847, "step": 2933 }, { "compression_loss": 84.13107299804688, "epoch": 1.06, "learning_rate": 1.4121792555113842e-05, "loss": 84.824, "step": 2934 }, { "compression_loss": 84.29344940185547, "epoch": 1.06, "learning_rate": 1.4116371521503435e-05, "loss": 84.7526, "step": 2935 }, { "compression_loss": 87.02485656738281, "epoch": 1.06, "learning_rate": 1.4110950487893025e-05, "loss": 87.4995, "step": 2936 }, { "compression_loss": 86.11140441894531, "epoch": 1.06, "learning_rate": 1.4105529454282616e-05, "loss": 86.8366, "step": 2937 }, { "compression_loss": 82.70210266113281, "epoch": 1.06, "learning_rate": 1.4100108420672207e-05, "loss": 83.0279, "step": 2938 }, { "compression_loss": 83.31371307373047, "epoch": 1.06, "learning_rate": 1.40946873870618e-05, "loss": 84.1426, "step": 2939 }, { "compression_loss": 86.76070404052734, "epoch": 1.06, "learning_rate": 1.4089266353451392e-05, "loss": 87.3417, "step": 2940 }, { "compression_loss": 83.3165283203125, "epoch": 1.06, "learning_rate": 1.4083845319840983e-05, "loss": 83.7251, "step": 2941 }, { "compression_loss": 84.28988647460938, "epoch": 1.06, "learning_rate": 1.4078424286230574e-05, "loss": 84.8058, "step": 2942 }, { "compression_loss": 85.50410461425781, "epoch": 1.06, "learning_rate": 1.4073003252620167e-05, "loss": 85.8204, "step": 2943 }, { "compression_loss": 84.80863952636719, "epoch": 1.06, "learning_rate": 1.4067582219009758e-05, "loss": 85.4636, "step": 2944 }, { "compression_loss": 84.97721862792969, "epoch": 1.06, "learning_rate": 1.406216118539935e-05, "loss": 85.2858, "step": 2945 }, { "compression_loss": 85.62033081054688, "epoch": 1.06, "learning_rate": 1.4056740151788941e-05, "loss": 86.1666, "step": 2946 }, { "compression_loss": 85.32841491699219, "epoch": 1.07, "learning_rate": 1.4051319118178534e-05, "loss": 85.7895, "step": 2947 }, { "compression_loss": 84.41902160644531, "epoch": 1.07, "learning_rate": 1.4045898084568125e-05, "loss": 84.7563, "step": 2948 }, { "compression_loss": 86.73323059082031, "epoch": 1.07, "learning_rate": 1.4040477050957716e-05, "loss": 87.1929, "step": 2949 }, { "compression_loss": 84.00015258789062, "epoch": 1.07, "learning_rate": 1.4035056017347308e-05, "loss": 84.2362, "step": 2950 }, { "compression_loss": 83.40521240234375, "epoch": 1.07, "learning_rate": 1.40296349837369e-05, "loss": 83.7468, "step": 2951 }, { "compression_loss": 85.07373809814453, "epoch": 1.07, "learning_rate": 1.4024213950126492e-05, "loss": 85.7614, "step": 2952 }, { "compression_loss": 83.30872344970703, "epoch": 1.07, "learning_rate": 1.4018792916516083e-05, "loss": 83.647, "step": 2953 }, { "compression_loss": 84.0908203125, "epoch": 1.07, "learning_rate": 1.4013371882905675e-05, "loss": 84.8248, "step": 2954 }, { "compression_loss": 84.97169494628906, "epoch": 1.07, "learning_rate": 1.4007950849295268e-05, "loss": 85.5228, "step": 2955 }, { "compression_loss": 85.18299865722656, "epoch": 1.07, "learning_rate": 1.4002529815684859e-05, "loss": 85.7413, "step": 2956 }, { "compression_loss": 84.44831848144531, "epoch": 1.07, "learning_rate": 1.399710878207445e-05, "loss": 84.9282, "step": 2957 }, { "compression_loss": 84.5221939086914, "epoch": 1.07, "learning_rate": 1.399168774846404e-05, "loss": 85.0116, "step": 2958 }, { "compression_loss": 84.62956237792969, "epoch": 1.07, "learning_rate": 1.3986266714853631e-05, "loss": 85.0474, "step": 2959 }, { "compression_loss": 84.73548889160156, "epoch": 1.07, "learning_rate": 1.3980845681243224e-05, "loss": 85.0959, "step": 2960 }, { "compression_loss": 85.0872802734375, "epoch": 1.07, "learning_rate": 1.3975424647632815e-05, "loss": 85.5176, "step": 2961 }, { "compression_loss": 85.42068481445312, "epoch": 1.07, "learning_rate": 1.3970003614022406e-05, "loss": 85.8493, "step": 2962 }, { "compression_loss": 86.57830810546875, "epoch": 1.07, "learning_rate": 1.3964582580411998e-05, "loss": 87.189, "step": 2963 }, { "compression_loss": 83.8611831665039, "epoch": 1.07, "learning_rate": 1.395916154680159e-05, "loss": 84.4183, "step": 2964 }, { "compression_loss": 83.48664093017578, "epoch": 1.07, "learning_rate": 1.3953740513191182e-05, "loss": 83.9658, "step": 2965 }, { "compression_loss": 82.44774627685547, "epoch": 1.07, "learning_rate": 1.3948319479580773e-05, "loss": 82.9498, "step": 2966 }, { "compression_loss": 83.30732727050781, "epoch": 1.07, "learning_rate": 1.3942898445970364e-05, "loss": 83.8446, "step": 2967 }, { "compression_loss": 84.74862670898438, "epoch": 1.07, "learning_rate": 1.3937477412359957e-05, "loss": 85.4452, "step": 2968 }, { "compression_loss": 83.52447509765625, "epoch": 1.07, "learning_rate": 1.3932056378749549e-05, "loss": 83.8992, "step": 2969 }, { "compression_loss": 84.23461151123047, "epoch": 1.07, "learning_rate": 1.392663534513914e-05, "loss": 84.5466, "step": 2970 }, { "compression_loss": 84.64386749267578, "epoch": 1.07, "learning_rate": 1.3921214311528731e-05, "loss": 84.9577, "step": 2971 }, { "compression_loss": 83.70697021484375, "epoch": 1.07, "learning_rate": 1.3915793277918324e-05, "loss": 84.2977, "step": 2972 }, { "compression_loss": 85.52058410644531, "epoch": 1.07, "learning_rate": 1.3910372244307915e-05, "loss": 85.9907, "step": 2973 }, { "compression_loss": 84.86650085449219, "epoch": 1.07, "learning_rate": 1.3904951210697507e-05, "loss": 85.3502, "step": 2974 }, { "compression_loss": 83.8707275390625, "epoch": 1.08, "learning_rate": 1.3899530177087098e-05, "loss": 84.1805, "step": 2975 }, { "compression_loss": 83.635498046875, "epoch": 1.08, "learning_rate": 1.3894109143476691e-05, "loss": 84.1297, "step": 2976 }, { "compression_loss": 84.86506652832031, "epoch": 1.08, "learning_rate": 1.3888688109866282e-05, "loss": 85.3568, "step": 2977 }, { "compression_loss": 86.39180755615234, "epoch": 1.08, "learning_rate": 1.3883267076255874e-05, "loss": 86.8218, "step": 2978 }, { "compression_loss": 84.69740295410156, "epoch": 1.08, "learning_rate": 1.3877846042645465e-05, "loss": 85.2543, "step": 2979 }, { "compression_loss": 85.34854888916016, "epoch": 1.08, "learning_rate": 1.3872425009035056e-05, "loss": 85.9428, "step": 2980 }, { "compression_loss": 82.47689819335938, "epoch": 1.08, "learning_rate": 1.3867003975424647e-05, "loss": 82.6558, "step": 2981 }, { "compression_loss": 83.2264404296875, "epoch": 1.08, "learning_rate": 1.3861582941814239e-05, "loss": 83.8993, "step": 2982 }, { "compression_loss": 84.31128692626953, "epoch": 1.08, "learning_rate": 1.385616190820383e-05, "loss": 84.6755, "step": 2983 }, { "compression_loss": 85.23301696777344, "epoch": 1.08, "learning_rate": 1.3850740874593423e-05, "loss": 85.592, "step": 2984 }, { "compression_loss": 86.44438171386719, "epoch": 1.08, "learning_rate": 1.3845319840983014e-05, "loss": 86.8352, "step": 2985 }, { "compression_loss": 83.15850067138672, "epoch": 1.08, "learning_rate": 1.3839898807372605e-05, "loss": 83.7406, "step": 2986 }, { "compression_loss": 85.42192077636719, "epoch": 1.08, "learning_rate": 1.3834477773762197e-05, "loss": 85.7602, "step": 2987 }, { "compression_loss": 84.81253814697266, "epoch": 1.08, "learning_rate": 1.382905674015179e-05, "loss": 85.1717, "step": 2988 }, { "compression_loss": 84.23101806640625, "epoch": 1.08, "learning_rate": 1.3823635706541381e-05, "loss": 84.7555, "step": 2989 }, { "compression_loss": 84.20108032226562, "epoch": 1.08, "learning_rate": 1.3818214672930972e-05, "loss": 84.6527, "step": 2990 }, { "compression_loss": 84.36338806152344, "epoch": 1.08, "learning_rate": 1.3812793639320563e-05, "loss": 84.9234, "step": 2991 }, { "compression_loss": 83.75840759277344, "epoch": 1.08, "learning_rate": 1.3807372605710156e-05, "loss": 84.1282, "step": 2992 }, { "compression_loss": 84.68180847167969, "epoch": 1.08, "learning_rate": 1.3801951572099748e-05, "loss": 85.0737, "step": 2993 }, { "compression_loss": 83.93624877929688, "epoch": 1.08, "learning_rate": 1.3796530538489339e-05, "loss": 84.3604, "step": 2994 }, { "compression_loss": 83.60005950927734, "epoch": 1.08, "learning_rate": 1.379110950487893e-05, "loss": 84.1401, "step": 2995 }, { "compression_loss": 83.97593688964844, "epoch": 1.08, "learning_rate": 1.3785688471268523e-05, "loss": 84.3696, "step": 2996 }, { "compression_loss": 85.314697265625, "epoch": 1.08, "learning_rate": 1.3780267437658114e-05, "loss": 85.7707, "step": 2997 }, { "compression_loss": 85.37812805175781, "epoch": 1.08, "learning_rate": 1.3774846404047706e-05, "loss": 85.8378, "step": 2998 }, { "compression_loss": 84.08372497558594, "epoch": 1.08, "learning_rate": 1.3769425370437297e-05, "loss": 84.5173, "step": 2999 }, { "compression_loss": 85.62816619873047, "epoch": 1.08, "learning_rate": 1.376400433682689e-05, "loss": 86.0579, "step": 3000 }, { "epoch": 1.08, "eval_exact_match": 86.74550614947965, "eval_f1": 93.01277070237498, "step": 3000 }, { "compression_loss": 85.24122619628906, "epoch": 1.08, "learning_rate": 1.375858330321648e-05, "loss": 85.7556, "step": 3001 }, { "compression_loss": 85.49301147460938, "epoch": 1.08, "learning_rate": 1.375316226960607e-05, "loss": 85.8835, "step": 3002 }, { "compression_loss": 82.93522644042969, "epoch": 1.09, "learning_rate": 1.3747741235995662e-05, "loss": 83.3073, "step": 3003 }, { "compression_loss": 83.3819580078125, "epoch": 1.09, "learning_rate": 1.3742320202385255e-05, "loss": 83.7841, "step": 3004 }, { "compression_loss": 84.5469970703125, "epoch": 1.09, "learning_rate": 1.3736899168774846e-05, "loss": 84.9973, "step": 3005 }, { "compression_loss": 85.66069030761719, "epoch": 1.09, "learning_rate": 1.3731478135164438e-05, "loss": 85.9167, "step": 3006 }, { "compression_loss": 84.35762023925781, "epoch": 1.09, "learning_rate": 1.3726057101554029e-05, "loss": 84.9251, "step": 3007 }, { "compression_loss": 83.99588012695312, "epoch": 1.09, "learning_rate": 1.3720636067943622e-05, "loss": 84.4642, "step": 3008 }, { "compression_loss": 82.92313385009766, "epoch": 1.09, "learning_rate": 1.3715215034333213e-05, "loss": 83.2526, "step": 3009 }, { "compression_loss": 83.7668228149414, "epoch": 1.09, "learning_rate": 1.3709794000722804e-05, "loss": 84.1842, "step": 3010 }, { "compression_loss": 83.60135650634766, "epoch": 1.09, "learning_rate": 1.3704372967112396e-05, "loss": 83.8169, "step": 3011 }, { "compression_loss": 83.26812744140625, "epoch": 1.09, "learning_rate": 1.3698951933501989e-05, "loss": 83.6902, "step": 3012 }, { "compression_loss": 85.63037109375, "epoch": 1.09, "learning_rate": 1.369353089989158e-05, "loss": 86.2356, "step": 3013 }, { "compression_loss": 84.55705261230469, "epoch": 1.09, "learning_rate": 1.3688109866281171e-05, "loss": 85.1086, "step": 3014 }, { "compression_loss": 85.3135986328125, "epoch": 1.09, "learning_rate": 1.3682688832670762e-05, "loss": 85.5481, "step": 3015 }, { "compression_loss": 84.41534423828125, "epoch": 1.09, "learning_rate": 1.3677267799060355e-05, "loss": 84.838, "step": 3016 }, { "compression_loss": 83.55533599853516, "epoch": 1.09, "learning_rate": 1.3671846765449947e-05, "loss": 83.8582, "step": 3017 }, { "compression_loss": 83.74551391601562, "epoch": 1.09, "learning_rate": 1.3666425731839538e-05, "loss": 84.2275, "step": 3018 }, { "compression_loss": 85.29449462890625, "epoch": 1.09, "learning_rate": 1.366100469822913e-05, "loss": 85.8674, "step": 3019 }, { "compression_loss": 82.69064331054688, "epoch": 1.09, "learning_rate": 1.3655583664618722e-05, "loss": 83.0048, "step": 3020 }, { "compression_loss": 85.119873046875, "epoch": 1.09, "learning_rate": 1.3650162631008313e-05, "loss": 85.5212, "step": 3021 }, { "compression_loss": 83.95759582519531, "epoch": 1.09, "learning_rate": 1.3644741597397905e-05, "loss": 84.1977, "step": 3022 }, { "compression_loss": 84.48818969726562, "epoch": 1.09, "learning_rate": 1.3639320563787494e-05, "loss": 85.1263, "step": 3023 }, { "compression_loss": 84.39642333984375, "epoch": 1.09, "learning_rate": 1.3633899530177087e-05, "loss": 84.8447, "step": 3024 }, { "compression_loss": 85.45311737060547, "epoch": 1.09, "learning_rate": 1.3628478496566679e-05, "loss": 86.5749, "step": 3025 }, { "compression_loss": 84.89197540283203, "epoch": 1.09, "learning_rate": 1.362305746295627e-05, "loss": 85.3779, "step": 3026 }, { "compression_loss": 82.80681610107422, "epoch": 1.09, "learning_rate": 1.3617636429345861e-05, "loss": 83.4218, "step": 3027 }, { "compression_loss": 84.30712127685547, "epoch": 1.09, "learning_rate": 1.3612215395735454e-05, "loss": 84.8297, "step": 3028 }, { "compression_loss": 84.32223510742188, "epoch": 1.09, "learning_rate": 1.3606794362125045e-05, "loss": 84.6491, "step": 3029 }, { "compression_loss": 85.003173828125, "epoch": 1.1, "learning_rate": 1.3601373328514637e-05, "loss": 85.3175, "step": 3030 }, { "compression_loss": 85.92866516113281, "epoch": 1.1, "learning_rate": 1.3595952294904228e-05, "loss": 86.6392, "step": 3031 }, { "compression_loss": 84.69273376464844, "epoch": 1.1, "learning_rate": 1.3590531261293821e-05, "loss": 85.1134, "step": 3032 }, { "compression_loss": 85.28746032714844, "epoch": 1.1, "learning_rate": 1.3585110227683412e-05, "loss": 85.9094, "step": 3033 }, { "compression_loss": 83.7210693359375, "epoch": 1.1, "learning_rate": 1.3579689194073003e-05, "loss": 84.08, "step": 3034 }, { "compression_loss": 83.09269714355469, "epoch": 1.1, "learning_rate": 1.3574268160462595e-05, "loss": 83.5586, "step": 3035 }, { "compression_loss": 84.56974029541016, "epoch": 1.1, "learning_rate": 1.3568847126852188e-05, "loss": 84.9402, "step": 3036 }, { "compression_loss": 84.80354309082031, "epoch": 1.1, "learning_rate": 1.3563426093241779e-05, "loss": 85.1664, "step": 3037 }, { "compression_loss": 85.05380249023438, "epoch": 1.1, "learning_rate": 1.355800505963137e-05, "loss": 85.8143, "step": 3038 }, { "compression_loss": 83.9307632446289, "epoch": 1.1, "learning_rate": 1.3552584026020961e-05, "loss": 84.4315, "step": 3039 }, { "compression_loss": 84.23851013183594, "epoch": 1.1, "learning_rate": 1.3547162992410554e-05, "loss": 84.8456, "step": 3040 }, { "compression_loss": 85.15492248535156, "epoch": 1.1, "learning_rate": 1.3541741958800146e-05, "loss": 85.6999, "step": 3041 }, { "compression_loss": 82.04786682128906, "epoch": 1.1, "learning_rate": 1.3536320925189737e-05, "loss": 82.5291, "step": 3042 }, { "compression_loss": 84.9727783203125, "epoch": 1.1, "learning_rate": 1.3530899891579328e-05, "loss": 85.382, "step": 3043 }, { "compression_loss": 83.95418548583984, "epoch": 1.1, "learning_rate": 1.3525478857968921e-05, "loss": 84.2629, "step": 3044 }, { "compression_loss": 83.79156494140625, "epoch": 1.1, "learning_rate": 1.352005782435851e-05, "loss": 84.1326, "step": 3045 }, { "compression_loss": 84.57151794433594, "epoch": 1.1, "learning_rate": 1.3514636790748102e-05, "loss": 85.009, "step": 3046 }, { "compression_loss": 84.78839874267578, "epoch": 1.1, "learning_rate": 1.3509215757137693e-05, "loss": 85.1403, "step": 3047 }, { "compression_loss": 84.62062072753906, "epoch": 1.1, "learning_rate": 1.3503794723527286e-05, "loss": 84.9976, "step": 3048 }, { "compression_loss": 86.33470916748047, "epoch": 1.1, "learning_rate": 1.3498373689916878e-05, "loss": 86.8547, "step": 3049 }, { "compression_loss": 85.2437515258789, "epoch": 1.1, "learning_rate": 1.3492952656306469e-05, "loss": 85.7969, "step": 3050 }, { "compression_loss": 86.47955322265625, "epoch": 1.1, "learning_rate": 1.348753162269606e-05, "loss": 87.408, "step": 3051 }, { "compression_loss": 84.644775390625, "epoch": 1.1, "learning_rate": 1.3482110589085653e-05, "loss": 85.184, "step": 3052 }, { "compression_loss": 82.85810089111328, "epoch": 1.1, "learning_rate": 1.3476689555475244e-05, "loss": 83.2719, "step": 3053 }, { "compression_loss": 84.2459716796875, "epoch": 1.1, "learning_rate": 1.3471268521864836e-05, "loss": 84.5719, "step": 3054 }, { "compression_loss": 83.54314422607422, "epoch": 1.1, "learning_rate": 1.3465847488254427e-05, "loss": 83.9061, "step": 3055 }, { "compression_loss": 83.96611022949219, "epoch": 1.1, "learning_rate": 1.346042645464402e-05, "loss": 84.2082, "step": 3056 }, { "compression_loss": 83.88355255126953, "epoch": 1.1, "learning_rate": 1.3455005421033611e-05, "loss": 84.3506, "step": 3057 }, { "compression_loss": 83.10604858398438, "epoch": 1.11, "learning_rate": 1.3449584387423202e-05, "loss": 83.5583, "step": 3058 }, { "compression_loss": 82.25830078125, "epoch": 1.11, "learning_rate": 1.3444163353812794e-05, "loss": 82.8164, "step": 3059 }, { "compression_loss": 84.00260162353516, "epoch": 1.11, "learning_rate": 1.3438742320202387e-05, "loss": 84.323, "step": 3060 }, { "compression_loss": 83.82035827636719, "epoch": 1.11, "learning_rate": 1.3433321286591978e-05, "loss": 84.3326, "step": 3061 }, { "compression_loss": 85.10726928710938, "epoch": 1.11, "learning_rate": 1.342790025298157e-05, "loss": 85.4733, "step": 3062 }, { "compression_loss": 84.33868408203125, "epoch": 1.11, "learning_rate": 1.342247921937116e-05, "loss": 84.7714, "step": 3063 }, { "compression_loss": 84.81829833984375, "epoch": 1.11, "learning_rate": 1.3417058185760753e-05, "loss": 85.2124, "step": 3064 }, { "compression_loss": 83.75497436523438, "epoch": 1.11, "learning_rate": 1.3411637152150345e-05, "loss": 84.2807, "step": 3065 }, { "compression_loss": 83.0518798828125, "epoch": 1.11, "learning_rate": 1.3406216118539936e-05, "loss": 83.4659, "step": 3066 }, { "compression_loss": 83.83297729492188, "epoch": 1.11, "learning_rate": 1.3400795084929526e-05, "loss": 84.3168, "step": 3067 }, { "compression_loss": 85.18501281738281, "epoch": 1.11, "learning_rate": 1.3395374051319118e-05, "loss": 85.6072, "step": 3068 }, { "compression_loss": 82.49799346923828, "epoch": 1.11, "learning_rate": 1.338995301770871e-05, "loss": 83.1545, "step": 3069 }, { "compression_loss": 84.30439758300781, "epoch": 1.11, "learning_rate": 1.3384531984098301e-05, "loss": 84.7087, "step": 3070 }, { "compression_loss": 83.88377380371094, "epoch": 1.11, "learning_rate": 1.3379110950487892e-05, "loss": 84.3728, "step": 3071 }, { "compression_loss": 85.37326049804688, "epoch": 1.11, "learning_rate": 1.3373689916877485e-05, "loss": 85.7974, "step": 3072 }, { "compression_loss": 84.3869857788086, "epoch": 1.11, "learning_rate": 1.3368268883267077e-05, "loss": 84.8664, "step": 3073 }, { "compression_loss": 85.5455322265625, "epoch": 1.11, "learning_rate": 1.3362847849656668e-05, "loss": 85.9511, "step": 3074 }, { "compression_loss": 85.69873809814453, "epoch": 1.11, "learning_rate": 1.3357426816046259e-05, "loss": 86.114, "step": 3075 }, { "compression_loss": 82.99395751953125, "epoch": 1.11, "learning_rate": 1.3352005782435852e-05, "loss": 83.4078, "step": 3076 }, { "compression_loss": 85.45220947265625, "epoch": 1.11, "learning_rate": 1.3346584748825443e-05, "loss": 85.8012, "step": 3077 }, { "compression_loss": 83.94070434570312, "epoch": 1.11, "learning_rate": 1.3341163715215035e-05, "loss": 84.3264, "step": 3078 }, { "compression_loss": 81.21469116210938, "epoch": 1.11, "learning_rate": 1.3335742681604626e-05, "loss": 81.623, "step": 3079 }, { "compression_loss": 83.52560424804688, "epoch": 1.11, "learning_rate": 1.3330321647994219e-05, "loss": 84.1193, "step": 3080 }, { "compression_loss": 84.53124237060547, "epoch": 1.11, "learning_rate": 1.332490061438381e-05, "loss": 84.9171, "step": 3081 }, { "compression_loss": 83.4873046875, "epoch": 1.11, "learning_rate": 1.3319479580773401e-05, "loss": 84.2321, "step": 3082 }, { "compression_loss": 84.61060333251953, "epoch": 1.11, "learning_rate": 1.3314058547162993e-05, "loss": 85.2594, "step": 3083 }, { "compression_loss": 83.9794921875, "epoch": 1.11, "learning_rate": 1.3308637513552586e-05, "loss": 84.3883, "step": 3084 }, { "compression_loss": 85.12339782714844, "epoch": 1.11, "learning_rate": 1.3303216479942177e-05, "loss": 85.6637, "step": 3085 }, { "compression_loss": 84.85408020019531, "epoch": 1.12, "learning_rate": 1.3297795446331768e-05, "loss": 85.1973, "step": 3086 }, { "compression_loss": 85.07456970214844, "epoch": 1.12, "learning_rate": 1.329237441272136e-05, "loss": 85.6961, "step": 3087 }, { "compression_loss": 82.79613494873047, "epoch": 1.12, "learning_rate": 1.328695337911095e-05, "loss": 83.2941, "step": 3088 }, { "compression_loss": 83.42536163330078, "epoch": 1.12, "learning_rate": 1.3281532345500542e-05, "loss": 83.7851, "step": 3089 }, { "compression_loss": 83.73077392578125, "epoch": 1.12, "learning_rate": 1.3276111311890133e-05, "loss": 84.0645, "step": 3090 }, { "compression_loss": 83.84534454345703, "epoch": 1.12, "learning_rate": 1.3270690278279725e-05, "loss": 84.2728, "step": 3091 }, { "compression_loss": 83.37130737304688, "epoch": 1.12, "learning_rate": 1.3265269244669317e-05, "loss": 83.6498, "step": 3092 }, { "compression_loss": 83.71308135986328, "epoch": 1.12, "learning_rate": 1.3259848211058909e-05, "loss": 84.0459, "step": 3093 }, { "compression_loss": 83.64363861083984, "epoch": 1.12, "learning_rate": 1.32544271774485e-05, "loss": 83.9819, "step": 3094 }, { "compression_loss": 85.39620971679688, "epoch": 1.12, "learning_rate": 1.3249006143838091e-05, "loss": 85.7915, "step": 3095 }, { "compression_loss": 82.63996124267578, "epoch": 1.12, "learning_rate": 1.3243585110227684e-05, "loss": 82.9164, "step": 3096 }, { "compression_loss": 84.93991088867188, "epoch": 1.12, "learning_rate": 1.3238164076617276e-05, "loss": 85.591, "step": 3097 }, { "compression_loss": 84.75241088867188, "epoch": 1.12, "learning_rate": 1.3232743043006867e-05, "loss": 85.1345, "step": 3098 }, { "compression_loss": 83.51719665527344, "epoch": 1.12, "learning_rate": 1.3227322009396458e-05, "loss": 84.0212, "step": 3099 }, { "compression_loss": 84.98817443847656, "epoch": 1.12, "learning_rate": 1.3221900975786051e-05, "loss": 85.3764, "step": 3100 }, { "compression_loss": 83.45755767822266, "epoch": 1.12, "learning_rate": 1.3216479942175642e-05, "loss": 83.6971, "step": 3101 }, { "compression_loss": 84.68125915527344, "epoch": 1.12, "learning_rate": 1.3211058908565234e-05, "loss": 84.977, "step": 3102 }, { "compression_loss": 84.22634887695312, "epoch": 1.12, "learning_rate": 1.3205637874954825e-05, "loss": 84.7342, "step": 3103 }, { "compression_loss": 84.01119995117188, "epoch": 1.12, "learning_rate": 1.3200216841344418e-05, "loss": 84.3363, "step": 3104 }, { "compression_loss": 84.10929107666016, "epoch": 1.12, "learning_rate": 1.3194795807734009e-05, "loss": 84.6203, "step": 3105 }, { "compression_loss": 85.52204895019531, "epoch": 1.12, "learning_rate": 1.31893747741236e-05, "loss": 86.1716, "step": 3106 }, { "compression_loss": 85.90097045898438, "epoch": 1.12, "learning_rate": 1.3183953740513192e-05, "loss": 86.3266, "step": 3107 }, { "compression_loss": 85.0535659790039, "epoch": 1.12, "learning_rate": 1.3178532706902785e-05, "loss": 85.6125, "step": 3108 }, { "compression_loss": 85.77066040039062, "epoch": 1.12, "learning_rate": 1.3173111673292376e-05, "loss": 86.1736, "step": 3109 }, { "compression_loss": 84.00677490234375, "epoch": 1.12, "learning_rate": 1.3167690639681965e-05, "loss": 84.557, "step": 3110 }, { "compression_loss": 86.26223754882812, "epoch": 1.12, "learning_rate": 1.3162269606071557e-05, "loss": 86.9512, "step": 3111 }, { "compression_loss": 85.17970275878906, "epoch": 1.12, "learning_rate": 1.315684857246115e-05, "loss": 85.6453, "step": 3112 }, { "compression_loss": 84.6007080078125, "epoch": 1.13, "learning_rate": 1.3151427538850741e-05, "loss": 85.005, "step": 3113 }, { "compression_loss": 82.64566802978516, "epoch": 1.13, "learning_rate": 1.3146006505240332e-05, "loss": 82.9846, "step": 3114 }, { "compression_loss": 85.73685455322266, "epoch": 1.13, "learning_rate": 1.3140585471629924e-05, "loss": 86.2297, "step": 3115 }, { "compression_loss": 84.39066314697266, "epoch": 1.13, "learning_rate": 1.3135164438019516e-05, "loss": 84.7866, "step": 3116 }, { "compression_loss": 85.11803436279297, "epoch": 1.13, "learning_rate": 1.3129743404409108e-05, "loss": 85.7951, "step": 3117 }, { "compression_loss": 83.99336242675781, "epoch": 1.13, "learning_rate": 1.3124322370798699e-05, "loss": 84.5087, "step": 3118 }, { "compression_loss": 85.05049133300781, "epoch": 1.13, "learning_rate": 1.311890133718829e-05, "loss": 85.4753, "step": 3119 }, { "compression_loss": 85.64225769042969, "epoch": 1.13, "learning_rate": 1.3113480303577883e-05, "loss": 85.9068, "step": 3120 }, { "compression_loss": 85.16877746582031, "epoch": 1.13, "learning_rate": 1.3108059269967475e-05, "loss": 85.4274, "step": 3121 }, { "compression_loss": 84.92070007324219, "epoch": 1.13, "learning_rate": 1.3102638236357066e-05, "loss": 85.5174, "step": 3122 }, { "compression_loss": 83.8578872680664, "epoch": 1.13, "learning_rate": 1.3097217202746657e-05, "loss": 84.3544, "step": 3123 }, { "compression_loss": 84.59818267822266, "epoch": 1.13, "learning_rate": 1.309179616913625e-05, "loss": 85.0839, "step": 3124 }, { "compression_loss": 85.58737182617188, "epoch": 1.13, "learning_rate": 1.3086375135525841e-05, "loss": 85.9883, "step": 3125 }, { "compression_loss": 83.51756286621094, "epoch": 1.13, "learning_rate": 1.3080954101915433e-05, "loss": 84.0091, "step": 3126 }, { "compression_loss": 84.71782684326172, "epoch": 1.13, "learning_rate": 1.3075533068305024e-05, "loss": 84.9101, "step": 3127 }, { "compression_loss": 84.11295318603516, "epoch": 1.13, "learning_rate": 1.3070112034694617e-05, "loss": 84.3989, "step": 3128 }, { "compression_loss": 85.15170288085938, "epoch": 1.13, "learning_rate": 1.3064691001084208e-05, "loss": 85.5328, "step": 3129 }, { "compression_loss": 84.00799560546875, "epoch": 1.13, "learning_rate": 1.30592699674738e-05, "loss": 84.3636, "step": 3130 }, { "compression_loss": 83.17755889892578, "epoch": 1.13, "learning_rate": 1.305384893386339e-05, "loss": 83.5972, "step": 3131 }, { "compression_loss": 83.28285217285156, "epoch": 1.13, "learning_rate": 1.3048427900252982e-05, "loss": 83.8386, "step": 3132 }, { "compression_loss": 86.12641143798828, "epoch": 1.13, "learning_rate": 1.3043006866642573e-05, "loss": 86.6175, "step": 3133 }, { "compression_loss": 83.87397766113281, "epoch": 1.13, "learning_rate": 1.3037585833032164e-05, "loss": 84.274, "step": 3134 }, { "compression_loss": 83.54618835449219, "epoch": 1.13, "learning_rate": 1.3032164799421756e-05, "loss": 83.9941, "step": 3135 }, { "compression_loss": 84.19532012939453, "epoch": 1.13, "learning_rate": 1.3026743765811349e-05, "loss": 84.6869, "step": 3136 }, { "compression_loss": 85.78475189208984, "epoch": 1.13, "learning_rate": 1.302132273220094e-05, "loss": 86.3349, "step": 3137 }, { "compression_loss": 83.76145935058594, "epoch": 1.13, "learning_rate": 1.3015901698590531e-05, "loss": 84.2594, "step": 3138 }, { "compression_loss": 85.96185302734375, "epoch": 1.13, "learning_rate": 1.3010480664980123e-05, "loss": 86.6502, "step": 3139 }, { "compression_loss": 84.95338439941406, "epoch": 1.13, "learning_rate": 1.3005059631369715e-05, "loss": 85.5575, "step": 3140 }, { "compression_loss": 84.70085144042969, "epoch": 1.14, "learning_rate": 1.2999638597759307e-05, "loss": 85.1539, "step": 3141 }, { "compression_loss": 85.67050170898438, "epoch": 1.14, "learning_rate": 1.2994217564148898e-05, "loss": 85.9462, "step": 3142 }, { "compression_loss": 83.82403564453125, "epoch": 1.14, "learning_rate": 1.298879653053849e-05, "loss": 84.3249, "step": 3143 }, { "compression_loss": 83.88925170898438, "epoch": 1.14, "learning_rate": 1.298337549692808e-05, "loss": 84.5947, "step": 3144 }, { "compression_loss": 84.10810852050781, "epoch": 1.14, "learning_rate": 1.2977954463317674e-05, "loss": 84.7796, "step": 3145 }, { "compression_loss": 83.240478515625, "epoch": 1.14, "learning_rate": 1.2972533429707265e-05, "loss": 83.6664, "step": 3146 }, { "compression_loss": 81.6862564086914, "epoch": 1.14, "learning_rate": 1.2967112396096856e-05, "loss": 81.9872, "step": 3147 }, { "compression_loss": 82.73991394042969, "epoch": 1.14, "learning_rate": 1.2961691362486447e-05, "loss": 83.3058, "step": 3148 }, { "compression_loss": 85.2216567993164, "epoch": 1.14, "learning_rate": 1.295627032887604e-05, "loss": 85.7184, "step": 3149 }, { "compression_loss": 84.81839752197266, "epoch": 1.14, "learning_rate": 1.2950849295265632e-05, "loss": 85.1661, "step": 3150 }, { "compression_loss": 85.47188568115234, "epoch": 1.14, "learning_rate": 1.2945428261655223e-05, "loss": 85.9522, "step": 3151 }, { "compression_loss": 85.25421905517578, "epoch": 1.14, "learning_rate": 1.2940007228044814e-05, "loss": 85.7574, "step": 3152 }, { "compression_loss": 81.74671173095703, "epoch": 1.14, "learning_rate": 1.2934586194434407e-05, "loss": 82.2057, "step": 3153 }, { "compression_loss": 84.60371398925781, "epoch": 1.14, "learning_rate": 1.2929165160823997e-05, "loss": 85.2283, "step": 3154 }, { "compression_loss": 84.14193725585938, "epoch": 1.14, "learning_rate": 1.2923744127213588e-05, "loss": 84.5397, "step": 3155 }, { "compression_loss": 84.82281494140625, "epoch": 1.14, "learning_rate": 1.291832309360318e-05, "loss": 85.3564, "step": 3156 }, { "compression_loss": 85.02328491210938, "epoch": 1.14, "learning_rate": 1.2912902059992772e-05, "loss": 85.9467, "step": 3157 }, { "compression_loss": 84.09620666503906, "epoch": 1.14, "learning_rate": 1.2907481026382363e-05, "loss": 84.5291, "step": 3158 }, { "compression_loss": 83.53526306152344, "epoch": 1.14, "learning_rate": 1.2902059992771955e-05, "loss": 83.8786, "step": 3159 }, { "compression_loss": 86.3653793334961, "epoch": 1.14, "learning_rate": 1.2896638959161546e-05, "loss": 86.8989, "step": 3160 }, { "compression_loss": 85.04621887207031, "epoch": 1.14, "learning_rate": 1.2891217925551139e-05, "loss": 85.6138, "step": 3161 }, { "compression_loss": 83.50263214111328, "epoch": 1.14, "learning_rate": 1.288579689194073e-05, "loss": 84.137, "step": 3162 }, { "compression_loss": 85.16021728515625, "epoch": 1.14, "learning_rate": 1.2880375858330322e-05, "loss": 85.6904, "step": 3163 }, { "compression_loss": 86.3647232055664, "epoch": 1.14, "learning_rate": 1.2874954824719913e-05, "loss": 86.8383, "step": 3164 }, { "compression_loss": 83.85147094726562, "epoch": 1.14, "learning_rate": 1.2869533791109506e-05, "loss": 84.4628, "step": 3165 }, { "compression_loss": 86.67350006103516, "epoch": 1.14, "learning_rate": 1.2864112757499097e-05, "loss": 87.0459, "step": 3166 }, { "compression_loss": 85.42356872558594, "epoch": 1.14, "learning_rate": 1.2858691723888688e-05, "loss": 85.8531, "step": 3167 }, { "compression_loss": 84.68905639648438, "epoch": 1.14, "learning_rate": 1.285327069027828e-05, "loss": 85.4039, "step": 3168 }, { "compression_loss": 86.5775146484375, "epoch": 1.15, "learning_rate": 1.2847849656667873e-05, "loss": 87.3383, "step": 3169 }, { "compression_loss": 84.28105163574219, "epoch": 1.15, "learning_rate": 1.2842428623057464e-05, "loss": 84.6208, "step": 3170 }, { "compression_loss": 85.15042114257812, "epoch": 1.15, "learning_rate": 1.2837007589447055e-05, "loss": 85.3482, "step": 3171 }, { "compression_loss": 84.55174255371094, "epoch": 1.15, "learning_rate": 1.2831586555836646e-05, "loss": 84.9434, "step": 3172 }, { "compression_loss": 82.98619079589844, "epoch": 1.15, "learning_rate": 1.282616552222624e-05, "loss": 83.4191, "step": 3173 }, { "compression_loss": 84.27827453613281, "epoch": 1.15, "learning_rate": 1.282074448861583e-05, "loss": 84.662, "step": 3174 }, { "compression_loss": 85.23233032226562, "epoch": 1.15, "learning_rate": 1.2815323455005422e-05, "loss": 85.6838, "step": 3175 }, { "compression_loss": 85.75962829589844, "epoch": 1.15, "learning_rate": 1.2809902421395011e-05, "loss": 86.1707, "step": 3176 }, { "compression_loss": 83.56849670410156, "epoch": 1.15, "learning_rate": 1.2804481387784604e-05, "loss": 83.919, "step": 3177 }, { "compression_loss": 83.51260375976562, "epoch": 1.15, "learning_rate": 1.2799060354174196e-05, "loss": 83.7889, "step": 3178 }, { "compression_loss": 84.34825134277344, "epoch": 1.15, "learning_rate": 1.2793639320563787e-05, "loss": 85.1193, "step": 3179 }, { "compression_loss": 83.95314025878906, "epoch": 1.15, "learning_rate": 1.2788218286953378e-05, "loss": 84.4111, "step": 3180 }, { "compression_loss": 83.93827819824219, "epoch": 1.15, "learning_rate": 1.2782797253342971e-05, "loss": 84.572, "step": 3181 }, { "compression_loss": 85.66908264160156, "epoch": 1.15, "learning_rate": 1.2777376219732562e-05, "loss": 86.0675, "step": 3182 }, { "compression_loss": 82.69390869140625, "epoch": 1.15, "learning_rate": 1.2771955186122154e-05, "loss": 82.9991, "step": 3183 }, { "compression_loss": 84.23934173583984, "epoch": 1.15, "learning_rate": 1.2766534152511745e-05, "loss": 84.7322, "step": 3184 }, { "compression_loss": 84.57084655761719, "epoch": 1.15, "learning_rate": 1.2761113118901338e-05, "loss": 84.789, "step": 3185 }, { "compression_loss": 83.47335052490234, "epoch": 1.15, "learning_rate": 1.275569208529093e-05, "loss": 83.788, "step": 3186 }, { "compression_loss": 85.58090209960938, "epoch": 1.15, "learning_rate": 1.275027105168052e-05, "loss": 86.0368, "step": 3187 }, { "compression_loss": 84.81010437011719, "epoch": 1.15, "learning_rate": 1.2744850018070112e-05, "loss": 85.2736, "step": 3188 }, { "compression_loss": 82.33219146728516, "epoch": 1.15, "learning_rate": 1.2739428984459705e-05, "loss": 82.5588, "step": 3189 }, { "compression_loss": 82.7366943359375, "epoch": 1.15, "learning_rate": 1.2734007950849296e-05, "loss": 83.3904, "step": 3190 }, { "compression_loss": 84.37208557128906, "epoch": 1.15, "learning_rate": 1.2728586917238887e-05, "loss": 84.6677, "step": 3191 }, { "compression_loss": 83.49188995361328, "epoch": 1.15, "learning_rate": 1.2723165883628479e-05, "loss": 83.7792, "step": 3192 }, { "compression_loss": 83.73384094238281, "epoch": 1.15, "learning_rate": 1.2717744850018072e-05, "loss": 84.0505, "step": 3193 }, { "compression_loss": 83.75567626953125, "epoch": 1.15, "learning_rate": 1.2712323816407663e-05, "loss": 84.0703, "step": 3194 }, { "compression_loss": 85.49605560302734, "epoch": 1.15, "learning_rate": 1.2706902782797254e-05, "loss": 86.1194, "step": 3195 }, { "compression_loss": 83.36190795898438, "epoch": 1.16, "learning_rate": 1.2701481749186845e-05, "loss": 83.8942, "step": 3196 }, { "compression_loss": 85.23403930664062, "epoch": 1.16, "learning_rate": 1.2696060715576437e-05, "loss": 85.9896, "step": 3197 }, { "compression_loss": 84.30136108398438, "epoch": 1.16, "learning_rate": 1.2690639681966028e-05, "loss": 84.7017, "step": 3198 }, { "compression_loss": 83.46647644042969, "epoch": 1.16, "learning_rate": 1.268521864835562e-05, "loss": 84.0259, "step": 3199 }, { "compression_loss": 84.71067810058594, "epoch": 1.16, "learning_rate": 1.267979761474521e-05, "loss": 85.3487, "step": 3200 }, { "compression_loss": 83.38348388671875, "epoch": 1.16, "learning_rate": 1.2674376581134803e-05, "loss": 83.8996, "step": 3201 }, { "compression_loss": 83.0428237915039, "epoch": 1.16, "learning_rate": 1.2668955547524395e-05, "loss": 83.4857, "step": 3202 }, { "compression_loss": 84.30940246582031, "epoch": 1.16, "learning_rate": 1.2663534513913986e-05, "loss": 84.7959, "step": 3203 }, { "compression_loss": 85.34274291992188, "epoch": 1.16, "learning_rate": 1.2658113480303577e-05, "loss": 85.8189, "step": 3204 }, { "compression_loss": 83.48893737792969, "epoch": 1.16, "learning_rate": 1.265269244669317e-05, "loss": 83.8783, "step": 3205 }, { "compression_loss": 85.03956604003906, "epoch": 1.16, "learning_rate": 1.2647271413082761e-05, "loss": 85.5769, "step": 3206 }, { "compression_loss": 83.76231384277344, "epoch": 1.16, "learning_rate": 1.2641850379472353e-05, "loss": 84.0341, "step": 3207 }, { "compression_loss": 85.78536987304688, "epoch": 1.16, "learning_rate": 1.2636429345861944e-05, "loss": 86.1294, "step": 3208 }, { "compression_loss": 85.92900085449219, "epoch": 1.16, "learning_rate": 1.2631008312251537e-05, "loss": 86.6347, "step": 3209 }, { "compression_loss": 84.98106384277344, "epoch": 1.16, "learning_rate": 1.2625587278641128e-05, "loss": 85.7083, "step": 3210 }, { "compression_loss": 83.831298828125, "epoch": 1.16, "learning_rate": 1.262016624503072e-05, "loss": 84.2212, "step": 3211 }, { "compression_loss": 84.66887664794922, "epoch": 1.16, "learning_rate": 1.261474521142031e-05, "loss": 85.0702, "step": 3212 }, { "compression_loss": 84.21649169921875, "epoch": 1.16, "learning_rate": 1.2609324177809904e-05, "loss": 84.6059, "step": 3213 }, { "compression_loss": 85.69984436035156, "epoch": 1.16, "learning_rate": 1.2603903144199495e-05, "loss": 86.1605, "step": 3214 }, { "compression_loss": 84.264892578125, "epoch": 1.16, "learning_rate": 1.2598482110589086e-05, "loss": 84.8041, "step": 3215 }, { "compression_loss": 83.7911376953125, "epoch": 1.16, "learning_rate": 1.2593061076978678e-05, "loss": 84.3195, "step": 3216 }, { "compression_loss": 85.54519653320312, "epoch": 1.16, "learning_rate": 1.258764004336827e-05, "loss": 85.931, "step": 3217 }, { "compression_loss": 85.41883087158203, "epoch": 1.16, "learning_rate": 1.2582219009757862e-05, "loss": 85.9851, "step": 3218 }, { "compression_loss": 85.76296997070312, "epoch": 1.16, "learning_rate": 1.2576797976147451e-05, "loss": 86.3085, "step": 3219 }, { "compression_loss": 85.449462890625, "epoch": 1.16, "learning_rate": 1.2571376942537043e-05, "loss": 86.2122, "step": 3220 }, { "compression_loss": 84.63139343261719, "epoch": 1.16, "learning_rate": 1.2565955908926636e-05, "loss": 84.944, "step": 3221 }, { "compression_loss": 83.59636688232422, "epoch": 1.16, "learning_rate": 1.2560534875316227e-05, "loss": 84.1581, "step": 3222 }, { "compression_loss": 84.89827728271484, "epoch": 1.16, "learning_rate": 1.2555113841705818e-05, "loss": 85.2352, "step": 3223 }, { "compression_loss": 84.55271911621094, "epoch": 1.17, "learning_rate": 1.254969280809541e-05, "loss": 84.955, "step": 3224 }, { "compression_loss": 83.32723999023438, "epoch": 1.17, "learning_rate": 1.2544271774485002e-05, "loss": 83.76, "step": 3225 }, { "compression_loss": 83.83110046386719, "epoch": 1.17, "learning_rate": 1.2538850740874594e-05, "loss": 84.0991, "step": 3226 }, { "compression_loss": 84.50524139404297, "epoch": 1.17, "learning_rate": 1.2533429707264185e-05, "loss": 85.0471, "step": 3227 }, { "compression_loss": 84.75717163085938, "epoch": 1.17, "learning_rate": 1.2528008673653776e-05, "loss": 85.1684, "step": 3228 }, { "compression_loss": 84.49281311035156, "epoch": 1.17, "learning_rate": 1.252258764004337e-05, "loss": 85.1301, "step": 3229 }, { "compression_loss": 85.88832092285156, "epoch": 1.17, "learning_rate": 1.251716660643296e-05, "loss": 86.132, "step": 3230 }, { "compression_loss": 83.60052490234375, "epoch": 1.17, "learning_rate": 1.2511745572822552e-05, "loss": 83.9318, "step": 3231 }, { "compression_loss": 85.31128692626953, "epoch": 1.17, "learning_rate": 1.2506324539212143e-05, "loss": 86.128, "step": 3232 }, { "compression_loss": 85.57643127441406, "epoch": 1.17, "learning_rate": 1.2500903505601736e-05, "loss": 86.1075, "step": 3233 }, { "compression_loss": 85.21629333496094, "epoch": 1.17, "learning_rate": 1.2495482471991327e-05, "loss": 85.6612, "step": 3234 }, { "compression_loss": 83.99674987792969, "epoch": 1.17, "learning_rate": 1.2490061438380918e-05, "loss": 84.7565, "step": 3235 }, { "compression_loss": 85.36502075195312, "epoch": 1.17, "learning_rate": 1.248464040477051e-05, "loss": 85.9096, "step": 3236 }, { "compression_loss": 83.58352661132812, "epoch": 1.17, "learning_rate": 1.2479219371160103e-05, "loss": 84.1351, "step": 3237 }, { "compression_loss": 84.00550842285156, "epoch": 1.17, "learning_rate": 1.2473798337549694e-05, "loss": 84.7056, "step": 3238 }, { "compression_loss": 84.64302062988281, "epoch": 1.17, "learning_rate": 1.2468377303939285e-05, "loss": 85.0957, "step": 3239 }, { "compression_loss": 84.4261703491211, "epoch": 1.17, "learning_rate": 1.2462956270328877e-05, "loss": 84.831, "step": 3240 }, { "compression_loss": 83.56932067871094, "epoch": 1.17, "learning_rate": 1.2457535236718468e-05, "loss": 83.9738, "step": 3241 }, { "compression_loss": 83.51958465576172, "epoch": 1.17, "learning_rate": 1.2452114203108059e-05, "loss": 83.9548, "step": 3242 }, { "compression_loss": 84.46684265136719, "epoch": 1.17, "learning_rate": 1.244669316949765e-05, "loss": 85.1613, "step": 3243 }, { "compression_loss": 86.40609741210938, "epoch": 1.17, "learning_rate": 1.2441272135887242e-05, "loss": 86.9787, "step": 3244 }, { "compression_loss": 82.59867858886719, "epoch": 1.17, "learning_rate": 1.2435851102276835e-05, "loss": 82.9932, "step": 3245 }, { "compression_loss": 83.7618408203125, "epoch": 1.17, "learning_rate": 1.2430430068666426e-05, "loss": 84.1517, "step": 3246 }, { "compression_loss": 84.77297973632812, "epoch": 1.17, "learning_rate": 1.2425009035056017e-05, "loss": 85.3063, "step": 3247 }, { "compression_loss": 84.54617309570312, "epoch": 1.17, "learning_rate": 1.2419588001445608e-05, "loss": 85.0971, "step": 3248 }, { "compression_loss": 86.19196319580078, "epoch": 1.17, "learning_rate": 1.2414166967835201e-05, "loss": 86.5801, "step": 3249 }, { "compression_loss": 82.76441955566406, "epoch": 1.17, "learning_rate": 1.2408745934224793e-05, "loss": 83.156, "step": 3250 }, { "epoch": 1.17, "eval_exact_match": 86.81173131504258, "eval_f1": 93.0060092981047, "step": 3250 }, { "compression_loss": 84.18540954589844, "epoch": 1.17, "learning_rate": 1.2403324900614384e-05, "loss": 84.6326, "step": 3251 }, { "compression_loss": 83.12358093261719, "epoch": 1.18, "learning_rate": 1.2397903867003975e-05, "loss": 83.4257, "step": 3252 }, { "compression_loss": 81.6822738647461, "epoch": 1.18, "learning_rate": 1.2392482833393568e-05, "loss": 82.012, "step": 3253 }, { "compression_loss": 83.14865112304688, "epoch": 1.18, "learning_rate": 1.238706179978316e-05, "loss": 83.5377, "step": 3254 }, { "compression_loss": 83.77715301513672, "epoch": 1.18, "learning_rate": 1.238164076617275e-05, "loss": 84.4125, "step": 3255 }, { "compression_loss": 83.70116424560547, "epoch": 1.18, "learning_rate": 1.2376219732562342e-05, "loss": 84.1695, "step": 3256 }, { "compression_loss": 85.05195617675781, "epoch": 1.18, "learning_rate": 1.2370798698951935e-05, "loss": 85.4357, "step": 3257 }, { "compression_loss": 82.72850036621094, "epoch": 1.18, "learning_rate": 1.2365377665341526e-05, "loss": 82.8996, "step": 3258 }, { "compression_loss": 84.15387725830078, "epoch": 1.18, "learning_rate": 1.2359956631731117e-05, "loss": 84.8011, "step": 3259 }, { "compression_loss": 84.97250366210938, "epoch": 1.18, "learning_rate": 1.2354535598120709e-05, "loss": 85.5311, "step": 3260 }, { "compression_loss": 83.6077880859375, "epoch": 1.18, "learning_rate": 1.2349114564510302e-05, "loss": 84.146, "step": 3261 }, { "compression_loss": 85.19189453125, "epoch": 1.18, "learning_rate": 1.2343693530899893e-05, "loss": 85.8956, "step": 3262 }, { "compression_loss": 85.570556640625, "epoch": 1.18, "learning_rate": 1.2338272497289483e-05, "loss": 85.9034, "step": 3263 }, { "compression_loss": 85.30094909667969, "epoch": 1.18, "learning_rate": 1.2332851463679074e-05, "loss": 85.9612, "step": 3264 }, { "compression_loss": 83.11266326904297, "epoch": 1.18, "learning_rate": 1.2327430430068667e-05, "loss": 83.6008, "step": 3265 }, { "compression_loss": 84.57057189941406, "epoch": 1.18, "learning_rate": 1.2322009396458258e-05, "loss": 84.7481, "step": 3266 }, { "compression_loss": 84.20283508300781, "epoch": 1.18, "learning_rate": 1.231658836284785e-05, "loss": 84.5402, "step": 3267 }, { "compression_loss": 82.6439208984375, "epoch": 1.18, "learning_rate": 1.231116732923744e-05, "loss": 82.8801, "step": 3268 }, { "compression_loss": 85.20193481445312, "epoch": 1.18, "learning_rate": 1.2305746295627034e-05, "loss": 85.6888, "step": 3269 }, { "compression_loss": 85.34956359863281, "epoch": 1.18, "learning_rate": 1.2300325262016625e-05, "loss": 85.6404, "step": 3270 }, { "compression_loss": 85.11868286132812, "epoch": 1.18, "learning_rate": 1.2294904228406216e-05, "loss": 85.5811, "step": 3271 }, { "compression_loss": 83.80540466308594, "epoch": 1.18, "learning_rate": 1.2289483194795807e-05, "loss": 84.3722, "step": 3272 }, { "compression_loss": 86.032470703125, "epoch": 1.18, "learning_rate": 1.22840621611854e-05, "loss": 86.5233, "step": 3273 }, { "compression_loss": 85.24299621582031, "epoch": 1.18, "learning_rate": 1.2278641127574992e-05, "loss": 85.969, "step": 3274 }, { "compression_loss": 86.13706970214844, "epoch": 1.18, "learning_rate": 1.2273220093964583e-05, "loss": 86.525, "step": 3275 }, { "compression_loss": 83.96217346191406, "epoch": 1.18, "learning_rate": 1.2267799060354174e-05, "loss": 84.2956, "step": 3276 }, { "compression_loss": 85.15731811523438, "epoch": 1.18, "learning_rate": 1.2262378026743767e-05, "loss": 85.5964, "step": 3277 }, { "compression_loss": 84.37813568115234, "epoch": 1.18, "learning_rate": 1.2256956993133358e-05, "loss": 84.8183, "step": 3278 }, { "compression_loss": 83.56631469726562, "epoch": 1.19, "learning_rate": 1.225153595952295e-05, "loss": 83.997, "step": 3279 }, { "compression_loss": 86.10673522949219, "epoch": 1.19, "learning_rate": 1.2246114925912541e-05, "loss": 86.5895, "step": 3280 }, { "compression_loss": 87.53738403320312, "epoch": 1.19, "learning_rate": 1.2240693892302134e-05, "loss": 88.3074, "step": 3281 }, { "compression_loss": 82.98892974853516, "epoch": 1.19, "learning_rate": 1.2235272858691725e-05, "loss": 83.2768, "step": 3282 }, { "compression_loss": 86.08692932128906, "epoch": 1.19, "learning_rate": 1.2229851825081316e-05, "loss": 86.5307, "step": 3283 }, { "compression_loss": 82.87510681152344, "epoch": 1.19, "learning_rate": 1.2224430791470908e-05, "loss": 83.2265, "step": 3284 }, { "compression_loss": 85.68463897705078, "epoch": 1.19, "learning_rate": 1.2219009757860499e-05, "loss": 86.1712, "step": 3285 }, { "compression_loss": 85.27470397949219, "epoch": 1.19, "learning_rate": 1.221358872425009e-05, "loss": 85.7492, "step": 3286 }, { "compression_loss": 85.33674621582031, "epoch": 1.19, "learning_rate": 1.2208167690639682e-05, "loss": 85.712, "step": 3287 }, { "compression_loss": 84.44965362548828, "epoch": 1.19, "learning_rate": 1.2202746657029273e-05, "loss": 85.374, "step": 3288 }, { "compression_loss": 84.00003814697266, "epoch": 1.19, "learning_rate": 1.2197325623418866e-05, "loss": 84.5109, "step": 3289 }, { "compression_loss": 83.82498168945312, "epoch": 1.19, "learning_rate": 1.2191904589808457e-05, "loss": 84.5643, "step": 3290 }, { "compression_loss": 82.95785522460938, "epoch": 1.19, "learning_rate": 1.2186483556198048e-05, "loss": 83.1456, "step": 3291 }, { "compression_loss": 81.94219970703125, "epoch": 1.19, "learning_rate": 1.218106252258764e-05, "loss": 82.3859, "step": 3292 }, { "compression_loss": 84.86549377441406, "epoch": 1.19, "learning_rate": 1.2175641488977233e-05, "loss": 85.1638, "step": 3293 }, { "compression_loss": 84.80221557617188, "epoch": 1.19, "learning_rate": 1.2170220455366824e-05, "loss": 85.3015, "step": 3294 }, { "compression_loss": 84.63156127929688, "epoch": 1.19, "learning_rate": 1.2164799421756415e-05, "loss": 84.9327, "step": 3295 }, { "compression_loss": 83.11874389648438, "epoch": 1.19, "learning_rate": 1.2159378388146006e-05, "loss": 83.5093, "step": 3296 }, { "compression_loss": 82.34150695800781, "epoch": 1.19, "learning_rate": 1.21539573545356e-05, "loss": 82.8201, "step": 3297 }, { "compression_loss": 84.00201416015625, "epoch": 1.19, "learning_rate": 1.214853632092519e-05, "loss": 84.5128, "step": 3298 }, { "compression_loss": 85.32437896728516, "epoch": 1.19, "learning_rate": 1.2143115287314782e-05, "loss": 85.756, "step": 3299 }, { "compression_loss": 85.3444595336914, "epoch": 1.19, "learning_rate": 1.2137694253704373e-05, "loss": 85.8333, "step": 3300 }, { "compression_loss": 84.84156799316406, "epoch": 1.19, "learning_rate": 1.2132273220093966e-05, "loss": 85.2826, "step": 3301 }, { "compression_loss": 84.65463256835938, "epoch": 1.19, "learning_rate": 1.2126852186483557e-05, "loss": 85.1477, "step": 3302 }, { "compression_loss": 85.11763000488281, "epoch": 1.19, "learning_rate": 1.2121431152873149e-05, "loss": 85.679, "step": 3303 }, { "compression_loss": 84.0965347290039, "epoch": 1.19, "learning_rate": 1.211601011926274e-05, "loss": 84.5553, "step": 3304 }, { "compression_loss": 83.29291534423828, "epoch": 1.19, "learning_rate": 1.2110589085652333e-05, "loss": 83.534, "step": 3305 }, { "compression_loss": 83.47953033447266, "epoch": 1.19, "learning_rate": 1.2105168052041923e-05, "loss": 83.9172, "step": 3306 }, { "compression_loss": 84.67040252685547, "epoch": 1.2, "learning_rate": 1.2099747018431514e-05, "loss": 84.9932, "step": 3307 }, { "compression_loss": 85.22293090820312, "epoch": 1.2, "learning_rate": 1.2094325984821105e-05, "loss": 85.5453, "step": 3308 }, { "compression_loss": 85.92212677001953, "epoch": 1.2, "learning_rate": 1.2088904951210698e-05, "loss": 86.7673, "step": 3309 }, { "compression_loss": 84.61209869384766, "epoch": 1.2, "learning_rate": 1.208348391760029e-05, "loss": 85.0563, "step": 3310 }, { "compression_loss": 83.74957275390625, "epoch": 1.2, "learning_rate": 1.207806288398988e-05, "loss": 84.0989, "step": 3311 }, { "compression_loss": 83.1948471069336, "epoch": 1.2, "learning_rate": 1.2072641850379472e-05, "loss": 83.6484, "step": 3312 }, { "compression_loss": 85.82395935058594, "epoch": 1.2, "learning_rate": 1.2067220816769065e-05, "loss": 86.2142, "step": 3313 }, { "compression_loss": 84.22000122070312, "epoch": 1.2, "learning_rate": 1.2061799783158656e-05, "loss": 84.7196, "step": 3314 }, { "compression_loss": 86.33499908447266, "epoch": 1.2, "learning_rate": 1.2056378749548247e-05, "loss": 86.8248, "step": 3315 }, { "compression_loss": 84.8702392578125, "epoch": 1.2, "learning_rate": 1.2050957715937839e-05, "loss": 85.1356, "step": 3316 }, { "compression_loss": 85.67382049560547, "epoch": 1.2, "learning_rate": 1.2045536682327432e-05, "loss": 86.3157, "step": 3317 }, { "compression_loss": 85.12398529052734, "epoch": 1.2, "learning_rate": 1.2040115648717023e-05, "loss": 85.4345, "step": 3318 }, { "compression_loss": 83.2242431640625, "epoch": 1.2, "learning_rate": 1.2034694615106614e-05, "loss": 83.5096, "step": 3319 }, { "compression_loss": 83.29721069335938, "epoch": 1.2, "learning_rate": 1.2029273581496205e-05, "loss": 83.8689, "step": 3320 }, { "compression_loss": 84.44229125976562, "epoch": 1.2, "learning_rate": 1.2023852547885798e-05, "loss": 85.0317, "step": 3321 }, { "compression_loss": 85.52388000488281, "epoch": 1.2, "learning_rate": 1.201843151427539e-05, "loss": 86.0652, "step": 3322 }, { "compression_loss": 83.75028228759766, "epoch": 1.2, "learning_rate": 1.2013010480664981e-05, "loss": 84.3138, "step": 3323 }, { "compression_loss": 85.28043365478516, "epoch": 1.2, "learning_rate": 1.2007589447054572e-05, "loss": 85.8511, "step": 3324 }, { "compression_loss": 84.07683563232422, "epoch": 1.2, "learning_rate": 1.2002168413444165e-05, "loss": 84.3143, "step": 3325 }, { "compression_loss": 86.29144287109375, "epoch": 1.2, "learning_rate": 1.1996747379833756e-05, "loss": 86.6781, "step": 3326 }, { "compression_loss": 85.22927856445312, "epoch": 1.2, "learning_rate": 1.1991326346223348e-05, "loss": 85.6449, "step": 3327 }, { "compression_loss": 84.2564468383789, "epoch": 1.2, "learning_rate": 1.1985905312612937e-05, "loss": 84.6786, "step": 3328 }, { "compression_loss": 84.63423156738281, "epoch": 1.2, "learning_rate": 1.1980484279002529e-05, "loss": 84.8581, "step": 3329 }, { "compression_loss": 84.67626953125, "epoch": 1.2, "learning_rate": 1.1975063245392122e-05, "loss": 85.409, "step": 3330 }, { "compression_loss": 84.47721862792969, "epoch": 1.2, "learning_rate": 1.1969642211781713e-05, "loss": 84.8992, "step": 3331 }, { "compression_loss": 86.16498565673828, "epoch": 1.2, "learning_rate": 1.1964221178171304e-05, "loss": 86.6976, "step": 3332 }, { "compression_loss": 83.70536041259766, "epoch": 1.2, "learning_rate": 1.1958800144560895e-05, "loss": 84.2509, "step": 3333 }, { "compression_loss": 83.56311798095703, "epoch": 1.2, "learning_rate": 1.1953379110950488e-05, "loss": 83.8074, "step": 3334 }, { "compression_loss": 86.09588623046875, "epoch": 1.21, "learning_rate": 1.194795807734008e-05, "loss": 86.5935, "step": 3335 }, { "compression_loss": 83.95164489746094, "epoch": 1.21, "learning_rate": 1.194253704372967e-05, "loss": 84.3458, "step": 3336 }, { "compression_loss": 84.89659118652344, "epoch": 1.21, "learning_rate": 1.1937116010119262e-05, "loss": 85.4484, "step": 3337 }, { "compression_loss": 84.67381286621094, "epoch": 1.21, "learning_rate": 1.1931694976508855e-05, "loss": 85.1372, "step": 3338 }, { "compression_loss": 84.00029754638672, "epoch": 1.21, "learning_rate": 1.1926273942898446e-05, "loss": 84.7785, "step": 3339 }, { "compression_loss": 84.81451416015625, "epoch": 1.21, "learning_rate": 1.1920852909288038e-05, "loss": 85.3054, "step": 3340 }, { "compression_loss": 82.61637878417969, "epoch": 1.21, "learning_rate": 1.1915431875677629e-05, "loss": 83.275, "step": 3341 }, { "compression_loss": 81.88284301757812, "epoch": 1.21, "learning_rate": 1.1910010842067222e-05, "loss": 82.1986, "step": 3342 }, { "compression_loss": 83.8918685913086, "epoch": 1.21, "learning_rate": 1.1904589808456813e-05, "loss": 84.4448, "step": 3343 }, { "compression_loss": 82.85616302490234, "epoch": 1.21, "learning_rate": 1.1899168774846404e-05, "loss": 83.1297, "step": 3344 }, { "compression_loss": 84.48793029785156, "epoch": 1.21, "learning_rate": 1.1893747741235996e-05, "loss": 84.9357, "step": 3345 }, { "compression_loss": 85.90104675292969, "epoch": 1.21, "learning_rate": 1.1888326707625589e-05, "loss": 86.5265, "step": 3346 }, { "compression_loss": 83.16570281982422, "epoch": 1.21, "learning_rate": 1.188290567401518e-05, "loss": 83.5984, "step": 3347 }, { "compression_loss": 84.36026000976562, "epoch": 1.21, "learning_rate": 1.1877484640404771e-05, "loss": 84.8178, "step": 3348 }, { "compression_loss": 85.45780181884766, "epoch": 1.21, "learning_rate": 1.1872063606794362e-05, "loss": 85.7145, "step": 3349 }, { "compression_loss": 84.12686157226562, "epoch": 1.21, "learning_rate": 1.1866642573183954e-05, "loss": 84.5269, "step": 3350 }, { "compression_loss": 85.20064544677734, "epoch": 1.21, "learning_rate": 1.1861221539573545e-05, "loss": 85.6012, "step": 3351 }, { "compression_loss": 84.27477264404297, "epoch": 1.21, "learning_rate": 1.1855800505963136e-05, "loss": 84.6938, "step": 3352 }, { "compression_loss": 85.20240783691406, "epoch": 1.21, "learning_rate": 1.1850379472352728e-05, "loss": 85.7293, "step": 3353 }, { "compression_loss": 84.49610900878906, "epoch": 1.21, "learning_rate": 1.184495843874232e-05, "loss": 84.7196, "step": 3354 }, { "compression_loss": 85.75595092773438, "epoch": 1.21, "learning_rate": 1.1839537405131912e-05, "loss": 86.2228, "step": 3355 }, { "compression_loss": 84.2415771484375, "epoch": 1.21, "learning_rate": 1.1834116371521503e-05, "loss": 85.2227, "step": 3356 }, { "compression_loss": 85.42139434814453, "epoch": 1.21, "learning_rate": 1.1828695337911094e-05, "loss": 85.8241, "step": 3357 }, { "compression_loss": 84.59713745117188, "epoch": 1.21, "learning_rate": 1.1823274304300687e-05, "loss": 84.8413, "step": 3358 }, { "compression_loss": 83.42183685302734, "epoch": 1.21, "learning_rate": 1.1817853270690279e-05, "loss": 83.655, "step": 3359 }, { "compression_loss": 84.9700698852539, "epoch": 1.21, "learning_rate": 1.181243223707987e-05, "loss": 85.4007, "step": 3360 }, { "compression_loss": 84.82536315917969, "epoch": 1.21, "learning_rate": 1.1807011203469461e-05, "loss": 85.1135, "step": 3361 }, { "compression_loss": 84.16038513183594, "epoch": 1.22, "learning_rate": 1.1801590169859054e-05, "loss": 84.4964, "step": 3362 }, { "compression_loss": 82.91300964355469, "epoch": 1.22, "learning_rate": 1.1796169136248645e-05, "loss": 83.2999, "step": 3363 }, { "compression_loss": 85.17079162597656, "epoch": 1.22, "learning_rate": 1.1790748102638237e-05, "loss": 85.4342, "step": 3364 }, { "compression_loss": 83.45320892333984, "epoch": 1.22, "learning_rate": 1.1785327069027828e-05, "loss": 83.8579, "step": 3365 }, { "compression_loss": 85.77392578125, "epoch": 1.22, "learning_rate": 1.1779906035417421e-05, "loss": 86.1938, "step": 3366 }, { "compression_loss": 83.49533081054688, "epoch": 1.22, "learning_rate": 1.1774485001807012e-05, "loss": 83.9191, "step": 3367 }, { "compression_loss": 83.50640106201172, "epoch": 1.22, "learning_rate": 1.1769063968196603e-05, "loss": 84.142, "step": 3368 }, { "compression_loss": 84.42646789550781, "epoch": 1.22, "learning_rate": 1.1763642934586195e-05, "loss": 84.7569, "step": 3369 }, { "compression_loss": 81.60501098632812, "epoch": 1.22, "learning_rate": 1.1758221900975788e-05, "loss": 81.93, "step": 3370 }, { "compression_loss": 83.938232421875, "epoch": 1.22, "learning_rate": 1.1752800867365379e-05, "loss": 84.6603, "step": 3371 }, { "compression_loss": 83.81153869628906, "epoch": 1.22, "learning_rate": 1.1747379833754968e-05, "loss": 84.1376, "step": 3372 }, { "compression_loss": 84.52661895751953, "epoch": 1.22, "learning_rate": 1.174195880014456e-05, "loss": 85.137, "step": 3373 }, { "compression_loss": 83.40473937988281, "epoch": 1.22, "learning_rate": 1.1736537766534153e-05, "loss": 83.9079, "step": 3374 }, { "compression_loss": 83.42662048339844, "epoch": 1.22, "learning_rate": 1.1731116732923744e-05, "loss": 83.7747, "step": 3375 }, { "compression_loss": 82.745361328125, "epoch": 1.22, "learning_rate": 1.1725695699313335e-05, "loss": 83.0883, "step": 3376 }, { "compression_loss": 84.08284759521484, "epoch": 1.22, "learning_rate": 1.1720274665702927e-05, "loss": 84.4559, "step": 3377 }, { "compression_loss": 83.93333435058594, "epoch": 1.22, "learning_rate": 1.171485363209252e-05, "loss": 84.2322, "step": 3378 }, { "compression_loss": 84.39238739013672, "epoch": 1.22, "learning_rate": 1.170943259848211e-05, "loss": 84.7202, "step": 3379 }, { "compression_loss": 84.5704116821289, "epoch": 1.22, "learning_rate": 1.1704011564871702e-05, "loss": 85.0678, "step": 3380 }, { "compression_loss": 84.99198913574219, "epoch": 1.22, "learning_rate": 1.1698590531261293e-05, "loss": 85.3755, "step": 3381 }, { "compression_loss": 84.81494140625, "epoch": 1.22, "learning_rate": 1.1693169497650886e-05, "loss": 85.1472, "step": 3382 }, { "compression_loss": 84.52352905273438, "epoch": 1.22, "learning_rate": 1.1687748464040478e-05, "loss": 84.8769, "step": 3383 }, { "compression_loss": 84.01907348632812, "epoch": 1.22, "learning_rate": 1.1682327430430069e-05, "loss": 84.6893, "step": 3384 }, { "compression_loss": 83.05330657958984, "epoch": 1.22, "learning_rate": 1.167690639681966e-05, "loss": 83.2542, "step": 3385 }, { "compression_loss": 84.40203094482422, "epoch": 1.22, "learning_rate": 1.1671485363209253e-05, "loss": 84.8954, "step": 3386 }, { "compression_loss": 85.21952819824219, "epoch": 1.22, "learning_rate": 1.1666064329598844e-05, "loss": 85.9048, "step": 3387 }, { "compression_loss": 84.6250228881836, "epoch": 1.22, "learning_rate": 1.1660643295988436e-05, "loss": 85.1321, "step": 3388 }, { "compression_loss": 85.2994613647461, "epoch": 1.22, "learning_rate": 1.1655222262378027e-05, "loss": 85.7167, "step": 3389 }, { "compression_loss": 84.50982666015625, "epoch": 1.23, "learning_rate": 1.164980122876762e-05, "loss": 85.1087, "step": 3390 }, { "compression_loss": 83.78416442871094, "epoch": 1.23, "learning_rate": 1.1644380195157211e-05, "loss": 84.2639, "step": 3391 }, { "compression_loss": 85.14986419677734, "epoch": 1.23, "learning_rate": 1.1638959161546802e-05, "loss": 85.4911, "step": 3392 }, { "compression_loss": 83.86639404296875, "epoch": 1.23, "learning_rate": 1.1633538127936394e-05, "loss": 84.3755, "step": 3393 }, { "compression_loss": 86.40637969970703, "epoch": 1.23, "learning_rate": 1.1628117094325985e-05, "loss": 87.0296, "step": 3394 }, { "compression_loss": 85.15936279296875, "epoch": 1.23, "learning_rate": 1.1622696060715576e-05, "loss": 85.634, "step": 3395 }, { "compression_loss": 84.510498046875, "epoch": 1.23, "learning_rate": 1.1617275027105167e-05, "loss": 84.78, "step": 3396 }, { "compression_loss": 84.62335968017578, "epoch": 1.23, "learning_rate": 1.1611853993494759e-05, "loss": 85.1525, "step": 3397 }, { "compression_loss": 83.51883697509766, "epoch": 1.23, "learning_rate": 1.1606432959884352e-05, "loss": 84.3858, "step": 3398 }, { "compression_loss": 84.69752502441406, "epoch": 1.23, "learning_rate": 1.1601011926273943e-05, "loss": 85.074, "step": 3399 }, { "compression_loss": 84.86839294433594, "epoch": 1.23, "learning_rate": 1.1595590892663534e-05, "loss": 85.3255, "step": 3400 }, { "compression_loss": 85.61363220214844, "epoch": 1.23, "learning_rate": 1.1590169859053126e-05, "loss": 86.265, "step": 3401 }, { "compression_loss": 85.3212890625, "epoch": 1.23, "learning_rate": 1.1584748825442718e-05, "loss": 85.6516, "step": 3402 }, { "compression_loss": 84.6114730834961, "epoch": 1.23, "learning_rate": 1.157932779183231e-05, "loss": 85.0753, "step": 3403 }, { "compression_loss": 84.31903839111328, "epoch": 1.23, "learning_rate": 1.1573906758221901e-05, "loss": 85.033, "step": 3404 }, { "compression_loss": 83.95038604736328, "epoch": 1.23, "learning_rate": 1.1568485724611492e-05, "loss": 84.4822, "step": 3405 }, { "compression_loss": 84.98823547363281, "epoch": 1.23, "learning_rate": 1.1563064691001085e-05, "loss": 85.5902, "step": 3406 }, { "compression_loss": 83.29793548583984, "epoch": 1.23, "learning_rate": 1.1557643657390677e-05, "loss": 83.6608, "step": 3407 }, { "compression_loss": 84.8351821899414, "epoch": 1.23, "learning_rate": 1.1552222623780268e-05, "loss": 85.5318, "step": 3408 }, { "compression_loss": 85.48023223876953, "epoch": 1.23, "learning_rate": 1.1546801590169859e-05, "loss": 86.0183, "step": 3409 }, { "compression_loss": 85.67987823486328, "epoch": 1.23, "learning_rate": 1.1541380556559452e-05, "loss": 86.3329, "step": 3410 }, { "compression_loss": 85.81775665283203, "epoch": 1.23, "learning_rate": 1.1535959522949043e-05, "loss": 86.5574, "step": 3411 }, { "compression_loss": 84.72953796386719, "epoch": 1.23, "learning_rate": 1.1530538489338635e-05, "loss": 85.0367, "step": 3412 }, { "compression_loss": 82.7982406616211, "epoch": 1.23, "learning_rate": 1.1525117455728226e-05, "loss": 83.1427, "step": 3413 }, { "compression_loss": 84.93203735351562, "epoch": 1.23, "learning_rate": 1.1519696422117819e-05, "loss": 85.277, "step": 3414 }, { "compression_loss": 83.53541564941406, "epoch": 1.23, "learning_rate": 1.1514275388507408e-05, "loss": 83.8309, "step": 3415 }, { "compression_loss": 83.9692611694336, "epoch": 1.23, "learning_rate": 1.1508854354897e-05, "loss": 84.4416, "step": 3416 }, { "compression_loss": 83.41175842285156, "epoch": 1.23, "learning_rate": 1.1503433321286591e-05, "loss": 83.7293, "step": 3417 }, { "compression_loss": 84.66459655761719, "epoch": 1.24, "learning_rate": 1.1498012287676184e-05, "loss": 85.2935, "step": 3418 }, { "compression_loss": 86.00741577148438, "epoch": 1.24, "learning_rate": 1.1492591254065775e-05, "loss": 86.6763, "step": 3419 }, { "compression_loss": 84.25001525878906, "epoch": 1.24, "learning_rate": 1.1487170220455366e-05, "loss": 84.7259, "step": 3420 }, { "compression_loss": 83.81009674072266, "epoch": 1.24, "learning_rate": 1.1481749186844958e-05, "loss": 84.2921, "step": 3421 }, { "compression_loss": 83.68539428710938, "epoch": 1.24, "learning_rate": 1.147632815323455e-05, "loss": 84.1813, "step": 3422 }, { "compression_loss": 85.36537170410156, "epoch": 1.24, "learning_rate": 1.1470907119624142e-05, "loss": 85.7243, "step": 3423 }, { "compression_loss": 84.82328033447266, "epoch": 1.24, "learning_rate": 1.1465486086013733e-05, "loss": 85.1933, "step": 3424 }, { "compression_loss": 86.57225036621094, "epoch": 1.24, "learning_rate": 1.1460065052403325e-05, "loss": 86.7611, "step": 3425 }, { "compression_loss": 84.1319351196289, "epoch": 1.24, "learning_rate": 1.1454644018792917e-05, "loss": 84.656, "step": 3426 }, { "compression_loss": 83.56260681152344, "epoch": 1.24, "learning_rate": 1.1449222985182509e-05, "loss": 83.885, "step": 3427 }, { "compression_loss": 84.70234680175781, "epoch": 1.24, "learning_rate": 1.14438019515721e-05, "loss": 85.2071, "step": 3428 }, { "compression_loss": 85.87274169921875, "epoch": 1.24, "learning_rate": 1.1438380917961691e-05, "loss": 86.5252, "step": 3429 }, { "compression_loss": 85.05928039550781, "epoch": 1.24, "learning_rate": 1.1432959884351284e-05, "loss": 85.4872, "step": 3430 }, { "compression_loss": 84.60183715820312, "epoch": 1.24, "learning_rate": 1.1427538850740876e-05, "loss": 85.1009, "step": 3431 }, { "compression_loss": 84.62139892578125, "epoch": 1.24, "learning_rate": 1.1422117817130467e-05, "loss": 85.1352, "step": 3432 }, { "compression_loss": 85.79670715332031, "epoch": 1.24, "learning_rate": 1.1416696783520058e-05, "loss": 86.2238, "step": 3433 }, { "compression_loss": 84.44918823242188, "epoch": 1.24, "learning_rate": 1.1411275749909651e-05, "loss": 84.7054, "step": 3434 }, { "compression_loss": 84.05970764160156, "epoch": 1.24, "learning_rate": 1.1405854716299242e-05, "loss": 84.5613, "step": 3435 }, { "compression_loss": 82.31610107421875, "epoch": 1.24, "learning_rate": 1.1400433682688834e-05, "loss": 82.7758, "step": 3436 }, { "compression_loss": 84.473876953125, "epoch": 1.24, "learning_rate": 1.1395012649078423e-05, "loss": 85.1039, "step": 3437 }, { "compression_loss": 83.71028900146484, "epoch": 1.24, "learning_rate": 1.1389591615468016e-05, "loss": 84.1098, "step": 3438 }, { "compression_loss": 83.67436218261719, "epoch": 1.24, "learning_rate": 1.1384170581857607e-05, "loss": 84.0013, "step": 3439 }, { "compression_loss": 85.23165130615234, "epoch": 1.24, "learning_rate": 1.1378749548247199e-05, "loss": 85.7195, "step": 3440 }, { "compression_loss": 85.22782897949219, "epoch": 1.24, "learning_rate": 1.137332851463679e-05, "loss": 85.5604, "step": 3441 }, { "compression_loss": 84.36695861816406, "epoch": 1.24, "learning_rate": 1.1367907481026383e-05, "loss": 84.9469, "step": 3442 }, { "compression_loss": 83.61699676513672, "epoch": 1.24, "learning_rate": 1.1362486447415974e-05, "loss": 84.0188, "step": 3443 }, { "compression_loss": 81.62344360351562, "epoch": 1.24, "learning_rate": 1.1357065413805565e-05, "loss": 82.309, "step": 3444 }, { "compression_loss": 86.35469055175781, "epoch": 1.25, "learning_rate": 1.1351644380195157e-05, "loss": 86.9264, "step": 3445 }, { "compression_loss": 84.68169403076172, "epoch": 1.25, "learning_rate": 1.134622334658475e-05, "loss": 85.0605, "step": 3446 }, { "compression_loss": 86.66766357421875, "epoch": 1.25, "learning_rate": 1.1340802312974341e-05, "loss": 87.1408, "step": 3447 }, { "compression_loss": 83.5614242553711, "epoch": 1.25, "learning_rate": 1.1335381279363932e-05, "loss": 83.9683, "step": 3448 }, { "compression_loss": 84.06671905517578, "epoch": 1.25, "learning_rate": 1.1329960245753524e-05, "loss": 84.3788, "step": 3449 }, { "compression_loss": 83.69650268554688, "epoch": 1.25, "learning_rate": 1.1324539212143116e-05, "loss": 84.062, "step": 3450 }, { "compression_loss": 84.22216033935547, "epoch": 1.25, "learning_rate": 1.1319118178532708e-05, "loss": 84.5516, "step": 3451 }, { "compression_loss": 82.4967041015625, "epoch": 1.25, "learning_rate": 1.1313697144922299e-05, "loss": 82.8811, "step": 3452 }, { "compression_loss": 84.70062255859375, "epoch": 1.25, "learning_rate": 1.130827611131189e-05, "loss": 84.9519, "step": 3453 }, { "compression_loss": 83.01275634765625, "epoch": 1.25, "learning_rate": 1.1302855077701483e-05, "loss": 83.6161, "step": 3454 }, { "compression_loss": 86.28160095214844, "epoch": 1.25, "learning_rate": 1.1297434044091075e-05, "loss": 87.3481, "step": 3455 }, { "compression_loss": 85.96086120605469, "epoch": 1.25, "learning_rate": 1.1292013010480666e-05, "loss": 86.6239, "step": 3456 }, { "compression_loss": 85.58323669433594, "epoch": 1.25, "learning_rate": 1.1286591976870257e-05, "loss": 86.1293, "step": 3457 }, { "compression_loss": 83.93083190917969, "epoch": 1.25, "learning_rate": 1.128117094325985e-05, "loss": 84.594, "step": 3458 }, { "compression_loss": 83.98885345458984, "epoch": 1.25, "learning_rate": 1.127574990964944e-05, "loss": 84.3083, "step": 3459 }, { "compression_loss": 83.81526184082031, "epoch": 1.25, "learning_rate": 1.1270328876039031e-05, "loss": 84.087, "step": 3460 }, { "compression_loss": 84.57123565673828, "epoch": 1.25, "learning_rate": 1.1264907842428622e-05, "loss": 85.0883, "step": 3461 }, { "compression_loss": 85.08076477050781, "epoch": 1.25, "learning_rate": 1.1259486808818215e-05, "loss": 85.8468, "step": 3462 }, { "compression_loss": 83.36766052246094, "epoch": 1.25, "learning_rate": 1.1254065775207806e-05, "loss": 83.6727, "step": 3463 }, { "compression_loss": 86.41777038574219, "epoch": 1.25, "learning_rate": 1.1248644741597398e-05, "loss": 86.8143, "step": 3464 }, { "compression_loss": 83.89444732666016, "epoch": 1.25, "learning_rate": 1.1243223707986989e-05, "loss": 84.3421, "step": 3465 }, { "compression_loss": 84.1681900024414, "epoch": 1.25, "learning_rate": 1.1237802674376582e-05, "loss": 84.8, "step": 3466 }, { "compression_loss": 85.73332214355469, "epoch": 1.25, "learning_rate": 1.1232381640766173e-05, "loss": 86.1411, "step": 3467 }, { "compression_loss": 84.35989379882812, "epoch": 1.25, "learning_rate": 1.1226960607155764e-05, "loss": 84.5956, "step": 3468 }, { "compression_loss": 85.87724304199219, "epoch": 1.25, "learning_rate": 1.1221539573545356e-05, "loss": 86.3505, "step": 3469 }, { "compression_loss": 84.76589965820312, "epoch": 1.25, "learning_rate": 1.1216118539934949e-05, "loss": 85.3263, "step": 3470 }, { "compression_loss": 84.59973907470703, "epoch": 1.25, "learning_rate": 1.121069750632454e-05, "loss": 85.0265, "step": 3471 }, { "compression_loss": 84.07748413085938, "epoch": 1.25, "learning_rate": 1.1205276472714131e-05, "loss": 84.5872, "step": 3472 }, { "compression_loss": 83.9686279296875, "epoch": 1.26, "learning_rate": 1.1199855439103723e-05, "loss": 84.4208, "step": 3473 }, { "compression_loss": 84.80836486816406, "epoch": 1.26, "learning_rate": 1.1194434405493315e-05, "loss": 85.3854, "step": 3474 }, { "compression_loss": 85.66698455810547, "epoch": 1.26, "learning_rate": 1.1189013371882907e-05, "loss": 86.0272, "step": 3475 }, { "compression_loss": 85.17256164550781, "epoch": 1.26, "learning_rate": 1.1183592338272498e-05, "loss": 85.3982, "step": 3476 }, { "compression_loss": 86.1850814819336, "epoch": 1.26, "learning_rate": 1.117817130466209e-05, "loss": 86.7503, "step": 3477 }, { "compression_loss": 83.34900665283203, "epoch": 1.26, "learning_rate": 1.1172750271051682e-05, "loss": 83.8242, "step": 3478 }, { "compression_loss": 85.18995666503906, "epoch": 1.26, "learning_rate": 1.1167329237441274e-05, "loss": 85.5441, "step": 3479 }, { "compression_loss": 84.32323455810547, "epoch": 1.26, "learning_rate": 1.1161908203830865e-05, "loss": 84.6428, "step": 3480 }, { "compression_loss": 83.86194610595703, "epoch": 1.26, "learning_rate": 1.1156487170220454e-05, "loss": 84.3411, "step": 3481 }, { "compression_loss": 84.54055786132812, "epoch": 1.26, "learning_rate": 1.1151066136610047e-05, "loss": 85.0775, "step": 3482 }, { "compression_loss": 83.740478515625, "epoch": 1.26, "learning_rate": 1.1145645102999639e-05, "loss": 84.0117, "step": 3483 }, { "compression_loss": 85.38082885742188, "epoch": 1.26, "learning_rate": 1.114022406938923e-05, "loss": 85.8185, "step": 3484 }, { "compression_loss": 84.77217864990234, "epoch": 1.26, "learning_rate": 1.1134803035778821e-05, "loss": 85.1548, "step": 3485 }, { "compression_loss": 83.90879821777344, "epoch": 1.26, "learning_rate": 1.1129382002168414e-05, "loss": 84.1855, "step": 3486 }, { "compression_loss": 82.84197998046875, "epoch": 1.26, "learning_rate": 1.1123960968558005e-05, "loss": 83.383, "step": 3487 }, { "compression_loss": 84.57230377197266, "epoch": 1.26, "learning_rate": 1.1118539934947597e-05, "loss": 85.0538, "step": 3488 }, { "compression_loss": 84.73078918457031, "epoch": 1.26, "learning_rate": 1.1113118901337188e-05, "loss": 85.1511, "step": 3489 }, { "compression_loss": 83.18800354003906, "epoch": 1.26, "learning_rate": 1.1107697867726781e-05, "loss": 83.5953, "step": 3490 }, { "compression_loss": 86.5123519897461, "epoch": 1.26, "learning_rate": 1.1102276834116372e-05, "loss": 87.1289, "step": 3491 }, { "compression_loss": 83.99392700195312, "epoch": 1.26, "learning_rate": 1.1096855800505963e-05, "loss": 84.4083, "step": 3492 }, { "compression_loss": 84.51089477539062, "epoch": 1.26, "learning_rate": 1.1091434766895555e-05, "loss": 84.8913, "step": 3493 }, { "compression_loss": 84.73395538330078, "epoch": 1.26, "learning_rate": 1.1086013733285148e-05, "loss": 85.1175, "step": 3494 }, { "compression_loss": 82.734130859375, "epoch": 1.26, "learning_rate": 1.1080592699674739e-05, "loss": 83.0027, "step": 3495 }, { "compression_loss": 86.18927001953125, "epoch": 1.26, "learning_rate": 1.107517166606433e-05, "loss": 86.5548, "step": 3496 }, { "compression_loss": 83.97632598876953, "epoch": 1.26, "learning_rate": 1.1069750632453922e-05, "loss": 84.3469, "step": 3497 }, { "compression_loss": 84.15090942382812, "epoch": 1.26, "learning_rate": 1.1064329598843514e-05, "loss": 84.471, "step": 3498 }, { "compression_loss": 83.22250366210938, "epoch": 1.26, "learning_rate": 1.1058908565233106e-05, "loss": 83.8352, "step": 3499 }, { "compression_loss": 85.25823974609375, "epoch": 1.26, "learning_rate": 1.1053487531622697e-05, "loss": 85.7763, "step": 3500 }, { "epoch": 1.26, "eval_exact_match": 86.70766319772942, "eval_f1": 92.99935642517451, "step": 3500 }, { "compression_loss": 84.11970520019531, "epoch": 1.27, "learning_rate": 1.1048066498012288e-05, "loss": 84.7812, "step": 3501 }, { "compression_loss": 83.62742614746094, "epoch": 1.27, "learning_rate": 1.104264546440188e-05, "loss": 83.9663, "step": 3502 }, { "compression_loss": 83.52582550048828, "epoch": 1.27, "learning_rate": 1.103722443079147e-05, "loss": 83.9228, "step": 3503 }, { "compression_loss": 84.38321685791016, "epoch": 1.27, "learning_rate": 1.1031803397181062e-05, "loss": 84.8148, "step": 3504 }, { "compression_loss": 82.85252380371094, "epoch": 1.27, "learning_rate": 1.1026382363570653e-05, "loss": 83.3087, "step": 3505 }, { "compression_loss": 84.00306701660156, "epoch": 1.27, "learning_rate": 1.1020961329960246e-05, "loss": 84.4434, "step": 3506 }, { "compression_loss": 84.58514404296875, "epoch": 1.27, "learning_rate": 1.1015540296349838e-05, "loss": 85.1103, "step": 3507 }, { "compression_loss": 83.64079284667969, "epoch": 1.27, "learning_rate": 1.1010119262739429e-05, "loss": 84.1099, "step": 3508 }, { "compression_loss": 83.72958374023438, "epoch": 1.27, "learning_rate": 1.100469822912902e-05, "loss": 84.2467, "step": 3509 }, { "compression_loss": 84.25227355957031, "epoch": 1.27, "learning_rate": 1.0999277195518611e-05, "loss": 84.6997, "step": 3510 }, { "compression_loss": 83.50452423095703, "epoch": 1.27, "learning_rate": 1.0993856161908204e-05, "loss": 83.8715, "step": 3511 }, { "compression_loss": 85.01307678222656, "epoch": 1.27, "learning_rate": 1.0988435128297796e-05, "loss": 85.6029, "step": 3512 }, { "compression_loss": 84.5574722290039, "epoch": 1.27, "learning_rate": 1.0983014094687387e-05, "loss": 84.8887, "step": 3513 }, { "compression_loss": 82.94038391113281, "epoch": 1.27, "learning_rate": 1.0977593061076978e-05, "loss": 83.2859, "step": 3514 }, { "compression_loss": 84.49879455566406, "epoch": 1.27, "learning_rate": 1.0972172027466571e-05, "loss": 84.9596, "step": 3515 }, { "compression_loss": 85.91400909423828, "epoch": 1.27, "learning_rate": 1.0966750993856162e-05, "loss": 86.6976, "step": 3516 }, { "compression_loss": 84.1169204711914, "epoch": 1.27, "learning_rate": 1.0961329960245754e-05, "loss": 84.6651, "step": 3517 }, { "compression_loss": 83.45372009277344, "epoch": 1.27, "learning_rate": 1.0955908926635345e-05, "loss": 83.795, "step": 3518 }, { "compression_loss": 84.14937591552734, "epoch": 1.27, "learning_rate": 1.0950487893024938e-05, "loss": 84.789, "step": 3519 }, { "compression_loss": 84.44397735595703, "epoch": 1.27, "learning_rate": 1.094506685941453e-05, "loss": 84.8435, "step": 3520 }, { "compression_loss": 85.89166259765625, "epoch": 1.27, "learning_rate": 1.093964582580412e-05, "loss": 86.6148, "step": 3521 }, { "compression_loss": 84.75487518310547, "epoch": 1.27, "learning_rate": 1.0934224792193712e-05, "loss": 85.2793, "step": 3522 }, { "compression_loss": 84.71232604980469, "epoch": 1.27, "learning_rate": 1.0928803758583305e-05, "loss": 85.4287, "step": 3523 }, { "compression_loss": 83.68263244628906, "epoch": 1.27, "learning_rate": 1.0923382724972894e-05, "loss": 84.1724, "step": 3524 }, { "compression_loss": 82.39749145507812, "epoch": 1.27, "learning_rate": 1.0917961691362486e-05, "loss": 82.7964, "step": 3525 }, { "compression_loss": 85.1941909790039, "epoch": 1.27, "learning_rate": 1.0912540657752077e-05, "loss": 85.8013, "step": 3526 }, { "compression_loss": 83.59522247314453, "epoch": 1.27, "learning_rate": 1.090711962414167e-05, "loss": 83.86, "step": 3527 }, { "compression_loss": 82.9810791015625, "epoch": 1.28, "learning_rate": 1.0901698590531261e-05, "loss": 83.3548, "step": 3528 }, { "compression_loss": 82.90887451171875, "epoch": 1.28, "learning_rate": 1.0896277556920852e-05, "loss": 83.2052, "step": 3529 }, { "compression_loss": 86.60338592529297, "epoch": 1.28, "learning_rate": 1.0890856523310444e-05, "loss": 87.0138, "step": 3530 }, { "compression_loss": 85.23201751708984, "epoch": 1.28, "learning_rate": 1.0885435489700037e-05, "loss": 85.7677, "step": 3531 }, { "compression_loss": 87.07669830322266, "epoch": 1.28, "learning_rate": 1.0880014456089628e-05, "loss": 88.0411, "step": 3532 }, { "compression_loss": 83.60739135742188, "epoch": 1.28, "learning_rate": 1.087459342247922e-05, "loss": 84.0045, "step": 3533 }, { "compression_loss": 84.14373779296875, "epoch": 1.28, "learning_rate": 1.086917238886881e-05, "loss": 84.6235, "step": 3534 }, { "compression_loss": 84.0421142578125, "epoch": 1.28, "learning_rate": 1.0863751355258403e-05, "loss": 84.3926, "step": 3535 }, { "compression_loss": 85.424560546875, "epoch": 1.28, "learning_rate": 1.0858330321647995e-05, "loss": 86.1056, "step": 3536 }, { "compression_loss": 82.74078369140625, "epoch": 1.28, "learning_rate": 1.0852909288037586e-05, "loss": 83.3499, "step": 3537 }, { "compression_loss": 85.93655395507812, "epoch": 1.28, "learning_rate": 1.0847488254427177e-05, "loss": 86.2639, "step": 3538 }, { "compression_loss": 85.00288391113281, "epoch": 1.28, "learning_rate": 1.084206722081677e-05, "loss": 85.5116, "step": 3539 }, { "compression_loss": 83.84801483154297, "epoch": 1.28, "learning_rate": 1.0836646187206361e-05, "loss": 84.2967, "step": 3540 }, { "compression_loss": 84.66778564453125, "epoch": 1.28, "learning_rate": 1.0831225153595953e-05, "loss": 85.0309, "step": 3541 }, { "compression_loss": 84.01271057128906, "epoch": 1.28, "learning_rate": 1.0825804119985544e-05, "loss": 84.3292, "step": 3542 }, { "compression_loss": 84.77529907226562, "epoch": 1.28, "learning_rate": 1.0820383086375137e-05, "loss": 85.3349, "step": 3543 }, { "compression_loss": 85.06401062011719, "epoch": 1.28, "learning_rate": 1.0814962052764728e-05, "loss": 85.4723, "step": 3544 }, { "compression_loss": 83.5679931640625, "epoch": 1.28, "learning_rate": 1.080954101915432e-05, "loss": 83.9329, "step": 3545 }, { "compression_loss": 83.2987060546875, "epoch": 1.28, "learning_rate": 1.0804119985543909e-05, "loss": 83.5763, "step": 3546 }, { "compression_loss": 83.07041931152344, "epoch": 1.28, "learning_rate": 1.0798698951933502e-05, "loss": 83.5883, "step": 3547 }, { "compression_loss": 83.49356079101562, "epoch": 1.28, "learning_rate": 1.0793277918323093e-05, "loss": 83.9713, "step": 3548 }, { "compression_loss": 85.45260620117188, "epoch": 1.28, "learning_rate": 1.0787856884712685e-05, "loss": 86.0292, "step": 3549 }, { "compression_loss": 84.77706909179688, "epoch": 1.28, "learning_rate": 1.0782435851102276e-05, "loss": 85.3139, "step": 3550 }, { "compression_loss": 84.50813293457031, "epoch": 1.28, "learning_rate": 1.0777014817491869e-05, "loss": 85.1006, "step": 3551 }, { "compression_loss": 85.21243286132812, "epoch": 1.28, "learning_rate": 1.077159378388146e-05, "loss": 85.6103, "step": 3552 }, { "compression_loss": 83.00790405273438, "epoch": 1.28, "learning_rate": 1.0766172750271051e-05, "loss": 83.2651, "step": 3553 }, { "compression_loss": 84.67882537841797, "epoch": 1.28, "learning_rate": 1.0760751716660643e-05, "loss": 85.1045, "step": 3554 }, { "compression_loss": 84.0499496459961, "epoch": 1.28, "learning_rate": 1.0755330683050236e-05, "loss": 84.6083, "step": 3555 }, { "compression_loss": 83.56746673583984, "epoch": 1.29, "learning_rate": 1.0749909649439827e-05, "loss": 83.8726, "step": 3556 }, { "compression_loss": 84.19435119628906, "epoch": 1.29, "learning_rate": 1.0744488615829418e-05, "loss": 84.6242, "step": 3557 }, { "compression_loss": 82.56289672851562, "epoch": 1.29, "learning_rate": 1.073906758221901e-05, "loss": 83.0226, "step": 3558 }, { "compression_loss": 85.01057434082031, "epoch": 1.29, "learning_rate": 1.0733646548608602e-05, "loss": 85.4968, "step": 3559 }, { "compression_loss": 85.50176239013672, "epoch": 1.29, "learning_rate": 1.0728225514998194e-05, "loss": 86.1296, "step": 3560 }, { "compression_loss": 85.40216827392578, "epoch": 1.29, "learning_rate": 1.0722804481387785e-05, "loss": 85.8846, "step": 3561 }, { "compression_loss": 87.05314636230469, "epoch": 1.29, "learning_rate": 1.0717383447777376e-05, "loss": 87.4237, "step": 3562 }, { "compression_loss": 84.52540588378906, "epoch": 1.29, "learning_rate": 1.071196241416697e-05, "loss": 85.4482, "step": 3563 }, { "compression_loss": 85.44386291503906, "epoch": 1.29, "learning_rate": 1.070654138055656e-05, "loss": 85.7261, "step": 3564 }, { "compression_loss": 84.30376434326172, "epoch": 1.29, "learning_rate": 1.0701120346946152e-05, "loss": 84.8643, "step": 3565 }, { "compression_loss": 84.59992218017578, "epoch": 1.29, "learning_rate": 1.0695699313335743e-05, "loss": 85.3403, "step": 3566 }, { "compression_loss": 83.55725860595703, "epoch": 1.29, "learning_rate": 1.0690278279725336e-05, "loss": 84.0282, "step": 3567 }, { "compression_loss": 84.51336669921875, "epoch": 1.29, "learning_rate": 1.0684857246114926e-05, "loss": 84.8114, "step": 3568 }, { "compression_loss": 84.37678527832031, "epoch": 1.29, "learning_rate": 1.0679436212504517e-05, "loss": 84.7636, "step": 3569 }, { "compression_loss": 84.34069061279297, "epoch": 1.29, "learning_rate": 1.0674015178894108e-05, "loss": 84.8488, "step": 3570 }, { "compression_loss": 85.26799011230469, "epoch": 1.29, "learning_rate": 1.0668594145283701e-05, "loss": 85.8977, "step": 3571 }, { "compression_loss": 84.58977508544922, "epoch": 1.29, "learning_rate": 1.0663173111673292e-05, "loss": 85.0092, "step": 3572 }, { "compression_loss": 85.42585754394531, "epoch": 1.29, "learning_rate": 1.0657752078062884e-05, "loss": 86.0925, "step": 3573 }, { "compression_loss": 85.95929718017578, "epoch": 1.29, "learning_rate": 1.0652331044452475e-05, "loss": 86.4293, "step": 3574 }, { "compression_loss": 83.7205581665039, "epoch": 1.29, "learning_rate": 1.0646910010842068e-05, "loss": 84.1366, "step": 3575 }, { "compression_loss": 83.12373352050781, "epoch": 1.29, "learning_rate": 1.0641488977231659e-05, "loss": 83.3747, "step": 3576 }, { "compression_loss": 84.725830078125, "epoch": 1.29, "learning_rate": 1.063606794362125e-05, "loss": 85.1268, "step": 3577 }, { "compression_loss": 83.47454833984375, "epoch": 1.29, "learning_rate": 1.0630646910010842e-05, "loss": 83.8563, "step": 3578 }, { "compression_loss": 83.92108917236328, "epoch": 1.29, "learning_rate": 1.0625225876400435e-05, "loss": 84.3446, "step": 3579 }, { "compression_loss": 82.24491882324219, "epoch": 1.29, "learning_rate": 1.0619804842790026e-05, "loss": 82.7087, "step": 3580 }, { "compression_loss": 85.37690734863281, "epoch": 1.29, "learning_rate": 1.0614383809179617e-05, "loss": 85.922, "step": 3581 }, { "compression_loss": 84.38057708740234, "epoch": 1.29, "learning_rate": 1.0608962775569208e-05, "loss": 84.9828, "step": 3582 }, { "compression_loss": 84.88043975830078, "epoch": 1.29, "learning_rate": 1.0603541741958801e-05, "loss": 85.4204, "step": 3583 }, { "compression_loss": 84.6439208984375, "epoch": 1.3, "learning_rate": 1.0598120708348393e-05, "loss": 85.2257, "step": 3584 }, { "compression_loss": 83.26849365234375, "epoch": 1.3, "learning_rate": 1.0592699674737984e-05, "loss": 83.5249, "step": 3585 }, { "compression_loss": 83.14602661132812, "epoch": 1.3, "learning_rate": 1.0587278641127575e-05, "loss": 83.6712, "step": 3586 }, { "compression_loss": 85.63227844238281, "epoch": 1.3, "learning_rate": 1.0581857607517168e-05, "loss": 86.0517, "step": 3587 }, { "compression_loss": 85.42958068847656, "epoch": 1.3, "learning_rate": 1.057643657390676e-05, "loss": 85.9051, "step": 3588 }, { "compression_loss": 83.31855773925781, "epoch": 1.3, "learning_rate": 1.057101554029635e-05, "loss": 83.6812, "step": 3589 }, { "compression_loss": 84.09323120117188, "epoch": 1.3, "learning_rate": 1.056559450668594e-05, "loss": 84.5694, "step": 3590 }, { "compression_loss": 85.02471923828125, "epoch": 1.3, "learning_rate": 1.0560173473075533e-05, "loss": 85.3652, "step": 3591 }, { "compression_loss": 83.83584594726562, "epoch": 1.3, "learning_rate": 1.0554752439465125e-05, "loss": 84.2968, "step": 3592 }, { "compression_loss": 84.07785034179688, "epoch": 1.3, "learning_rate": 1.0549331405854716e-05, "loss": 84.4977, "step": 3593 }, { "compression_loss": 83.75323486328125, "epoch": 1.3, "learning_rate": 1.0543910372244307e-05, "loss": 84.2231, "step": 3594 }, { "compression_loss": 82.48321533203125, "epoch": 1.3, "learning_rate": 1.05384893386339e-05, "loss": 82.8482, "step": 3595 }, { "compression_loss": 81.97915649414062, "epoch": 1.3, "learning_rate": 1.0533068305023491e-05, "loss": 82.3521, "step": 3596 }, { "compression_loss": 83.57582092285156, "epoch": 1.3, "learning_rate": 1.0527647271413083e-05, "loss": 83.9337, "step": 3597 }, { "compression_loss": 86.2541275024414, "epoch": 1.3, "learning_rate": 1.0522226237802674e-05, "loss": 86.9836, "step": 3598 }, { "compression_loss": 85.00252532958984, "epoch": 1.3, "learning_rate": 1.0516805204192267e-05, "loss": 85.3542, "step": 3599 }, { "compression_loss": 84.80767822265625, "epoch": 1.3, "learning_rate": 1.0511384170581858e-05, "loss": 85.2697, "step": 3600 }, { "compression_loss": 83.06170654296875, "epoch": 1.3, "learning_rate": 1.050596313697145e-05, "loss": 83.5071, "step": 3601 }, { "compression_loss": 83.02691650390625, "epoch": 1.3, "learning_rate": 1.050054210336104e-05, "loss": 83.5072, "step": 3602 }, { "compression_loss": 83.6644515991211, "epoch": 1.3, "learning_rate": 1.0495121069750634e-05, "loss": 84.1434, "step": 3603 }, { "compression_loss": 85.3459701538086, "epoch": 1.3, "learning_rate": 1.0489700036140225e-05, "loss": 85.7357, "step": 3604 }, { "compression_loss": 84.24710083007812, "epoch": 1.3, "learning_rate": 1.0484279002529816e-05, "loss": 84.5267, "step": 3605 }, { "compression_loss": 84.2022705078125, "epoch": 1.3, "learning_rate": 1.0478857968919407e-05, "loss": 84.8484, "step": 3606 }, { "compression_loss": 85.019775390625, "epoch": 1.3, "learning_rate": 1.0473436935309e-05, "loss": 85.8667, "step": 3607 }, { "compression_loss": 84.92035675048828, "epoch": 1.3, "learning_rate": 1.0468015901698592e-05, "loss": 85.6072, "step": 3608 }, { "compression_loss": 85.05960845947266, "epoch": 1.3, "learning_rate": 1.0462594868088183e-05, "loss": 85.3706, "step": 3609 }, { "compression_loss": 84.4212646484375, "epoch": 1.3, "learning_rate": 1.0457173834477774e-05, "loss": 84.9215, "step": 3610 }, { "compression_loss": 84.37789154052734, "epoch": 1.31, "learning_rate": 1.0451752800867365e-05, "loss": 84.9607, "step": 3611 }, { "compression_loss": 84.98193359375, "epoch": 1.31, "learning_rate": 1.0446331767256957e-05, "loss": 85.5216, "step": 3612 }, { "compression_loss": 85.48027038574219, "epoch": 1.31, "learning_rate": 1.0440910733646548e-05, "loss": 85.9271, "step": 3613 }, { "compression_loss": 84.71879577636719, "epoch": 1.31, "learning_rate": 1.043548970003614e-05, "loss": 85.1138, "step": 3614 }, { "compression_loss": 85.6099853515625, "epoch": 1.31, "learning_rate": 1.0430068666425732e-05, "loss": 85.9621, "step": 3615 }, { "compression_loss": 87.22080993652344, "epoch": 1.31, "learning_rate": 1.0424647632815324e-05, "loss": 87.7545, "step": 3616 }, { "compression_loss": 84.80223083496094, "epoch": 1.31, "learning_rate": 1.0419226599204915e-05, "loss": 85.0946, "step": 3617 }, { "compression_loss": 85.28007507324219, "epoch": 1.31, "learning_rate": 1.0413805565594506e-05, "loss": 85.8375, "step": 3618 }, { "compression_loss": 83.89905548095703, "epoch": 1.31, "learning_rate": 1.0408384531984099e-05, "loss": 84.4051, "step": 3619 }, { "compression_loss": 83.68495178222656, "epoch": 1.31, "learning_rate": 1.040296349837369e-05, "loss": 84.0322, "step": 3620 }, { "compression_loss": 85.07087707519531, "epoch": 1.31, "learning_rate": 1.0397542464763282e-05, "loss": 85.7118, "step": 3621 }, { "compression_loss": 85.66688537597656, "epoch": 1.31, "learning_rate": 1.0392121431152873e-05, "loss": 86.1705, "step": 3622 }, { "compression_loss": 82.62074279785156, "epoch": 1.31, "learning_rate": 1.0386700397542466e-05, "loss": 83.1544, "step": 3623 }, { "compression_loss": 85.50341796875, "epoch": 1.31, "learning_rate": 1.0381279363932057e-05, "loss": 85.9897, "step": 3624 }, { "compression_loss": 83.72431182861328, "epoch": 1.31, "learning_rate": 1.0375858330321648e-05, "loss": 84.049, "step": 3625 }, { "compression_loss": 83.8370361328125, "epoch": 1.31, "learning_rate": 1.037043729671124e-05, "loss": 84.1239, "step": 3626 }, { "compression_loss": 84.33780670166016, "epoch": 1.31, "learning_rate": 1.0365016263100833e-05, "loss": 84.7554, "step": 3627 }, { "compression_loss": 85.49871826171875, "epoch": 1.31, "learning_rate": 1.0359595229490424e-05, "loss": 86.0036, "step": 3628 }, { "compression_loss": 82.59005737304688, "epoch": 1.31, "learning_rate": 1.0354174195880015e-05, "loss": 82.9769, "step": 3629 }, { "compression_loss": 85.76317596435547, "epoch": 1.31, "learning_rate": 1.0348753162269606e-05, "loss": 86.1977, "step": 3630 }, { "compression_loss": 84.97223663330078, "epoch": 1.31, "learning_rate": 1.03433321286592e-05, "loss": 85.2517, "step": 3631 }, { "compression_loss": 85.17121887207031, "epoch": 1.31, "learning_rate": 1.033791109504879e-05, "loss": 85.4536, "step": 3632 }, { "compression_loss": 84.00199890136719, "epoch": 1.31, "learning_rate": 1.033249006143838e-05, "loss": 84.6276, "step": 3633 }, { "compression_loss": 84.20877075195312, "epoch": 1.31, "learning_rate": 1.0327069027827971e-05, "loss": 84.6279, "step": 3634 }, { "compression_loss": 84.21354675292969, "epoch": 1.31, "learning_rate": 1.0321647994217564e-05, "loss": 84.9994, "step": 3635 }, { "compression_loss": 83.28919219970703, "epoch": 1.31, "learning_rate": 1.0316226960607156e-05, "loss": 83.811, "step": 3636 }, { "compression_loss": 83.6545181274414, "epoch": 1.31, "learning_rate": 1.0310805926996747e-05, "loss": 83.9274, "step": 3637 }, { "compression_loss": 84.63636779785156, "epoch": 1.31, "learning_rate": 1.0305384893386338e-05, "loss": 85.2021, "step": 3638 }, { "compression_loss": 86.15887451171875, "epoch": 1.32, "learning_rate": 1.0299963859775931e-05, "loss": 86.5601, "step": 3639 }, { "compression_loss": 81.29051971435547, "epoch": 1.32, "learning_rate": 1.0294542826165523e-05, "loss": 81.5894, "step": 3640 }, { "compression_loss": 83.78595733642578, "epoch": 1.32, "learning_rate": 1.0289121792555114e-05, "loss": 84.2421, "step": 3641 }, { "compression_loss": 84.8052978515625, "epoch": 1.32, "learning_rate": 1.0283700758944705e-05, "loss": 85.2486, "step": 3642 }, { "compression_loss": 85.188232421875, "epoch": 1.32, "learning_rate": 1.0278279725334298e-05, "loss": 85.644, "step": 3643 }, { "compression_loss": 82.71797943115234, "epoch": 1.32, "learning_rate": 1.027285869172389e-05, "loss": 83.0426, "step": 3644 }, { "compression_loss": 84.71566009521484, "epoch": 1.32, "learning_rate": 1.026743765811348e-05, "loss": 85.5396, "step": 3645 }, { "compression_loss": 85.56320190429688, "epoch": 1.32, "learning_rate": 1.0262016624503072e-05, "loss": 85.9082, "step": 3646 }, { "compression_loss": 82.90523529052734, "epoch": 1.32, "learning_rate": 1.0256595590892665e-05, "loss": 83.2553, "step": 3647 }, { "compression_loss": 83.76258850097656, "epoch": 1.32, "learning_rate": 1.0251174557282256e-05, "loss": 84.3515, "step": 3648 }, { "compression_loss": 84.33489990234375, "epoch": 1.32, "learning_rate": 1.0245753523671847e-05, "loss": 85.1445, "step": 3649 }, { "compression_loss": 83.79826354980469, "epoch": 1.32, "learning_rate": 1.0240332490061439e-05, "loss": 84.5979, "step": 3650 }, { "compression_loss": 86.15802001953125, "epoch": 1.32, "learning_rate": 1.0234911456451032e-05, "loss": 86.6124, "step": 3651 }, { "compression_loss": 85.0745849609375, "epoch": 1.32, "learning_rate": 1.0229490422840623e-05, "loss": 85.5016, "step": 3652 }, { "compression_loss": 82.84780883789062, "epoch": 1.32, "learning_rate": 1.0224069389230214e-05, "loss": 83.1746, "step": 3653 }, { "compression_loss": 85.00255584716797, "epoch": 1.32, "learning_rate": 1.0218648355619805e-05, "loss": 85.479, "step": 3654 }, { "compression_loss": 83.51516723632812, "epoch": 1.32, "learning_rate": 1.0213227322009397e-05, "loss": 83.9798, "step": 3655 }, { "compression_loss": 82.56458282470703, "epoch": 1.32, "learning_rate": 1.0207806288398988e-05, "loss": 83.1644, "step": 3656 }, { "compression_loss": 83.73922729492188, "epoch": 1.32, "learning_rate": 1.020238525478858e-05, "loss": 84.1815, "step": 3657 }, { "compression_loss": 85.27091979980469, "epoch": 1.32, "learning_rate": 1.019696422117817e-05, "loss": 85.7368, "step": 3658 }, { "compression_loss": 84.65528106689453, "epoch": 1.32, "learning_rate": 1.0191543187567763e-05, "loss": 85.1164, "step": 3659 }, { "compression_loss": 85.7332534790039, "epoch": 1.32, "learning_rate": 1.0186122153957355e-05, "loss": 86.0987, "step": 3660 }, { "compression_loss": 85.80215454101562, "epoch": 1.32, "learning_rate": 1.0180701120346946e-05, "loss": 86.4183, "step": 3661 }, { "compression_loss": 83.66017150878906, "epoch": 1.32, "learning_rate": 1.0175280086736537e-05, "loss": 84.0743, "step": 3662 }, { "compression_loss": 84.0906753540039, "epoch": 1.32, "learning_rate": 1.016985905312613e-05, "loss": 84.4542, "step": 3663 }, { "compression_loss": 82.74700927734375, "epoch": 1.32, "learning_rate": 1.0164438019515722e-05, "loss": 83.2991, "step": 3664 }, { "compression_loss": 84.88774871826172, "epoch": 1.32, "learning_rate": 1.0159016985905313e-05, "loss": 85.5122, "step": 3665 }, { "compression_loss": 84.6222915649414, "epoch": 1.32, "learning_rate": 1.0153595952294904e-05, "loss": 85.1135, "step": 3666 }, { "compression_loss": 85.0327377319336, "epoch": 1.33, "learning_rate": 1.0148174918684497e-05, "loss": 85.7795, "step": 3667 }, { "compression_loss": 84.24813842773438, "epoch": 1.33, "learning_rate": 1.0142753885074088e-05, "loss": 84.6901, "step": 3668 }, { "compression_loss": 85.12907409667969, "epoch": 1.33, "learning_rate": 1.013733285146368e-05, "loss": 85.6242, "step": 3669 }, { "compression_loss": 83.24705505371094, "epoch": 1.33, "learning_rate": 1.013191181785327e-05, "loss": 83.7903, "step": 3670 }, { "compression_loss": 84.75935363769531, "epoch": 1.33, "learning_rate": 1.0126490784242864e-05, "loss": 85.1132, "step": 3671 }, { "compression_loss": 85.28514862060547, "epoch": 1.33, "learning_rate": 1.0121069750632455e-05, "loss": 85.8177, "step": 3672 }, { "compression_loss": 84.63656616210938, "epoch": 1.33, "learning_rate": 1.0115648717022046e-05, "loss": 85.0219, "step": 3673 }, { "compression_loss": 84.71219635009766, "epoch": 1.33, "learning_rate": 1.0110227683411638e-05, "loss": 85.2448, "step": 3674 }, { "compression_loss": 85.36238098144531, "epoch": 1.33, "learning_rate": 1.010480664980123e-05, "loss": 85.756, "step": 3675 }, { "compression_loss": 82.39231872558594, "epoch": 1.33, "learning_rate": 1.0099385616190822e-05, "loss": 82.7088, "step": 3676 }, { "compression_loss": 84.7494888305664, "epoch": 1.33, "learning_rate": 1.0093964582580411e-05, "loss": 85.4257, "step": 3677 }, { "compression_loss": 83.1866455078125, "epoch": 1.33, "learning_rate": 1.0088543548970003e-05, "loss": 83.5576, "step": 3678 }, { "compression_loss": 83.62969970703125, "epoch": 1.33, "learning_rate": 1.0083122515359596e-05, "loss": 84.1503, "step": 3679 }, { "compression_loss": 83.94132232666016, "epoch": 1.33, "learning_rate": 1.0077701481749187e-05, "loss": 84.4995, "step": 3680 }, { "compression_loss": 83.87639617919922, "epoch": 1.33, "learning_rate": 1.0072280448138778e-05, "loss": 84.6665, "step": 3681 }, { "compression_loss": 81.57571411132812, "epoch": 1.33, "learning_rate": 1.006685941452837e-05, "loss": 82.0359, "step": 3682 }, { "compression_loss": 85.96829223632812, "epoch": 1.33, "learning_rate": 1.0061438380917962e-05, "loss": 86.5333, "step": 3683 }, { "compression_loss": 82.44389343261719, "epoch": 1.33, "learning_rate": 1.0056017347307554e-05, "loss": 82.8085, "step": 3684 }, { "compression_loss": 85.0392837524414, "epoch": 1.33, "learning_rate": 1.0050596313697145e-05, "loss": 85.4392, "step": 3685 }, { "compression_loss": 83.91789245605469, "epoch": 1.33, "learning_rate": 1.0045175280086736e-05, "loss": 84.4227, "step": 3686 }, { "compression_loss": 85.42337036132812, "epoch": 1.33, "learning_rate": 1.003975424647633e-05, "loss": 85.8445, "step": 3687 }, { "compression_loss": 83.78303527832031, "epoch": 1.33, "learning_rate": 1.003433321286592e-05, "loss": 84.1373, "step": 3688 }, { "compression_loss": 84.866455078125, "epoch": 1.33, "learning_rate": 1.0028912179255512e-05, "loss": 85.2751, "step": 3689 }, { "compression_loss": 85.34493255615234, "epoch": 1.33, "learning_rate": 1.0023491145645103e-05, "loss": 85.931, "step": 3690 }, { "compression_loss": 84.50406646728516, "epoch": 1.33, "learning_rate": 1.0018070112034696e-05, "loss": 85.0942, "step": 3691 }, { "compression_loss": 83.51632690429688, "epoch": 1.33, "learning_rate": 1.0012649078424287e-05, "loss": 84.1502, "step": 3692 }, { "compression_loss": 83.62913513183594, "epoch": 1.33, "learning_rate": 1.0007228044813879e-05, "loss": 83.9573, "step": 3693 }, { "compression_loss": 85.36549377441406, "epoch": 1.34, "learning_rate": 1.000180701120347e-05, "loss": 85.5988, "step": 3694 }, { "compression_loss": 82.72151184082031, "epoch": 1.34, "learning_rate": 9.996385977593061e-06, "loss": 83.1491, "step": 3695 }, { "compression_loss": 84.48648071289062, "epoch": 1.34, "learning_rate": 9.990964943982654e-06, "loss": 84.853, "step": 3696 }, { "compression_loss": 85.66096496582031, "epoch": 1.34, "learning_rate": 9.985543910372245e-06, "loss": 85.9617, "step": 3697 }, { "compression_loss": 84.0414047241211, "epoch": 1.34, "learning_rate": 9.980122876761837e-06, "loss": 84.2678, "step": 3698 }, { "compression_loss": 84.24807739257812, "epoch": 1.34, "learning_rate": 9.974701843151426e-06, "loss": 84.8637, "step": 3699 }, { "compression_loss": 84.76577758789062, "epoch": 1.34, "learning_rate": 9.96928080954102e-06, "loss": 85.0911, "step": 3700 }, { "compression_loss": 84.35369873046875, "epoch": 1.34, "learning_rate": 9.96385977593061e-06, "loss": 84.9215, "step": 3701 }, { "compression_loss": 85.4842300415039, "epoch": 1.34, "learning_rate": 9.958438742320202e-06, "loss": 86.0174, "step": 3702 }, { "compression_loss": 82.16529846191406, "epoch": 1.34, "learning_rate": 9.953017708709793e-06, "loss": 82.4808, "step": 3703 }, { "compression_loss": 84.08577728271484, "epoch": 1.34, "learning_rate": 9.947596675099386e-06, "loss": 84.4349, "step": 3704 }, { "compression_loss": 85.3551025390625, "epoch": 1.34, "learning_rate": 9.942175641488977e-06, "loss": 85.8819, "step": 3705 }, { "compression_loss": 84.36538696289062, "epoch": 1.34, "learning_rate": 9.936754607878568e-06, "loss": 84.6801, "step": 3706 }, { "compression_loss": 84.34591674804688, "epoch": 1.34, "learning_rate": 9.93133357426816e-06, "loss": 84.7103, "step": 3707 }, { "compression_loss": 84.51746368408203, "epoch": 1.34, "learning_rate": 9.925912540657753e-06, "loss": 84.9292, "step": 3708 }, { "compression_loss": 83.54912567138672, "epoch": 1.34, "learning_rate": 9.920491507047344e-06, "loss": 83.7859, "step": 3709 }, { "compression_loss": 84.50050354003906, "epoch": 1.34, "learning_rate": 9.915070473436935e-06, "loss": 85.0477, "step": 3710 }, { "compression_loss": 82.836181640625, "epoch": 1.34, "learning_rate": 9.909649439826527e-06, "loss": 83.163, "step": 3711 }, { "compression_loss": 84.73744201660156, "epoch": 1.34, "learning_rate": 9.90422840621612e-06, "loss": 85.2851, "step": 3712 }, { "compression_loss": 83.2080078125, "epoch": 1.34, "learning_rate": 9.89880737260571e-06, "loss": 84.0167, "step": 3713 }, { "compression_loss": 83.66252136230469, "epoch": 1.34, "learning_rate": 9.893386338995302e-06, "loss": 83.9477, "step": 3714 }, { "compression_loss": 85.41973876953125, "epoch": 1.34, "learning_rate": 9.887965305384893e-06, "loss": 85.9855, "step": 3715 }, { "compression_loss": 84.54460144042969, "epoch": 1.34, "learning_rate": 9.882544271774486e-06, "loss": 84.8952, "step": 3716 }, { "compression_loss": 83.83798217773438, "epoch": 1.34, "learning_rate": 9.877123238164078e-06, "loss": 84.1019, "step": 3717 }, { "compression_loss": 83.99032592773438, "epoch": 1.34, "learning_rate": 9.871702204553669e-06, "loss": 84.6167, "step": 3718 }, { "compression_loss": 83.95463562011719, "epoch": 1.34, "learning_rate": 9.86628117094326e-06, "loss": 84.2194, "step": 3719 }, { "compression_loss": 84.3765869140625, "epoch": 1.34, "learning_rate": 9.860860137332851e-06, "loss": 84.9519, "step": 3720 }, { "compression_loss": 85.83338928222656, "epoch": 1.34, "learning_rate": 9.855439103722443e-06, "loss": 86.34, "step": 3721 }, { "compression_loss": 83.06006622314453, "epoch": 1.35, "learning_rate": 9.850018070112034e-06, "loss": 83.2753, "step": 3722 }, { "compression_loss": 82.68071746826172, "epoch": 1.35, "learning_rate": 9.844597036501625e-06, "loss": 82.9167, "step": 3723 }, { "compression_loss": 83.98880767822266, "epoch": 1.35, "learning_rate": 9.839176002891218e-06, "loss": 84.8254, "step": 3724 }, { "compression_loss": 83.49212646484375, "epoch": 1.35, "learning_rate": 9.83375496928081e-06, "loss": 83.8289, "step": 3725 }, { "compression_loss": 84.71220397949219, "epoch": 1.35, "learning_rate": 9.8283339356704e-06, "loss": 85.1914, "step": 3726 }, { "compression_loss": 82.72416687011719, "epoch": 1.35, "learning_rate": 9.822912902059992e-06, "loss": 83.2416, "step": 3727 }, { "compression_loss": 84.67028045654297, "epoch": 1.35, "learning_rate": 9.817491868449585e-06, "loss": 85.1672, "step": 3728 }, { "compression_loss": 84.87496948242188, "epoch": 1.35, "learning_rate": 9.812070834839176e-06, "loss": 85.3402, "step": 3729 }, { "compression_loss": 85.90469360351562, "epoch": 1.35, "learning_rate": 9.806649801228767e-06, "loss": 86.4043, "step": 3730 }, { "compression_loss": 84.80760192871094, "epoch": 1.35, "learning_rate": 9.801228767618359e-06, "loss": 85.0255, "step": 3731 }, { "compression_loss": 84.99714660644531, "epoch": 1.35, "learning_rate": 9.795807734007952e-06, "loss": 85.5209, "step": 3732 }, { "compression_loss": 83.40718841552734, "epoch": 1.35, "learning_rate": 9.790386700397543e-06, "loss": 83.8632, "step": 3733 }, { "compression_loss": 84.6900405883789, "epoch": 1.35, "learning_rate": 9.784965666787134e-06, "loss": 85.1443, "step": 3734 }, { "compression_loss": 84.04792785644531, "epoch": 1.35, "learning_rate": 9.779544633176726e-06, "loss": 84.5179, "step": 3735 }, { "compression_loss": 83.11720275878906, "epoch": 1.35, "learning_rate": 9.774123599566318e-06, "loss": 83.4596, "step": 3736 }, { "compression_loss": 84.69760131835938, "epoch": 1.35, "learning_rate": 9.76870256595591e-06, "loss": 85.0803, "step": 3737 }, { "compression_loss": 84.53578186035156, "epoch": 1.35, "learning_rate": 9.763281532345501e-06, "loss": 84.9578, "step": 3738 }, { "compression_loss": 84.43344116210938, "epoch": 1.35, "learning_rate": 9.757860498735092e-06, "loss": 85.0489, "step": 3739 }, { "compression_loss": 83.05608367919922, "epoch": 1.35, "learning_rate": 9.752439465124685e-06, "loss": 83.3814, "step": 3740 }, { "compression_loss": 83.57554626464844, "epoch": 1.35, "learning_rate": 9.747018431514277e-06, "loss": 84.0218, "step": 3741 }, { "compression_loss": 85.28316497802734, "epoch": 1.35, "learning_rate": 9.741597397903866e-06, "loss": 86.1491, "step": 3742 }, { "compression_loss": 83.17778015136719, "epoch": 1.35, "learning_rate": 9.736176364293457e-06, "loss": 83.7553, "step": 3743 }, { "compression_loss": 83.18510437011719, "epoch": 1.35, "learning_rate": 9.73075533068305e-06, "loss": 83.5869, "step": 3744 }, { "compression_loss": 85.6637954711914, "epoch": 1.35, "learning_rate": 9.725334297072642e-06, "loss": 85.9457, "step": 3745 }, { "compression_loss": 83.2144775390625, "epoch": 1.35, "learning_rate": 9.719913263462233e-06, "loss": 83.5303, "step": 3746 }, { "compression_loss": 84.31389617919922, "epoch": 1.35, "learning_rate": 9.714492229851824e-06, "loss": 84.7478, "step": 3747 }, { "compression_loss": 83.38842010498047, "epoch": 1.35, "learning_rate": 9.709071196241417e-06, "loss": 84.111, "step": 3748 }, { "compression_loss": 85.8895263671875, "epoch": 1.35, "learning_rate": 9.703650162631008e-06, "loss": 86.2138, "step": 3749 }, { "compression_loss": 82.51734924316406, "epoch": 1.36, "learning_rate": 9.6982291290206e-06, "loss": 82.8514, "step": 3750 }, { "epoch": 1.36, "eval_exact_match": 86.89687795648061, "eval_f1": 93.08039314041818, "step": 3750 }, { "compression_loss": 83.72856903076172, "epoch": 1.36, "learning_rate": 9.692808095410191e-06, "loss": 84.4078, "step": 3751 }, { "compression_loss": 84.90547180175781, "epoch": 1.36, "learning_rate": 9.687387061799784e-06, "loss": 85.4154, "step": 3752 }, { "compression_loss": 85.24283599853516, "epoch": 1.36, "learning_rate": 9.681966028189375e-06, "loss": 85.7565, "step": 3753 }, { "compression_loss": 82.9663314819336, "epoch": 1.36, "learning_rate": 9.676544994578966e-06, "loss": 83.4167, "step": 3754 }, { "compression_loss": 83.36123657226562, "epoch": 1.36, "learning_rate": 9.671123960968558e-06, "loss": 83.5849, "step": 3755 }, { "compression_loss": 84.45468139648438, "epoch": 1.36, "learning_rate": 9.66570292735815e-06, "loss": 84.8901, "step": 3756 }, { "compression_loss": 85.11773681640625, "epoch": 1.36, "learning_rate": 9.660281893747742e-06, "loss": 85.5145, "step": 3757 }, { "compression_loss": 85.27791595458984, "epoch": 1.36, "learning_rate": 9.654860860137333e-06, "loss": 85.557, "step": 3758 }, { "compression_loss": 85.8443374633789, "epoch": 1.36, "learning_rate": 9.649439826526925e-06, "loss": 86.349, "step": 3759 }, { "compression_loss": 85.17090606689453, "epoch": 1.36, "learning_rate": 9.644018792916517e-06, "loss": 85.5821, "step": 3760 }, { "compression_loss": 85.75142669677734, "epoch": 1.36, "learning_rate": 9.638597759306109e-06, "loss": 86.2506, "step": 3761 }, { "compression_loss": 84.02753448486328, "epoch": 1.36, "learning_rate": 9.6331767256957e-06, "loss": 84.5796, "step": 3762 }, { "compression_loss": 83.5096435546875, "epoch": 1.36, "learning_rate": 9.627755692085291e-06, "loss": 83.9967, "step": 3763 }, { "compression_loss": 84.49453735351562, "epoch": 1.36, "learning_rate": 9.622334658474883e-06, "loss": 85.1233, "step": 3764 }, { "compression_loss": 85.0802993774414, "epoch": 1.36, "learning_rate": 9.616913624864474e-06, "loss": 85.634, "step": 3765 }, { "compression_loss": 85.9207763671875, "epoch": 1.36, "learning_rate": 9.611492591254065e-06, "loss": 86.4335, "step": 3766 }, { "compression_loss": 85.13861083984375, "epoch": 1.36, "learning_rate": 9.606071557643656e-06, "loss": 85.5344, "step": 3767 }, { "compression_loss": 84.63101959228516, "epoch": 1.36, "learning_rate": 9.60065052403325e-06, "loss": 85.075, "step": 3768 }, { "compression_loss": 84.3357162475586, "epoch": 1.36, "learning_rate": 9.59522949042284e-06, "loss": 84.6333, "step": 3769 }, { "compression_loss": 84.60942840576172, "epoch": 1.36, "learning_rate": 9.589808456812432e-06, "loss": 85.1406, "step": 3770 }, { "compression_loss": 83.47845458984375, "epoch": 1.36, "learning_rate": 9.584387423202023e-06, "loss": 83.962, "step": 3771 }, { "compression_loss": 83.69942474365234, "epoch": 1.36, "learning_rate": 9.578966389591616e-06, "loss": 84.1424, "step": 3772 }, { "compression_loss": 85.1849365234375, "epoch": 1.36, "learning_rate": 9.573545355981207e-06, "loss": 85.8495, "step": 3773 }, { "compression_loss": 84.13873291015625, "epoch": 1.36, "learning_rate": 9.568124322370799e-06, "loss": 84.6678, "step": 3774 }, { "compression_loss": 83.74205017089844, "epoch": 1.36, "learning_rate": 9.56270328876039e-06, "loss": 84.5224, "step": 3775 }, { "compression_loss": 85.78439331054688, "epoch": 1.36, "learning_rate": 9.557282255149983e-06, "loss": 86.4032, "step": 3776 }, { "compression_loss": 85.63371276855469, "epoch": 1.37, "learning_rate": 9.551861221539574e-06, "loss": 86.0895, "step": 3777 }, { "compression_loss": 84.27432250976562, "epoch": 1.37, "learning_rate": 9.546440187929165e-06, "loss": 84.9763, "step": 3778 }, { "compression_loss": 84.5081787109375, "epoch": 1.37, "learning_rate": 9.541019154318757e-06, "loss": 84.8299, "step": 3779 }, { "compression_loss": 86.11967468261719, "epoch": 1.37, "learning_rate": 9.53559812070835e-06, "loss": 86.5479, "step": 3780 }, { "compression_loss": 85.14151000976562, "epoch": 1.37, "learning_rate": 9.530177087097941e-06, "loss": 85.8159, "step": 3781 }, { "compression_loss": 86.10209655761719, "epoch": 1.37, "learning_rate": 9.524756053487532e-06, "loss": 86.4132, "step": 3782 }, { "compression_loss": 83.5233154296875, "epoch": 1.37, "learning_rate": 9.519335019877124e-06, "loss": 83.9793, "step": 3783 }, { "compression_loss": 86.53842163085938, "epoch": 1.37, "learning_rate": 9.513913986266716e-06, "loss": 86.9354, "step": 3784 }, { "compression_loss": 85.31137084960938, "epoch": 1.37, "learning_rate": 9.508492952656308e-06, "loss": 85.7872, "step": 3785 }, { "compression_loss": 83.24424743652344, "epoch": 1.37, "learning_rate": 9.503071919045897e-06, "loss": 83.6572, "step": 3786 }, { "compression_loss": 84.03741455078125, "epoch": 1.37, "learning_rate": 9.497650885435489e-06, "loss": 84.4155, "step": 3787 }, { "compression_loss": 85.08415222167969, "epoch": 1.37, "learning_rate": 9.492229851825082e-06, "loss": 85.5386, "step": 3788 }, { "compression_loss": 85.07785034179688, "epoch": 1.37, "learning_rate": 9.486808818214673e-06, "loss": 85.6029, "step": 3789 }, { "compression_loss": 84.60717010498047, "epoch": 1.37, "learning_rate": 9.481387784604264e-06, "loss": 85.0693, "step": 3790 }, { "compression_loss": 83.63288879394531, "epoch": 1.37, "learning_rate": 9.475966750993855e-06, "loss": 84.0267, "step": 3791 }, { "compression_loss": 84.31886291503906, "epoch": 1.37, "learning_rate": 9.470545717383448e-06, "loss": 85.022, "step": 3792 }, { "compression_loss": 85.41567993164062, "epoch": 1.37, "learning_rate": 9.46512468377304e-06, "loss": 85.6194, "step": 3793 }, { "compression_loss": 82.75641632080078, "epoch": 1.37, "learning_rate": 9.459703650162631e-06, "loss": 83.1484, "step": 3794 }, { "compression_loss": 83.2170181274414, "epoch": 1.37, "learning_rate": 9.454282616552222e-06, "loss": 83.6421, "step": 3795 }, { "compression_loss": 84.80410766601562, "epoch": 1.37, "learning_rate": 9.448861582941815e-06, "loss": 85.274, "step": 3796 }, { "compression_loss": 84.66669464111328, "epoch": 1.37, "learning_rate": 9.443440549331406e-06, "loss": 85.0572, "step": 3797 }, { "compression_loss": 84.82056427001953, "epoch": 1.37, "learning_rate": 9.438019515720998e-06, "loss": 85.2592, "step": 3798 }, { "compression_loss": 84.75690460205078, "epoch": 1.37, "learning_rate": 9.432598482110589e-06, "loss": 85.1964, "step": 3799 }, { "compression_loss": 84.02991485595703, "epoch": 1.37, "learning_rate": 9.427177448500182e-06, "loss": 84.4544, "step": 3800 }, { "compression_loss": 84.57453918457031, "epoch": 1.37, "learning_rate": 9.421756414889773e-06, "loss": 85.0261, "step": 3801 }, { "compression_loss": 84.10264587402344, "epoch": 1.37, "learning_rate": 9.416335381279364e-06, "loss": 84.4027, "step": 3802 }, { "compression_loss": 82.42658996582031, "epoch": 1.37, "learning_rate": 9.410914347668956e-06, "loss": 82.6936, "step": 3803 }, { "compression_loss": 85.9658432006836, "epoch": 1.37, "learning_rate": 9.405493314058549e-06, "loss": 86.4543, "step": 3804 }, { "compression_loss": 84.55440521240234, "epoch": 1.38, "learning_rate": 9.40007228044814e-06, "loss": 85.0066, "step": 3805 }, { "compression_loss": 84.48025512695312, "epoch": 1.38, "learning_rate": 9.394651246837731e-06, "loss": 85.1769, "step": 3806 }, { "compression_loss": 84.81002807617188, "epoch": 1.38, "learning_rate": 9.389230213227323e-06, "loss": 85.1946, "step": 3807 }, { "compression_loss": 82.92750549316406, "epoch": 1.38, "learning_rate": 9.383809179616914e-06, "loss": 83.6395, "step": 3808 }, { "compression_loss": 84.23051452636719, "epoch": 1.38, "learning_rate": 9.378388146006505e-06, "loss": 84.792, "step": 3809 }, { "compression_loss": 84.40792846679688, "epoch": 1.38, "learning_rate": 9.372967112396096e-06, "loss": 84.8221, "step": 3810 }, { "compression_loss": 84.71355438232422, "epoch": 1.38, "learning_rate": 9.367546078785688e-06, "loss": 85.2541, "step": 3811 }, { "compression_loss": 87.34042358398438, "epoch": 1.38, "learning_rate": 9.36212504517528e-06, "loss": 88.2382, "step": 3812 }, { "compression_loss": 82.91107940673828, "epoch": 1.38, "learning_rate": 9.356704011564872e-06, "loss": 83.2329, "step": 3813 }, { "compression_loss": 84.5301284790039, "epoch": 1.38, "learning_rate": 9.351282977954463e-06, "loss": 85.0409, "step": 3814 }, { "compression_loss": 85.06419372558594, "epoch": 1.38, "learning_rate": 9.345861944344054e-06, "loss": 85.535, "step": 3815 }, { "compression_loss": 84.21695709228516, "epoch": 1.38, "learning_rate": 9.340440910733647e-06, "loss": 84.648, "step": 3816 }, { "compression_loss": 85.90885162353516, "epoch": 1.38, "learning_rate": 9.335019877123239e-06, "loss": 86.5896, "step": 3817 }, { "compression_loss": 85.07200622558594, "epoch": 1.38, "learning_rate": 9.32959884351283e-06, "loss": 85.5316, "step": 3818 }, { "compression_loss": 81.63005065917969, "epoch": 1.38, "learning_rate": 9.324177809902421e-06, "loss": 81.9975, "step": 3819 }, { "compression_loss": 84.39459991455078, "epoch": 1.38, "learning_rate": 9.318756776292014e-06, "loss": 85.0589, "step": 3820 }, { "compression_loss": 84.40219116210938, "epoch": 1.38, "learning_rate": 9.313335742681605e-06, "loss": 84.6813, "step": 3821 }, { "compression_loss": 85.27606201171875, "epoch": 1.38, "learning_rate": 9.307914709071197e-06, "loss": 86.0496, "step": 3822 }, { "compression_loss": 84.00828552246094, "epoch": 1.38, "learning_rate": 9.302493675460788e-06, "loss": 84.5098, "step": 3823 }, { "compression_loss": 83.8882064819336, "epoch": 1.38, "learning_rate": 9.297072641850381e-06, "loss": 84.3102, "step": 3824 }, { "compression_loss": 83.38645935058594, "epoch": 1.38, "learning_rate": 9.291651608239972e-06, "loss": 83.7882, "step": 3825 }, { "compression_loss": 84.15283966064453, "epoch": 1.38, "learning_rate": 9.286230574629563e-06, "loss": 84.5477, "step": 3826 }, { "compression_loss": 85.40009307861328, "epoch": 1.38, "learning_rate": 9.280809541019155e-06, "loss": 85.7159, "step": 3827 }, { "compression_loss": 84.05039978027344, "epoch": 1.38, "learning_rate": 9.275388507408748e-06, "loss": 84.4376, "step": 3828 }, { "compression_loss": 84.8417739868164, "epoch": 1.38, "learning_rate": 9.269967473798337e-06, "loss": 85.2159, "step": 3829 }, { "compression_loss": 82.91542053222656, "epoch": 1.38, "learning_rate": 9.264546440187929e-06, "loss": 83.2949, "step": 3830 }, { "compression_loss": 83.7037353515625, "epoch": 1.38, "learning_rate": 9.25912540657752e-06, "loss": 84.0905, "step": 3831 }, { "compression_loss": 84.6889419555664, "epoch": 1.38, "learning_rate": 9.253704372967113e-06, "loss": 85.3976, "step": 3832 }, { "compression_loss": 84.19816589355469, "epoch": 1.39, "learning_rate": 9.248283339356704e-06, "loss": 84.6372, "step": 3833 }, { "compression_loss": 83.02635192871094, "epoch": 1.39, "learning_rate": 9.242862305746295e-06, "loss": 83.3941, "step": 3834 }, { "compression_loss": 82.564697265625, "epoch": 1.39, "learning_rate": 9.237441272135887e-06, "loss": 83.1712, "step": 3835 }, { "compression_loss": 85.9029769897461, "epoch": 1.39, "learning_rate": 9.23202023852548e-06, "loss": 86.7011, "step": 3836 }, { "compression_loss": 81.98381042480469, "epoch": 1.39, "learning_rate": 9.22659920491507e-06, "loss": 82.4853, "step": 3837 }, { "compression_loss": 84.41291809082031, "epoch": 1.39, "learning_rate": 9.221178171304662e-06, "loss": 84.7687, "step": 3838 }, { "compression_loss": 84.48403930664062, "epoch": 1.39, "learning_rate": 9.215757137694253e-06, "loss": 84.9742, "step": 3839 }, { "compression_loss": 84.88001251220703, "epoch": 1.39, "learning_rate": 9.210336104083846e-06, "loss": 85.4532, "step": 3840 }, { "compression_loss": 85.62043762207031, "epoch": 1.39, "learning_rate": 9.204915070473438e-06, "loss": 85.9337, "step": 3841 }, { "compression_loss": 84.76739501953125, "epoch": 1.39, "learning_rate": 9.199494036863029e-06, "loss": 85.2793, "step": 3842 }, { "compression_loss": 84.13458251953125, "epoch": 1.39, "learning_rate": 9.19407300325262e-06, "loss": 84.6985, "step": 3843 }, { "compression_loss": 84.9151840209961, "epoch": 1.39, "learning_rate": 9.188651969642213e-06, "loss": 85.6925, "step": 3844 }, { "compression_loss": 82.84558868408203, "epoch": 1.39, "learning_rate": 9.183230936031804e-06, "loss": 83.1992, "step": 3845 }, { "compression_loss": 83.42591094970703, "epoch": 1.39, "learning_rate": 9.177809902421396e-06, "loss": 83.6618, "step": 3846 }, { "compression_loss": 84.28321075439453, "epoch": 1.39, "learning_rate": 9.172388868810987e-06, "loss": 84.6546, "step": 3847 }, { "compression_loss": 84.76947021484375, "epoch": 1.39, "learning_rate": 9.16696783520058e-06, "loss": 85.2073, "step": 3848 }, { "compression_loss": 83.4404525756836, "epoch": 1.39, "learning_rate": 9.161546801590171e-06, "loss": 83.9781, "step": 3849 }, { "compression_loss": 84.03407287597656, "epoch": 1.39, "learning_rate": 9.156125767979762e-06, "loss": 84.2705, "step": 3850 }, { "compression_loss": 85.16173553466797, "epoch": 1.39, "learning_rate": 9.150704734369352e-06, "loss": 85.4905, "step": 3851 }, { "compression_loss": 85.50360107421875, "epoch": 1.39, "learning_rate": 9.145283700758945e-06, "loss": 85.9773, "step": 3852 }, { "compression_loss": 85.61245727539062, "epoch": 1.39, "learning_rate": 9.139862667148536e-06, "loss": 86.311, "step": 3853 }, { "compression_loss": 86.15628051757812, "epoch": 1.39, "learning_rate": 9.134441633538128e-06, "loss": 86.763, "step": 3854 }, { "compression_loss": 85.08892822265625, "epoch": 1.39, "learning_rate": 9.129020599927719e-06, "loss": 85.5536, "step": 3855 }, { "compression_loss": 85.1852035522461, "epoch": 1.39, "learning_rate": 9.123599566317312e-06, "loss": 85.6192, "step": 3856 }, { "compression_loss": 83.22222900390625, "epoch": 1.39, "learning_rate": 9.118178532706903e-06, "loss": 83.8018, "step": 3857 }, { "compression_loss": 83.78692626953125, "epoch": 1.39, "learning_rate": 9.112757499096494e-06, "loss": 84.4223, "step": 3858 }, { "compression_loss": 83.85791015625, "epoch": 1.39, "learning_rate": 9.107336465486086e-06, "loss": 84.1513, "step": 3859 }, { "compression_loss": 82.55868530273438, "epoch": 1.4, "learning_rate": 9.101915431875679e-06, "loss": 83.0293, "step": 3860 }, { "compression_loss": 84.9184799194336, "epoch": 1.4, "learning_rate": 9.09649439826527e-06, "loss": 85.1898, "step": 3861 }, { "compression_loss": 84.83248901367188, "epoch": 1.4, "learning_rate": 9.091073364654861e-06, "loss": 85.3725, "step": 3862 }, { "compression_loss": 85.20587158203125, "epoch": 1.4, "learning_rate": 9.085652331044452e-06, "loss": 85.7204, "step": 3863 }, { "compression_loss": 82.90220642089844, "epoch": 1.4, "learning_rate": 9.080231297434045e-06, "loss": 83.3787, "step": 3864 }, { "compression_loss": 83.1277084350586, "epoch": 1.4, "learning_rate": 9.074810263823637e-06, "loss": 83.6685, "step": 3865 }, { "compression_loss": 85.16709899902344, "epoch": 1.4, "learning_rate": 9.069389230213228e-06, "loss": 85.8624, "step": 3866 }, { "compression_loss": 85.038330078125, "epoch": 1.4, "learning_rate": 9.06396819660282e-06, "loss": 85.4203, "step": 3867 }, { "compression_loss": 85.15396881103516, "epoch": 1.4, "learning_rate": 9.058547162992412e-06, "loss": 85.61, "step": 3868 }, { "compression_loss": 83.79204559326172, "epoch": 1.4, "learning_rate": 9.053126129382003e-06, "loss": 84.2012, "step": 3869 }, { "compression_loss": 87.03388977050781, "epoch": 1.4, "learning_rate": 9.047705095771595e-06, "loss": 87.7574, "step": 3870 }, { "compression_loss": 84.4718246459961, "epoch": 1.4, "learning_rate": 9.042284062161186e-06, "loss": 84.8115, "step": 3871 }, { "compression_loss": 84.13262176513672, "epoch": 1.4, "learning_rate": 9.036863028550779e-06, "loss": 84.5727, "step": 3872 }, { "compression_loss": 84.94479370117188, "epoch": 1.4, "learning_rate": 9.031441994940368e-06, "loss": 85.1712, "step": 3873 }, { "compression_loss": 83.2808609008789, "epoch": 1.4, "learning_rate": 9.02602096132996e-06, "loss": 83.6155, "step": 3874 }, { "compression_loss": 85.25289916992188, "epoch": 1.4, "learning_rate": 9.020599927719551e-06, "loss": 85.6763, "step": 3875 }, { "compression_loss": 85.03849792480469, "epoch": 1.4, "learning_rate": 9.015178894109144e-06, "loss": 85.4643, "step": 3876 }, { "compression_loss": 84.84613800048828, "epoch": 1.4, "learning_rate": 9.009757860498735e-06, "loss": 85.3015, "step": 3877 }, { "compression_loss": 84.6060791015625, "epoch": 1.4, "learning_rate": 9.004336826888327e-06, "loss": 85.1359, "step": 3878 }, { "compression_loss": 84.71659088134766, "epoch": 1.4, "learning_rate": 8.998915793277918e-06, "loss": 85.0956, "step": 3879 }, { "compression_loss": 84.97225189208984, "epoch": 1.4, "learning_rate": 8.993494759667509e-06, "loss": 85.6471, "step": 3880 }, { "compression_loss": 84.76024627685547, "epoch": 1.4, "learning_rate": 8.988073726057102e-06, "loss": 85.1509, "step": 3881 }, { "compression_loss": 85.27033996582031, "epoch": 1.4, "learning_rate": 8.982652692446693e-06, "loss": 86.1312, "step": 3882 }, { "compression_loss": 85.07810974121094, "epoch": 1.4, "learning_rate": 8.977231658836285e-06, "loss": 85.5266, "step": 3883 }, { "compression_loss": 84.5230712890625, "epoch": 1.4, "learning_rate": 8.971810625225876e-06, "loss": 85.1896, "step": 3884 }, { "compression_loss": 84.57251739501953, "epoch": 1.4, "learning_rate": 8.966389591615469e-06, "loss": 85.068, "step": 3885 }, { "compression_loss": 83.71711730957031, "epoch": 1.4, "learning_rate": 8.96096855800506e-06, "loss": 83.9906, "step": 3886 }, { "compression_loss": 84.65901184082031, "epoch": 1.4, "learning_rate": 8.955547524394651e-06, "loss": 85.2373, "step": 3887 }, { "compression_loss": 84.25010681152344, "epoch": 1.41, "learning_rate": 8.950126490784243e-06, "loss": 85.0552, "step": 3888 }, { "compression_loss": 83.99101257324219, "epoch": 1.41, "learning_rate": 8.944705457173836e-06, "loss": 84.308, "step": 3889 }, { "compression_loss": 84.99383544921875, "epoch": 1.41, "learning_rate": 8.939284423563427e-06, "loss": 85.219, "step": 3890 }, { "compression_loss": 84.51451110839844, "epoch": 1.41, "learning_rate": 8.933863389953018e-06, "loss": 85.0218, "step": 3891 }, { "compression_loss": 84.43331909179688, "epoch": 1.41, "learning_rate": 8.92844235634261e-06, "loss": 84.8327, "step": 3892 }, { "compression_loss": 83.44764709472656, "epoch": 1.41, "learning_rate": 8.923021322732202e-06, "loss": 83.8928, "step": 3893 }, { "compression_loss": 83.95429992675781, "epoch": 1.41, "learning_rate": 8.917600289121794e-06, "loss": 84.287, "step": 3894 }, { "compression_loss": 86.37821197509766, "epoch": 1.41, "learning_rate": 8.912179255511383e-06, "loss": 86.6548, "step": 3895 }, { "compression_loss": 85.06775665283203, "epoch": 1.41, "learning_rate": 8.906758221900975e-06, "loss": 85.4632, "step": 3896 }, { "compression_loss": 84.83525848388672, "epoch": 1.41, "learning_rate": 8.901337188290567e-06, "loss": 85.8222, "step": 3897 }, { "compression_loss": 85.97613525390625, "epoch": 1.41, "learning_rate": 8.895916154680159e-06, "loss": 86.5372, "step": 3898 }, { "compression_loss": 83.51284790039062, "epoch": 1.41, "learning_rate": 8.89049512106975e-06, "loss": 83.9577, "step": 3899 }, { "compression_loss": 84.29283905029297, "epoch": 1.41, "learning_rate": 8.885074087459341e-06, "loss": 85.0543, "step": 3900 }, { "compression_loss": 85.47571563720703, "epoch": 1.41, "learning_rate": 8.879653053848934e-06, "loss": 85.9106, "step": 3901 }, { "compression_loss": 84.99441528320312, "epoch": 1.41, "learning_rate": 8.874232020238526e-06, "loss": 85.62, "step": 3902 }, { "compression_loss": 84.94246673583984, "epoch": 1.41, "learning_rate": 8.868810986628117e-06, "loss": 85.5149, "step": 3903 }, { "compression_loss": 84.29930114746094, "epoch": 1.41, "learning_rate": 8.863389953017708e-06, "loss": 84.7612, "step": 3904 }, { "compression_loss": 84.30475616455078, "epoch": 1.41, "learning_rate": 8.857968919407301e-06, "loss": 84.6308, "step": 3905 }, { "compression_loss": 84.45211791992188, "epoch": 1.41, "learning_rate": 8.852547885796892e-06, "loss": 84.8535, "step": 3906 }, { "compression_loss": 84.78570556640625, "epoch": 1.41, "learning_rate": 8.847126852186484e-06, "loss": 85.2461, "step": 3907 }, { "compression_loss": 84.03820037841797, "epoch": 1.41, "learning_rate": 8.841705818576075e-06, "loss": 84.6518, "step": 3908 }, { "compression_loss": 85.98155212402344, "epoch": 1.41, "learning_rate": 8.836284784965668e-06, "loss": 86.5578, "step": 3909 }, { "compression_loss": 84.52909088134766, "epoch": 1.41, "learning_rate": 8.830863751355259e-06, "loss": 84.9741, "step": 3910 }, { "compression_loss": 83.39754486083984, "epoch": 1.41, "learning_rate": 8.82544271774485e-06, "loss": 83.6572, "step": 3911 }, { "compression_loss": 83.626953125, "epoch": 1.41, "learning_rate": 8.820021684134442e-06, "loss": 84.0418, "step": 3912 }, { "compression_loss": 85.052490234375, "epoch": 1.41, "learning_rate": 8.814600650524035e-06, "loss": 85.4419, "step": 3913 }, { "compression_loss": 84.65938568115234, "epoch": 1.41, "learning_rate": 8.809179616913626e-06, "loss": 85.1635, "step": 3914 }, { "compression_loss": 83.49308013916016, "epoch": 1.41, "learning_rate": 8.803758583303217e-06, "loss": 83.8669, "step": 3915 }, { "compression_loss": 86.2838134765625, "epoch": 1.42, "learning_rate": 8.798337549692808e-06, "loss": 86.651, "step": 3916 }, { "compression_loss": 84.87432861328125, "epoch": 1.42, "learning_rate": 8.7929165160824e-06, "loss": 85.3342, "step": 3917 }, { "compression_loss": 85.1725845336914, "epoch": 1.42, "learning_rate": 8.787495482471991e-06, "loss": 85.6667, "step": 3918 }, { "compression_loss": 84.45075988769531, "epoch": 1.42, "learning_rate": 8.782074448861582e-06, "loss": 84.8624, "step": 3919 }, { "compression_loss": 82.92894744873047, "epoch": 1.42, "learning_rate": 8.776653415251174e-06, "loss": 83.2784, "step": 3920 }, { "compression_loss": 85.18988037109375, "epoch": 1.42, "learning_rate": 8.771232381640766e-06, "loss": 85.7546, "step": 3921 }, { "compression_loss": 83.85250091552734, "epoch": 1.42, "learning_rate": 8.765811348030358e-06, "loss": 84.4574, "step": 3922 }, { "compression_loss": 85.40020751953125, "epoch": 1.42, "learning_rate": 8.760390314419949e-06, "loss": 85.9645, "step": 3923 }, { "compression_loss": 83.19412994384766, "epoch": 1.42, "learning_rate": 8.75496928080954e-06, "loss": 83.8169, "step": 3924 }, { "compression_loss": 83.98321533203125, "epoch": 1.42, "learning_rate": 8.749548247199133e-06, "loss": 84.6636, "step": 3925 }, { "compression_loss": 86.39547729492188, "epoch": 1.42, "learning_rate": 8.744127213588725e-06, "loss": 86.7592, "step": 3926 }, { "compression_loss": 85.96058654785156, "epoch": 1.42, "learning_rate": 8.738706179978316e-06, "loss": 86.2308, "step": 3927 }, { "compression_loss": 86.65756225585938, "epoch": 1.42, "learning_rate": 8.733285146367907e-06, "loss": 87.0861, "step": 3928 }, { "compression_loss": 84.54176330566406, "epoch": 1.42, "learning_rate": 8.7278641127575e-06, "loss": 84.9402, "step": 3929 }, { "compression_loss": 82.84429931640625, "epoch": 1.42, "learning_rate": 8.722443079147091e-06, "loss": 83.2058, "step": 3930 }, { "compression_loss": 83.8807144165039, "epoch": 1.42, "learning_rate": 8.717022045536683e-06, "loss": 84.3342, "step": 3931 }, { "compression_loss": 83.30438232421875, "epoch": 1.42, "learning_rate": 8.711601011926274e-06, "loss": 83.7062, "step": 3932 }, { "compression_loss": 86.01090240478516, "epoch": 1.42, "learning_rate": 8.706179978315867e-06, "loss": 86.5278, "step": 3933 }, { "compression_loss": 85.89822387695312, "epoch": 1.42, "learning_rate": 8.700758944705458e-06, "loss": 86.5411, "step": 3934 }, { "compression_loss": 83.62055969238281, "epoch": 1.42, "learning_rate": 8.69533791109505e-06, "loss": 83.9421, "step": 3935 }, { "compression_loss": 85.47767639160156, "epoch": 1.42, "learning_rate": 8.68991687748464e-06, "loss": 86.123, "step": 3936 }, { "compression_loss": 83.41288757324219, "epoch": 1.42, "learning_rate": 8.684495843874234e-06, "loss": 84.027, "step": 3937 }, { "compression_loss": 82.74314880371094, "epoch": 1.42, "learning_rate": 8.679074810263823e-06, "loss": 83.1356, "step": 3938 }, { "compression_loss": 83.94454956054688, "epoch": 1.42, "learning_rate": 8.673653776653414e-06, "loss": 84.2384, "step": 3939 }, { "compression_loss": 86.16217041015625, "epoch": 1.42, "learning_rate": 8.668232743043006e-06, "loss": 86.6436, "step": 3940 }, { "compression_loss": 84.24346923828125, "epoch": 1.42, "learning_rate": 8.662811709432599e-06, "loss": 84.7077, "step": 3941 }, { "compression_loss": 83.8691635131836, "epoch": 1.42, "learning_rate": 8.65739067582219e-06, "loss": 84.3287, "step": 3942 }, { "compression_loss": 83.93840026855469, "epoch": 1.43, "learning_rate": 8.651969642211781e-06, "loss": 84.3366, "step": 3943 }, { "compression_loss": 82.91901397705078, "epoch": 1.43, "learning_rate": 8.646548608601372e-06, "loss": 83.6539, "step": 3944 }, { "compression_loss": 83.87096405029297, "epoch": 1.43, "learning_rate": 8.641127574990965e-06, "loss": 84.2023, "step": 3945 }, { "compression_loss": 83.32964324951172, "epoch": 1.43, "learning_rate": 8.635706541380557e-06, "loss": 83.7449, "step": 3946 }, { "compression_loss": 84.92835998535156, "epoch": 1.43, "learning_rate": 8.630285507770148e-06, "loss": 85.767, "step": 3947 }, { "compression_loss": 84.30836486816406, "epoch": 1.43, "learning_rate": 8.62486447415974e-06, "loss": 84.6329, "step": 3948 }, { "compression_loss": 84.57074737548828, "epoch": 1.43, "learning_rate": 8.619443440549332e-06, "loss": 84.9642, "step": 3949 }, { "compression_loss": 84.14796447753906, "epoch": 1.43, "learning_rate": 8.614022406938924e-06, "loss": 84.8218, "step": 3950 }, { "compression_loss": 83.11653900146484, "epoch": 1.43, "learning_rate": 8.608601373328515e-06, "loss": 83.4758, "step": 3951 }, { "compression_loss": 83.74425506591797, "epoch": 1.43, "learning_rate": 8.603180339718106e-06, "loss": 84.2258, "step": 3952 }, { "compression_loss": 83.39820861816406, "epoch": 1.43, "learning_rate": 8.597759306107699e-06, "loss": 83.6662, "step": 3953 }, { "compression_loss": 82.92877197265625, "epoch": 1.43, "learning_rate": 8.59233827249729e-06, "loss": 83.3741, "step": 3954 }, { "compression_loss": 82.42304229736328, "epoch": 1.43, "learning_rate": 8.586917238886882e-06, "loss": 82.8866, "step": 3955 }, { "compression_loss": 85.62490844726562, "epoch": 1.43, "learning_rate": 8.581496205276473e-06, "loss": 86.1107, "step": 3956 }, { "compression_loss": 84.21881103515625, "epoch": 1.43, "learning_rate": 8.576075171666066e-06, "loss": 84.6802, "step": 3957 }, { "compression_loss": 84.01060485839844, "epoch": 1.43, "learning_rate": 8.570654138055657e-06, "loss": 84.3783, "step": 3958 }, { "compression_loss": 84.29607391357422, "epoch": 1.43, "learning_rate": 8.565233104445248e-06, "loss": 84.6135, "step": 3959 }, { "compression_loss": 84.83271789550781, "epoch": 1.43, "learning_rate": 8.559812070834838e-06, "loss": 85.15, "step": 3960 }, { "compression_loss": 82.87215423583984, "epoch": 1.43, "learning_rate": 8.554391037224431e-06, "loss": 83.2891, "step": 3961 }, { "compression_loss": 82.71231079101562, "epoch": 1.43, "learning_rate": 8.548970003614022e-06, "loss": 83.2045, "step": 3962 }, { "compression_loss": 83.80242919921875, "epoch": 1.43, "learning_rate": 8.543548970003613e-06, "loss": 84.2762, "step": 3963 }, { "compression_loss": 83.247802734375, "epoch": 1.43, "learning_rate": 8.538127936393205e-06, "loss": 83.6, "step": 3964 }, { "compression_loss": 85.89624786376953, "epoch": 1.43, "learning_rate": 8.532706902782798e-06, "loss": 86.0873, "step": 3965 }, { "compression_loss": 83.81170654296875, "epoch": 1.43, "learning_rate": 8.527285869172389e-06, "loss": 84.5213, "step": 3966 }, { "compression_loss": 83.13780212402344, "epoch": 1.43, "learning_rate": 8.52186483556198e-06, "loss": 83.5167, "step": 3967 }, { "compression_loss": 83.71549987792969, "epoch": 1.43, "learning_rate": 8.516443801951571e-06, "loss": 84.3261, "step": 3968 }, { "compression_loss": 83.72008514404297, "epoch": 1.43, "learning_rate": 8.511022768341164e-06, "loss": 84.144, "step": 3969 }, { "compression_loss": 82.50204467773438, "epoch": 1.43, "learning_rate": 8.505601734730756e-06, "loss": 82.87, "step": 3970 }, { "compression_loss": 84.20863342285156, "epoch": 1.44, "learning_rate": 8.500180701120347e-06, "loss": 84.3458, "step": 3971 }, { "compression_loss": 85.60448455810547, "epoch": 1.44, "learning_rate": 8.494759667509938e-06, "loss": 86.1417, "step": 3972 }, { "compression_loss": 83.0594253540039, "epoch": 1.44, "learning_rate": 8.489338633899531e-06, "loss": 83.5362, "step": 3973 }, { "compression_loss": 84.28858947753906, "epoch": 1.44, "learning_rate": 8.483917600289123e-06, "loss": 84.8507, "step": 3974 }, { "compression_loss": 86.880126953125, "epoch": 1.44, "learning_rate": 8.478496566678714e-06, "loss": 87.4866, "step": 3975 }, { "compression_loss": 85.49738311767578, "epoch": 1.44, "learning_rate": 8.473075533068305e-06, "loss": 85.8552, "step": 3976 }, { "compression_loss": 84.00480651855469, "epoch": 1.44, "learning_rate": 8.467654499457898e-06, "loss": 84.4306, "step": 3977 }, { "compression_loss": 85.17153930664062, "epoch": 1.44, "learning_rate": 8.46223346584749e-06, "loss": 85.5433, "step": 3978 }, { "compression_loss": 83.27301025390625, "epoch": 1.44, "learning_rate": 8.45681243223708e-06, "loss": 83.8965, "step": 3979 }, { "compression_loss": 86.82574462890625, "epoch": 1.44, "learning_rate": 8.451391398626672e-06, "loss": 87.3229, "step": 3980 }, { "compression_loss": 84.46373748779297, "epoch": 1.44, "learning_rate": 8.445970365016265e-06, "loss": 85.0093, "step": 3981 }, { "compression_loss": 82.2380599975586, "epoch": 1.44, "learning_rate": 8.440549331405854e-06, "loss": 82.728, "step": 3982 }, { "compression_loss": 85.78958129882812, "epoch": 1.44, "learning_rate": 8.435128297795446e-06, "loss": 86.0631, "step": 3983 }, { "compression_loss": 85.04903411865234, "epoch": 1.44, "learning_rate": 8.429707264185037e-06, "loss": 85.5539, "step": 3984 }, { "compression_loss": 83.09712219238281, "epoch": 1.44, "learning_rate": 8.42428623057463e-06, "loss": 83.4231, "step": 3985 }, { "compression_loss": 84.97795104980469, "epoch": 1.44, "learning_rate": 8.418865196964221e-06, "loss": 85.4928, "step": 3986 }, { "compression_loss": 84.97150421142578, "epoch": 1.44, "learning_rate": 8.413444163353812e-06, "loss": 85.5271, "step": 3987 }, { "compression_loss": 84.44046020507812, "epoch": 1.44, "learning_rate": 8.408023129743404e-06, "loss": 84.8392, "step": 3988 }, { "compression_loss": 83.7213363647461, "epoch": 1.44, "learning_rate": 8.402602096132997e-06, "loss": 84.191, "step": 3989 }, { "compression_loss": 83.71678161621094, "epoch": 1.44, "learning_rate": 8.397181062522588e-06, "loss": 84.2382, "step": 3990 }, { "compression_loss": 83.94697570800781, "epoch": 1.44, "learning_rate": 8.39176002891218e-06, "loss": 84.395, "step": 3991 }, { "compression_loss": 81.7148208618164, "epoch": 1.44, "learning_rate": 8.38633899530177e-06, "loss": 82.2996, "step": 3992 }, { "compression_loss": 84.55486297607422, "epoch": 1.44, "learning_rate": 8.380917961691363e-06, "loss": 85.0848, "step": 3993 }, { "compression_loss": 83.48002624511719, "epoch": 1.44, "learning_rate": 8.375496928080955e-06, "loss": 84.0419, "step": 3994 }, { "compression_loss": 85.67312622070312, "epoch": 1.44, "learning_rate": 8.370075894470546e-06, "loss": 86.1211, "step": 3995 }, { "compression_loss": 85.05614471435547, "epoch": 1.44, "learning_rate": 8.364654860860137e-06, "loss": 85.6487, "step": 3996 }, { "compression_loss": 84.2533950805664, "epoch": 1.44, "learning_rate": 8.35923382724973e-06, "loss": 84.618, "step": 3997 }, { "compression_loss": 82.65507507324219, "epoch": 1.44, "learning_rate": 8.353812793639322e-06, "loss": 83.0368, "step": 3998 }, { "compression_loss": 82.13822937011719, "epoch": 1.45, "learning_rate": 8.348391760028913e-06, "loss": 82.8541, "step": 3999 }, { "compression_loss": 83.52559661865234, "epoch": 1.45, "learning_rate": 8.342970726418504e-06, "loss": 84.1548, "step": 4000 }, { "epoch": 1.45, "eval_exact_match": 86.87795648060549, "eval_f1": 93.07019737772445, "step": 4000 }, { "compression_loss": 84.20699310302734, "epoch": 1.45, "learning_rate": 8.337549692808097e-06, "loss": 84.5399, "step": 4001 }, { "compression_loss": 85.1815414428711, "epoch": 1.45, "learning_rate": 8.332128659197688e-06, "loss": 85.4593, "step": 4002 }, { "compression_loss": 83.04866027832031, "epoch": 1.45, "learning_rate": 8.32670762558728e-06, "loss": 83.2805, "step": 4003 }, { "compression_loss": 84.53594970703125, "epoch": 1.45, "learning_rate": 8.321286591976869e-06, "loss": 85.1584, "step": 4004 }, { "compression_loss": 86.68606567382812, "epoch": 1.45, "learning_rate": 8.315865558366462e-06, "loss": 87.0329, "step": 4005 }, { "compression_loss": 83.46957397460938, "epoch": 1.45, "learning_rate": 8.310444524756053e-06, "loss": 84.1427, "step": 4006 }, { "compression_loss": 82.72642517089844, "epoch": 1.45, "learning_rate": 8.305023491145645e-06, "loss": 82.9723, "step": 4007 }, { "compression_loss": 82.67564392089844, "epoch": 1.45, "learning_rate": 8.299602457535236e-06, "loss": 83.1387, "step": 4008 }, { "compression_loss": 84.80158233642578, "epoch": 1.45, "learning_rate": 8.294181423924829e-06, "loss": 85.1467, "step": 4009 }, { "compression_loss": 84.83604431152344, "epoch": 1.45, "learning_rate": 8.28876039031442e-06, "loss": 85.4972, "step": 4010 }, { "compression_loss": 82.23638153076172, "epoch": 1.45, "learning_rate": 8.283339356704011e-06, "loss": 82.8644, "step": 4011 }, { "compression_loss": 84.90838623046875, "epoch": 1.45, "learning_rate": 8.277918323093603e-06, "loss": 85.4247, "step": 4012 }, { "compression_loss": 85.91305541992188, "epoch": 1.45, "learning_rate": 8.272497289483196e-06, "loss": 86.6121, "step": 4013 }, { "compression_loss": 85.23828887939453, "epoch": 1.45, "learning_rate": 8.267076255872787e-06, "loss": 85.5183, "step": 4014 }, { "compression_loss": 84.93142700195312, "epoch": 1.45, "learning_rate": 8.261655222262378e-06, "loss": 85.3488, "step": 4015 }, { "compression_loss": 83.66349029541016, "epoch": 1.45, "learning_rate": 8.25623418865197e-06, "loss": 84.1001, "step": 4016 }, { "compression_loss": 85.13432312011719, "epoch": 1.45, "learning_rate": 8.250813155041562e-06, "loss": 85.9507, "step": 4017 }, { "compression_loss": 84.31802368164062, "epoch": 1.45, "learning_rate": 8.245392121431154e-06, "loss": 84.8198, "step": 4018 }, { "compression_loss": 84.5308609008789, "epoch": 1.45, "learning_rate": 8.239971087820745e-06, "loss": 84.7895, "step": 4019 }, { "compression_loss": 81.9653091430664, "epoch": 1.45, "learning_rate": 8.234550054210336e-06, "loss": 82.4517, "step": 4020 }, { "compression_loss": 87.27034759521484, "epoch": 1.45, "learning_rate": 8.22912902059993e-06, "loss": 87.8091, "step": 4021 }, { "compression_loss": 85.53529357910156, "epoch": 1.45, "learning_rate": 8.22370798698952e-06, "loss": 86.1268, "step": 4022 }, { "compression_loss": 84.23667907714844, "epoch": 1.45, "learning_rate": 8.218286953379112e-06, "loss": 84.4997, "step": 4023 }, { "compression_loss": 85.2034912109375, "epoch": 1.45, "learning_rate": 8.212865919768703e-06, "loss": 85.6267, "step": 4024 }, { "compression_loss": 83.14260864257812, "epoch": 1.45, "learning_rate": 8.207444886158294e-06, "loss": 83.2965, "step": 4025 }, { "compression_loss": 83.63822174072266, "epoch": 1.46, "learning_rate": 8.202023852547886e-06, "loss": 84.0163, "step": 4026 }, { "compression_loss": 82.3258056640625, "epoch": 1.46, "learning_rate": 8.196602818937477e-06, "loss": 82.5605, "step": 4027 }, { "compression_loss": 83.94435119628906, "epoch": 1.46, "learning_rate": 8.191181785327068e-06, "loss": 84.2216, "step": 4028 }, { "compression_loss": 84.61820983886719, "epoch": 1.46, "learning_rate": 8.185760751716661e-06, "loss": 85.2461, "step": 4029 }, { "compression_loss": 83.10720825195312, "epoch": 1.46, "learning_rate": 8.180339718106252e-06, "loss": 83.4069, "step": 4030 }, { "compression_loss": 82.53894805908203, "epoch": 1.46, "learning_rate": 8.174918684495844e-06, "loss": 83.1292, "step": 4031 }, { "compression_loss": 83.81884002685547, "epoch": 1.46, "learning_rate": 8.169497650885435e-06, "loss": 84.1354, "step": 4032 }, { "compression_loss": 84.06332397460938, "epoch": 1.46, "learning_rate": 8.164076617275028e-06, "loss": 84.5185, "step": 4033 }, { "compression_loss": 83.80744934082031, "epoch": 1.46, "learning_rate": 8.15865558366462e-06, "loss": 84.3224, "step": 4034 }, { "compression_loss": 83.89786529541016, "epoch": 1.46, "learning_rate": 8.15323455005421e-06, "loss": 84.4148, "step": 4035 }, { "compression_loss": 84.9400634765625, "epoch": 1.46, "learning_rate": 8.147813516443802e-06, "loss": 85.2481, "step": 4036 }, { "compression_loss": 84.6898422241211, "epoch": 1.46, "learning_rate": 8.142392482833395e-06, "loss": 84.9968, "step": 4037 }, { "compression_loss": 86.45538330078125, "epoch": 1.46, "learning_rate": 8.136971449222986e-06, "loss": 86.9927, "step": 4038 }, { "compression_loss": 84.2423095703125, "epoch": 1.46, "learning_rate": 8.131550415612577e-06, "loss": 84.9153, "step": 4039 }, { "compression_loss": 83.65057373046875, "epoch": 1.46, "learning_rate": 8.126129382002168e-06, "loss": 84.0597, "step": 4040 }, { "compression_loss": 84.64603424072266, "epoch": 1.46, "learning_rate": 8.120708348391761e-06, "loss": 85.0094, "step": 4041 }, { "compression_loss": 84.98812103271484, "epoch": 1.46, "learning_rate": 8.115287314781353e-06, "loss": 85.7065, "step": 4042 }, { "compression_loss": 84.72602081298828, "epoch": 1.46, "learning_rate": 8.109866281170944e-06, "loss": 85.087, "step": 4043 }, { "compression_loss": 84.0633544921875, "epoch": 1.46, "learning_rate": 8.104445247560535e-06, "loss": 84.4393, "step": 4044 }, { "compression_loss": 85.22610473632812, "epoch": 1.46, "learning_rate": 8.099024213950128e-06, "loss": 85.5856, "step": 4045 }, { "compression_loss": 85.74645233154297, "epoch": 1.46, "learning_rate": 8.09360318033972e-06, "loss": 86.3071, "step": 4046 }, { "compression_loss": 85.23565673828125, "epoch": 1.46, "learning_rate": 8.088182146729309e-06, "loss": 85.6751, "step": 4047 }, { "compression_loss": 84.4311752319336, "epoch": 1.46, "learning_rate": 8.0827611131189e-06, "loss": 84.7239, "step": 4048 }, { "compression_loss": 86.68878173828125, "epoch": 1.46, "learning_rate": 8.077340079508493e-06, "loss": 87.2174, "step": 4049 }, { "compression_loss": 86.76515197753906, "epoch": 1.46, "learning_rate": 8.071919045898085e-06, "loss": 87.267, "step": 4050 }, { "compression_loss": 84.09957122802734, "epoch": 1.46, "learning_rate": 8.066498012287676e-06, "loss": 84.3724, "step": 4051 }, { "compression_loss": 83.9018325805664, "epoch": 1.46, "learning_rate": 8.061076978677267e-06, "loss": 84.2961, "step": 4052 }, { "compression_loss": 83.05427551269531, "epoch": 1.46, "learning_rate": 8.05565594506686e-06, "loss": 83.5478, "step": 4053 }, { "compression_loss": 83.84539794921875, "epoch": 1.47, "learning_rate": 8.050234911456451e-06, "loss": 84.2137, "step": 4054 }, { "compression_loss": 84.42646789550781, "epoch": 1.47, "learning_rate": 8.044813877846043e-06, "loss": 84.9811, "step": 4055 }, { "compression_loss": 86.38864135742188, "epoch": 1.47, "learning_rate": 8.039392844235634e-06, "loss": 87.3617, "step": 4056 }, { "compression_loss": 82.35159301757812, "epoch": 1.47, "learning_rate": 8.033971810625227e-06, "loss": 82.6102, "step": 4057 }, { "compression_loss": 84.04497528076172, "epoch": 1.47, "learning_rate": 8.028550777014818e-06, "loss": 84.6041, "step": 4058 }, { "compression_loss": 83.76835632324219, "epoch": 1.47, "learning_rate": 8.02312974340441e-06, "loss": 84.0813, "step": 4059 }, { "compression_loss": 84.22916412353516, "epoch": 1.47, "learning_rate": 8.017708709794e-06, "loss": 84.5935, "step": 4060 }, { "compression_loss": 83.72262573242188, "epoch": 1.47, "learning_rate": 8.012287676183594e-06, "loss": 84.0956, "step": 4061 }, { "compression_loss": 85.93157958984375, "epoch": 1.47, "learning_rate": 8.006866642573185e-06, "loss": 86.4008, "step": 4062 }, { "compression_loss": 84.14985656738281, "epoch": 1.47, "learning_rate": 8.001445608962776e-06, "loss": 84.6666, "step": 4063 }, { "compression_loss": 83.35596466064453, "epoch": 1.47, "learning_rate": 7.996024575352367e-06, "loss": 83.7236, "step": 4064 }, { "compression_loss": 83.1849136352539, "epoch": 1.47, "learning_rate": 7.990603541741959e-06, "loss": 83.4584, "step": 4065 }, { "compression_loss": 84.30570983886719, "epoch": 1.47, "learning_rate": 7.985182508131552e-06, "loss": 84.7668, "step": 4066 }, { "compression_loss": 85.98112487792969, "epoch": 1.47, "learning_rate": 7.979761474521143e-06, "loss": 86.608, "step": 4067 }, { "compression_loss": 83.80265808105469, "epoch": 1.47, "learning_rate": 7.974340440910734e-06, "loss": 84.1725, "step": 4068 }, { "compression_loss": 84.44356536865234, "epoch": 1.47, "learning_rate": 7.968919407300324e-06, "loss": 84.9022, "step": 4069 }, { "compression_loss": 85.44694519042969, "epoch": 1.47, "learning_rate": 7.963498373689917e-06, "loss": 85.7124, "step": 4070 }, { "compression_loss": 83.75543212890625, "epoch": 1.47, "learning_rate": 7.958077340079508e-06, "loss": 84.2146, "step": 4071 }, { "compression_loss": 83.97711944580078, "epoch": 1.47, "learning_rate": 7.9526563064691e-06, "loss": 84.543, "step": 4072 }, { "compression_loss": 83.3121337890625, "epoch": 1.47, "learning_rate": 7.94723527285869e-06, "loss": 83.7309, "step": 4073 }, { "compression_loss": 83.106689453125, "epoch": 1.47, "learning_rate": 7.941814239248284e-06, "loss": 83.4197, "step": 4074 }, { "compression_loss": 83.97299194335938, "epoch": 1.47, "learning_rate": 7.936393205637875e-06, "loss": 84.3727, "step": 4075 }, { "compression_loss": 84.12020111083984, "epoch": 1.47, "learning_rate": 7.930972172027466e-06, "loss": 84.4475, "step": 4076 }, { "compression_loss": 84.66810607910156, "epoch": 1.47, "learning_rate": 7.925551138417057e-06, "loss": 85.2952, "step": 4077 }, { "compression_loss": 84.34071350097656, "epoch": 1.47, "learning_rate": 7.92013010480665e-06, "loss": 84.7618, "step": 4078 }, { "compression_loss": 84.90535736083984, "epoch": 1.47, "learning_rate": 7.914709071196242e-06, "loss": 85.3758, "step": 4079 }, { "compression_loss": 83.96652221679688, "epoch": 1.47, "learning_rate": 7.909288037585833e-06, "loss": 84.2042, "step": 4080 }, { "compression_loss": 83.34143829345703, "epoch": 1.47, "learning_rate": 7.903867003975424e-06, "loss": 83.6699, "step": 4081 }, { "compression_loss": 83.21316528320312, "epoch": 1.48, "learning_rate": 7.898445970365017e-06, "loss": 83.674, "step": 4082 }, { "compression_loss": 84.7431640625, "epoch": 1.48, "learning_rate": 7.893024936754608e-06, "loss": 85.2379, "step": 4083 }, { "compression_loss": 83.69664001464844, "epoch": 1.48, "learning_rate": 7.8876039031442e-06, "loss": 84.2398, "step": 4084 }, { "compression_loss": 84.48026275634766, "epoch": 1.48, "learning_rate": 7.882182869533791e-06, "loss": 84.7959, "step": 4085 }, { "compression_loss": 85.26801300048828, "epoch": 1.48, "learning_rate": 7.876761835923384e-06, "loss": 85.6758, "step": 4086 }, { "compression_loss": 84.190185546875, "epoch": 1.48, "learning_rate": 7.871340802312975e-06, "loss": 84.6382, "step": 4087 }, { "compression_loss": 84.3678970336914, "epoch": 1.48, "learning_rate": 7.865919768702566e-06, "loss": 84.7846, "step": 4088 }, { "compression_loss": 85.16661071777344, "epoch": 1.48, "learning_rate": 7.860498735092158e-06, "loss": 85.6468, "step": 4089 }, { "compression_loss": 84.19512176513672, "epoch": 1.48, "learning_rate": 7.85507770148175e-06, "loss": 84.6644, "step": 4090 }, { "compression_loss": 83.22587585449219, "epoch": 1.48, "learning_rate": 7.84965666787134e-06, "loss": 83.5417, "step": 4091 }, { "compression_loss": 83.82479858398438, "epoch": 1.48, "learning_rate": 7.844235634260932e-06, "loss": 84.2599, "step": 4092 }, { "compression_loss": 83.42521667480469, "epoch": 1.48, "learning_rate": 7.838814600650523e-06, "loss": 83.8648, "step": 4093 }, { "compression_loss": 83.29210662841797, "epoch": 1.48, "learning_rate": 7.833393567040116e-06, "loss": 83.7398, "step": 4094 }, { "compression_loss": 84.50491333007812, "epoch": 1.48, "learning_rate": 7.827972533429707e-06, "loss": 85.0162, "step": 4095 }, { "compression_loss": 84.71755981445312, "epoch": 1.48, "learning_rate": 7.822551499819298e-06, "loss": 85.0866, "step": 4096 }, { "compression_loss": 84.08486938476562, "epoch": 1.48, "learning_rate": 7.81713046620889e-06, "loss": 84.464, "step": 4097 }, { "compression_loss": 85.35917663574219, "epoch": 1.48, "learning_rate": 7.811709432598483e-06, "loss": 85.8778, "step": 4098 }, { "compression_loss": 84.0057373046875, "epoch": 1.48, "learning_rate": 7.806288398988074e-06, "loss": 84.4123, "step": 4099 }, { "compression_loss": 84.34674072265625, "epoch": 1.48, "learning_rate": 7.800867365377665e-06, "loss": 84.9516, "step": 4100 }, { "compression_loss": 85.44699096679688, "epoch": 1.48, "learning_rate": 7.795446331767256e-06, "loss": 86.0949, "step": 4101 }, { "compression_loss": 84.78144073486328, "epoch": 1.48, "learning_rate": 7.79002529815685e-06, "loss": 85.2772, "step": 4102 }, { "compression_loss": 85.17686462402344, "epoch": 1.48, "learning_rate": 7.78460426454644e-06, "loss": 85.5547, "step": 4103 }, { "compression_loss": 84.61339569091797, "epoch": 1.48, "learning_rate": 7.779183230936032e-06, "loss": 84.935, "step": 4104 }, { "compression_loss": 86.321533203125, "epoch": 1.48, "learning_rate": 7.773762197325623e-06, "loss": 86.7813, "step": 4105 }, { "compression_loss": 83.47721862792969, "epoch": 1.48, "learning_rate": 7.768341163715216e-06, "loss": 83.9782, "step": 4106 }, { "compression_loss": 83.16188049316406, "epoch": 1.48, "learning_rate": 7.762920130104807e-06, "loss": 83.7374, "step": 4107 }, { "compression_loss": 82.55229949951172, "epoch": 1.48, "learning_rate": 7.757499096494399e-06, "loss": 82.8608, "step": 4108 }, { "compression_loss": 82.78018188476562, "epoch": 1.49, "learning_rate": 7.75207806288399e-06, "loss": 83.1379, "step": 4109 }, { "compression_loss": 84.50894165039062, "epoch": 1.49, "learning_rate": 7.746657029273583e-06, "loss": 84.9787, "step": 4110 }, { "compression_loss": 83.69976806640625, "epoch": 1.49, "learning_rate": 7.741235995663174e-06, "loss": 84.1478, "step": 4111 }, { "compression_loss": 83.51434326171875, "epoch": 1.49, "learning_rate": 7.735814962052764e-06, "loss": 83.9964, "step": 4112 }, { "compression_loss": 84.54598999023438, "epoch": 1.49, "learning_rate": 7.730393928442355e-06, "loss": 85.2131, "step": 4113 }, { "compression_loss": 83.72018432617188, "epoch": 1.49, "learning_rate": 7.724972894831948e-06, "loss": 84.2257, "step": 4114 }, { "compression_loss": 85.78227996826172, "epoch": 1.49, "learning_rate": 7.71955186122154e-06, "loss": 86.3152, "step": 4115 }, { "compression_loss": 84.51412200927734, "epoch": 1.49, "learning_rate": 7.71413082761113e-06, "loss": 84.9721, "step": 4116 }, { "compression_loss": 83.58002471923828, "epoch": 1.49, "learning_rate": 7.708709794000722e-06, "loss": 84.0281, "step": 4117 }, { "compression_loss": 83.94598388671875, "epoch": 1.49, "learning_rate": 7.703288760390315e-06, "loss": 84.3427, "step": 4118 }, { "compression_loss": 83.69276428222656, "epoch": 1.49, "learning_rate": 7.697867726779906e-06, "loss": 84.4303, "step": 4119 }, { "compression_loss": 82.92465209960938, "epoch": 1.49, "learning_rate": 7.692446693169497e-06, "loss": 83.161, "step": 4120 }, { "compression_loss": 84.66635131835938, "epoch": 1.49, "learning_rate": 7.687025659559089e-06, "loss": 85.0945, "step": 4121 }, { "compression_loss": 84.3950424194336, "epoch": 1.49, "learning_rate": 7.681604625948682e-06, "loss": 84.9133, "step": 4122 }, { "compression_loss": 83.861328125, "epoch": 1.49, "learning_rate": 7.676183592338273e-06, "loss": 84.5376, "step": 4123 }, { "compression_loss": 84.00787353515625, "epoch": 1.49, "learning_rate": 7.670762558727864e-06, "loss": 84.528, "step": 4124 }, { "compression_loss": 86.3774185180664, "epoch": 1.49, "learning_rate": 7.665341525117455e-06, "loss": 87.0004, "step": 4125 }, { "compression_loss": 84.990234375, "epoch": 1.49, "learning_rate": 7.659920491507048e-06, "loss": 85.2937, "step": 4126 }, { "compression_loss": 84.89926147460938, "epoch": 1.49, "learning_rate": 7.65449945789664e-06, "loss": 85.2964, "step": 4127 }, { "compression_loss": 83.03691101074219, "epoch": 1.49, "learning_rate": 7.649078424286231e-06, "loss": 83.4597, "step": 4128 }, { "compression_loss": 85.71488189697266, "epoch": 1.49, "learning_rate": 7.643657390675822e-06, "loss": 86.2857, "step": 4129 }, { "compression_loss": 82.6429443359375, "epoch": 1.49, "learning_rate": 7.638236357065415e-06, "loss": 83.1427, "step": 4130 }, { "compression_loss": 85.90995788574219, "epoch": 1.49, "learning_rate": 7.632815323455006e-06, "loss": 86.5778, "step": 4131 }, { "compression_loss": 84.97187805175781, "epoch": 1.49, "learning_rate": 7.627394289844597e-06, "loss": 85.5194, "step": 4132 }, { "compression_loss": 83.875, "epoch": 1.49, "learning_rate": 7.621973256234188e-06, "loss": 84.2415, "step": 4133 }, { "compression_loss": 83.95108032226562, "epoch": 1.49, "learning_rate": 7.616552222623781e-06, "loss": 84.3174, "step": 4134 }, { "compression_loss": 85.2117919921875, "epoch": 1.49, "learning_rate": 7.611131189013372e-06, "loss": 85.6868, "step": 4135 }, { "compression_loss": 84.7712173461914, "epoch": 1.49, "learning_rate": 7.605710155402964e-06, "loss": 85.1837, "step": 4136 }, { "compression_loss": 83.74012756347656, "epoch": 1.5, "learning_rate": 7.600289121792555e-06, "loss": 84.1865, "step": 4137 }, { "compression_loss": 84.11788177490234, "epoch": 1.5, "learning_rate": 7.594868088182148e-06, "loss": 84.6171, "step": 4138 }, { "compression_loss": 84.36737823486328, "epoch": 1.5, "learning_rate": 7.589447054571739e-06, "loss": 84.8213, "step": 4139 }, { "compression_loss": 83.55162811279297, "epoch": 1.5, "learning_rate": 7.5840260209613295e-06, "loss": 84.0913, "step": 4140 }, { "compression_loss": 84.35498046875, "epoch": 1.5, "learning_rate": 7.578604987350921e-06, "loss": 84.9908, "step": 4141 }, { "compression_loss": 84.5593032836914, "epoch": 1.5, "learning_rate": 7.573183953740514e-06, "loss": 84.8647, "step": 4142 }, { "compression_loss": 83.7229995727539, "epoch": 1.5, "learning_rate": 7.567762920130105e-06, "loss": 84.2947, "step": 4143 }, { "compression_loss": 85.64812469482422, "epoch": 1.5, "learning_rate": 7.562341886519696e-06, "loss": 85.9917, "step": 4144 }, { "compression_loss": 85.04165649414062, "epoch": 1.5, "learning_rate": 7.556920852909288e-06, "loss": 85.6971, "step": 4145 }, { "compression_loss": 85.32398986816406, "epoch": 1.5, "learning_rate": 7.5514998192988806e-06, "loss": 85.7875, "step": 4146 }, { "compression_loss": 85.8176040649414, "epoch": 1.5, "learning_rate": 7.546078785688472e-06, "loss": 86.4493, "step": 4147 }, { "compression_loss": 83.05913543701172, "epoch": 1.5, "learning_rate": 7.540657752078063e-06, "loss": 83.4566, "step": 4148 }, { "compression_loss": 84.7332763671875, "epoch": 1.5, "learning_rate": 7.535236718467654e-06, "loss": 85.2065, "step": 4149 }, { "compression_loss": 84.46998596191406, "epoch": 1.5, "learning_rate": 7.5298156848572465e-06, "loss": 85.159, "step": 4150 }, { "compression_loss": 85.6349105834961, "epoch": 1.5, "learning_rate": 7.524394651246838e-06, "loss": 86.243, "step": 4151 }, { "compression_loss": 85.1932373046875, "epoch": 1.5, "learning_rate": 7.518973617636429e-06, "loss": 85.7067, "step": 4152 }, { "compression_loss": 85.05801391601562, "epoch": 1.5, "learning_rate": 7.51355258402602e-06, "loss": 85.605, "step": 4153 }, { "compression_loss": 82.98634338378906, "epoch": 1.5, "learning_rate": 7.508131550415613e-06, "loss": 83.2682, "step": 4154 }, { "compression_loss": 82.97935485839844, "epoch": 1.5, "learning_rate": 7.5027105168052046e-06, "loss": 83.3566, "step": 4155 }, { "compression_loss": 84.20658874511719, "epoch": 1.5, "learning_rate": 7.497289483194796e-06, "loss": 84.9628, "step": 4156 }, { "compression_loss": 84.90476989746094, "epoch": 1.5, "learning_rate": 7.491868449584388e-06, "loss": 85.2036, "step": 4157 }, { "compression_loss": 84.061767578125, "epoch": 1.5, "learning_rate": 7.486447415973979e-06, "loss": 84.5154, "step": 4158 }, { "compression_loss": 83.78985595703125, "epoch": 1.5, "learning_rate": 7.481026382363571e-06, "loss": 84.2637, "step": 4159 }, { "compression_loss": 84.43829345703125, "epoch": 1.5, "learning_rate": 7.475605348753163e-06, "loss": 84.713, "step": 4160 }, { "compression_loss": 83.61848449707031, "epoch": 1.5, "learning_rate": 7.470184315142754e-06, "loss": 84.1421, "step": 4161 }, { "compression_loss": 83.87386322021484, "epoch": 1.5, "learning_rate": 7.464763281532345e-06, "loss": 84.175, "step": 4162 }, { "compression_loss": 85.52848815917969, "epoch": 1.5, "learning_rate": 7.459342247921937e-06, "loss": 85.8063, "step": 4163 }, { "compression_loss": 84.18986511230469, "epoch": 1.5, "learning_rate": 7.4539212143115285e-06, "loss": 84.4992, "step": 4164 }, { "compression_loss": 84.47888946533203, "epoch": 1.51, "learning_rate": 7.448500180701121e-06, "loss": 84.9501, "step": 4165 }, { "compression_loss": 82.28089141845703, "epoch": 1.51, "learning_rate": 7.443079147090712e-06, "loss": 82.5038, "step": 4166 }, { "compression_loss": 82.85760498046875, "epoch": 1.51, "learning_rate": 7.437658113480304e-06, "loss": 83.4534, "step": 4167 }, { "compression_loss": 85.00981140136719, "epoch": 1.51, "learning_rate": 7.432237079869895e-06, "loss": 85.2969, "step": 4168 }, { "compression_loss": 84.9703140258789, "epoch": 1.51, "learning_rate": 7.4268160462594874e-06, "loss": 85.3701, "step": 4169 }, { "compression_loss": 84.36604309082031, "epoch": 1.51, "learning_rate": 7.421395012649079e-06, "loss": 84.8112, "step": 4170 }, { "compression_loss": 83.15068054199219, "epoch": 1.51, "learning_rate": 7.415973979038671e-06, "loss": 83.5994, "step": 4171 }, { "compression_loss": 85.64854431152344, "epoch": 1.51, "learning_rate": 7.410552945428261e-06, "loss": 86.2309, "step": 4172 }, { "compression_loss": 83.34843444824219, "epoch": 1.51, "learning_rate": 7.405131911817853e-06, "loss": 83.706, "step": 4173 }, { "compression_loss": 84.83253479003906, "epoch": 1.51, "learning_rate": 7.399710878207445e-06, "loss": 85.2621, "step": 4174 }, { "compression_loss": 83.69495391845703, "epoch": 1.51, "learning_rate": 7.394289844597037e-06, "loss": 84.1511, "step": 4175 }, { "compression_loss": 82.84793090820312, "epoch": 1.51, "learning_rate": 7.388868810986628e-06, "loss": 83.2717, "step": 4176 }, { "compression_loss": 85.87879943847656, "epoch": 1.51, "learning_rate": 7.38344777737622e-06, "loss": 86.5449, "step": 4177 }, { "compression_loss": 85.67095184326172, "epoch": 1.51, "learning_rate": 7.3780267437658114e-06, "loss": 86.096, "step": 4178 }, { "compression_loss": 83.74486541748047, "epoch": 1.51, "learning_rate": 7.3726057101554035e-06, "loss": 84.3183, "step": 4179 }, { "compression_loss": 85.16305541992188, "epoch": 1.51, "learning_rate": 7.367184676544995e-06, "loss": 85.5665, "step": 4180 }, { "compression_loss": 85.97380065917969, "epoch": 1.51, "learning_rate": 7.361763642934587e-06, "loss": 86.3647, "step": 4181 }, { "compression_loss": 83.22087860107422, "epoch": 1.51, "learning_rate": 7.356342609324178e-06, "loss": 83.5282, "step": 4182 }, { "compression_loss": 84.17113494873047, "epoch": 1.51, "learning_rate": 7.3509215757137695e-06, "loss": 84.6918, "step": 4183 }, { "compression_loss": 84.16770935058594, "epoch": 1.51, "learning_rate": 7.345500542103361e-06, "loss": 84.4551, "step": 4184 }, { "compression_loss": 85.62242126464844, "epoch": 1.51, "learning_rate": 7.340079508492953e-06, "loss": 86.1977, "step": 4185 }, { "compression_loss": 84.42884826660156, "epoch": 1.51, "learning_rate": 7.334658474882544e-06, "loss": 84.821, "step": 4186 }, { "compression_loss": 82.99464416503906, "epoch": 1.51, "learning_rate": 7.329237441272136e-06, "loss": 83.4543, "step": 4187 }, { "compression_loss": 84.06112670898438, "epoch": 1.51, "learning_rate": 7.3238164076617275e-06, "loss": 84.5002, "step": 4188 }, { "compression_loss": 86.3520278930664, "epoch": 1.51, "learning_rate": 7.31839537405132e-06, "loss": 86.5216, "step": 4189 }, { "compression_loss": 85.39099884033203, "epoch": 1.51, "learning_rate": 7.312974340440911e-06, "loss": 86.1066, "step": 4190 }, { "compression_loss": 85.9451904296875, "epoch": 1.51, "learning_rate": 7.307553306830503e-06, "loss": 86.4572, "step": 4191 }, { "compression_loss": 85.0627670288086, "epoch": 1.51, "learning_rate": 7.302132273220094e-06, "loss": 85.7133, "step": 4192 }, { "compression_loss": 83.85413360595703, "epoch": 1.52, "learning_rate": 7.2967112396096864e-06, "loss": 84.2629, "step": 4193 }, { "compression_loss": 84.65206909179688, "epoch": 1.52, "learning_rate": 7.291290205999277e-06, "loss": 85.1789, "step": 4194 }, { "compression_loss": 85.3380355834961, "epoch": 1.52, "learning_rate": 7.285869172388869e-06, "loss": 85.719, "step": 4195 }, { "compression_loss": 84.48039245605469, "epoch": 1.52, "learning_rate": 7.28044813877846e-06, "loss": 84.9185, "step": 4196 }, { "compression_loss": 83.56059265136719, "epoch": 1.52, "learning_rate": 7.275027105168052e-06, "loss": 84.0092, "step": 4197 }, { "compression_loss": 85.57681274414062, "epoch": 1.52, "learning_rate": 7.269606071557644e-06, "loss": 85.8997, "step": 4198 }, { "compression_loss": 83.11849975585938, "epoch": 1.52, "learning_rate": 7.264185037947236e-06, "loss": 83.5564, "step": 4199 }, { "compression_loss": 83.57369995117188, "epoch": 1.52, "learning_rate": 7.258764004336827e-06, "loss": 83.8927, "step": 4200 }, { "compression_loss": 83.53242492675781, "epoch": 1.52, "learning_rate": 7.253342970726419e-06, "loss": 83.8292, "step": 4201 }, { "compression_loss": 84.96507263183594, "epoch": 1.52, "learning_rate": 7.2479219371160104e-06, "loss": 85.6954, "step": 4202 }, { "compression_loss": 84.96894836425781, "epoch": 1.52, "learning_rate": 7.2425009035056025e-06, "loss": 85.3063, "step": 4203 }, { "compression_loss": 84.10121154785156, "epoch": 1.52, "learning_rate": 7.237079869895194e-06, "loss": 84.3656, "step": 4204 }, { "compression_loss": 82.69316101074219, "epoch": 1.52, "learning_rate": 7.231658836284785e-06, "loss": 83.1468, "step": 4205 }, { "compression_loss": 82.81097412109375, "epoch": 1.52, "learning_rate": 7.226237802674376e-06, "loss": 83.2363, "step": 4206 }, { "compression_loss": 84.11637878417969, "epoch": 1.52, "learning_rate": 7.2208167690639685e-06, "loss": 84.571, "step": 4207 }, { "compression_loss": 83.9804458618164, "epoch": 1.52, "learning_rate": 7.21539573545356e-06, "loss": 84.6008, "step": 4208 }, { "compression_loss": 83.10740661621094, "epoch": 1.52, "learning_rate": 7.209974701843152e-06, "loss": 83.684, "step": 4209 }, { "compression_loss": 85.22389221191406, "epoch": 1.52, "learning_rate": 7.204553668232743e-06, "loss": 85.7066, "step": 4210 }, { "compression_loss": 84.77960205078125, "epoch": 1.52, "learning_rate": 7.199132634622335e-06, "loss": 85.5416, "step": 4211 }, { "compression_loss": 85.4298095703125, "epoch": 1.52, "learning_rate": 7.1937116010119265e-06, "loss": 85.7827, "step": 4212 }, { "compression_loss": 84.57968139648438, "epoch": 1.52, "learning_rate": 7.188290567401519e-06, "loss": 85.2412, "step": 4213 }, { "compression_loss": 84.69526672363281, "epoch": 1.52, "learning_rate": 7.18286953379111e-06, "loss": 85.1609, "step": 4214 }, { "compression_loss": 86.50425720214844, "epoch": 1.52, "learning_rate": 7.177448500180702e-06, "loss": 87.1702, "step": 4215 }, { "compression_loss": 83.31285858154297, "epoch": 1.52, "learning_rate": 7.1720274665702925e-06, "loss": 83.8841, "step": 4216 }, { "compression_loss": 82.08328247070312, "epoch": 1.52, "learning_rate": 7.166606432959885e-06, "loss": 82.2997, "step": 4217 }, { "compression_loss": 84.94579315185547, "epoch": 1.52, "learning_rate": 7.161185399349476e-06, "loss": 85.5446, "step": 4218 }, { "compression_loss": 85.27500915527344, "epoch": 1.52, "learning_rate": 7.155764365739068e-06, "loss": 85.77, "step": 4219 }, { "compression_loss": 84.51460266113281, "epoch": 1.53, "learning_rate": 7.150343332128659e-06, "loss": 84.8382, "step": 4220 }, { "compression_loss": 84.24658203125, "epoch": 1.53, "learning_rate": 7.144922298518251e-06, "loss": 84.8704, "step": 4221 }, { "compression_loss": 84.25602722167969, "epoch": 1.53, "learning_rate": 7.139501264907843e-06, "loss": 84.8392, "step": 4222 }, { "compression_loss": 83.64955139160156, "epoch": 1.53, "learning_rate": 7.134080231297435e-06, "loss": 84.1835, "step": 4223 }, { "compression_loss": 85.28619384765625, "epoch": 1.53, "learning_rate": 7.128659197687026e-06, "loss": 85.7262, "step": 4224 }, { "compression_loss": 85.56928253173828, "epoch": 1.53, "learning_rate": 7.123238164076618e-06, "loss": 86.0104, "step": 4225 }, { "compression_loss": 83.66720581054688, "epoch": 1.53, "learning_rate": 7.117817130466209e-06, "loss": 84.0129, "step": 4226 }, { "compression_loss": 84.62701416015625, "epoch": 1.53, "learning_rate": 7.112396096855801e-06, "loss": 85.2192, "step": 4227 }, { "compression_loss": 84.08779907226562, "epoch": 1.53, "learning_rate": 7.106975063245392e-06, "loss": 84.4039, "step": 4228 }, { "compression_loss": 84.78340148925781, "epoch": 1.53, "learning_rate": 7.101554029634984e-06, "loss": 85.4559, "step": 4229 }, { "compression_loss": 83.44871520996094, "epoch": 1.53, "learning_rate": 7.096132996024575e-06, "loss": 83.9832, "step": 4230 }, { "compression_loss": 84.2694320678711, "epoch": 1.53, "learning_rate": 7.0907119624141675e-06, "loss": 84.656, "step": 4231 }, { "compression_loss": 83.93109130859375, "epoch": 1.53, "learning_rate": 7.085290928803759e-06, "loss": 84.5963, "step": 4232 }, { "compression_loss": 85.60037231445312, "epoch": 1.53, "learning_rate": 7.079869895193351e-06, "loss": 86.3048, "step": 4233 }, { "compression_loss": 82.5978012084961, "epoch": 1.53, "learning_rate": 7.074448861582942e-06, "loss": 83.0427, "step": 4234 }, { "compression_loss": 85.05426025390625, "epoch": 1.53, "learning_rate": 7.069027827972534e-06, "loss": 85.6201, "step": 4235 }, { "compression_loss": 85.98454284667969, "epoch": 1.53, "learning_rate": 7.0636067943621255e-06, "loss": 86.9583, "step": 4236 }, { "compression_loss": 83.1739501953125, "epoch": 1.53, "learning_rate": 7.058185760751718e-06, "loss": 83.5078, "step": 4237 }, { "compression_loss": 82.91748046875, "epoch": 1.53, "learning_rate": 7.052764727141308e-06, "loss": 83.2339, "step": 4238 }, { "compression_loss": 83.79327392578125, "epoch": 1.53, "learning_rate": 7.0473436935309e-06, "loss": 84.3544, "step": 4239 }, { "compression_loss": 85.08955383300781, "epoch": 1.53, "learning_rate": 7.0419226599204915e-06, "loss": 85.6044, "step": 4240 }, { "compression_loss": 86.49604797363281, "epoch": 1.53, "learning_rate": 7.036501626310084e-06, "loss": 86.8617, "step": 4241 }, { "compression_loss": 86.0322265625, "epoch": 1.53, "learning_rate": 7.031080592699675e-06, "loss": 86.6274, "step": 4242 }, { "compression_loss": 84.39532470703125, "epoch": 1.53, "learning_rate": 7.025659559089267e-06, "loss": 85.2226, "step": 4243 }, { "compression_loss": 82.14447021484375, "epoch": 1.53, "learning_rate": 7.020238525478858e-06, "loss": 82.5796, "step": 4244 }, { "compression_loss": 85.71812438964844, "epoch": 1.53, "learning_rate": 7.01481749186845e-06, "loss": 85.9376, "step": 4245 }, { "compression_loss": 84.32402038574219, "epoch": 1.53, "learning_rate": 7.009396458258042e-06, "loss": 84.9007, "step": 4246 }, { "compression_loss": 85.81352996826172, "epoch": 1.53, "learning_rate": 7.003975424647634e-06, "loss": 86.2411, "step": 4247 }, { "compression_loss": 83.18635559082031, "epoch": 1.54, "learning_rate": 6.998554391037225e-06, "loss": 83.655, "step": 4248 }, { "compression_loss": 83.34678649902344, "epoch": 1.54, "learning_rate": 6.9931333574268154e-06, "loss": 83.7027, "step": 4249 }, { "compression_loss": 84.06138610839844, "epoch": 1.54, "learning_rate": 6.9877123238164076e-06, "loss": 84.7076, "step": 4250 }, { "epoch": 1.54, "eval_exact_match": 87.17123935666982, "eval_f1": 93.12569508301915, "step": 4250 }, { "compression_loss": 84.37246704101562, "epoch": 1.54, "learning_rate": 6.982291290205999e-06, "loss": 84.9613, "step": 4251 }, { "compression_loss": 84.21487426757812, "epoch": 1.54, "learning_rate": 6.976870256595591e-06, "loss": 84.5258, "step": 4252 }, { "compression_loss": 84.4444351196289, "epoch": 1.54, "learning_rate": 6.971449222985182e-06, "loss": 84.8303, "step": 4253 }, { "compression_loss": 85.79457092285156, "epoch": 1.54, "learning_rate": 6.966028189374774e-06, "loss": 86.3041, "step": 4254 }, { "compression_loss": 84.95614624023438, "epoch": 1.54, "learning_rate": 6.960607155764366e-06, "loss": 85.326, "step": 4255 }, { "compression_loss": 84.678466796875, "epoch": 1.54, "learning_rate": 6.955186122153958e-06, "loss": 84.9591, "step": 4256 }, { "compression_loss": 83.4192886352539, "epoch": 1.54, "learning_rate": 6.949765088543549e-06, "loss": 83.9619, "step": 4257 }, { "compression_loss": 84.39080047607422, "epoch": 1.54, "learning_rate": 6.944344054933141e-06, "loss": 85.0857, "step": 4258 }, { "compression_loss": 83.48342895507812, "epoch": 1.54, "learning_rate": 6.938923021322732e-06, "loss": 83.9622, "step": 4259 }, { "compression_loss": 86.52684020996094, "epoch": 1.54, "learning_rate": 6.933501987712324e-06, "loss": 86.8857, "step": 4260 }, { "compression_loss": 85.39425659179688, "epoch": 1.54, "learning_rate": 6.928080954101915e-06, "loss": 86.3334, "step": 4261 }, { "compression_loss": 86.18074035644531, "epoch": 1.54, "learning_rate": 6.922659920491507e-06, "loss": 86.7277, "step": 4262 }, { "compression_loss": 83.94447326660156, "epoch": 1.54, "learning_rate": 6.917238886881098e-06, "loss": 84.3119, "step": 4263 }, { "compression_loss": 84.2496337890625, "epoch": 1.54, "learning_rate": 6.9118178532706905e-06, "loss": 84.5901, "step": 4264 }, { "compression_loss": 86.57044982910156, "epoch": 1.54, "learning_rate": 6.906396819660282e-06, "loss": 87.3323, "step": 4265 }, { "compression_loss": 84.41173553466797, "epoch": 1.54, "learning_rate": 6.900975786049874e-06, "loss": 85.4731, "step": 4266 }, { "compression_loss": 83.36248779296875, "epoch": 1.54, "learning_rate": 6.895554752439465e-06, "loss": 83.915, "step": 4267 }, { "compression_loss": 84.58380126953125, "epoch": 1.54, "learning_rate": 6.890133718829057e-06, "loss": 84.9835, "step": 4268 }, { "compression_loss": 84.24484252929688, "epoch": 1.54, "learning_rate": 6.8847126852186485e-06, "loss": 84.6169, "step": 4269 }, { "compression_loss": 83.56214904785156, "epoch": 1.54, "learning_rate": 6.87929165160824e-06, "loss": 84.0357, "step": 4270 }, { "compression_loss": 87.32505798339844, "epoch": 1.54, "learning_rate": 6.873870617997831e-06, "loss": 87.646, "step": 4271 }, { "compression_loss": 83.42570495605469, "epoch": 1.54, "learning_rate": 6.868449584387423e-06, "loss": 83.7631, "step": 4272 }, { "compression_loss": 83.59465026855469, "epoch": 1.54, "learning_rate": 6.8630285507770144e-06, "loss": 84.1366, "step": 4273 }, { "compression_loss": 83.37367248535156, "epoch": 1.54, "learning_rate": 6.8576075171666066e-06, "loss": 83.8607, "step": 4274 }, { "compression_loss": 84.6141357421875, "epoch": 1.54, "learning_rate": 6.852186483556198e-06, "loss": 85.057, "step": 4275 }, { "compression_loss": 84.49378967285156, "epoch": 1.55, "learning_rate": 6.84676544994579e-06, "loss": 85.1756, "step": 4276 }, { "compression_loss": 83.71463012695312, "epoch": 1.55, "learning_rate": 6.841344416335381e-06, "loss": 83.9735, "step": 4277 }, { "compression_loss": 83.98036193847656, "epoch": 1.55, "learning_rate": 6.835923382724973e-06, "loss": 84.6661, "step": 4278 }, { "compression_loss": 84.47419738769531, "epoch": 1.55, "learning_rate": 6.830502349114565e-06, "loss": 85.0752, "step": 4279 }, { "compression_loss": 84.56100463867188, "epoch": 1.55, "learning_rate": 6.825081315504157e-06, "loss": 84.8499, "step": 4280 }, { "compression_loss": 84.16417694091797, "epoch": 1.55, "learning_rate": 6.819660281893747e-06, "loss": 84.5028, "step": 4281 }, { "compression_loss": 85.14495849609375, "epoch": 1.55, "learning_rate": 6.814239248283339e-06, "loss": 85.6812, "step": 4282 }, { "compression_loss": 84.75975036621094, "epoch": 1.55, "learning_rate": 6.8088182146729306e-06, "loss": 85.051, "step": 4283 }, { "compression_loss": 82.09970092773438, "epoch": 1.55, "learning_rate": 6.803397181062523e-06, "loss": 82.4113, "step": 4284 }, { "compression_loss": 83.60957336425781, "epoch": 1.55, "learning_rate": 6.797976147452114e-06, "loss": 83.9092, "step": 4285 }, { "compression_loss": 85.90326690673828, "epoch": 1.55, "learning_rate": 6.792555113841706e-06, "loss": 86.392, "step": 4286 }, { "compression_loss": 84.8404769897461, "epoch": 1.55, "learning_rate": 6.787134080231297e-06, "loss": 85.2611, "step": 4287 }, { "compression_loss": 83.099853515625, "epoch": 1.55, "learning_rate": 6.7817130466208895e-06, "loss": 83.609, "step": 4288 }, { "compression_loss": 83.52099609375, "epoch": 1.55, "learning_rate": 6.776292013010481e-06, "loss": 83.9067, "step": 4289 }, { "compression_loss": 84.42080688476562, "epoch": 1.55, "learning_rate": 6.770870979400073e-06, "loss": 85.2473, "step": 4290 }, { "compression_loss": 85.5128173828125, "epoch": 1.55, "learning_rate": 6.765449945789664e-06, "loss": 85.816, "step": 4291 }, { "compression_loss": 83.5681381225586, "epoch": 1.55, "learning_rate": 6.760028912179255e-06, "loss": 83.9581, "step": 4292 }, { "compression_loss": 84.08907318115234, "epoch": 1.55, "learning_rate": 6.754607878568847e-06, "loss": 84.5994, "step": 4293 }, { "compression_loss": 84.26785278320312, "epoch": 1.55, "learning_rate": 6.749186844958439e-06, "loss": 84.6845, "step": 4294 }, { "compression_loss": 83.89270782470703, "epoch": 1.55, "learning_rate": 6.74376581134803e-06, "loss": 84.2106, "step": 4295 }, { "compression_loss": 84.36470031738281, "epoch": 1.55, "learning_rate": 6.738344777737622e-06, "loss": 84.7915, "step": 4296 }, { "compression_loss": 84.88508605957031, "epoch": 1.55, "learning_rate": 6.7329237441272134e-06, "loss": 85.3305, "step": 4297 }, { "compression_loss": 83.19110107421875, "epoch": 1.55, "learning_rate": 6.7275027105168056e-06, "loss": 83.702, "step": 4298 }, { "compression_loss": 83.95147705078125, "epoch": 1.55, "learning_rate": 6.722081676906397e-06, "loss": 84.3934, "step": 4299 }, { "compression_loss": 86.65744018554688, "epoch": 1.55, "learning_rate": 6.716660643295989e-06, "loss": 87.0728, "step": 4300 }, { "compression_loss": 83.78594970703125, "epoch": 1.55, "learning_rate": 6.71123960968558e-06, "loss": 84.2652, "step": 4301 }, { "compression_loss": 82.8739013671875, "epoch": 1.55, "learning_rate": 6.705818576075172e-06, "loss": 83.2876, "step": 4302 }, { "compression_loss": 84.21435546875, "epoch": 1.56, "learning_rate": 6.700397542464763e-06, "loss": 84.7112, "step": 4303 }, { "compression_loss": 84.05743408203125, "epoch": 1.56, "learning_rate": 6.694976508854355e-06, "loss": 84.4948, "step": 4304 }, { "compression_loss": 84.54399108886719, "epoch": 1.56, "learning_rate": 6.689555475243946e-06, "loss": 84.9706, "step": 4305 }, { "compression_loss": 85.25593566894531, "epoch": 1.56, "learning_rate": 6.684134441633538e-06, "loss": 85.7926, "step": 4306 }, { "compression_loss": 84.23869323730469, "epoch": 1.56, "learning_rate": 6.6787134080231295e-06, "loss": 84.8995, "step": 4307 }, { "compression_loss": 84.71075439453125, "epoch": 1.56, "learning_rate": 6.673292374412722e-06, "loss": 85.2335, "step": 4308 }, { "compression_loss": 84.49385070800781, "epoch": 1.56, "learning_rate": 6.667871340802313e-06, "loss": 84.9272, "step": 4309 }, { "compression_loss": 83.2086410522461, "epoch": 1.56, "learning_rate": 6.662450307191905e-06, "loss": 83.6281, "step": 4310 }, { "compression_loss": 84.18250274658203, "epoch": 1.56, "learning_rate": 6.657029273581496e-06, "loss": 84.5403, "step": 4311 }, { "compression_loss": 84.45907592773438, "epoch": 1.56, "learning_rate": 6.6516082399710884e-06, "loss": 85.0638, "step": 4312 }, { "compression_loss": 85.66183471679688, "epoch": 1.56, "learning_rate": 6.64618720636068e-06, "loss": 86.2027, "step": 4313 }, { "compression_loss": 82.3314208984375, "epoch": 1.56, "learning_rate": 6.640766172750271e-06, "loss": 82.9648, "step": 4314 }, { "compression_loss": 83.96451568603516, "epoch": 1.56, "learning_rate": 6.635345139139862e-06, "loss": 84.514, "step": 4315 }, { "compression_loss": 84.67874145507812, "epoch": 1.56, "learning_rate": 6.629924105529454e-06, "loss": 85.1683, "step": 4316 }, { "compression_loss": 84.0499496459961, "epoch": 1.56, "learning_rate": 6.624503071919046e-06, "loss": 84.3325, "step": 4317 }, { "compression_loss": 82.58716583251953, "epoch": 1.56, "learning_rate": 6.619082038308638e-06, "loss": 82.9914, "step": 4318 }, { "compression_loss": 84.25757598876953, "epoch": 1.56, "learning_rate": 6.613661004698229e-06, "loss": 84.54, "step": 4319 }, { "compression_loss": 85.1716537475586, "epoch": 1.56, "learning_rate": 6.608239971087821e-06, "loss": 85.5631, "step": 4320 }, { "compression_loss": 84.54910278320312, "epoch": 1.56, "learning_rate": 6.6028189374774124e-06, "loss": 85.2967, "step": 4321 }, { "compression_loss": 84.30066680908203, "epoch": 1.56, "learning_rate": 6.5973979038670046e-06, "loss": 84.7975, "step": 4322 }, { "compression_loss": 83.3212890625, "epoch": 1.56, "learning_rate": 6.591976870256596e-06, "loss": 83.5607, "step": 4323 }, { "compression_loss": 84.80174255371094, "epoch": 1.56, "learning_rate": 6.586555836646188e-06, "loss": 85.0227, "step": 4324 }, { "compression_loss": 85.22737884521484, "epoch": 1.56, "learning_rate": 6.581134803035778e-06, "loss": 85.6422, "step": 4325 }, { "compression_loss": 84.02705383300781, "epoch": 1.56, "learning_rate": 6.5757137694253705e-06, "loss": 84.4588, "step": 4326 }, { "compression_loss": 85.07199096679688, "epoch": 1.56, "learning_rate": 6.570292735814962e-06, "loss": 85.4696, "step": 4327 }, { "compression_loss": 83.14896392822266, "epoch": 1.56, "learning_rate": 6.564871702204554e-06, "loss": 83.4432, "step": 4328 }, { "compression_loss": 83.18856811523438, "epoch": 1.56, "learning_rate": 6.559450668594145e-06, "loss": 83.6191, "step": 4329 }, { "compression_loss": 83.75977325439453, "epoch": 1.56, "learning_rate": 6.554029634983737e-06, "loss": 84.3772, "step": 4330 }, { "compression_loss": 85.21619415283203, "epoch": 1.57, "learning_rate": 6.5486086013733285e-06, "loss": 85.8749, "step": 4331 }, { "compression_loss": 85.005126953125, "epoch": 1.57, "learning_rate": 6.543187567762921e-06, "loss": 85.2868, "step": 4332 }, { "compression_loss": 84.16862487792969, "epoch": 1.57, "learning_rate": 6.537766534152512e-06, "loss": 84.6962, "step": 4333 }, { "compression_loss": 83.30032348632812, "epoch": 1.57, "learning_rate": 6.532345500542104e-06, "loss": 83.7681, "step": 4334 }, { "compression_loss": 84.5848617553711, "epoch": 1.57, "learning_rate": 6.526924466931695e-06, "loss": 85.0221, "step": 4335 }, { "compression_loss": 85.40020751953125, "epoch": 1.57, "learning_rate": 6.521503433321287e-06, "loss": 85.8705, "step": 4336 }, { "compression_loss": 84.81546020507812, "epoch": 1.57, "learning_rate": 6.516082399710878e-06, "loss": 85.2887, "step": 4337 }, { "compression_loss": 84.58857727050781, "epoch": 1.57, "learning_rate": 6.51066136610047e-06, "loss": 84.9217, "step": 4338 }, { "compression_loss": 85.99188232421875, "epoch": 1.57, "learning_rate": 6.505240332490061e-06, "loss": 86.5735, "step": 4339 }, { "compression_loss": 86.8344497680664, "epoch": 1.57, "learning_rate": 6.499819298879653e-06, "loss": 87.3518, "step": 4340 }, { "compression_loss": 85.65615844726562, "epoch": 1.57, "learning_rate": 6.494398265269245e-06, "loss": 86.2295, "step": 4341 }, { "compression_loss": 84.58001708984375, "epoch": 1.57, "learning_rate": 6.488977231658837e-06, "loss": 85.1025, "step": 4342 }, { "compression_loss": 85.04861450195312, "epoch": 1.57, "learning_rate": 6.483556198048428e-06, "loss": 85.543, "step": 4343 }, { "compression_loss": 84.31163024902344, "epoch": 1.57, "learning_rate": 6.47813516443802e-06, "loss": 84.6439, "step": 4344 }, { "compression_loss": 86.25788879394531, "epoch": 1.57, "learning_rate": 6.4727141308276114e-06, "loss": 86.9527, "step": 4345 }, { "compression_loss": 82.88775634765625, "epoch": 1.57, "learning_rate": 6.4672930972172035e-06, "loss": 83.31, "step": 4346 }, { "compression_loss": 84.10405731201172, "epoch": 1.57, "learning_rate": 6.461872063606794e-06, "loss": 84.9215, "step": 4347 }, { "compression_loss": 83.1607894897461, "epoch": 1.57, "learning_rate": 6.456451029996386e-06, "loss": 83.6169, "step": 4348 }, { "compression_loss": 84.23210144042969, "epoch": 1.57, "learning_rate": 6.451029996385977e-06, "loss": 84.5031, "step": 4349 }, { "compression_loss": 84.1427001953125, "epoch": 1.57, "learning_rate": 6.4456089627755695e-06, "loss": 84.586, "step": 4350 }, { "compression_loss": 82.51959991455078, "epoch": 1.57, "learning_rate": 6.440187929165161e-06, "loss": 82.793, "step": 4351 }, { "compression_loss": 84.05170440673828, "epoch": 1.57, "learning_rate": 6.434766895554753e-06, "loss": 84.5762, "step": 4352 }, { "compression_loss": 82.06014251708984, "epoch": 1.57, "learning_rate": 6.429345861944344e-06, "loss": 82.4009, "step": 4353 }, { "compression_loss": 85.45187377929688, "epoch": 1.57, "learning_rate": 6.423924828333936e-06, "loss": 86.0119, "step": 4354 }, { "compression_loss": 83.34030151367188, "epoch": 1.57, "learning_rate": 6.4185037947235275e-06, "loss": 83.6531, "step": 4355 }, { "compression_loss": 85.1798324584961, "epoch": 1.57, "learning_rate": 6.41308276111312e-06, "loss": 86.0249, "step": 4356 }, { "compression_loss": 84.16606140136719, "epoch": 1.57, "learning_rate": 6.407661727502711e-06, "loss": 84.7301, "step": 4357 }, { "compression_loss": 84.24560546875, "epoch": 1.57, "learning_rate": 6.402240693892302e-06, "loss": 84.7207, "step": 4358 }, { "compression_loss": 86.1479263305664, "epoch": 1.58, "learning_rate": 6.3968196602818935e-06, "loss": 86.4744, "step": 4359 }, { "compression_loss": 84.87982940673828, "epoch": 1.58, "learning_rate": 6.391398626671486e-06, "loss": 85.5826, "step": 4360 }, { "compression_loss": 84.96533966064453, "epoch": 1.58, "learning_rate": 6.385977593061077e-06, "loss": 85.4528, "step": 4361 }, { "compression_loss": 85.38995361328125, "epoch": 1.58, "learning_rate": 6.380556559450669e-06, "loss": 85.9357, "step": 4362 }, { "compression_loss": 84.38207244873047, "epoch": 1.58, "learning_rate": 6.37513552584026e-06, "loss": 84.7329, "step": 4363 }, { "compression_loss": 85.20207977294922, "epoch": 1.58, "learning_rate": 6.369714492229852e-06, "loss": 85.779, "step": 4364 }, { "compression_loss": 85.79154968261719, "epoch": 1.58, "learning_rate": 6.364293458619444e-06, "loss": 86.2824, "step": 4365 }, { "compression_loss": 85.05816650390625, "epoch": 1.58, "learning_rate": 6.358872425009036e-06, "loss": 85.4676, "step": 4366 }, { "compression_loss": 84.66744995117188, "epoch": 1.58, "learning_rate": 6.353451391398627e-06, "loss": 85.0334, "step": 4367 }, { "compression_loss": 84.1911849975586, "epoch": 1.58, "learning_rate": 6.348030357788218e-06, "loss": 84.6912, "step": 4368 }, { "compression_loss": 82.87725830078125, "epoch": 1.58, "learning_rate": 6.34260932417781e-06, "loss": 83.2181, "step": 4369 }, { "compression_loss": 84.3596420288086, "epoch": 1.58, "learning_rate": 6.337188290567402e-06, "loss": 84.8472, "step": 4370 }, { "compression_loss": 82.79164123535156, "epoch": 1.58, "learning_rate": 6.331767256956993e-06, "loss": 83.056, "step": 4371 }, { "compression_loss": 83.236083984375, "epoch": 1.58, "learning_rate": 6.326346223346585e-06, "loss": 83.5593, "step": 4372 }, { "compression_loss": 85.13523864746094, "epoch": 1.58, "learning_rate": 6.320925189736176e-06, "loss": 85.7848, "step": 4373 }, { "compression_loss": 84.32610321044922, "epoch": 1.58, "learning_rate": 6.3155041561257685e-06, "loss": 84.7505, "step": 4374 }, { "compression_loss": 82.625, "epoch": 1.58, "learning_rate": 6.31008312251536e-06, "loss": 82.8876, "step": 4375 }, { "compression_loss": 83.93357849121094, "epoch": 1.58, "learning_rate": 6.304662088904952e-06, "loss": 84.4619, "step": 4376 }, { "compression_loss": 83.38511657714844, "epoch": 1.58, "learning_rate": 6.299241055294543e-06, "loss": 83.8873, "step": 4377 }, { "compression_loss": 84.30220794677734, "epoch": 1.58, "learning_rate": 6.293820021684135e-06, "loss": 84.7832, "step": 4378 }, { "compression_loss": 83.97061157226562, "epoch": 1.58, "learning_rate": 6.288398988073726e-06, "loss": 84.2932, "step": 4379 }, { "compression_loss": 84.39257049560547, "epoch": 1.58, "learning_rate": 6.282977954463318e-06, "loss": 84.9535, "step": 4380 }, { "compression_loss": 85.01484680175781, "epoch": 1.58, "learning_rate": 6.277556920852909e-06, "loss": 85.6758, "step": 4381 }, { "compression_loss": 86.26907348632812, "epoch": 1.58, "learning_rate": 6.272135887242501e-06, "loss": 86.6295, "step": 4382 }, { "compression_loss": 83.27208709716797, "epoch": 1.58, "learning_rate": 6.2667148536320925e-06, "loss": 83.6599, "step": 4383 }, { "compression_loss": 83.24041748046875, "epoch": 1.58, "learning_rate": 6.261293820021685e-06, "loss": 83.4946, "step": 4384 }, { "compression_loss": 85.25054931640625, "epoch": 1.58, "learning_rate": 6.255872786411276e-06, "loss": 85.5309, "step": 4385 }, { "compression_loss": 86.15338134765625, "epoch": 1.59, "learning_rate": 6.250451752800868e-06, "loss": 86.5286, "step": 4386 }, { "compression_loss": 85.91581726074219, "epoch": 1.59, "learning_rate": 6.245030719190459e-06, "loss": 86.2742, "step": 4387 }, { "compression_loss": 83.96426391601562, "epoch": 1.59, "learning_rate": 6.239609685580051e-06, "loss": 84.2927, "step": 4388 }, { "compression_loss": 84.08906555175781, "epoch": 1.59, "learning_rate": 6.234188651969643e-06, "loss": 84.8714, "step": 4389 }, { "compression_loss": 84.10812377929688, "epoch": 1.59, "learning_rate": 6.228767618359234e-06, "loss": 84.5254, "step": 4390 }, { "compression_loss": 83.57270812988281, "epoch": 1.59, "learning_rate": 6.223346584748825e-06, "loss": 83.9264, "step": 4391 }, { "compression_loss": 85.05470275878906, "epoch": 1.59, "learning_rate": 6.217925551138417e-06, "loss": 85.4909, "step": 4392 }, { "compression_loss": 84.74610137939453, "epoch": 1.59, "learning_rate": 6.2125045175280086e-06, "loss": 85.3322, "step": 4393 }, { "compression_loss": 84.82450866699219, "epoch": 1.59, "learning_rate": 6.207083483917601e-06, "loss": 85.2553, "step": 4394 }, { "compression_loss": 84.70452117919922, "epoch": 1.59, "learning_rate": 6.201662450307192e-06, "loss": 85.2773, "step": 4395 }, { "compression_loss": 84.01200866699219, "epoch": 1.59, "learning_rate": 6.196241416696784e-06, "loss": 84.4442, "step": 4396 }, { "compression_loss": 85.4088134765625, "epoch": 1.59, "learning_rate": 6.190820383086375e-06, "loss": 85.789, "step": 4397 }, { "compression_loss": 84.51757049560547, "epoch": 1.59, "learning_rate": 6.1853993494759675e-06, "loss": 84.7713, "step": 4398 }, { "compression_loss": 86.02439880371094, "epoch": 1.59, "learning_rate": 6.179978315865559e-06, "loss": 86.3821, "step": 4399 }, { "compression_loss": 83.95036315917969, "epoch": 1.59, "learning_rate": 6.174557282255151e-06, "loss": 84.3087, "step": 4400 }, { "compression_loss": 83.68955993652344, "epoch": 1.59, "learning_rate": 6.169136248644741e-06, "loss": 84.1056, "step": 4401 }, { "compression_loss": 86.24751281738281, "epoch": 1.59, "learning_rate": 6.163715215034333e-06, "loss": 86.5609, "step": 4402 }, { "compression_loss": 84.11647033691406, "epoch": 1.59, "learning_rate": 6.158294181423925e-06, "loss": 84.4862, "step": 4403 }, { "compression_loss": 84.48848724365234, "epoch": 1.59, "learning_rate": 6.152873147813517e-06, "loss": 84.9152, "step": 4404 }, { "compression_loss": 84.95620727539062, "epoch": 1.59, "learning_rate": 6.147452114203108e-06, "loss": 85.3287, "step": 4405 }, { "compression_loss": 83.92778778076172, "epoch": 1.59, "learning_rate": 6.1420310805927e-06, "loss": 84.3625, "step": 4406 }, { "compression_loss": 84.20362854003906, "epoch": 1.59, "learning_rate": 6.1366100469822915e-06, "loss": 84.6538, "step": 4407 }, { "compression_loss": 83.17412567138672, "epoch": 1.59, "learning_rate": 6.131189013371884e-06, "loss": 83.5863, "step": 4408 }, { "compression_loss": 84.42547607421875, "epoch": 1.59, "learning_rate": 6.125767979761475e-06, "loss": 84.9058, "step": 4409 }, { "compression_loss": 83.75273132324219, "epoch": 1.59, "learning_rate": 6.120346946151067e-06, "loss": 84.0693, "step": 4410 }, { "compression_loss": 84.23875427246094, "epoch": 1.59, "learning_rate": 6.114925912540658e-06, "loss": 84.7078, "step": 4411 }, { "compression_loss": 85.5019760131836, "epoch": 1.59, "learning_rate": 6.1095048789302495e-06, "loss": 86.0626, "step": 4412 }, { "compression_loss": 85.15748596191406, "epoch": 1.59, "learning_rate": 6.104083845319841e-06, "loss": 85.6964, "step": 4413 }, { "compression_loss": 83.58642578125, "epoch": 1.6, "learning_rate": 6.098662811709433e-06, "loss": 84.0128, "step": 4414 }, { "compression_loss": 84.48757934570312, "epoch": 1.6, "learning_rate": 6.093241778099024e-06, "loss": 85.1028, "step": 4415 }, { "compression_loss": 84.53317260742188, "epoch": 1.6, "learning_rate": 6.087820744488616e-06, "loss": 84.9494, "step": 4416 }, { "compression_loss": 83.78404235839844, "epoch": 1.6, "learning_rate": 6.0823997108782076e-06, "loss": 84.3468, "step": 4417 }, { "compression_loss": 85.2468490600586, "epoch": 1.6, "learning_rate": 6.0769786772678e-06, "loss": 85.6456, "step": 4418 }, { "compression_loss": 84.5328369140625, "epoch": 1.6, "learning_rate": 6.071557643657391e-06, "loss": 85.2111, "step": 4419 }, { "compression_loss": 86.48554229736328, "epoch": 1.6, "learning_rate": 6.066136610046983e-06, "loss": 86.8844, "step": 4420 }, { "compression_loss": 83.53033447265625, "epoch": 1.6, "learning_rate": 6.060715576436574e-06, "loss": 83.9938, "step": 4421 }, { "compression_loss": 86.12250518798828, "epoch": 1.6, "learning_rate": 6.0552945428261665e-06, "loss": 86.6973, "step": 4422 }, { "compression_loss": 83.43004608154297, "epoch": 1.6, "learning_rate": 6.049873509215757e-06, "loss": 83.8049, "step": 4423 }, { "compression_loss": 84.80084991455078, "epoch": 1.6, "learning_rate": 6.044452475605349e-06, "loss": 85.449, "step": 4424 }, { "compression_loss": 81.94508361816406, "epoch": 1.6, "learning_rate": 6.03903144199494e-06, "loss": 82.4264, "step": 4425 }, { "compression_loss": 86.16468811035156, "epoch": 1.6, "learning_rate": 6.033610408384532e-06, "loss": 86.6291, "step": 4426 }, { "compression_loss": 85.40706634521484, "epoch": 1.6, "learning_rate": 6.028189374774124e-06, "loss": 85.9192, "step": 4427 }, { "compression_loss": 84.99600982666016, "epoch": 1.6, "learning_rate": 6.022768341163716e-06, "loss": 85.7668, "step": 4428 }, { "compression_loss": 85.56069946289062, "epoch": 1.6, "learning_rate": 6.017347307553307e-06, "loss": 86.1119, "step": 4429 }, { "compression_loss": 84.62446594238281, "epoch": 1.6, "learning_rate": 6.011926273942899e-06, "loss": 84.8419, "step": 4430 }, { "compression_loss": 85.40669250488281, "epoch": 1.6, "learning_rate": 6.0065052403324905e-06, "loss": 85.8476, "step": 4431 }, { "compression_loss": 83.8492660522461, "epoch": 1.6, "learning_rate": 6.0010842067220826e-06, "loss": 84.4746, "step": 4432 }, { "compression_loss": 85.41355895996094, "epoch": 1.6, "learning_rate": 5.995663173111674e-06, "loss": 85.9486, "step": 4433 }, { "compression_loss": 83.29714965820312, "epoch": 1.6, "learning_rate": 5.990242139501264e-06, "loss": 83.5204, "step": 4434 }, { "compression_loss": 83.84888458251953, "epoch": 1.6, "learning_rate": 5.984821105890856e-06, "loss": 84.248, "step": 4435 }, { "compression_loss": 84.2508544921875, "epoch": 1.6, "learning_rate": 5.979400072280448e-06, "loss": 84.7646, "step": 4436 }, { "compression_loss": 83.26252746582031, "epoch": 1.6, "learning_rate": 5.97397903867004e-06, "loss": 83.7836, "step": 4437 }, { "compression_loss": 84.20896911621094, "epoch": 1.6, "learning_rate": 5.968558005059631e-06, "loss": 84.7782, "step": 4438 }, { "compression_loss": 83.85652923583984, "epoch": 1.6, "learning_rate": 5.963136971449223e-06, "loss": 84.2748, "step": 4439 }, { "compression_loss": 82.56119537353516, "epoch": 1.6, "learning_rate": 5.9577159378388144e-06, "loss": 82.9999, "step": 4440 }, { "compression_loss": 82.88704681396484, "epoch": 1.6, "learning_rate": 5.9522949042284066e-06, "loss": 83.3175, "step": 4441 }, { "compression_loss": 81.88257598876953, "epoch": 1.61, "learning_rate": 5.946873870617998e-06, "loss": 82.199, "step": 4442 }, { "compression_loss": 85.3689956665039, "epoch": 1.61, "learning_rate": 5.94145283700759e-06, "loss": 85.8815, "step": 4443 }, { "compression_loss": 84.13499450683594, "epoch": 1.61, "learning_rate": 5.936031803397181e-06, "loss": 84.5194, "step": 4444 }, { "compression_loss": 83.89508819580078, "epoch": 1.61, "learning_rate": 5.9306107697867725e-06, "loss": 84.3042, "step": 4445 }, { "compression_loss": 84.4230728149414, "epoch": 1.61, "learning_rate": 5.925189736176364e-06, "loss": 84.8102, "step": 4446 }, { "compression_loss": 84.43812561035156, "epoch": 1.61, "learning_rate": 5.919768702565956e-06, "loss": 85.078, "step": 4447 }, { "compression_loss": 84.91868591308594, "epoch": 1.61, "learning_rate": 5.914347668955547e-06, "loss": 85.261, "step": 4448 }, { "compression_loss": 83.62246704101562, "epoch": 1.61, "learning_rate": 5.908926635345139e-06, "loss": 83.8883, "step": 4449 }, { "compression_loss": 84.89716339111328, "epoch": 1.61, "learning_rate": 5.9035056017347306e-06, "loss": 85.4487, "step": 4450 }, { "compression_loss": 84.15158081054688, "epoch": 1.61, "learning_rate": 5.898084568124323e-06, "loss": 84.4292, "step": 4451 }, { "compression_loss": 84.59947967529297, "epoch": 1.61, "learning_rate": 5.892663534513914e-06, "loss": 85.1154, "step": 4452 }, { "compression_loss": 84.59846496582031, "epoch": 1.61, "learning_rate": 5.887242500903506e-06, "loss": 85.0791, "step": 4453 }, { "compression_loss": 83.61441802978516, "epoch": 1.61, "learning_rate": 5.881821467293097e-06, "loss": 84.1609, "step": 4454 }, { "compression_loss": 82.28875732421875, "epoch": 1.61, "learning_rate": 5.8764004336826895e-06, "loss": 82.7227, "step": 4455 }, { "compression_loss": 83.47512817382812, "epoch": 1.61, "learning_rate": 5.87097940007228e-06, "loss": 83.8001, "step": 4456 }, { "compression_loss": 84.3064956665039, "epoch": 1.61, "learning_rate": 5.865558366461872e-06, "loss": 84.8812, "step": 4457 }, { "compression_loss": 84.64766693115234, "epoch": 1.61, "learning_rate": 5.860137332851463e-06, "loss": 85.1329, "step": 4458 }, { "compression_loss": 83.05990600585938, "epoch": 1.61, "learning_rate": 5.854716299241055e-06, "loss": 83.4565, "step": 4459 }, { "compression_loss": 83.16302490234375, "epoch": 1.61, "learning_rate": 5.849295265630647e-06, "loss": 83.4425, "step": 4460 }, { "compression_loss": 85.87881469726562, "epoch": 1.61, "learning_rate": 5.843874232020239e-06, "loss": 86.5217, "step": 4461 }, { "compression_loss": 85.4247055053711, "epoch": 1.61, "learning_rate": 5.83845319840983e-06, "loss": 85.8088, "step": 4462 }, { "compression_loss": 85.34129333496094, "epoch": 1.61, "learning_rate": 5.833032164799422e-06, "loss": 85.9138, "step": 4463 }, { "compression_loss": 84.41445922851562, "epoch": 1.61, "learning_rate": 5.8276111311890134e-06, "loss": 84.8129, "step": 4464 }, { "compression_loss": 86.00776672363281, "epoch": 1.61, "learning_rate": 5.8221900975786056e-06, "loss": 86.5544, "step": 4465 }, { "compression_loss": 83.13853454589844, "epoch": 1.61, "learning_rate": 5.816769063968197e-06, "loss": 83.4395, "step": 4466 }, { "compression_loss": 84.55667114257812, "epoch": 1.61, "learning_rate": 5.811348030357788e-06, "loss": 85.0603, "step": 4467 }, { "compression_loss": 84.02098083496094, "epoch": 1.61, "learning_rate": 5.805926996747379e-06, "loss": 84.4312, "step": 4468 }, { "compression_loss": 82.01451110839844, "epoch": 1.62, "learning_rate": 5.8005059631369715e-06, "loss": 82.2498, "step": 4469 }, { "compression_loss": 83.76995849609375, "epoch": 1.62, "learning_rate": 5.795084929526563e-06, "loss": 84.4154, "step": 4470 }, { "compression_loss": 84.0306167602539, "epoch": 1.62, "learning_rate": 5.789663895916155e-06, "loss": 84.648, "step": 4471 }, { "compression_loss": 83.473876953125, "epoch": 1.62, "learning_rate": 5.784242862305746e-06, "loss": 84.047, "step": 4472 }, { "compression_loss": 83.81723022460938, "epoch": 1.62, "learning_rate": 5.778821828695338e-06, "loss": 84.2198, "step": 4473 }, { "compression_loss": 86.08004760742188, "epoch": 1.62, "learning_rate": 5.7734007950849295e-06, "loss": 86.4174, "step": 4474 }, { "compression_loss": 83.80192565917969, "epoch": 1.62, "learning_rate": 5.767979761474522e-06, "loss": 84.3218, "step": 4475 }, { "compression_loss": 84.42925262451172, "epoch": 1.62, "learning_rate": 5.762558727864113e-06, "loss": 84.8029, "step": 4476 }, { "compression_loss": 85.1735610961914, "epoch": 1.62, "learning_rate": 5.757137694253704e-06, "loss": 85.3816, "step": 4477 }, { "compression_loss": 84.66022491455078, "epoch": 1.62, "learning_rate": 5.7517166606432955e-06, "loss": 84.9883, "step": 4478 }, { "compression_loss": 85.5367660522461, "epoch": 1.62, "learning_rate": 5.746295627032888e-06, "loss": 85.8255, "step": 4479 }, { "compression_loss": 84.16128540039062, "epoch": 1.62, "learning_rate": 5.740874593422479e-06, "loss": 84.6406, "step": 4480 }, { "compression_loss": 84.9254379272461, "epoch": 1.62, "learning_rate": 5.735453559812071e-06, "loss": 85.4321, "step": 4481 }, { "compression_loss": 84.66195678710938, "epoch": 1.62, "learning_rate": 5.730032526201662e-06, "loss": 85.052, "step": 4482 }, { "compression_loss": 83.67108154296875, "epoch": 1.62, "learning_rate": 5.724611492591254e-06, "loss": 84.5551, "step": 4483 }, { "compression_loss": 84.22047424316406, "epoch": 1.62, "learning_rate": 5.719190458980846e-06, "loss": 84.9004, "step": 4484 }, { "compression_loss": 83.962158203125, "epoch": 1.62, "learning_rate": 5.713769425370438e-06, "loss": 84.433, "step": 4485 }, { "compression_loss": 84.75350952148438, "epoch": 1.62, "learning_rate": 5.708348391760029e-06, "loss": 85.0212, "step": 4486 }, { "compression_loss": 83.54232788085938, "epoch": 1.62, "learning_rate": 5.702927358149621e-06, "loss": 84.0879, "step": 4487 }, { "compression_loss": 82.19601440429688, "epoch": 1.62, "learning_rate": 5.697506324539212e-06, "loss": 82.6544, "step": 4488 }, { "compression_loss": 85.43292236328125, "epoch": 1.62, "learning_rate": 5.692085290928804e-06, "loss": 86.0323, "step": 4489 }, { "compression_loss": 81.71952819824219, "epoch": 1.62, "learning_rate": 5.686664257318395e-06, "loss": 82.0731, "step": 4490 }, { "compression_loss": 83.4806137084961, "epoch": 1.62, "learning_rate": 5.681243223707987e-06, "loss": 83.8395, "step": 4491 }, { "compression_loss": 83.08865356445312, "epoch": 1.62, "learning_rate": 5.675822190097578e-06, "loss": 83.5847, "step": 4492 }, { "compression_loss": 84.90946960449219, "epoch": 1.62, "learning_rate": 5.6704011564871705e-06, "loss": 85.5774, "step": 4493 }, { "compression_loss": 81.98015594482422, "epoch": 1.62, "learning_rate": 5.664980122876762e-06, "loss": 82.2651, "step": 4494 }, { "compression_loss": 85.17068481445312, "epoch": 1.62, "learning_rate": 5.659559089266354e-06, "loss": 85.6636, "step": 4495 }, { "compression_loss": 84.9537582397461, "epoch": 1.62, "learning_rate": 5.654138055655945e-06, "loss": 85.3382, "step": 4496 }, { "compression_loss": 83.46089935302734, "epoch": 1.63, "learning_rate": 5.648717022045537e-06, "loss": 83.9411, "step": 4497 }, { "compression_loss": 83.1009750366211, "epoch": 1.63, "learning_rate": 5.6432959884351285e-06, "loss": 83.497, "step": 4498 }, { "compression_loss": 83.1397705078125, "epoch": 1.63, "learning_rate": 5.63787495482472e-06, "loss": 83.5185, "step": 4499 }, { "compression_loss": 83.49226379394531, "epoch": 1.63, "learning_rate": 5.632453921214311e-06, "loss": 84.0394, "step": 4500 }, { "epoch": 1.63, "eval_exact_match": 86.8306527909177, "eval_f1": 93.04610605683578, "step": 4500 }, { "compression_loss": 85.03704833984375, "epoch": 1.63, "learning_rate": 5.627032887603903e-06, "loss": 85.2871, "step": 4501 }, { "compression_loss": 84.07638549804688, "epoch": 1.63, "learning_rate": 5.6216118539934945e-06, "loss": 84.6903, "step": 4502 }, { "compression_loss": 84.61043548583984, "epoch": 1.63, "learning_rate": 5.616190820383087e-06, "loss": 84.9528, "step": 4503 }, { "compression_loss": 84.25270080566406, "epoch": 1.63, "learning_rate": 5.610769786772678e-06, "loss": 84.6775, "step": 4504 }, { "compression_loss": 82.67353057861328, "epoch": 1.63, "learning_rate": 5.60534875316227e-06, "loss": 82.8875, "step": 4505 }, { "compression_loss": 84.02488708496094, "epoch": 1.63, "learning_rate": 5.599927719551861e-06, "loss": 84.4016, "step": 4506 }, { "compression_loss": 83.69808959960938, "epoch": 1.63, "learning_rate": 5.594506685941453e-06, "loss": 83.9954, "step": 4507 }, { "compression_loss": 81.33856201171875, "epoch": 1.63, "learning_rate": 5.589085652331045e-06, "loss": 81.7135, "step": 4508 }, { "compression_loss": 84.869873046875, "epoch": 1.63, "learning_rate": 5.583664618720637e-06, "loss": 85.422, "step": 4509 }, { "compression_loss": 86.2708511352539, "epoch": 1.63, "learning_rate": 5.578243585110227e-06, "loss": 86.8535, "step": 4510 }, { "compression_loss": 83.68544006347656, "epoch": 1.63, "learning_rate": 5.572822551499819e-06, "loss": 84.1059, "step": 4511 }, { "compression_loss": 85.59261322021484, "epoch": 1.63, "learning_rate": 5.567401517889411e-06, "loss": 85.9441, "step": 4512 }, { "compression_loss": 84.08993530273438, "epoch": 1.63, "learning_rate": 5.561980484279003e-06, "loss": 84.4268, "step": 4513 }, { "compression_loss": 85.60344696044922, "epoch": 1.63, "learning_rate": 5.556559450668594e-06, "loss": 85.9356, "step": 4514 }, { "compression_loss": 86.04902648925781, "epoch": 1.63, "learning_rate": 5.551138417058186e-06, "loss": 86.3545, "step": 4515 }, { "compression_loss": 84.27214050292969, "epoch": 1.63, "learning_rate": 5.545717383447777e-06, "loss": 84.6248, "step": 4516 }, { "compression_loss": 84.29252624511719, "epoch": 1.63, "learning_rate": 5.5402963498373695e-06, "loss": 84.781, "step": 4517 }, { "compression_loss": 84.78264617919922, "epoch": 1.63, "learning_rate": 5.534875316226961e-06, "loss": 85.3711, "step": 4518 }, { "compression_loss": 83.91012573242188, "epoch": 1.63, "learning_rate": 5.529454282616553e-06, "loss": 84.3062, "step": 4519 }, { "compression_loss": 83.98025512695312, "epoch": 1.63, "learning_rate": 5.524033249006144e-06, "loss": 84.5648, "step": 4520 }, { "compression_loss": 83.80416870117188, "epoch": 1.63, "learning_rate": 5.518612215395735e-06, "loss": 84.1862, "step": 4521 }, { "compression_loss": 82.79188537597656, "epoch": 1.63, "learning_rate": 5.513191181785327e-06, "loss": 83.1559, "step": 4522 }, { "compression_loss": 83.12934875488281, "epoch": 1.63, "learning_rate": 5.507770148174919e-06, "loss": 83.4349, "step": 4523 }, { "compression_loss": 83.64649200439453, "epoch": 1.63, "learning_rate": 5.50234911456451e-06, "loss": 84.1956, "step": 4524 }, { "compression_loss": 86.98468017578125, "epoch": 1.64, "learning_rate": 5.496928080954102e-06, "loss": 87.5488, "step": 4525 }, { "compression_loss": 84.5727767944336, "epoch": 1.64, "learning_rate": 5.4915070473436935e-06, "loss": 84.8795, "step": 4526 }, { "compression_loss": 84.68663024902344, "epoch": 1.64, "learning_rate": 5.486086013733286e-06, "loss": 85.2179, "step": 4527 }, { "compression_loss": 85.04229736328125, "epoch": 1.64, "learning_rate": 5.480664980122877e-06, "loss": 85.3641, "step": 4528 }, { "compression_loss": 83.74241638183594, "epoch": 1.64, "learning_rate": 5.475243946512469e-06, "loss": 84.0977, "step": 4529 }, { "compression_loss": 84.89299011230469, "epoch": 1.64, "learning_rate": 5.46982291290206e-06, "loss": 85.3302, "step": 4530 }, { "compression_loss": 84.65225982666016, "epoch": 1.64, "learning_rate": 5.464401879291652e-06, "loss": 85.4169, "step": 4531 }, { "compression_loss": 83.32814025878906, "epoch": 1.64, "learning_rate": 5.458980845681243e-06, "loss": 84.0153, "step": 4532 }, { "compression_loss": 84.23301696777344, "epoch": 1.64, "learning_rate": 5.453559812070835e-06, "loss": 84.5457, "step": 4533 }, { "compression_loss": 84.50230407714844, "epoch": 1.64, "learning_rate": 5.448138778460426e-06, "loss": 85.0464, "step": 4534 }, { "compression_loss": 84.26825714111328, "epoch": 1.64, "learning_rate": 5.442717744850018e-06, "loss": 84.8331, "step": 4535 }, { "compression_loss": 83.61663055419922, "epoch": 1.64, "learning_rate": 5.43729671123961e-06, "loss": 84.1054, "step": 4536 }, { "compression_loss": 83.46326446533203, "epoch": 1.64, "learning_rate": 5.431875677629202e-06, "loss": 83.6822, "step": 4537 }, { "compression_loss": 84.32089233398438, "epoch": 1.64, "learning_rate": 5.426454644018793e-06, "loss": 84.7424, "step": 4538 }, { "compression_loss": 84.64498901367188, "epoch": 1.64, "learning_rate": 5.421033610408385e-06, "loss": 85.075, "step": 4539 }, { "compression_loss": 83.73770141601562, "epoch": 1.64, "learning_rate": 5.415612576797976e-06, "loss": 84.1046, "step": 4540 }, { "compression_loss": 83.48295593261719, "epoch": 1.64, "learning_rate": 5.4101915431875685e-06, "loss": 83.7603, "step": 4541 }, { "compression_loss": 80.16690826416016, "epoch": 1.64, "learning_rate": 5.40477050957716e-06, "loss": 80.3147, "step": 4542 }, { "compression_loss": 86.43572235107422, "epoch": 1.64, "learning_rate": 5.399349475966751e-06, "loss": 87.0771, "step": 4543 }, { "compression_loss": 83.73062133789062, "epoch": 1.64, "learning_rate": 5.393928442356342e-06, "loss": 84.2071, "step": 4544 }, { "compression_loss": 84.46521759033203, "epoch": 1.64, "learning_rate": 5.388507408745934e-06, "loss": 84.6912, "step": 4545 }, { "compression_loss": 83.53128051757812, "epoch": 1.64, "learning_rate": 5.383086375135526e-06, "loss": 83.7928, "step": 4546 }, { "compression_loss": 85.27873229980469, "epoch": 1.64, "learning_rate": 5.377665341525118e-06, "loss": 85.7172, "step": 4547 }, { "compression_loss": 83.89088439941406, "epoch": 1.64, "learning_rate": 5.372244307914709e-06, "loss": 84.314, "step": 4548 }, { "compression_loss": 85.27840423583984, "epoch": 1.64, "learning_rate": 5.366823274304301e-06, "loss": 85.7675, "step": 4549 }, { "compression_loss": 85.71819305419922, "epoch": 1.64, "learning_rate": 5.3614022406938925e-06, "loss": 86.1286, "step": 4550 }, { "compression_loss": 83.93060302734375, "epoch": 1.64, "learning_rate": 5.355981207083485e-06, "loss": 84.6154, "step": 4551 }, { "compression_loss": 84.4078369140625, "epoch": 1.65, "learning_rate": 5.350560173473076e-06, "loss": 84.8231, "step": 4552 }, { "compression_loss": 85.06936645507812, "epoch": 1.65, "learning_rate": 5.345139139862668e-06, "loss": 85.5966, "step": 4553 }, { "compression_loss": 83.19556427001953, "epoch": 1.65, "learning_rate": 5.339718106252258e-06, "loss": 83.5287, "step": 4554 }, { "compression_loss": 83.48285675048828, "epoch": 1.65, "learning_rate": 5.3342970726418505e-06, "loss": 84.114, "step": 4555 }, { "compression_loss": 85.50411987304688, "epoch": 1.65, "learning_rate": 5.328876039031442e-06, "loss": 85.7676, "step": 4556 }, { "compression_loss": 84.67024230957031, "epoch": 1.65, "learning_rate": 5.323455005421034e-06, "loss": 85.072, "step": 4557 }, { "compression_loss": 84.16261291503906, "epoch": 1.65, "learning_rate": 5.318033971810625e-06, "loss": 84.8344, "step": 4558 }, { "compression_loss": 83.91922760009766, "epoch": 1.65, "learning_rate": 5.312612938200217e-06, "loss": 84.3256, "step": 4559 }, { "compression_loss": 84.4638671875, "epoch": 1.65, "learning_rate": 5.3071919045898086e-06, "loss": 84.809, "step": 4560 }, { "compression_loss": 84.32569885253906, "epoch": 1.65, "learning_rate": 5.301770870979401e-06, "loss": 84.8058, "step": 4561 }, { "compression_loss": 83.24870300292969, "epoch": 1.65, "learning_rate": 5.296349837368992e-06, "loss": 83.6631, "step": 4562 }, { "compression_loss": 85.58285522460938, "epoch": 1.65, "learning_rate": 5.290928803758584e-06, "loss": 86.3577, "step": 4563 }, { "compression_loss": 84.56734466552734, "epoch": 1.65, "learning_rate": 5.285507770148175e-06, "loss": 85.2506, "step": 4564 }, { "compression_loss": 84.07396697998047, "epoch": 1.65, "learning_rate": 5.280086736537767e-06, "loss": 84.4819, "step": 4565 }, { "compression_loss": 85.02131652832031, "epoch": 1.65, "learning_rate": 5.274665702927358e-06, "loss": 85.8336, "step": 4566 }, { "compression_loss": 83.63230895996094, "epoch": 1.65, "learning_rate": 5.26924466931695e-06, "loss": 84.2245, "step": 4567 }, { "compression_loss": 82.60566711425781, "epoch": 1.65, "learning_rate": 5.263823635706541e-06, "loss": 83.1344, "step": 4568 }, { "compression_loss": 85.92538452148438, "epoch": 1.65, "learning_rate": 5.258402602096133e-06, "loss": 86.3009, "step": 4569 }, { "compression_loss": 82.35035705566406, "epoch": 1.65, "learning_rate": 5.252981568485725e-06, "loss": 83.0419, "step": 4570 }, { "compression_loss": 86.00631713867188, "epoch": 1.65, "learning_rate": 5.247560534875317e-06, "loss": 86.5839, "step": 4571 }, { "compression_loss": 85.01715087890625, "epoch": 1.65, "learning_rate": 5.242139501264908e-06, "loss": 85.5049, "step": 4572 }, { "compression_loss": 83.82421875, "epoch": 1.65, "learning_rate": 5.2367184676545e-06, "loss": 84.1816, "step": 4573 }, { "compression_loss": 85.67448425292969, "epoch": 1.65, "learning_rate": 5.2312974340440915e-06, "loss": 86.2129, "step": 4574 }, { "compression_loss": 85.12570190429688, "epoch": 1.65, "learning_rate": 5.225876400433683e-06, "loss": 85.5502, "step": 4575 }, { "compression_loss": 84.0328140258789, "epoch": 1.65, "learning_rate": 5.220455366823274e-06, "loss": 84.6106, "step": 4576 }, { "compression_loss": 83.3507308959961, "epoch": 1.65, "learning_rate": 5.215034333212866e-06, "loss": 83.608, "step": 4577 }, { "compression_loss": 84.45549774169922, "epoch": 1.65, "learning_rate": 5.209613299602457e-06, "loss": 84.972, "step": 4578 }, { "compression_loss": 83.53520965576172, "epoch": 1.65, "learning_rate": 5.2041922659920495e-06, "loss": 83.9265, "step": 4579 }, { "compression_loss": 84.29884338378906, "epoch": 1.66, "learning_rate": 5.198771232381641e-06, "loss": 84.7504, "step": 4580 }, { "compression_loss": 85.27120208740234, "epoch": 1.66, "learning_rate": 5.193350198771233e-06, "loss": 85.5876, "step": 4581 }, { "compression_loss": 83.38606262207031, "epoch": 1.66, "learning_rate": 5.187929165160824e-06, "loss": 83.7502, "step": 4582 }, { "compression_loss": 83.51995849609375, "epoch": 1.66, "learning_rate": 5.182508131550416e-06, "loss": 83.9236, "step": 4583 }, { "compression_loss": 84.18932342529297, "epoch": 1.66, "learning_rate": 5.1770870979400076e-06, "loss": 84.7649, "step": 4584 }, { "compression_loss": 84.1220474243164, "epoch": 1.66, "learning_rate": 5.1716660643296e-06, "loss": 84.7696, "step": 4585 }, { "compression_loss": 84.93924713134766, "epoch": 1.66, "learning_rate": 5.16624503071919e-06, "loss": 85.5308, "step": 4586 }, { "compression_loss": 84.31796264648438, "epoch": 1.66, "learning_rate": 5.160823997108782e-06, "loss": 84.7074, "step": 4587 }, { "compression_loss": 83.9197998046875, "epoch": 1.66, "learning_rate": 5.1554029634983735e-06, "loss": 84.2867, "step": 4588 }, { "compression_loss": 85.04492950439453, "epoch": 1.66, "learning_rate": 5.149981929887966e-06, "loss": 85.5266, "step": 4589 }, { "compression_loss": 85.95570373535156, "epoch": 1.66, "learning_rate": 5.144560896277557e-06, "loss": 86.3552, "step": 4590 }, { "compression_loss": 83.53469848632812, "epoch": 1.66, "learning_rate": 5.139139862667149e-06, "loss": 83.8252, "step": 4591 }, { "compression_loss": 84.60273742675781, "epoch": 1.66, "learning_rate": 5.13371882905674e-06, "loss": 85.2823, "step": 4592 }, { "compression_loss": 84.95268249511719, "epoch": 1.66, "learning_rate": 5.128297795446332e-06, "loss": 85.2661, "step": 4593 }, { "compression_loss": 82.92835235595703, "epoch": 1.66, "learning_rate": 5.122876761835924e-06, "loss": 83.3567, "step": 4594 }, { "compression_loss": 84.28042602539062, "epoch": 1.66, "learning_rate": 5.117455728225516e-06, "loss": 84.7682, "step": 4595 }, { "compression_loss": 83.00334167480469, "epoch": 1.66, "learning_rate": 5.112034694615107e-06, "loss": 83.5928, "step": 4596 }, { "compression_loss": 86.06433868408203, "epoch": 1.66, "learning_rate": 5.106613661004698e-06, "loss": 86.5665, "step": 4597 }, { "compression_loss": 85.6677474975586, "epoch": 1.66, "learning_rate": 5.10119262739429e-06, "loss": 86.2727, "step": 4598 }, { "compression_loss": 84.292236328125, "epoch": 1.66, "learning_rate": 5.095771593783882e-06, "loss": 84.6648, "step": 4599 }, { "compression_loss": 84.41452026367188, "epoch": 1.66, "learning_rate": 5.090350560173473e-06, "loss": 84.7735, "step": 4600 }, { "compression_loss": 82.855224609375, "epoch": 1.66, "learning_rate": 5.084929526563065e-06, "loss": 83.0598, "step": 4601 }, { "compression_loss": 84.93716430664062, "epoch": 1.66, "learning_rate": 5.079508492952656e-06, "loss": 85.3333, "step": 4602 }, { "compression_loss": 82.22898864746094, "epoch": 1.66, "learning_rate": 5.0740874593422485e-06, "loss": 82.5765, "step": 4603 }, { "compression_loss": 85.45199584960938, "epoch": 1.66, "learning_rate": 5.06866642573184e-06, "loss": 85.6203, "step": 4604 }, { "compression_loss": 84.2002182006836, "epoch": 1.66, "learning_rate": 5.063245392121432e-06, "loss": 84.6159, "step": 4605 }, { "compression_loss": 84.8787841796875, "epoch": 1.66, "learning_rate": 5.057824358511023e-06, "loss": 85.5953, "step": 4606 }, { "compression_loss": 84.80404663085938, "epoch": 1.66, "learning_rate": 5.052403324900615e-06, "loss": 85.1657, "step": 4607 }, { "compression_loss": 83.34716796875, "epoch": 1.67, "learning_rate": 5.046982291290206e-06, "loss": 83.6999, "step": 4608 }, { "compression_loss": 82.25489807128906, "epoch": 1.67, "learning_rate": 5.041561257679798e-06, "loss": 82.6144, "step": 4609 }, { "compression_loss": 83.180908203125, "epoch": 1.67, "learning_rate": 5.036140224069389e-06, "loss": 83.8078, "step": 4610 }, { "compression_loss": 83.88737487792969, "epoch": 1.67, "learning_rate": 5.030719190458981e-06, "loss": 84.3962, "step": 4611 }, { "compression_loss": 84.16629028320312, "epoch": 1.67, "learning_rate": 5.0252981568485725e-06, "loss": 84.8386, "step": 4612 }, { "compression_loss": 83.82044982910156, "epoch": 1.67, "learning_rate": 5.019877123238165e-06, "loss": 84.424, "step": 4613 }, { "compression_loss": 84.01032257080078, "epoch": 1.67, "learning_rate": 5.014456089627756e-06, "loss": 84.4919, "step": 4614 }, { "compression_loss": 84.62464904785156, "epoch": 1.67, "learning_rate": 5.009035056017348e-06, "loss": 85.1046, "step": 4615 }, { "compression_loss": 84.79363250732422, "epoch": 1.67, "learning_rate": 5.003614022406939e-06, "loss": 85.1416, "step": 4616 }, { "compression_loss": 85.93853759765625, "epoch": 1.67, "learning_rate": 4.9981929887965306e-06, "loss": 86.2917, "step": 4617 }, { "compression_loss": 86.1389389038086, "epoch": 1.67, "learning_rate": 4.992771955186123e-06, "loss": 86.86, "step": 4618 }, { "compression_loss": 83.8095703125, "epoch": 1.67, "learning_rate": 4.987350921575713e-06, "loss": 84.1682, "step": 4619 }, { "compression_loss": 83.27833557128906, "epoch": 1.67, "learning_rate": 4.981929887965305e-06, "loss": 83.5041, "step": 4620 }, { "compression_loss": 86.62464141845703, "epoch": 1.67, "learning_rate": 4.9765088543548965e-06, "loss": 87.1914, "step": 4621 }, { "compression_loss": 84.29396057128906, "epoch": 1.67, "learning_rate": 4.971087820744489e-06, "loss": 84.8544, "step": 4622 }, { "compression_loss": 84.35529327392578, "epoch": 1.67, "learning_rate": 4.96566678713408e-06, "loss": 84.6853, "step": 4623 }, { "compression_loss": 84.89104461669922, "epoch": 1.67, "learning_rate": 4.960245753523672e-06, "loss": 85.2459, "step": 4624 }, { "compression_loss": 84.21368408203125, "epoch": 1.67, "learning_rate": 4.954824719913263e-06, "loss": 84.8913, "step": 4625 }, { "compression_loss": 85.86959075927734, "epoch": 1.67, "learning_rate": 4.949403686302855e-06, "loss": 86.5162, "step": 4626 }, { "compression_loss": 85.39744567871094, "epoch": 1.67, "learning_rate": 4.943982652692447e-06, "loss": 85.6869, "step": 4627 }, { "compression_loss": 84.01493835449219, "epoch": 1.67, "learning_rate": 4.938561619082039e-06, "loss": 84.6804, "step": 4628 }, { "compression_loss": 85.26992797851562, "epoch": 1.67, "learning_rate": 4.93314058547163e-06, "loss": 85.7098, "step": 4629 }, { "compression_loss": 83.30977630615234, "epoch": 1.67, "learning_rate": 4.927719551861221e-06, "loss": 84.0707, "step": 4630 }, { "compression_loss": 85.48861694335938, "epoch": 1.67, "learning_rate": 4.922298518250813e-06, "loss": 85.987, "step": 4631 }, { "compression_loss": 84.140869140625, "epoch": 1.67, "learning_rate": 4.916877484640405e-06, "loss": 84.7439, "step": 4632 }, { "compression_loss": 82.62840270996094, "epoch": 1.67, "learning_rate": 4.911456451029996e-06, "loss": 82.9981, "step": 4633 }, { "compression_loss": 85.45828247070312, "epoch": 1.67, "learning_rate": 4.906035417419588e-06, "loss": 85.9362, "step": 4634 }, { "compression_loss": 84.22843170166016, "epoch": 1.68, "learning_rate": 4.900614383809179e-06, "loss": 84.6427, "step": 4635 }, { "compression_loss": 84.51242065429688, "epoch": 1.68, "learning_rate": 4.8951933501987715e-06, "loss": 84.9554, "step": 4636 }, { "compression_loss": 84.00154876708984, "epoch": 1.68, "learning_rate": 4.889772316588363e-06, "loss": 84.4684, "step": 4637 }, { "compression_loss": 84.18183898925781, "epoch": 1.68, "learning_rate": 4.884351282977955e-06, "loss": 84.6597, "step": 4638 }, { "compression_loss": 86.31611633300781, "epoch": 1.68, "learning_rate": 4.878930249367546e-06, "loss": 86.7743, "step": 4639 }, { "compression_loss": 85.46321105957031, "epoch": 1.68, "learning_rate": 4.873509215757138e-06, "loss": 85.8337, "step": 4640 }, { "compression_loss": 83.02317810058594, "epoch": 1.68, "learning_rate": 4.868088182146729e-06, "loss": 83.4801, "step": 4641 }, { "compression_loss": 85.92756652832031, "epoch": 1.68, "learning_rate": 4.862667148536321e-06, "loss": 86.3518, "step": 4642 }, { "compression_loss": 84.7008056640625, "epoch": 1.68, "learning_rate": 4.857246114925912e-06, "loss": 85.3159, "step": 4643 }, { "compression_loss": 85.26362609863281, "epoch": 1.68, "learning_rate": 4.851825081315504e-06, "loss": 85.634, "step": 4644 }, { "compression_loss": 84.05354309082031, "epoch": 1.68, "learning_rate": 4.8464040477050955e-06, "loss": 84.4854, "step": 4645 }, { "compression_loss": 84.10597229003906, "epoch": 1.68, "learning_rate": 4.840983014094688e-06, "loss": 84.37, "step": 4646 }, { "compression_loss": 83.23973083496094, "epoch": 1.68, "learning_rate": 4.835561980484279e-06, "loss": 83.7212, "step": 4647 }, { "compression_loss": 82.76280212402344, "epoch": 1.68, "learning_rate": 4.830140946873871e-06, "loss": 83.0648, "step": 4648 }, { "compression_loss": 84.88619995117188, "epoch": 1.68, "learning_rate": 4.824719913263462e-06, "loss": 85.2193, "step": 4649 }, { "compression_loss": 87.81295776367188, "epoch": 1.68, "learning_rate": 4.819298879653054e-06, "loss": 88.3553, "step": 4650 }, { "compression_loss": 84.55364990234375, "epoch": 1.68, "learning_rate": 4.813877846042646e-06, "loss": 84.8695, "step": 4651 }, { "compression_loss": 83.73706817626953, "epoch": 1.68, "learning_rate": 4.808456812432237e-06, "loss": 84.2294, "step": 4652 }, { "compression_loss": 84.46356201171875, "epoch": 1.68, "learning_rate": 4.803035778821828e-06, "loss": 84.9545, "step": 4653 }, { "compression_loss": 84.52176666259766, "epoch": 1.68, "learning_rate": 4.79761474521142e-06, "loss": 85.0858, "step": 4654 }, { "compression_loss": 82.87591552734375, "epoch": 1.68, "learning_rate": 4.792193711601012e-06, "loss": 83.2638, "step": 4655 }, { "compression_loss": 83.83106231689453, "epoch": 1.68, "learning_rate": 4.786772677990604e-06, "loss": 84.1695, "step": 4656 }, { "compression_loss": 82.89830780029297, "epoch": 1.68, "learning_rate": 4.781351644380195e-06, "loss": 83.524, "step": 4657 }, { "compression_loss": 84.56838989257812, "epoch": 1.68, "learning_rate": 4.775930610769787e-06, "loss": 84.9907, "step": 4658 }, { "compression_loss": 86.170654296875, "epoch": 1.68, "learning_rate": 4.770509577159378e-06, "loss": 86.423, "step": 4659 }, { "compression_loss": 83.81865692138672, "epoch": 1.68, "learning_rate": 4.7650885435489705e-06, "loss": 84.0436, "step": 4660 }, { "compression_loss": 84.67786407470703, "epoch": 1.68, "learning_rate": 4.759667509938562e-06, "loss": 85.0571, "step": 4661 }, { "compression_loss": 85.39823913574219, "epoch": 1.68, "learning_rate": 4.754246476328154e-06, "loss": 85.6902, "step": 4662 }, { "compression_loss": 83.02615356445312, "epoch": 1.69, "learning_rate": 4.748825442717744e-06, "loss": 83.404, "step": 4663 }, { "compression_loss": 83.25593566894531, "epoch": 1.69, "learning_rate": 4.743404409107336e-06, "loss": 83.9759, "step": 4664 }, { "compression_loss": 85.44314575195312, "epoch": 1.69, "learning_rate": 4.737983375496928e-06, "loss": 86.2232, "step": 4665 }, { "compression_loss": 84.3984146118164, "epoch": 1.69, "learning_rate": 4.73256234188652e-06, "loss": 84.7969, "step": 4666 }, { "compression_loss": 85.1722412109375, "epoch": 1.69, "learning_rate": 4.727141308276111e-06, "loss": 85.4304, "step": 4667 }, { "compression_loss": 84.67024230957031, "epoch": 1.69, "learning_rate": 4.721720274665703e-06, "loss": 85.1172, "step": 4668 }, { "compression_loss": 86.2514877319336, "epoch": 1.69, "learning_rate": 4.7162992410552945e-06, "loss": 86.8115, "step": 4669 }, { "compression_loss": 83.94650268554688, "epoch": 1.69, "learning_rate": 4.710878207444887e-06, "loss": 84.2779, "step": 4670 }, { "compression_loss": 84.49799346923828, "epoch": 1.69, "learning_rate": 4.705457173834478e-06, "loss": 84.8337, "step": 4671 }, { "compression_loss": 85.15515899658203, "epoch": 1.69, "learning_rate": 4.70003614022407e-06, "loss": 85.4704, "step": 4672 }, { "compression_loss": 84.0567855834961, "epoch": 1.69, "learning_rate": 4.694615106613661e-06, "loss": 84.7269, "step": 4673 }, { "compression_loss": 84.68289947509766, "epoch": 1.69, "learning_rate": 4.6891940730032525e-06, "loss": 85.035, "step": 4674 }, { "compression_loss": 81.94111633300781, "epoch": 1.69, "learning_rate": 4.683773039392844e-06, "loss": 82.3635, "step": 4675 }, { "compression_loss": 82.77164459228516, "epoch": 1.69, "learning_rate": 4.678352005782436e-06, "loss": 83.1094, "step": 4676 }, { "compression_loss": 85.06012725830078, "epoch": 1.69, "learning_rate": 4.672930972172027e-06, "loss": 85.7295, "step": 4677 }, { "compression_loss": 84.08378601074219, "epoch": 1.69, "learning_rate": 4.667509938561619e-06, "loss": 84.5196, "step": 4678 }, { "compression_loss": 84.75318145751953, "epoch": 1.69, "learning_rate": 4.662088904951211e-06, "loss": 85.2075, "step": 4679 }, { "compression_loss": 84.15013122558594, "epoch": 1.69, "learning_rate": 4.656667871340803e-06, "loss": 84.4488, "step": 4680 }, { "compression_loss": 82.01994323730469, "epoch": 1.69, "learning_rate": 4.651246837730394e-06, "loss": 82.4966, "step": 4681 }, { "compression_loss": 84.09535217285156, "epoch": 1.69, "learning_rate": 4.645825804119986e-06, "loss": 84.6601, "step": 4682 }, { "compression_loss": 82.25831604003906, "epoch": 1.69, "learning_rate": 4.640404770509577e-06, "loss": 82.6094, "step": 4683 }, { "compression_loss": 85.36128997802734, "epoch": 1.69, "learning_rate": 4.634983736899169e-06, "loss": 85.9786, "step": 4684 }, { "compression_loss": 86.31708526611328, "epoch": 1.69, "learning_rate": 4.62956270328876e-06, "loss": 86.979, "step": 4685 }, { "compression_loss": 83.54177856445312, "epoch": 1.69, "learning_rate": 4.624141669678352e-06, "loss": 83.8795, "step": 4686 }, { "compression_loss": 84.69700622558594, "epoch": 1.69, "learning_rate": 4.618720636067943e-06, "loss": 85.0518, "step": 4687 }, { "compression_loss": 84.870361328125, "epoch": 1.69, "learning_rate": 4.613299602457535e-06, "loss": 85.4223, "step": 4688 }, { "compression_loss": 84.60578918457031, "epoch": 1.69, "learning_rate": 4.607878568847127e-06, "loss": 85.2873, "step": 4689 }, { "compression_loss": 84.24470520019531, "epoch": 1.69, "learning_rate": 4.602457535236719e-06, "loss": 84.7539, "step": 4690 }, { "compression_loss": 84.19108581542969, "epoch": 1.7, "learning_rate": 4.59703650162631e-06, "loss": 84.4365, "step": 4691 }, { "compression_loss": 83.93917083740234, "epoch": 1.7, "learning_rate": 4.591615468015902e-06, "loss": 84.6067, "step": 4692 }, { "compression_loss": 84.40070343017578, "epoch": 1.7, "learning_rate": 4.5861944344054935e-06, "loss": 84.8529, "step": 4693 }, { "compression_loss": 85.12678527832031, "epoch": 1.7, "learning_rate": 4.580773400795086e-06, "loss": 85.479, "step": 4694 }, { "compression_loss": 83.86681365966797, "epoch": 1.7, "learning_rate": 4.575352367184676e-06, "loss": 84.2303, "step": 4695 }, { "compression_loss": 83.87916564941406, "epoch": 1.7, "learning_rate": 4.569931333574268e-06, "loss": 84.2081, "step": 4696 }, { "compression_loss": 83.41020202636719, "epoch": 1.7, "learning_rate": 4.564510299963859e-06, "loss": 83.7459, "step": 4697 }, { "compression_loss": 84.40946960449219, "epoch": 1.7, "learning_rate": 4.5590892663534515e-06, "loss": 84.6488, "step": 4698 }, { "compression_loss": 83.37504577636719, "epoch": 1.7, "learning_rate": 4.553668232743043e-06, "loss": 83.6912, "step": 4699 }, { "compression_loss": 83.41012573242188, "epoch": 1.7, "learning_rate": 4.548247199132635e-06, "loss": 83.7669, "step": 4700 }, { "compression_loss": 82.14036560058594, "epoch": 1.7, "learning_rate": 4.542826165522226e-06, "loss": 82.4862, "step": 4701 }, { "compression_loss": 84.82402038574219, "epoch": 1.7, "learning_rate": 4.537405131911818e-06, "loss": 85.2694, "step": 4702 }, { "compression_loss": 84.86920928955078, "epoch": 1.7, "learning_rate": 4.53198409830141e-06, "loss": 85.3126, "step": 4703 }, { "compression_loss": 84.74264526367188, "epoch": 1.7, "learning_rate": 4.526563064691002e-06, "loss": 85.1603, "step": 4704 }, { "compression_loss": 83.147705078125, "epoch": 1.7, "learning_rate": 4.521142031080593e-06, "loss": 83.441, "step": 4705 }, { "compression_loss": 86.72520446777344, "epoch": 1.7, "learning_rate": 4.515720997470184e-06, "loss": 87.14, "step": 4706 }, { "compression_loss": 84.9607925415039, "epoch": 1.7, "learning_rate": 4.5102999638597755e-06, "loss": 85.5179, "step": 4707 }, { "compression_loss": 84.59970092773438, "epoch": 1.7, "learning_rate": 4.504878930249368e-06, "loss": 84.8851, "step": 4708 }, { "compression_loss": 86.21480560302734, "epoch": 1.7, "learning_rate": 4.499457896638959e-06, "loss": 86.6574, "step": 4709 }, { "compression_loss": 83.23676300048828, "epoch": 1.7, "learning_rate": 4.494036863028551e-06, "loss": 83.6523, "step": 4710 }, { "compression_loss": 84.16630554199219, "epoch": 1.7, "learning_rate": 4.488615829418142e-06, "loss": 84.6575, "step": 4711 }, { "compression_loss": 85.70159912109375, "epoch": 1.7, "learning_rate": 4.483194795807734e-06, "loss": 86.0908, "step": 4712 }, { "compression_loss": 84.73641967773438, "epoch": 1.7, "learning_rate": 4.477773762197326e-06, "loss": 84.9331, "step": 4713 }, { "compression_loss": 84.09339904785156, "epoch": 1.7, "learning_rate": 4.472352728586918e-06, "loss": 84.6013, "step": 4714 }, { "compression_loss": 82.50173950195312, "epoch": 1.7, "learning_rate": 4.466931694976509e-06, "loss": 82.7878, "step": 4715 }, { "compression_loss": 84.69560241699219, "epoch": 1.7, "learning_rate": 4.461510661366101e-06, "loss": 85.109, "step": 4716 }, { "compression_loss": 82.77528381347656, "epoch": 1.7, "learning_rate": 4.456089627755692e-06, "loss": 83.0191, "step": 4717 }, { "compression_loss": 85.2891845703125, "epoch": 1.71, "learning_rate": 4.450668594145284e-06, "loss": 85.7555, "step": 4718 }, { "compression_loss": 82.48625183105469, "epoch": 1.71, "learning_rate": 4.445247560534875e-06, "loss": 82.8322, "step": 4719 }, { "compression_loss": 84.13972473144531, "epoch": 1.71, "learning_rate": 4.439826526924467e-06, "loss": 84.688, "step": 4720 }, { "compression_loss": 83.74716186523438, "epoch": 1.71, "learning_rate": 4.434405493314058e-06, "loss": 84.0725, "step": 4721 }, { "compression_loss": 84.23812866210938, "epoch": 1.71, "learning_rate": 4.4289844597036505e-06, "loss": 84.8344, "step": 4722 }, { "compression_loss": 84.01078796386719, "epoch": 1.71, "learning_rate": 4.423563426093242e-06, "loss": 84.3488, "step": 4723 }, { "compression_loss": 83.60574340820312, "epoch": 1.71, "learning_rate": 4.418142392482834e-06, "loss": 84.0099, "step": 4724 }, { "compression_loss": 85.8226318359375, "epoch": 1.71, "learning_rate": 4.412721358872425e-06, "loss": 86.3477, "step": 4725 }, { "compression_loss": 82.78450012207031, "epoch": 1.71, "learning_rate": 4.407300325262017e-06, "loss": 82.996, "step": 4726 }, { "compression_loss": 83.22823333740234, "epoch": 1.71, "learning_rate": 4.4018792916516086e-06, "loss": 83.5053, "step": 4727 }, { "compression_loss": 85.12444305419922, "epoch": 1.71, "learning_rate": 4.3964582580412e-06, "loss": 85.4902, "step": 4728 }, { "compression_loss": 84.8774642944336, "epoch": 1.71, "learning_rate": 4.391037224430791e-06, "loss": 85.3223, "step": 4729 }, { "compression_loss": 83.50605773925781, "epoch": 1.71, "learning_rate": 4.385616190820383e-06, "loss": 83.8938, "step": 4730 }, { "compression_loss": 83.34027099609375, "epoch": 1.71, "learning_rate": 4.3801951572099745e-06, "loss": 83.7509, "step": 4731 }, { "compression_loss": 83.96371459960938, "epoch": 1.71, "learning_rate": 4.374774123599567e-06, "loss": 84.2833, "step": 4732 }, { "compression_loss": 84.6282958984375, "epoch": 1.71, "learning_rate": 4.369353089989158e-06, "loss": 84.8922, "step": 4733 }, { "compression_loss": 84.19281005859375, "epoch": 1.71, "learning_rate": 4.36393205637875e-06, "loss": 84.6927, "step": 4734 }, { "compression_loss": 84.14250183105469, "epoch": 1.71, "learning_rate": 4.358511022768341e-06, "loss": 84.612, "step": 4735 }, { "compression_loss": 83.74836730957031, "epoch": 1.71, "learning_rate": 4.353089989157933e-06, "loss": 84.0136, "step": 4736 }, { "compression_loss": 84.07868957519531, "epoch": 1.71, "learning_rate": 4.347668955547525e-06, "loss": 84.5157, "step": 4737 }, { "compression_loss": 83.96209716796875, "epoch": 1.71, "learning_rate": 4.342247921937117e-06, "loss": 84.4164, "step": 4738 }, { "compression_loss": 83.53910827636719, "epoch": 1.71, "learning_rate": 4.336826888326707e-06, "loss": 84.106, "step": 4739 }, { "compression_loss": 84.43408203125, "epoch": 1.71, "learning_rate": 4.331405854716299e-06, "loss": 84.7023, "step": 4740 }, { "compression_loss": 87.10944366455078, "epoch": 1.71, "learning_rate": 4.325984821105891e-06, "loss": 87.6138, "step": 4741 }, { "compression_loss": 83.78767395019531, "epoch": 1.71, "learning_rate": 4.320563787495483e-06, "loss": 84.2049, "step": 4742 }, { "compression_loss": 84.37434387207031, "epoch": 1.71, "learning_rate": 4.315142753885074e-06, "loss": 84.7542, "step": 4743 }, { "compression_loss": 84.19564819335938, "epoch": 1.71, "learning_rate": 4.309721720274666e-06, "loss": 84.6416, "step": 4744 }, { "compression_loss": 81.95895385742188, "epoch": 1.71, "learning_rate": 4.304300686664257e-06, "loss": 82.3329, "step": 4745 }, { "compression_loss": 84.15055847167969, "epoch": 1.72, "learning_rate": 4.2988796530538495e-06, "loss": 84.582, "step": 4746 }, { "compression_loss": 85.14801788330078, "epoch": 1.72, "learning_rate": 4.293458619443441e-06, "loss": 85.5825, "step": 4747 }, { "compression_loss": 84.07235717773438, "epoch": 1.72, "learning_rate": 4.288037585833033e-06, "loss": 84.5467, "step": 4748 }, { "compression_loss": 83.78080749511719, "epoch": 1.72, "learning_rate": 4.282616552222624e-06, "loss": 84.3322, "step": 4749 }, { "compression_loss": 85.99089050292969, "epoch": 1.72, "learning_rate": 4.2771955186122154e-06, "loss": 86.4868, "step": 4750 }, { "epoch": 1.72, "eval_exact_match": 86.80227057710502, "eval_f1": 92.99721219929512, "step": 4750 }, { "compression_loss": 83.54489135742188, "epoch": 1.72, "learning_rate": 4.271774485001807e-06, "loss": 84.0248, "step": 4751 }, { "compression_loss": 83.26707458496094, "epoch": 1.72, "learning_rate": 4.266353451391399e-06, "loss": 83.7586, "step": 4752 }, { "compression_loss": 84.1233139038086, "epoch": 1.72, "learning_rate": 4.26093241778099e-06, "loss": 84.5434, "step": 4753 }, { "compression_loss": 84.29887390136719, "epoch": 1.72, "learning_rate": 4.255511384170582e-06, "loss": 84.8816, "step": 4754 }, { "compression_loss": 84.58505249023438, "epoch": 1.72, "learning_rate": 4.2500903505601735e-06, "loss": 85.1381, "step": 4755 }, { "compression_loss": 84.49358367919922, "epoch": 1.72, "learning_rate": 4.244669316949766e-06, "loss": 84.9703, "step": 4756 }, { "compression_loss": 83.89286804199219, "epoch": 1.72, "learning_rate": 4.239248283339357e-06, "loss": 84.1159, "step": 4757 }, { "compression_loss": 84.92727661132812, "epoch": 1.72, "learning_rate": 4.233827249728949e-06, "loss": 85.4422, "step": 4758 }, { "compression_loss": 82.69305419921875, "epoch": 1.72, "learning_rate": 4.22840621611854e-06, "loss": 83.2692, "step": 4759 }, { "compression_loss": 83.64778137207031, "epoch": 1.72, "learning_rate": 4.222985182508132e-06, "loss": 83.9725, "step": 4760 }, { "compression_loss": 83.79583740234375, "epoch": 1.72, "learning_rate": 4.217564148897723e-06, "loss": 84.1262, "step": 4761 }, { "compression_loss": 83.56455993652344, "epoch": 1.72, "learning_rate": 4.212143115287315e-06, "loss": 83.8712, "step": 4762 }, { "compression_loss": 84.89485168457031, "epoch": 1.72, "learning_rate": 4.206722081676906e-06, "loss": 85.2409, "step": 4763 }, { "compression_loss": 85.802978515625, "epoch": 1.72, "learning_rate": 4.201301048066498e-06, "loss": 86.2052, "step": 4764 }, { "compression_loss": 82.76862335205078, "epoch": 1.72, "learning_rate": 4.19588001445609e-06, "loss": 83.4149, "step": 4765 }, { "compression_loss": 82.44746398925781, "epoch": 1.72, "learning_rate": 4.190458980845682e-06, "loss": 82.7471, "step": 4766 }, { "compression_loss": 85.39002227783203, "epoch": 1.72, "learning_rate": 4.185037947235273e-06, "loss": 85.9338, "step": 4767 }, { "compression_loss": 84.65354919433594, "epoch": 1.72, "learning_rate": 4.179616913624865e-06, "loss": 85.0868, "step": 4768 }, { "compression_loss": 86.0341796875, "epoch": 1.72, "learning_rate": 4.174195880014456e-06, "loss": 86.557, "step": 4769 }, { "compression_loss": 84.93050384521484, "epoch": 1.72, "learning_rate": 4.1687748464040485e-06, "loss": 85.2036, "step": 4770 }, { "compression_loss": 85.0279541015625, "epoch": 1.72, "learning_rate": 4.16335381279364e-06, "loss": 85.5801, "step": 4771 }, { "compression_loss": 87.18782043457031, "epoch": 1.72, "learning_rate": 4.157932779183231e-06, "loss": 87.4784, "step": 4772 }, { "compression_loss": 84.69825744628906, "epoch": 1.72, "learning_rate": 4.152511745572822e-06, "loss": 85.4931, "step": 4773 }, { "compression_loss": 82.59623718261719, "epoch": 1.73, "learning_rate": 4.1470907119624144e-06, "loss": 82.9255, "step": 4774 }, { "compression_loss": 84.89976501464844, "epoch": 1.73, "learning_rate": 4.141669678352006e-06, "loss": 85.1225, "step": 4775 }, { "compression_loss": 84.22711181640625, "epoch": 1.73, "learning_rate": 4.136248644741598e-06, "loss": 84.5504, "step": 4776 }, { "compression_loss": 85.92288208007812, "epoch": 1.73, "learning_rate": 4.130827611131189e-06, "loss": 86.5106, "step": 4777 }, { "compression_loss": 85.68038177490234, "epoch": 1.73, "learning_rate": 4.125406577520781e-06, "loss": 86.297, "step": 4778 }, { "compression_loss": 83.87895202636719, "epoch": 1.73, "learning_rate": 4.1199855439103725e-06, "loss": 84.4588, "step": 4779 }, { "compression_loss": 84.57766723632812, "epoch": 1.73, "learning_rate": 4.114564510299965e-06, "loss": 84.7695, "step": 4780 }, { "compression_loss": 83.4533462524414, "epoch": 1.73, "learning_rate": 4.109143476689556e-06, "loss": 83.9276, "step": 4781 }, { "compression_loss": 83.31437683105469, "epoch": 1.73, "learning_rate": 4.103722443079147e-06, "loss": 83.6015, "step": 4782 }, { "compression_loss": 82.88479614257812, "epoch": 1.73, "learning_rate": 4.0983014094687384e-06, "loss": 83.243, "step": 4783 }, { "compression_loss": 85.29518127441406, "epoch": 1.73, "learning_rate": 4.0928803758583306e-06, "loss": 85.8202, "step": 4784 }, { "compression_loss": 84.17626953125, "epoch": 1.73, "learning_rate": 4.087459342247922e-06, "loss": 84.5218, "step": 4785 }, { "compression_loss": 83.40519714355469, "epoch": 1.73, "learning_rate": 4.082038308637514e-06, "loss": 83.8646, "step": 4786 }, { "compression_loss": 85.31529235839844, "epoch": 1.73, "learning_rate": 4.076617275027105e-06, "loss": 86.2054, "step": 4787 }, { "compression_loss": 83.63915252685547, "epoch": 1.73, "learning_rate": 4.071196241416697e-06, "loss": 84.0133, "step": 4788 }, { "compression_loss": 82.98318481445312, "epoch": 1.73, "learning_rate": 4.065775207806289e-06, "loss": 83.2601, "step": 4789 }, { "compression_loss": 84.42015075683594, "epoch": 1.73, "learning_rate": 4.060354174195881e-06, "loss": 84.6838, "step": 4790 }, { "compression_loss": 85.07064819335938, "epoch": 1.73, "learning_rate": 4.054933140585472e-06, "loss": 85.5566, "step": 4791 }, { "compression_loss": 85.91633605957031, "epoch": 1.73, "learning_rate": 4.049512106975064e-06, "loss": 86.1914, "step": 4792 }, { "compression_loss": 85.63693237304688, "epoch": 1.73, "learning_rate": 4.0440910733646545e-06, "loss": 86.4651, "step": 4793 }, { "compression_loss": 85.5908432006836, "epoch": 1.73, "learning_rate": 4.038670039754247e-06, "loss": 86.2161, "step": 4794 }, { "compression_loss": 84.37847900390625, "epoch": 1.73, "learning_rate": 4.033249006143838e-06, "loss": 84.7699, "step": 4795 }, { "compression_loss": 84.48114013671875, "epoch": 1.73, "learning_rate": 4.02782797253343e-06, "loss": 85.1366, "step": 4796 }, { "compression_loss": 85.03009796142578, "epoch": 1.73, "learning_rate": 4.022406938923021e-06, "loss": 85.5077, "step": 4797 }, { "compression_loss": 84.0054931640625, "epoch": 1.73, "learning_rate": 4.0169859053126134e-06, "loss": 84.3408, "step": 4798 }, { "compression_loss": 83.29653930664062, "epoch": 1.73, "learning_rate": 4.011564871702205e-06, "loss": 83.8731, "step": 4799 }, { "compression_loss": 84.2811279296875, "epoch": 1.73, "learning_rate": 4.006143838091797e-06, "loss": 84.7356, "step": 4800 }, { "compression_loss": 86.24737548828125, "epoch": 1.74, "learning_rate": 4.000722804481388e-06, "loss": 86.5462, "step": 4801 }, { "compression_loss": 82.7101058959961, "epoch": 1.74, "learning_rate": 3.995301770870979e-06, "loss": 83.2302, "step": 4802 }, { "compression_loss": 84.64677429199219, "epoch": 1.74, "learning_rate": 3.9898807372605715e-06, "loss": 85.2907, "step": 4803 }, { "compression_loss": 85.04749298095703, "epoch": 1.74, "learning_rate": 3.984459703650162e-06, "loss": 85.7313, "step": 4804 }, { "compression_loss": 84.58692169189453, "epoch": 1.74, "learning_rate": 3.979038670039754e-06, "loss": 84.9428, "step": 4805 }, { "compression_loss": 86.01240539550781, "epoch": 1.74, "learning_rate": 3.973617636429345e-06, "loss": 86.4507, "step": 4806 }, { "compression_loss": 84.39657592773438, "epoch": 1.74, "learning_rate": 3.9681966028189374e-06, "loss": 84.7818, "step": 4807 }, { "compression_loss": 83.39445495605469, "epoch": 1.74, "learning_rate": 3.962775569208529e-06, "loss": 83.7886, "step": 4808 }, { "compression_loss": 83.28352355957031, "epoch": 1.74, "learning_rate": 3.957354535598121e-06, "loss": 83.5163, "step": 4809 }, { "compression_loss": 84.2302017211914, "epoch": 1.74, "learning_rate": 3.951933501987712e-06, "loss": 84.9543, "step": 4810 }, { "compression_loss": 84.3680419921875, "epoch": 1.74, "learning_rate": 3.946512468377304e-06, "loss": 84.7461, "step": 4811 }, { "compression_loss": 85.12946319580078, "epoch": 1.74, "learning_rate": 3.9410914347668955e-06, "loss": 85.6167, "step": 4812 }, { "compression_loss": 85.15052032470703, "epoch": 1.74, "learning_rate": 3.935670401156488e-06, "loss": 85.4763, "step": 4813 }, { "compression_loss": 85.43580627441406, "epoch": 1.74, "learning_rate": 3.930249367546079e-06, "loss": 85.9366, "step": 4814 }, { "compression_loss": 85.2403564453125, "epoch": 1.74, "learning_rate": 3.92482833393567e-06, "loss": 85.8213, "step": 4815 }, { "compression_loss": 85.82389831542969, "epoch": 1.74, "learning_rate": 3.919407300325261e-06, "loss": 86.5001, "step": 4816 }, { "compression_loss": 83.8302001953125, "epoch": 1.74, "learning_rate": 3.9139862667148535e-06, "loss": 84.6824, "step": 4817 }, { "compression_loss": 85.28693389892578, "epoch": 1.74, "learning_rate": 3.908565233104445e-06, "loss": 85.7293, "step": 4818 }, { "compression_loss": 84.67855072021484, "epoch": 1.74, "learning_rate": 3.903144199494037e-06, "loss": 85.1735, "step": 4819 }, { "compression_loss": 84.71288299560547, "epoch": 1.74, "learning_rate": 3.897723165883628e-06, "loss": 85.1042, "step": 4820 }, { "compression_loss": 85.58149719238281, "epoch": 1.74, "learning_rate": 3.89230213227322e-06, "loss": 86.0015, "step": 4821 }, { "compression_loss": 85.15618133544922, "epoch": 1.74, "learning_rate": 3.886881098662812e-06, "loss": 86.2481, "step": 4822 }, { "compression_loss": 84.49093627929688, "epoch": 1.74, "learning_rate": 3.881460065052404e-06, "loss": 84.7812, "step": 4823 }, { "compression_loss": 84.98591613769531, "epoch": 1.74, "learning_rate": 3.876039031441995e-06, "loss": 85.5561, "step": 4824 }, { "compression_loss": 85.0373764038086, "epoch": 1.74, "learning_rate": 3.870617997831587e-06, "loss": 85.6632, "step": 4825 }, { "compression_loss": 85.91403198242188, "epoch": 1.74, "learning_rate": 3.8651969642211775e-06, "loss": 86.3592, "step": 4826 }, { "compression_loss": 84.55520629882812, "epoch": 1.74, "learning_rate": 3.85977593061077e-06, "loss": 85.1462, "step": 4827 }, { "compression_loss": 84.25028991699219, "epoch": 1.74, "learning_rate": 3.854354897000361e-06, "loss": 84.5123, "step": 4828 }, { "compression_loss": 84.05672454833984, "epoch": 1.75, "learning_rate": 3.848933863389953e-06, "loss": 84.3497, "step": 4829 }, { "compression_loss": 83.33908081054688, "epoch": 1.75, "learning_rate": 3.843512829779544e-06, "loss": 83.8013, "step": 4830 }, { "compression_loss": 83.77081298828125, "epoch": 1.75, "learning_rate": 3.838091796169136e-06, "loss": 84.2255, "step": 4831 }, { "compression_loss": 84.50975036621094, "epoch": 1.75, "learning_rate": 3.832670762558728e-06, "loss": 85.0026, "step": 4832 }, { "compression_loss": 83.01042175292969, "epoch": 1.75, "learning_rate": 3.82724972894832e-06, "loss": 83.4332, "step": 4833 }, { "compression_loss": 83.89034271240234, "epoch": 1.75, "learning_rate": 3.821828695337911e-06, "loss": 84.3919, "step": 4834 }, { "compression_loss": 83.95402526855469, "epoch": 1.75, "learning_rate": 3.816407661727503e-06, "loss": 84.3262, "step": 4835 }, { "compression_loss": 86.59982299804688, "epoch": 1.75, "learning_rate": 3.810986628117094e-06, "loss": 86.9785, "step": 4836 }, { "compression_loss": 84.39083862304688, "epoch": 1.75, "learning_rate": 3.805565594506686e-06, "loss": 84.9376, "step": 4837 }, { "compression_loss": 84.46060180664062, "epoch": 1.75, "learning_rate": 3.8001445608962774e-06, "loss": 84.8907, "step": 4838 }, { "compression_loss": 83.38675689697266, "epoch": 1.75, "learning_rate": 3.7947235272858696e-06, "loss": 83.6984, "step": 4839 }, { "compression_loss": 83.5687255859375, "epoch": 1.75, "learning_rate": 3.7893024936754604e-06, "loss": 83.7894, "step": 4840 }, { "compression_loss": 83.85179901123047, "epoch": 1.75, "learning_rate": 3.7838814600650525e-06, "loss": 84.3856, "step": 4841 }, { "compression_loss": 85.62089538574219, "epoch": 1.75, "learning_rate": 3.778460426454644e-06, "loss": 86.3229, "step": 4842 }, { "compression_loss": 84.91302490234375, "epoch": 1.75, "learning_rate": 3.773039392844236e-06, "loss": 85.3073, "step": 4843 }, { "compression_loss": 84.79898834228516, "epoch": 1.75, "learning_rate": 3.767618359233827e-06, "loss": 85.3233, "step": 4844 }, { "compression_loss": 83.88037109375, "epoch": 1.75, "learning_rate": 3.762197325623419e-06, "loss": 84.3283, "step": 4845 }, { "compression_loss": 83.36732482910156, "epoch": 1.75, "learning_rate": 3.75677629201301e-06, "loss": 83.7591, "step": 4846 }, { "compression_loss": 84.99640655517578, "epoch": 1.75, "learning_rate": 3.7513552584026023e-06, "loss": 85.4066, "step": 4847 }, { "compression_loss": 83.62551879882812, "epoch": 1.75, "learning_rate": 3.745934224792194e-06, "loss": 84.1024, "step": 4848 }, { "compression_loss": 83.62303924560547, "epoch": 1.75, "learning_rate": 3.7405131911817857e-06, "loss": 84.1331, "step": 4849 }, { "compression_loss": 84.95481872558594, "epoch": 1.75, "learning_rate": 3.735092157571377e-06, "loss": 85.448, "step": 4850 }, { "compression_loss": 84.07666778564453, "epoch": 1.75, "learning_rate": 3.7296711239609686e-06, "loss": 84.7143, "step": 4851 }, { "compression_loss": 83.5962142944336, "epoch": 1.75, "learning_rate": 3.7242500903505603e-06, "loss": 84.1836, "step": 4852 }, { "compression_loss": 81.15711975097656, "epoch": 1.75, "learning_rate": 3.718829056740152e-06, "loss": 81.4477, "step": 4853 }, { "compression_loss": 85.56214904785156, "epoch": 1.75, "learning_rate": 3.7134080231297437e-06, "loss": 86.0469, "step": 4854 }, { "compression_loss": 85.75727844238281, "epoch": 1.75, "learning_rate": 3.7079869895193354e-06, "loss": 86.1928, "step": 4855 }, { "compression_loss": 84.65705871582031, "epoch": 1.75, "learning_rate": 3.7025659559089267e-06, "loss": 84.8492, "step": 4856 }, { "compression_loss": 84.71553802490234, "epoch": 1.76, "learning_rate": 3.6971449222985184e-06, "loss": 85.3985, "step": 4857 }, { "compression_loss": 84.34925842285156, "epoch": 1.76, "learning_rate": 3.69172388868811e-06, "loss": 84.777, "step": 4858 }, { "compression_loss": 85.29422760009766, "epoch": 1.76, "learning_rate": 3.6863028550777018e-06, "loss": 85.8893, "step": 4859 }, { "compression_loss": 84.22360229492188, "epoch": 1.76, "learning_rate": 3.6808818214672935e-06, "loss": 84.5317, "step": 4860 }, { "compression_loss": 82.6104965209961, "epoch": 1.76, "learning_rate": 3.6754607878568847e-06, "loss": 83.2325, "step": 4861 }, { "compression_loss": 83.56382751464844, "epoch": 1.76, "learning_rate": 3.6700397542464764e-06, "loss": 83.9345, "step": 4862 }, { "compression_loss": 82.842529296875, "epoch": 1.76, "learning_rate": 3.664618720636068e-06, "loss": 83.1117, "step": 4863 }, { "compression_loss": 82.95127868652344, "epoch": 1.76, "learning_rate": 3.65919768702566e-06, "loss": 83.4824, "step": 4864 }, { "compression_loss": 85.12435913085938, "epoch": 1.76, "learning_rate": 3.6537766534152515e-06, "loss": 85.6619, "step": 4865 }, { "compression_loss": 84.13191223144531, "epoch": 1.76, "learning_rate": 3.6483556198048432e-06, "loss": 84.6846, "step": 4866 }, { "compression_loss": 83.64645385742188, "epoch": 1.76, "learning_rate": 3.6429345861944345e-06, "loss": 84.0423, "step": 4867 }, { "compression_loss": 83.07315063476562, "epoch": 1.76, "learning_rate": 3.637513552584026e-06, "loss": 83.4185, "step": 4868 }, { "compression_loss": 84.49266052246094, "epoch": 1.76, "learning_rate": 3.632092518973618e-06, "loss": 84.7886, "step": 4869 }, { "compression_loss": 85.60992431640625, "epoch": 1.76, "learning_rate": 3.6266714853632096e-06, "loss": 85.8614, "step": 4870 }, { "compression_loss": 84.3157958984375, "epoch": 1.76, "learning_rate": 3.6212504517528013e-06, "loss": 84.9668, "step": 4871 }, { "compression_loss": 83.895751953125, "epoch": 1.76, "learning_rate": 3.6158294181423925e-06, "loss": 84.2377, "step": 4872 }, { "compression_loss": 84.190185546875, "epoch": 1.76, "learning_rate": 3.6104083845319842e-06, "loss": 84.5822, "step": 4873 }, { "compression_loss": 82.92377471923828, "epoch": 1.76, "learning_rate": 3.604987350921576e-06, "loss": 83.2807, "step": 4874 }, { "compression_loss": 84.44184875488281, "epoch": 1.76, "learning_rate": 3.5995663173111676e-06, "loss": 84.9174, "step": 4875 }, { "compression_loss": 82.89837646484375, "epoch": 1.76, "learning_rate": 3.5941452837007593e-06, "loss": 83.225, "step": 4876 }, { "compression_loss": 83.17391967773438, "epoch": 1.76, "learning_rate": 3.588724250090351e-06, "loss": 83.572, "step": 4877 }, { "compression_loss": 85.34197998046875, "epoch": 1.76, "learning_rate": 3.5833032164799423e-06, "loss": 85.815, "step": 4878 }, { "compression_loss": 84.85678100585938, "epoch": 1.76, "learning_rate": 3.577882182869534e-06, "loss": 85.4135, "step": 4879 }, { "compression_loss": 84.09506225585938, "epoch": 1.76, "learning_rate": 3.5724611492591257e-06, "loss": 84.5348, "step": 4880 }, { "compression_loss": 83.94429779052734, "epoch": 1.76, "learning_rate": 3.5670401156487174e-06, "loss": 84.2867, "step": 4881 }, { "compression_loss": 85.4085922241211, "epoch": 1.76, "learning_rate": 3.561619082038309e-06, "loss": 86.1222, "step": 4882 }, { "compression_loss": 85.30845642089844, "epoch": 1.76, "learning_rate": 3.5561980484279003e-06, "loss": 86.0447, "step": 4883 }, { "compression_loss": 83.63043212890625, "epoch": 1.77, "learning_rate": 3.550777014817492e-06, "loss": 84.0967, "step": 4884 }, { "compression_loss": 85.95726013183594, "epoch": 1.77, "learning_rate": 3.5453559812070837e-06, "loss": 86.5965, "step": 4885 }, { "compression_loss": 84.2717514038086, "epoch": 1.77, "learning_rate": 3.5399349475966754e-06, "loss": 84.7379, "step": 4886 }, { "compression_loss": 85.61868286132812, "epoch": 1.77, "learning_rate": 3.534513913986267e-06, "loss": 86.2359, "step": 4887 }, { "compression_loss": 85.04740905761719, "epoch": 1.77, "learning_rate": 3.529092880375859e-06, "loss": 85.6911, "step": 4888 }, { "compression_loss": 83.86273193359375, "epoch": 1.77, "learning_rate": 3.52367184676545e-06, "loss": 84.2625, "step": 4889 }, { "compression_loss": 84.4487075805664, "epoch": 1.77, "learning_rate": 3.518250813155042e-06, "loss": 85.1333, "step": 4890 }, { "compression_loss": 84.55152893066406, "epoch": 1.77, "learning_rate": 3.5128297795446335e-06, "loss": 84.8556, "step": 4891 }, { "compression_loss": 85.02310180664062, "epoch": 1.77, "learning_rate": 3.507408745934225e-06, "loss": 85.3967, "step": 4892 }, { "compression_loss": 84.75741577148438, "epoch": 1.77, "learning_rate": 3.501987712323817e-06, "loss": 85.4028, "step": 4893 }, { "compression_loss": 83.6153564453125, "epoch": 1.77, "learning_rate": 3.4965666787134077e-06, "loss": 84.0519, "step": 4894 }, { "compression_loss": 83.10910034179688, "epoch": 1.77, "learning_rate": 3.4911456451029994e-06, "loss": 83.6704, "step": 4895 }, { "compression_loss": 83.91162872314453, "epoch": 1.77, "learning_rate": 3.485724611492591e-06, "loss": 84.4075, "step": 4896 }, { "compression_loss": 85.59516906738281, "epoch": 1.77, "learning_rate": 3.480303577882183e-06, "loss": 86.0182, "step": 4897 }, { "compression_loss": 84.75481414794922, "epoch": 1.77, "learning_rate": 3.4748825442717745e-06, "loss": 85.2645, "step": 4898 }, { "compression_loss": 83.12774658203125, "epoch": 1.77, "learning_rate": 3.469461510661366e-06, "loss": 83.5611, "step": 4899 }, { "compression_loss": 85.55765533447266, "epoch": 1.77, "learning_rate": 3.4640404770509575e-06, "loss": 85.8972, "step": 4900 }, { "compression_loss": 86.04878234863281, "epoch": 1.77, "learning_rate": 3.458619443440549e-06, "loss": 86.4745, "step": 4901 }, { "compression_loss": 85.70013427734375, "epoch": 1.77, "learning_rate": 3.453198409830141e-06, "loss": 86.1164, "step": 4902 }, { "compression_loss": 85.74191284179688, "epoch": 1.77, "learning_rate": 3.4477773762197326e-06, "loss": 86.3825, "step": 4903 }, { "compression_loss": 82.61593627929688, "epoch": 1.77, "learning_rate": 3.4423563426093243e-06, "loss": 83.0333, "step": 4904 }, { "compression_loss": 83.81391906738281, "epoch": 1.77, "learning_rate": 3.4369353089989155e-06, "loss": 84.1953, "step": 4905 }, { "compression_loss": 84.79761505126953, "epoch": 1.77, "learning_rate": 3.4315142753885072e-06, "loss": 85.1629, "step": 4906 }, { "compression_loss": 83.30171203613281, "epoch": 1.77, "learning_rate": 3.426093241778099e-06, "loss": 83.6694, "step": 4907 }, { "compression_loss": 84.81153869628906, "epoch": 1.77, "learning_rate": 3.4206722081676906e-06, "loss": 85.1477, "step": 4908 }, { "compression_loss": 85.84178161621094, "epoch": 1.77, "learning_rate": 3.4152511745572823e-06, "loss": 86.5954, "step": 4909 }, { "compression_loss": 84.42070007324219, "epoch": 1.77, "learning_rate": 3.4098301409468736e-06, "loss": 84.6167, "step": 4910 }, { "compression_loss": 84.43539428710938, "epoch": 1.77, "learning_rate": 3.4044091073364653e-06, "loss": 84.7907, "step": 4911 }, { "compression_loss": 84.7283935546875, "epoch": 1.78, "learning_rate": 3.398988073726057e-06, "loss": 85.0151, "step": 4912 }, { "compression_loss": 83.76754760742188, "epoch": 1.78, "learning_rate": 3.3935670401156487e-06, "loss": 84.2085, "step": 4913 }, { "compression_loss": 83.2187728881836, "epoch": 1.78, "learning_rate": 3.3881460065052404e-06, "loss": 83.5623, "step": 4914 }, { "compression_loss": 82.42314147949219, "epoch": 1.78, "learning_rate": 3.382724972894832e-06, "loss": 82.999, "step": 4915 }, { "compression_loss": 83.527099609375, "epoch": 1.78, "learning_rate": 3.3773039392844233e-06, "loss": 84.0195, "step": 4916 }, { "compression_loss": 85.09567260742188, "epoch": 1.78, "learning_rate": 3.371882905674015e-06, "loss": 85.6021, "step": 4917 }, { "compression_loss": 83.43325805664062, "epoch": 1.78, "learning_rate": 3.3664618720636067e-06, "loss": 84.1002, "step": 4918 }, { "compression_loss": 85.1772689819336, "epoch": 1.78, "learning_rate": 3.3610408384531984e-06, "loss": 85.5016, "step": 4919 }, { "compression_loss": 85.58851623535156, "epoch": 1.78, "learning_rate": 3.35561980484279e-06, "loss": 86.1236, "step": 4920 }, { "compression_loss": 81.95909118652344, "epoch": 1.78, "learning_rate": 3.3501987712323814e-06, "loss": 82.2782, "step": 4921 }, { "compression_loss": 83.12557983398438, "epoch": 1.78, "learning_rate": 3.344777737621973e-06, "loss": 83.4163, "step": 4922 }, { "compression_loss": 83.84040832519531, "epoch": 1.78, "learning_rate": 3.3393567040115648e-06, "loss": 84.3379, "step": 4923 }, { "compression_loss": 84.67635345458984, "epoch": 1.78, "learning_rate": 3.3339356704011565e-06, "loss": 85.2965, "step": 4924 }, { "compression_loss": 85.1047134399414, "epoch": 1.78, "learning_rate": 3.328514636790748e-06, "loss": 85.4585, "step": 4925 }, { "compression_loss": 85.14051818847656, "epoch": 1.78, "learning_rate": 3.32309360318034e-06, "loss": 85.4722, "step": 4926 }, { "compression_loss": 84.37193298339844, "epoch": 1.78, "learning_rate": 3.317672569569931e-06, "loss": 84.673, "step": 4927 }, { "compression_loss": 84.20831298828125, "epoch": 1.78, "learning_rate": 3.312251535959523e-06, "loss": 84.9085, "step": 4928 }, { "compression_loss": 86.41293334960938, "epoch": 1.78, "learning_rate": 3.3068305023491145e-06, "loss": 86.8952, "step": 4929 }, { "compression_loss": 83.82962036132812, "epoch": 1.78, "learning_rate": 3.3014094687387062e-06, "loss": 84.2101, "step": 4930 }, { "compression_loss": 83.42195129394531, "epoch": 1.78, "learning_rate": 3.295988435128298e-06, "loss": 83.9387, "step": 4931 }, { "compression_loss": 85.92594146728516, "epoch": 1.78, "learning_rate": 3.290567401517889e-06, "loss": 86.6615, "step": 4932 }, { "compression_loss": 85.49809265136719, "epoch": 1.78, "learning_rate": 3.285146367907481e-06, "loss": 85.9769, "step": 4933 }, { "compression_loss": 83.80370330810547, "epoch": 1.78, "learning_rate": 3.2797253342970726e-06, "loss": 84.0099, "step": 4934 }, { "compression_loss": 85.53515625, "epoch": 1.78, "learning_rate": 3.2743043006866643e-06, "loss": 86.2167, "step": 4935 }, { "compression_loss": 83.6906509399414, "epoch": 1.78, "learning_rate": 3.268883267076256e-06, "loss": 84.1814, "step": 4936 }, { "compression_loss": 84.4205322265625, "epoch": 1.78, "learning_rate": 3.2634622334658477e-06, "loss": 84.9084, "step": 4937 }, { "compression_loss": 83.7797622680664, "epoch": 1.78, "learning_rate": 3.258041199855439e-06, "loss": 84.1884, "step": 4938 }, { "compression_loss": 83.9122314453125, "epoch": 1.78, "learning_rate": 3.2526201662450306e-06, "loss": 84.4832, "step": 4939 }, { "compression_loss": 85.39883422851562, "epoch": 1.79, "learning_rate": 3.2471991326346223e-06, "loss": 85.9343, "step": 4940 }, { "compression_loss": 84.16316986083984, "epoch": 1.79, "learning_rate": 3.241778099024214e-06, "loss": 84.524, "step": 4941 }, { "compression_loss": 85.41569519042969, "epoch": 1.79, "learning_rate": 3.2363570654138057e-06, "loss": 86.0366, "step": 4942 }, { "compression_loss": 84.74054718017578, "epoch": 1.79, "learning_rate": 3.230936031803397e-06, "loss": 85.3464, "step": 4943 }, { "compression_loss": 85.56779479980469, "epoch": 1.79, "learning_rate": 3.2255149981929887e-06, "loss": 85.9443, "step": 4944 }, { "compression_loss": 82.7872085571289, "epoch": 1.79, "learning_rate": 3.2200939645825804e-06, "loss": 83.2445, "step": 4945 }, { "compression_loss": 85.61741638183594, "epoch": 1.79, "learning_rate": 3.214672930972172e-06, "loss": 85.8817, "step": 4946 }, { "compression_loss": 85.22872924804688, "epoch": 1.79, "learning_rate": 3.2092518973617638e-06, "loss": 85.6116, "step": 4947 }, { "compression_loss": 84.129638671875, "epoch": 1.79, "learning_rate": 3.2038308637513555e-06, "loss": 84.4633, "step": 4948 }, { "compression_loss": 85.08732604980469, "epoch": 1.79, "learning_rate": 3.1984098301409467e-06, "loss": 85.6033, "step": 4949 }, { "compression_loss": 84.63491821289062, "epoch": 1.79, "learning_rate": 3.1929887965305384e-06, "loss": 85.1246, "step": 4950 }, { "compression_loss": 81.86555480957031, "epoch": 1.79, "learning_rate": 3.18756776292013e-06, "loss": 82.1135, "step": 4951 }, { "compression_loss": 84.6061782836914, "epoch": 1.79, "learning_rate": 3.182146729309722e-06, "loss": 84.871, "step": 4952 }, { "compression_loss": 83.67278289794922, "epoch": 1.79, "learning_rate": 3.1767256956993135e-06, "loss": 84.2446, "step": 4953 }, { "compression_loss": 83.55349731445312, "epoch": 1.79, "learning_rate": 3.171304662088905e-06, "loss": 83.8932, "step": 4954 }, { "compression_loss": 84.02301025390625, "epoch": 1.79, "learning_rate": 3.1658836284784965e-06, "loss": 84.3805, "step": 4955 }, { "compression_loss": 85.79179382324219, "epoch": 1.79, "learning_rate": 3.160462594868088e-06, "loss": 86.2892, "step": 4956 }, { "compression_loss": 82.55451965332031, "epoch": 1.79, "learning_rate": 3.15504156125768e-06, "loss": 82.8711, "step": 4957 }, { "compression_loss": 83.21891021728516, "epoch": 1.79, "learning_rate": 3.1496205276472716e-06, "loss": 83.8043, "step": 4958 }, { "compression_loss": 86.52754974365234, "epoch": 1.79, "learning_rate": 3.144199494036863e-06, "loss": 87.3181, "step": 4959 }, { "compression_loss": 86.08201599121094, "epoch": 1.79, "learning_rate": 3.1387784604264545e-06, "loss": 86.6845, "step": 4960 }, { "compression_loss": 83.13629150390625, "epoch": 1.79, "learning_rate": 3.1333574268160462e-06, "loss": 83.6072, "step": 4961 }, { "compression_loss": 83.46026611328125, "epoch": 1.79, "learning_rate": 3.127936393205638e-06, "loss": 83.9328, "step": 4962 }, { "compression_loss": 84.31141662597656, "epoch": 1.79, "learning_rate": 3.1225153595952296e-06, "loss": 85.0153, "step": 4963 }, { "compression_loss": 84.76290130615234, "epoch": 1.79, "learning_rate": 3.1170943259848213e-06, "loss": 85.3044, "step": 4964 }, { "compression_loss": 83.7203598022461, "epoch": 1.79, "learning_rate": 3.1116732923744126e-06, "loss": 84.0941, "step": 4965 }, { "compression_loss": 86.58711242675781, "epoch": 1.79, "learning_rate": 3.1062522587640043e-06, "loss": 86.9901, "step": 4966 }, { "compression_loss": 83.892333984375, "epoch": 1.8, "learning_rate": 3.100831225153596e-06, "loss": 84.396, "step": 4967 }, { "compression_loss": 86.09978485107422, "epoch": 1.8, "learning_rate": 3.0954101915431877e-06, "loss": 86.4287, "step": 4968 }, { "compression_loss": 85.1236343383789, "epoch": 1.8, "learning_rate": 3.0899891579327794e-06, "loss": 85.6679, "step": 4969 }, { "compression_loss": 82.47232055664062, "epoch": 1.8, "learning_rate": 3.0845681243223706e-06, "loss": 83.0249, "step": 4970 }, { "compression_loss": 85.07943725585938, "epoch": 1.8, "learning_rate": 3.0791470907119623e-06, "loss": 85.569, "step": 4971 }, { "compression_loss": 83.08389282226562, "epoch": 1.8, "learning_rate": 3.073726057101554e-06, "loss": 83.6931, "step": 4972 }, { "compression_loss": 83.08837890625, "epoch": 1.8, "learning_rate": 3.0683050234911457e-06, "loss": 83.5584, "step": 4973 }, { "compression_loss": 84.11943817138672, "epoch": 1.8, "learning_rate": 3.0628839898807374e-06, "loss": 84.356, "step": 4974 }, { "compression_loss": 84.7439193725586, "epoch": 1.8, "learning_rate": 3.057462956270329e-06, "loss": 85.0015, "step": 4975 }, { "compression_loss": 82.15594482421875, "epoch": 1.8, "learning_rate": 3.0520419226599204e-06, "loss": 82.481, "step": 4976 }, { "compression_loss": 84.52217864990234, "epoch": 1.8, "learning_rate": 3.046620889049512e-06, "loss": 85.2326, "step": 4977 }, { "compression_loss": 84.25770568847656, "epoch": 1.8, "learning_rate": 3.0411998554391038e-06, "loss": 84.463, "step": 4978 }, { "compression_loss": 83.64149475097656, "epoch": 1.8, "learning_rate": 3.0357788218286955e-06, "loss": 84.2579, "step": 4979 }, { "compression_loss": 83.57002258300781, "epoch": 1.8, "learning_rate": 3.030357788218287e-06, "loss": 83.9873, "step": 4980 }, { "compression_loss": 84.47989654541016, "epoch": 1.8, "learning_rate": 3.0249367546078784e-06, "loss": 85.0929, "step": 4981 }, { "compression_loss": 86.4017333984375, "epoch": 1.8, "learning_rate": 3.01951572099747e-06, "loss": 86.7972, "step": 4982 }, { "compression_loss": 83.04631042480469, "epoch": 1.8, "learning_rate": 3.014094687387062e-06, "loss": 83.385, "step": 4983 }, { "compression_loss": 83.63174438476562, "epoch": 1.8, "learning_rate": 3.0086736537766535e-06, "loss": 84.0837, "step": 4984 }, { "compression_loss": 82.74339294433594, "epoch": 1.8, "learning_rate": 3.0032526201662452e-06, "loss": 83.2375, "step": 4985 }, { "compression_loss": 84.24252319335938, "epoch": 1.8, "learning_rate": 2.997831586555837e-06, "loss": 84.5674, "step": 4986 }, { "compression_loss": 85.12165832519531, "epoch": 1.8, "learning_rate": 2.992410552945428e-06, "loss": 85.485, "step": 4987 }, { "compression_loss": 83.58695983886719, "epoch": 1.8, "learning_rate": 2.98698951933502e-06, "loss": 84.0021, "step": 4988 }, { "compression_loss": 85.81317138671875, "epoch": 1.8, "learning_rate": 2.9815684857246116e-06, "loss": 86.7129, "step": 4989 }, { "compression_loss": 84.59523010253906, "epoch": 1.8, "learning_rate": 2.9761474521142033e-06, "loss": 85.0636, "step": 4990 }, { "compression_loss": 85.22412872314453, "epoch": 1.8, "learning_rate": 2.970726418503795e-06, "loss": 85.683, "step": 4991 }, { "compression_loss": 82.87490844726562, "epoch": 1.8, "learning_rate": 2.9653053848933862e-06, "loss": 83.3456, "step": 4992 }, { "compression_loss": 84.65518188476562, "epoch": 1.8, "learning_rate": 2.959884351282978e-06, "loss": 85.0024, "step": 4993 }, { "compression_loss": 84.96782684326172, "epoch": 1.8, "learning_rate": 2.9544633176725696e-06, "loss": 85.6776, "step": 4994 }, { "compression_loss": 85.80345153808594, "epoch": 1.81, "learning_rate": 2.9490422840621613e-06, "loss": 86.3282, "step": 4995 }, { "compression_loss": 85.07913970947266, "epoch": 1.81, "learning_rate": 2.943621250451753e-06, "loss": 85.4054, "step": 4996 }, { "compression_loss": 82.7734603881836, "epoch": 1.81, "learning_rate": 2.9382002168413447e-06, "loss": 83.2907, "step": 4997 }, { "compression_loss": 83.57418060302734, "epoch": 1.81, "learning_rate": 2.932779183230936e-06, "loss": 84.0048, "step": 4998 }, { "compression_loss": 84.7027587890625, "epoch": 1.81, "learning_rate": 2.9273581496205277e-06, "loss": 85.1967, "step": 4999 }, { "compression_loss": 84.81615447998047, "epoch": 1.81, "learning_rate": 2.9219371160101194e-06, "loss": 85.215, "step": 5000 }, { "epoch": 1.81, "eval_exact_match": 87.01986754966887, "eval_f1": 93.04996423082547, "step": 5000 }, { "compression_loss": 83.13479614257812, "epoch": 1.81, "learning_rate": 2.916516082399711e-06, "loss": 83.5458, "step": 5001 }, { "compression_loss": 83.93461608886719, "epoch": 1.81, "learning_rate": 2.9110950487893028e-06, "loss": 84.3387, "step": 5002 }, { "compression_loss": 84.92842102050781, "epoch": 1.81, "learning_rate": 2.905674015178894e-06, "loss": 85.4777, "step": 5003 }, { "compression_loss": 86.13031768798828, "epoch": 1.81, "learning_rate": 2.9002529815684857e-06, "loss": 86.9006, "step": 5004 }, { "compression_loss": 84.68158721923828, "epoch": 1.81, "learning_rate": 2.8948319479580774e-06, "loss": 85.1952, "step": 5005 }, { "compression_loss": 82.37509155273438, "epoch": 1.81, "learning_rate": 2.889410914347669e-06, "loss": 82.6138, "step": 5006 }, { "compression_loss": 85.32783508300781, "epoch": 1.81, "learning_rate": 2.883989880737261e-06, "loss": 85.7626, "step": 5007 }, { "compression_loss": 84.49185180664062, "epoch": 1.81, "learning_rate": 2.878568847126852e-06, "loss": 84.9196, "step": 5008 }, { "compression_loss": 84.08384704589844, "epoch": 1.81, "learning_rate": 2.873147813516444e-06, "loss": 84.7076, "step": 5009 }, { "compression_loss": 84.76118469238281, "epoch": 1.81, "learning_rate": 2.8677267799060355e-06, "loss": 85.1108, "step": 5010 }, { "compression_loss": 82.54667663574219, "epoch": 1.81, "learning_rate": 2.862305746295627e-06, "loss": 82.7394, "step": 5011 }, { "compression_loss": 85.91447448730469, "epoch": 1.81, "learning_rate": 2.856884712685219e-06, "loss": 86.2645, "step": 5012 }, { "compression_loss": 82.61993408203125, "epoch": 1.81, "learning_rate": 2.8514636790748106e-06, "loss": 83.1193, "step": 5013 }, { "compression_loss": 84.98204040527344, "epoch": 1.81, "learning_rate": 2.846042645464402e-06, "loss": 85.6668, "step": 5014 }, { "compression_loss": 85.73323059082031, "epoch": 1.81, "learning_rate": 2.8406216118539935e-06, "loss": 86.0855, "step": 5015 }, { "compression_loss": 85.46923828125, "epoch": 1.81, "learning_rate": 2.8352005782435852e-06, "loss": 85.7136, "step": 5016 }, { "compression_loss": 85.46027374267578, "epoch": 1.81, "learning_rate": 2.829779544633177e-06, "loss": 85.9537, "step": 5017 }, { "compression_loss": 84.97296142578125, "epoch": 1.81, "learning_rate": 2.8243585110227686e-06, "loss": 85.4225, "step": 5018 }, { "compression_loss": 86.11302185058594, "epoch": 1.81, "learning_rate": 2.81893747741236e-06, "loss": 86.4809, "step": 5019 }, { "compression_loss": 85.70308685302734, "epoch": 1.81, "learning_rate": 2.8135164438019516e-06, "loss": 86.072, "step": 5020 }, { "compression_loss": 84.10507202148438, "epoch": 1.81, "learning_rate": 2.8080954101915433e-06, "loss": 84.3386, "step": 5021 }, { "compression_loss": 85.31726837158203, "epoch": 1.81, "learning_rate": 2.802674376581135e-06, "loss": 86.0206, "step": 5022 }, { "compression_loss": 84.53475952148438, "epoch": 1.82, "learning_rate": 2.7972533429707267e-06, "loss": 84.8885, "step": 5023 }, { "compression_loss": 83.19717407226562, "epoch": 1.82, "learning_rate": 2.7918323093603184e-06, "loss": 83.5028, "step": 5024 }, { "compression_loss": 82.60554504394531, "epoch": 1.82, "learning_rate": 2.7864112757499097e-06, "loss": 83.0181, "step": 5025 }, { "compression_loss": 85.16320037841797, "epoch": 1.82, "learning_rate": 2.7809902421395014e-06, "loss": 85.6476, "step": 5026 }, { "compression_loss": 83.52336883544922, "epoch": 1.82, "learning_rate": 2.775569208529093e-06, "loss": 83.8717, "step": 5027 }, { "compression_loss": 81.73724365234375, "epoch": 1.82, "learning_rate": 2.7701481749186847e-06, "loss": 81.9806, "step": 5028 }, { "compression_loss": 84.05340576171875, "epoch": 1.82, "learning_rate": 2.7647271413082764e-06, "loss": 84.4056, "step": 5029 }, { "compression_loss": 85.91999816894531, "epoch": 1.82, "learning_rate": 2.7593061076978677e-06, "loss": 86.352, "step": 5030 }, { "compression_loss": 85.9990234375, "epoch": 1.82, "learning_rate": 2.7538850740874594e-06, "loss": 86.5733, "step": 5031 }, { "compression_loss": 83.26490783691406, "epoch": 1.82, "learning_rate": 2.748464040477051e-06, "loss": 83.5927, "step": 5032 }, { "compression_loss": 83.64949035644531, "epoch": 1.82, "learning_rate": 2.743043006866643e-06, "loss": 83.9944, "step": 5033 }, { "compression_loss": 82.71629333496094, "epoch": 1.82, "learning_rate": 2.7376219732562345e-06, "loss": 83.3958, "step": 5034 }, { "compression_loss": 85.731689453125, "epoch": 1.82, "learning_rate": 2.732200939645826e-06, "loss": 86.1431, "step": 5035 }, { "compression_loss": 85.55563354492188, "epoch": 1.82, "learning_rate": 2.7267799060354175e-06, "loss": 86.0902, "step": 5036 }, { "compression_loss": 84.8470458984375, "epoch": 1.82, "learning_rate": 2.721358872425009e-06, "loss": 85.224, "step": 5037 }, { "compression_loss": 83.00881958007812, "epoch": 1.82, "learning_rate": 2.715937838814601e-06, "loss": 83.4579, "step": 5038 }, { "compression_loss": 84.68240356445312, "epoch": 1.82, "learning_rate": 2.7105168052041925e-06, "loss": 85.28, "step": 5039 }, { "compression_loss": 86.61293029785156, "epoch": 1.82, "learning_rate": 2.7050957715937842e-06, "loss": 87.2419, "step": 5040 }, { "compression_loss": 84.04751586914062, "epoch": 1.82, "learning_rate": 2.6996747379833755e-06, "loss": 84.4164, "step": 5041 }, { "compression_loss": 85.43464660644531, "epoch": 1.82, "learning_rate": 2.694253704372967e-06, "loss": 85.9963, "step": 5042 }, { "compression_loss": 84.78092956542969, "epoch": 1.82, "learning_rate": 2.688832670762559e-06, "loss": 85.4394, "step": 5043 }, { "compression_loss": 83.46574401855469, "epoch": 1.82, "learning_rate": 2.6834116371521506e-06, "loss": 83.7704, "step": 5044 }, { "compression_loss": 84.71670532226562, "epoch": 1.82, "learning_rate": 2.6779906035417423e-06, "loss": 85.143, "step": 5045 }, { "compression_loss": 86.45530700683594, "epoch": 1.82, "learning_rate": 2.672569569931334e-06, "loss": 86.965, "step": 5046 }, { "compression_loss": 84.41008758544922, "epoch": 1.82, "learning_rate": 2.6671485363209253e-06, "loss": 84.9333, "step": 5047 }, { "compression_loss": 84.27894592285156, "epoch": 1.82, "learning_rate": 2.661727502710517e-06, "loss": 84.6652, "step": 5048 }, { "compression_loss": 84.5547866821289, "epoch": 1.82, "learning_rate": 2.6563064691001087e-06, "loss": 84.7447, "step": 5049 }, { "compression_loss": 84.98930358886719, "epoch": 1.83, "learning_rate": 2.6508854354897003e-06, "loss": 85.3568, "step": 5050 }, { "compression_loss": 84.50233459472656, "epoch": 1.83, "learning_rate": 2.645464401879292e-06, "loss": 84.9045, "step": 5051 }, { "compression_loss": 85.45899963378906, "epoch": 1.83, "learning_rate": 2.6400433682688833e-06, "loss": 85.979, "step": 5052 }, { "compression_loss": 83.15390014648438, "epoch": 1.83, "learning_rate": 2.634622334658475e-06, "loss": 83.4871, "step": 5053 }, { "compression_loss": 84.47720336914062, "epoch": 1.83, "learning_rate": 2.6292013010480667e-06, "loss": 84.9548, "step": 5054 }, { "compression_loss": 84.5205078125, "epoch": 1.83, "learning_rate": 2.6237802674376584e-06, "loss": 84.8127, "step": 5055 }, { "compression_loss": 85.55118560791016, "epoch": 1.83, "learning_rate": 2.61835923382725e-06, "loss": 85.9636, "step": 5056 }, { "compression_loss": 84.88414001464844, "epoch": 1.83, "learning_rate": 2.6129382002168414e-06, "loss": 85.3455, "step": 5057 }, { "compression_loss": 83.22509765625, "epoch": 1.83, "learning_rate": 2.607517166606433e-06, "loss": 83.4643, "step": 5058 }, { "compression_loss": 83.47562408447266, "epoch": 1.83, "learning_rate": 2.6020961329960248e-06, "loss": 84.1351, "step": 5059 }, { "compression_loss": 84.63810729980469, "epoch": 1.83, "learning_rate": 2.5966750993856165e-06, "loss": 84.993, "step": 5060 }, { "compression_loss": 84.4217529296875, "epoch": 1.83, "learning_rate": 2.591254065775208e-06, "loss": 85.1597, "step": 5061 }, { "compression_loss": 84.37971496582031, "epoch": 1.83, "learning_rate": 2.5858330321648e-06, "loss": 84.9036, "step": 5062 }, { "compression_loss": 85.16922760009766, "epoch": 1.83, "learning_rate": 2.580411998554391e-06, "loss": 85.6901, "step": 5063 }, { "compression_loss": 83.30452728271484, "epoch": 1.83, "learning_rate": 2.574990964943983e-06, "loss": 83.8305, "step": 5064 }, { "compression_loss": 84.70979309082031, "epoch": 1.83, "learning_rate": 2.5695699313335745e-06, "loss": 84.9531, "step": 5065 }, { "compression_loss": 83.71734619140625, "epoch": 1.83, "learning_rate": 2.564148897723166e-06, "loss": 84.1077, "step": 5066 }, { "compression_loss": 82.07180786132812, "epoch": 1.83, "learning_rate": 2.558727864112758e-06, "loss": 82.4183, "step": 5067 }, { "compression_loss": 84.38040161132812, "epoch": 1.83, "learning_rate": 2.553306830502349e-06, "loss": 84.9038, "step": 5068 }, { "compression_loss": 83.88589477539062, "epoch": 1.83, "learning_rate": 2.547885796891941e-06, "loss": 84.4715, "step": 5069 }, { "compression_loss": 82.7934799194336, "epoch": 1.83, "learning_rate": 2.5424647632815326e-06, "loss": 83.0412, "step": 5070 }, { "compression_loss": 83.27378845214844, "epoch": 1.83, "learning_rate": 2.5370437296711243e-06, "loss": 83.6058, "step": 5071 }, { "compression_loss": 83.66277313232422, "epoch": 1.83, "learning_rate": 2.531622696060716e-06, "loss": 84.3223, "step": 5072 }, { "compression_loss": 84.884765625, "epoch": 1.83, "learning_rate": 2.5262016624503076e-06, "loss": 85.2871, "step": 5073 }, { "compression_loss": 84.14277648925781, "epoch": 1.83, "learning_rate": 2.520780628839899e-06, "loss": 84.4053, "step": 5074 }, { "compression_loss": 83.51011657714844, "epoch": 1.83, "learning_rate": 2.5153595952294906e-06, "loss": 84.0295, "step": 5075 }, { "compression_loss": 83.85493469238281, "epoch": 1.83, "learning_rate": 2.5099385616190823e-06, "loss": 84.1607, "step": 5076 }, { "compression_loss": 85.13236999511719, "epoch": 1.83, "learning_rate": 2.504517528008674e-06, "loss": 85.6532, "step": 5077 }, { "compression_loss": 83.38753509521484, "epoch": 1.84, "learning_rate": 2.4990964943982653e-06, "loss": 83.6988, "step": 5078 }, { "compression_loss": 86.03367614746094, "epoch": 1.84, "learning_rate": 2.4936754607878565e-06, "loss": 86.4047, "step": 5079 }, { "compression_loss": 85.69792175292969, "epoch": 1.84, "learning_rate": 2.4882544271774482e-06, "loss": 86.3573, "step": 5080 }, { "compression_loss": 83.48198699951172, "epoch": 1.84, "learning_rate": 2.48283339356704e-06, "loss": 84.0547, "step": 5081 }, { "compression_loss": 85.25068664550781, "epoch": 1.84, "learning_rate": 2.4774123599566316e-06, "loss": 85.8784, "step": 5082 }, { "compression_loss": 81.95541381835938, "epoch": 1.84, "learning_rate": 2.4719913263462233e-06, "loss": 82.2094, "step": 5083 }, { "compression_loss": 84.70333862304688, "epoch": 1.84, "learning_rate": 2.466570292735815e-06, "loss": 85.2888, "step": 5084 }, { "compression_loss": 84.1020278930664, "epoch": 1.84, "learning_rate": 2.4611492591254063e-06, "loss": 84.5069, "step": 5085 }, { "compression_loss": 84.42620849609375, "epoch": 1.84, "learning_rate": 2.455728225514998e-06, "loss": 84.799, "step": 5086 }, { "compression_loss": 84.48311614990234, "epoch": 1.84, "learning_rate": 2.4503071919045897e-06, "loss": 84.8429, "step": 5087 }, { "compression_loss": 82.6475830078125, "epoch": 1.84, "learning_rate": 2.4448861582941814e-06, "loss": 83.0346, "step": 5088 }, { "compression_loss": 84.80424499511719, "epoch": 1.84, "learning_rate": 2.439465124683773e-06, "loss": 85.2674, "step": 5089 }, { "compression_loss": 83.67465209960938, "epoch": 1.84, "learning_rate": 2.4340440910733643e-06, "loss": 84.0416, "step": 5090 }, { "compression_loss": 84.09317016601562, "epoch": 1.84, "learning_rate": 2.428623057462956e-06, "loss": 84.45, "step": 5091 }, { "compression_loss": 83.20781707763672, "epoch": 1.84, "learning_rate": 2.4232020238525477e-06, "loss": 83.6349, "step": 5092 }, { "compression_loss": 84.9693603515625, "epoch": 1.84, "learning_rate": 2.4177809902421394e-06, "loss": 85.7929, "step": 5093 }, { "compression_loss": 84.76388549804688, "epoch": 1.84, "learning_rate": 2.412359956631731e-06, "loss": 85.2198, "step": 5094 }, { "compression_loss": 82.1331787109375, "epoch": 1.84, "learning_rate": 2.406938923021323e-06, "loss": 82.49, "step": 5095 }, { "compression_loss": 83.88948059082031, "epoch": 1.84, "learning_rate": 2.401517889410914e-06, "loss": 84.226, "step": 5096 }, { "compression_loss": 84.22456359863281, "epoch": 1.84, "learning_rate": 2.396096855800506e-06, "loss": 84.7842, "step": 5097 }, { "compression_loss": 84.41085052490234, "epoch": 1.84, "learning_rate": 2.3906758221900975e-06, "loss": 84.5643, "step": 5098 }, { "compression_loss": 84.47520446777344, "epoch": 1.84, "learning_rate": 2.385254788579689e-06, "loss": 85.1302, "step": 5099 }, { "compression_loss": 83.20240020751953, "epoch": 1.84, "learning_rate": 2.379833754969281e-06, "loss": 83.5383, "step": 5100 }, { "compression_loss": 84.40839385986328, "epoch": 1.84, "learning_rate": 2.374412721358872e-06, "loss": 84.751, "step": 5101 }, { "compression_loss": 83.82778930664062, "epoch": 1.84, "learning_rate": 2.368991687748464e-06, "loss": 84.1983, "step": 5102 }, { "compression_loss": 82.60311126708984, "epoch": 1.84, "learning_rate": 2.3635706541380555e-06, "loss": 83.0919, "step": 5103 }, { "compression_loss": 83.65975952148438, "epoch": 1.84, "learning_rate": 2.3581496205276472e-06, "loss": 84.0306, "step": 5104 }, { "compression_loss": 85.06914520263672, "epoch": 1.84, "learning_rate": 2.352728586917239e-06, "loss": 85.6338, "step": 5105 }, { "compression_loss": 85.28092956542969, "epoch": 1.85, "learning_rate": 2.3473075533068306e-06, "loss": 85.6312, "step": 5106 }, { "compression_loss": 83.90097045898438, "epoch": 1.85, "learning_rate": 2.341886519696422e-06, "loss": 84.2464, "step": 5107 }, { "compression_loss": 84.040283203125, "epoch": 1.85, "learning_rate": 2.3364654860860136e-06, "loss": 84.4657, "step": 5108 }, { "compression_loss": 85.88288116455078, "epoch": 1.85, "learning_rate": 2.3310444524756053e-06, "loss": 86.1854, "step": 5109 }, { "compression_loss": 84.46951293945312, "epoch": 1.85, "learning_rate": 2.325623418865197e-06, "loss": 84.8175, "step": 5110 }, { "compression_loss": 83.51242065429688, "epoch": 1.85, "learning_rate": 2.3202023852547887e-06, "loss": 84.0206, "step": 5111 }, { "compression_loss": 83.76654052734375, "epoch": 1.85, "learning_rate": 2.31478135164438e-06, "loss": 84.1551, "step": 5112 }, { "compression_loss": 86.80799865722656, "epoch": 1.85, "learning_rate": 2.3093603180339716e-06, "loss": 87.3116, "step": 5113 }, { "compression_loss": 85.09480285644531, "epoch": 1.85, "learning_rate": 2.3039392844235633e-06, "loss": 85.6884, "step": 5114 }, { "compression_loss": 83.85456085205078, "epoch": 1.85, "learning_rate": 2.298518250813155e-06, "loss": 84.5704, "step": 5115 }, { "compression_loss": 84.697998046875, "epoch": 1.85, "learning_rate": 2.2930972172027467e-06, "loss": 85.1775, "step": 5116 }, { "compression_loss": 83.30374908447266, "epoch": 1.85, "learning_rate": 2.287676183592338e-06, "loss": 83.5835, "step": 5117 }, { "compression_loss": 85.37895202636719, "epoch": 1.85, "learning_rate": 2.2822551499819297e-06, "loss": 86.1463, "step": 5118 }, { "compression_loss": 84.39446258544922, "epoch": 1.85, "learning_rate": 2.2768341163715214e-06, "loss": 84.8061, "step": 5119 }, { "compression_loss": 85.18910217285156, "epoch": 1.85, "learning_rate": 2.271413082761113e-06, "loss": 86.1705, "step": 5120 }, { "compression_loss": 84.42825317382812, "epoch": 1.85, "learning_rate": 2.265992049150705e-06, "loss": 84.6936, "step": 5121 }, { "compression_loss": 82.75773620605469, "epoch": 1.85, "learning_rate": 2.2605710155402965e-06, "loss": 82.9397, "step": 5122 }, { "compression_loss": 84.12013244628906, "epoch": 1.85, "learning_rate": 2.2551499819298878e-06, "loss": 84.5111, "step": 5123 }, { "compression_loss": 84.6436538696289, "epoch": 1.85, "learning_rate": 2.2497289483194795e-06, "loss": 85.0479, "step": 5124 }, { "compression_loss": 86.64517211914062, "epoch": 1.85, "learning_rate": 2.244307914709071e-06, "loss": 87.0488, "step": 5125 }, { "compression_loss": 85.07737731933594, "epoch": 1.85, "learning_rate": 2.238886881098663e-06, "loss": 85.3054, "step": 5126 }, { "compression_loss": 84.25984191894531, "epoch": 1.85, "learning_rate": 2.2334658474882545e-06, "loss": 84.9174, "step": 5127 }, { "compression_loss": 83.16179656982422, "epoch": 1.85, "learning_rate": 2.228044813877846e-06, "loss": 83.5222, "step": 5128 }, { "compression_loss": 86.39753723144531, "epoch": 1.85, "learning_rate": 2.2226237802674375e-06, "loss": 87.0598, "step": 5129 }, { "compression_loss": 84.0789794921875, "epoch": 1.85, "learning_rate": 2.217202746657029e-06, "loss": 84.7214, "step": 5130 }, { "compression_loss": 84.62161254882812, "epoch": 1.85, "learning_rate": 2.211781713046621e-06, "loss": 85.0941, "step": 5131 }, { "compression_loss": 86.97555541992188, "epoch": 1.85, "learning_rate": 2.2063606794362126e-06, "loss": 87.6036, "step": 5132 }, { "compression_loss": 83.7607192993164, "epoch": 1.86, "learning_rate": 2.2009396458258043e-06, "loss": 84.0383, "step": 5133 }, { "compression_loss": 83.53692626953125, "epoch": 1.86, "learning_rate": 2.1955186122153956e-06, "loss": 84.2348, "step": 5134 }, { "compression_loss": 85.93544006347656, "epoch": 1.86, "learning_rate": 2.1900975786049873e-06, "loss": 86.4962, "step": 5135 }, { "compression_loss": 82.4791259765625, "epoch": 1.86, "learning_rate": 2.184676544994579e-06, "loss": 82.7902, "step": 5136 }, { "compression_loss": 84.0892333984375, "epoch": 1.86, "learning_rate": 2.1792555113841706e-06, "loss": 84.5598, "step": 5137 }, { "compression_loss": 85.18917846679688, "epoch": 1.86, "learning_rate": 2.1738344777737623e-06, "loss": 85.5872, "step": 5138 }, { "compression_loss": 83.89434814453125, "epoch": 1.86, "learning_rate": 2.1684134441633536e-06, "loss": 84.5367, "step": 5139 }, { "compression_loss": 83.30622863769531, "epoch": 1.86, "learning_rate": 2.1629924105529453e-06, "loss": 83.6235, "step": 5140 }, { "compression_loss": 82.89225006103516, "epoch": 1.86, "learning_rate": 2.157571376942537e-06, "loss": 83.3539, "step": 5141 }, { "compression_loss": 84.39359283447266, "epoch": 1.86, "learning_rate": 2.1521503433321287e-06, "loss": 84.7402, "step": 5142 }, { "compression_loss": 85.97853088378906, "epoch": 1.86, "learning_rate": 2.1467293097217204e-06, "loss": 86.411, "step": 5143 }, { "compression_loss": 84.78240203857422, "epoch": 1.86, "learning_rate": 2.141308276111312e-06, "loss": 85.2306, "step": 5144 }, { "compression_loss": 84.53328704833984, "epoch": 1.86, "learning_rate": 2.1358872425009034e-06, "loss": 85.186, "step": 5145 }, { "compression_loss": 85.80314636230469, "epoch": 1.86, "learning_rate": 2.130466208890495e-06, "loss": 86.2062, "step": 5146 }, { "compression_loss": 84.52401733398438, "epoch": 1.86, "learning_rate": 2.1250451752800868e-06, "loss": 84.9482, "step": 5147 }, { "compression_loss": 86.13526916503906, "epoch": 1.86, "learning_rate": 2.1196241416696784e-06, "loss": 86.5786, "step": 5148 }, { "compression_loss": 84.7569580078125, "epoch": 1.86, "learning_rate": 2.11420310805927e-06, "loss": 85.3429, "step": 5149 }, { "compression_loss": 85.84697723388672, "epoch": 1.86, "learning_rate": 2.1087820744488614e-06, "loss": 86.1159, "step": 5150 }, { "compression_loss": 84.61393737792969, "epoch": 1.86, "learning_rate": 2.103361040838453e-06, "loss": 85.1971, "step": 5151 }, { "compression_loss": 83.98394775390625, "epoch": 1.86, "learning_rate": 2.097940007228045e-06, "loss": 84.4311, "step": 5152 }, { "compression_loss": 84.4199447631836, "epoch": 1.86, "learning_rate": 2.0925189736176365e-06, "loss": 85.0657, "step": 5153 }, { "compression_loss": 84.08464813232422, "epoch": 1.86, "learning_rate": 2.087097940007228e-06, "loss": 84.6956, "step": 5154 }, { "compression_loss": 84.7255859375, "epoch": 1.86, "learning_rate": 2.08167690639682e-06, "loss": 85.0378, "step": 5155 }, { "compression_loss": 83.59844970703125, "epoch": 1.86, "learning_rate": 2.076255872786411e-06, "loss": 84.0236, "step": 5156 }, { "compression_loss": 85.16426086425781, "epoch": 1.86, "learning_rate": 2.070834839176003e-06, "loss": 85.6919, "step": 5157 }, { "compression_loss": 84.42120361328125, "epoch": 1.86, "learning_rate": 2.0654138055655946e-06, "loss": 84.8761, "step": 5158 }, { "compression_loss": 85.23419189453125, "epoch": 1.86, "learning_rate": 2.0599927719551862e-06, "loss": 85.4994, "step": 5159 }, { "compression_loss": 84.0060806274414, "epoch": 1.86, "learning_rate": 2.054571738344778e-06, "loss": 84.6571, "step": 5160 }, { "compression_loss": 82.94085693359375, "epoch": 1.87, "learning_rate": 2.0491507047343692e-06, "loss": 83.3606, "step": 5161 }, { "compression_loss": 84.98664855957031, "epoch": 1.87, "learning_rate": 2.043729671123961e-06, "loss": 85.7012, "step": 5162 }, { "compression_loss": 84.20574188232422, "epoch": 1.87, "learning_rate": 2.0383086375135526e-06, "loss": 84.7122, "step": 5163 }, { "compression_loss": 84.4449691772461, "epoch": 1.87, "learning_rate": 2.0328876039031443e-06, "loss": 84.9141, "step": 5164 }, { "compression_loss": 84.13429260253906, "epoch": 1.87, "learning_rate": 2.027466570292736e-06, "loss": 84.2745, "step": 5165 }, { "compression_loss": 84.75352478027344, "epoch": 1.87, "learning_rate": 2.0220455366823273e-06, "loss": 85.2327, "step": 5166 }, { "compression_loss": 85.03324127197266, "epoch": 1.87, "learning_rate": 2.016624503071919e-06, "loss": 85.3398, "step": 5167 }, { "compression_loss": 83.52214050292969, "epoch": 1.87, "learning_rate": 2.0112034694615107e-06, "loss": 83.9868, "step": 5168 }, { "compression_loss": 84.66023254394531, "epoch": 1.87, "learning_rate": 2.0057824358511024e-06, "loss": 85.2392, "step": 5169 }, { "compression_loss": 84.68492126464844, "epoch": 1.87, "learning_rate": 2.000361402240694e-06, "loss": 85.0857, "step": 5170 }, { "compression_loss": 84.85665893554688, "epoch": 1.87, "learning_rate": 1.9949403686302857e-06, "loss": 85.1908, "step": 5171 }, { "compression_loss": 82.2969970703125, "epoch": 1.87, "learning_rate": 1.989519335019877e-06, "loss": 82.64, "step": 5172 }, { "compression_loss": 85.45921325683594, "epoch": 1.87, "learning_rate": 1.9840983014094687e-06, "loss": 85.7697, "step": 5173 }, { "compression_loss": 85.11338806152344, "epoch": 1.87, "learning_rate": 1.9786772677990604e-06, "loss": 85.5037, "step": 5174 }, { "compression_loss": 84.03945922851562, "epoch": 1.87, "learning_rate": 1.973256234188652e-06, "loss": 84.279, "step": 5175 }, { "compression_loss": 85.26441192626953, "epoch": 1.87, "learning_rate": 1.967835200578244e-06, "loss": 85.9645, "step": 5176 }, { "compression_loss": 84.853271484375, "epoch": 1.87, "learning_rate": 1.962414166967835e-06, "loss": 85.4151, "step": 5177 }, { "compression_loss": 83.84138488769531, "epoch": 1.87, "learning_rate": 1.9569931333574268e-06, "loss": 84.2503, "step": 5178 }, { "compression_loss": 84.23727416992188, "epoch": 1.87, "learning_rate": 1.9515720997470185e-06, "loss": 84.516, "step": 5179 }, { "compression_loss": 83.72808837890625, "epoch": 1.87, "learning_rate": 1.94615106613661e-06, "loss": 83.9558, "step": 5180 }, { "compression_loss": 84.37151336669922, "epoch": 1.87, "learning_rate": 1.940730032526202e-06, "loss": 84.9075, "step": 5181 }, { "compression_loss": 83.05225372314453, "epoch": 1.87, "learning_rate": 1.9353089989157935e-06, "loss": 83.4681, "step": 5182 }, { "compression_loss": 84.98907470703125, "epoch": 1.87, "learning_rate": 1.929887965305385e-06, "loss": 85.7389, "step": 5183 }, { "compression_loss": 83.15480041503906, "epoch": 1.87, "learning_rate": 1.9244669316949765e-06, "loss": 83.4115, "step": 5184 }, { "compression_loss": 84.010009765625, "epoch": 1.87, "learning_rate": 1.919045898084568e-06, "loss": 84.4343, "step": 5185 }, { "compression_loss": 83.66948699951172, "epoch": 1.87, "learning_rate": 1.91362486447416e-06, "loss": 84.1133, "step": 5186 }, { "compression_loss": 83.4967041015625, "epoch": 1.87, "learning_rate": 1.9082038308637516e-06, "loss": 83.811, "step": 5187 }, { "compression_loss": 84.68780517578125, "epoch": 1.87, "learning_rate": 1.902782797253343e-06, "loss": 85.1669, "step": 5188 }, { "compression_loss": 81.56187438964844, "epoch": 1.88, "learning_rate": 1.8973617636429348e-06, "loss": 81.9055, "step": 5189 }, { "compression_loss": 83.68443298339844, "epoch": 1.88, "learning_rate": 1.8919407300325263e-06, "loss": 84.2363, "step": 5190 }, { "compression_loss": 86.17975616455078, "epoch": 1.88, "learning_rate": 1.886519696422118e-06, "loss": 86.6498, "step": 5191 }, { "compression_loss": 86.12361145019531, "epoch": 1.88, "learning_rate": 1.8810986628117094e-06, "loss": 86.8731, "step": 5192 }, { "compression_loss": 82.29022216796875, "epoch": 1.88, "learning_rate": 1.8756776292013011e-06, "loss": 82.798, "step": 5193 }, { "compression_loss": 84.62800598144531, "epoch": 1.88, "learning_rate": 1.8702565955908928e-06, "loss": 84.8591, "step": 5194 }, { "compression_loss": 84.13597106933594, "epoch": 1.88, "learning_rate": 1.8648355619804843e-06, "loss": 84.4211, "step": 5195 }, { "compression_loss": 85.37933349609375, "epoch": 1.88, "learning_rate": 1.859414528370076e-06, "loss": 85.9016, "step": 5196 }, { "compression_loss": 85.8585205078125, "epoch": 1.88, "learning_rate": 1.8539934947596677e-06, "loss": 86.172, "step": 5197 }, { "compression_loss": 84.09346771240234, "epoch": 1.88, "learning_rate": 1.8485724611492592e-06, "loss": 84.7861, "step": 5198 }, { "compression_loss": 84.62179565429688, "epoch": 1.88, "learning_rate": 1.8431514275388509e-06, "loss": 85.1356, "step": 5199 }, { "compression_loss": 84.46028137207031, "epoch": 1.88, "learning_rate": 1.8377303939284424e-06, "loss": 84.8706, "step": 5200 }, { "compression_loss": 84.5265884399414, "epoch": 1.88, "learning_rate": 1.832309360318034e-06, "loss": 84.9725, "step": 5201 }, { "compression_loss": 83.72950744628906, "epoch": 1.88, "learning_rate": 1.8268883267076258e-06, "loss": 84.0602, "step": 5202 }, { "compression_loss": 85.15480041503906, "epoch": 1.88, "learning_rate": 1.8214672930972172e-06, "loss": 85.7369, "step": 5203 }, { "compression_loss": 83.10861206054688, "epoch": 1.88, "learning_rate": 1.816046259486809e-06, "loss": 83.5143, "step": 5204 }, { "compression_loss": 84.82488250732422, "epoch": 1.88, "learning_rate": 1.8106252258764006e-06, "loss": 85.032, "step": 5205 }, { "compression_loss": 84.25774383544922, "epoch": 1.88, "learning_rate": 1.8052041922659921e-06, "loss": 84.5681, "step": 5206 }, { "compression_loss": 85.61302185058594, "epoch": 1.88, "learning_rate": 1.7997831586555838e-06, "loss": 86.4025, "step": 5207 }, { "compression_loss": 84.55972290039062, "epoch": 1.88, "learning_rate": 1.7943621250451755e-06, "loss": 85.2583, "step": 5208 }, { "compression_loss": 83.21979522705078, "epoch": 1.88, "learning_rate": 1.788941091434767e-06, "loss": 83.5146, "step": 5209 }, { "compression_loss": 84.07831573486328, "epoch": 1.88, "learning_rate": 1.7835200578243587e-06, "loss": 84.4972, "step": 5210 }, { "compression_loss": 82.97498321533203, "epoch": 1.88, "learning_rate": 1.7780990242139502e-06, "loss": 83.6881, "step": 5211 }, { "compression_loss": 84.21823120117188, "epoch": 1.88, "learning_rate": 1.7726779906035419e-06, "loss": 84.7128, "step": 5212 }, { "compression_loss": 84.88064575195312, "epoch": 1.88, "learning_rate": 1.7672569569931336e-06, "loss": 85.3531, "step": 5213 }, { "compression_loss": 82.55425262451172, "epoch": 1.88, "learning_rate": 1.761835923382725e-06, "loss": 82.9694, "step": 5214 }, { "compression_loss": 84.52182006835938, "epoch": 1.88, "learning_rate": 1.7564148897723167e-06, "loss": 84.9784, "step": 5215 }, { "compression_loss": 83.91931915283203, "epoch": 1.89, "learning_rate": 1.7509938561619084e-06, "loss": 84.3612, "step": 5216 }, { "compression_loss": 85.95442199707031, "epoch": 1.89, "learning_rate": 1.7455728225514997e-06, "loss": 86.5882, "step": 5217 }, { "compression_loss": 84.0294189453125, "epoch": 1.89, "learning_rate": 1.7401517889410914e-06, "loss": 84.4781, "step": 5218 }, { "compression_loss": 83.80740356445312, "epoch": 1.89, "learning_rate": 1.734730755330683e-06, "loss": 84.3612, "step": 5219 }, { "compression_loss": 82.59794616699219, "epoch": 1.89, "learning_rate": 1.7293097217202746e-06, "loss": 83.1051, "step": 5220 }, { "compression_loss": 85.10484313964844, "epoch": 1.89, "learning_rate": 1.7238886881098663e-06, "loss": 85.8067, "step": 5221 }, { "compression_loss": 84.50897216796875, "epoch": 1.89, "learning_rate": 1.7184676544994578e-06, "loss": 85.0031, "step": 5222 }, { "compression_loss": 85.96158599853516, "epoch": 1.89, "learning_rate": 1.7130466208890495e-06, "loss": 86.5478, "step": 5223 }, { "compression_loss": 84.15354919433594, "epoch": 1.89, "learning_rate": 1.7076255872786412e-06, "loss": 85.0702, "step": 5224 }, { "compression_loss": 84.16473388671875, "epoch": 1.89, "learning_rate": 1.7022045536682326e-06, "loss": 84.4469, "step": 5225 }, { "compression_loss": 85.0155029296875, "epoch": 1.89, "learning_rate": 1.6967835200578243e-06, "loss": 85.5088, "step": 5226 }, { "compression_loss": 84.47453308105469, "epoch": 1.89, "learning_rate": 1.691362486447416e-06, "loss": 84.9363, "step": 5227 }, { "compression_loss": 83.65486907958984, "epoch": 1.89, "learning_rate": 1.6859414528370075e-06, "loss": 83.9098, "step": 5228 }, { "compression_loss": 85.80350494384766, "epoch": 1.89, "learning_rate": 1.6805204192265992e-06, "loss": 86.2916, "step": 5229 }, { "compression_loss": 85.18937683105469, "epoch": 1.89, "learning_rate": 1.6750993856161907e-06, "loss": 85.6466, "step": 5230 }, { "compression_loss": 84.23900604248047, "epoch": 1.89, "learning_rate": 1.6696783520057824e-06, "loss": 84.5759, "step": 5231 }, { "compression_loss": 84.4910888671875, "epoch": 1.89, "learning_rate": 1.664257318395374e-06, "loss": 84.8439, "step": 5232 }, { "compression_loss": 83.91574096679688, "epoch": 1.89, "learning_rate": 1.6588362847849656e-06, "loss": 84.2491, "step": 5233 }, { "compression_loss": 84.26199340820312, "epoch": 1.89, "learning_rate": 1.6534152511745573e-06, "loss": 84.6176, "step": 5234 }, { "compression_loss": 84.30204772949219, "epoch": 1.89, "learning_rate": 1.647994217564149e-06, "loss": 84.8032, "step": 5235 }, { "compression_loss": 84.20314025878906, "epoch": 1.89, "learning_rate": 1.6425731839537404e-06, "loss": 84.5373, "step": 5236 }, { "compression_loss": 84.9151382446289, "epoch": 1.89, "learning_rate": 1.6371521503433321e-06, "loss": 85.3383, "step": 5237 }, { "compression_loss": 85.01280212402344, "epoch": 1.89, "learning_rate": 1.6317311167329238e-06, "loss": 85.246, "step": 5238 }, { "compression_loss": 83.73825073242188, "epoch": 1.89, "learning_rate": 1.6263100831225153e-06, "loss": 84.2339, "step": 5239 }, { "compression_loss": 83.17703247070312, "epoch": 1.89, "learning_rate": 1.620889049512107e-06, "loss": 83.9147, "step": 5240 }, { "compression_loss": 83.48990631103516, "epoch": 1.89, "learning_rate": 1.6154680159016985e-06, "loss": 83.8836, "step": 5241 }, { "compression_loss": 84.02942657470703, "epoch": 1.89, "learning_rate": 1.6100469822912902e-06, "loss": 84.6787, "step": 5242 }, { "compression_loss": 85.19207000732422, "epoch": 1.89, "learning_rate": 1.6046259486808819e-06, "loss": 85.8438, "step": 5243 }, { "compression_loss": 85.16749572753906, "epoch": 1.9, "learning_rate": 1.5992049150704734e-06, "loss": 85.5064, "step": 5244 }, { "compression_loss": 83.82640075683594, "epoch": 1.9, "learning_rate": 1.593783881460065e-06, "loss": 84.1329, "step": 5245 }, { "compression_loss": 83.3785400390625, "epoch": 1.9, "learning_rate": 1.5883628478496568e-06, "loss": 83.6323, "step": 5246 }, { "compression_loss": 85.71038055419922, "epoch": 1.9, "learning_rate": 1.5829418142392482e-06, "loss": 86.1946, "step": 5247 }, { "compression_loss": 86.27008056640625, "epoch": 1.9, "learning_rate": 1.57752078062884e-06, "loss": 86.663, "step": 5248 }, { "compression_loss": 83.87751770019531, "epoch": 1.9, "learning_rate": 1.5720997470184314e-06, "loss": 84.4207, "step": 5249 }, { "compression_loss": 83.47270202636719, "epoch": 1.9, "learning_rate": 1.5666787134080231e-06, "loss": 83.8451, "step": 5250 }, { "epoch": 1.9, "eval_exact_match": 87.1050141911069, "eval_f1": 93.18974326106118, "step": 5250 }, { "compression_loss": 83.84115600585938, "epoch": 1.9, "learning_rate": 1.5612576797976148e-06, "loss": 84.4654, "step": 5251 }, { "compression_loss": 85.08777618408203, "epoch": 1.9, "learning_rate": 1.5558366461872063e-06, "loss": 85.562, "step": 5252 }, { "compression_loss": 84.71387481689453, "epoch": 1.9, "learning_rate": 1.550415612576798e-06, "loss": 85.3102, "step": 5253 }, { "compression_loss": 82.5277099609375, "epoch": 1.9, "learning_rate": 1.5449945789663897e-06, "loss": 82.9076, "step": 5254 }, { "compression_loss": 84.92041015625, "epoch": 1.9, "learning_rate": 1.5395735453559812e-06, "loss": 85.2941, "step": 5255 }, { "compression_loss": 84.06568908691406, "epoch": 1.9, "learning_rate": 1.5341525117455729e-06, "loss": 84.3073, "step": 5256 }, { "compression_loss": 86.12088012695312, "epoch": 1.9, "learning_rate": 1.5287314781351646e-06, "loss": 86.3848, "step": 5257 }, { "compression_loss": 86.62882232666016, "epoch": 1.9, "learning_rate": 1.523310444524756e-06, "loss": 87.2324, "step": 5258 }, { "compression_loss": 85.69722747802734, "epoch": 1.9, "learning_rate": 1.5178894109143477e-06, "loss": 86.0666, "step": 5259 }, { "compression_loss": 83.97392272949219, "epoch": 1.9, "learning_rate": 1.5124683773039392e-06, "loss": 84.4883, "step": 5260 }, { "compression_loss": 84.19926452636719, "epoch": 1.9, "learning_rate": 1.507047343693531e-06, "loss": 84.6344, "step": 5261 }, { "compression_loss": 83.26715087890625, "epoch": 1.9, "learning_rate": 1.5016263100831226e-06, "loss": 83.582, "step": 5262 }, { "compression_loss": 85.85771179199219, "epoch": 1.9, "learning_rate": 1.496205276472714e-06, "loss": 86.1775, "step": 5263 }, { "compression_loss": 84.52201843261719, "epoch": 1.9, "learning_rate": 1.4907842428623058e-06, "loss": 85.0384, "step": 5264 }, { "compression_loss": 84.17476654052734, "epoch": 1.9, "learning_rate": 1.4853632092518975e-06, "loss": 84.5687, "step": 5265 }, { "compression_loss": 83.0846176147461, "epoch": 1.9, "learning_rate": 1.479942175641489e-06, "loss": 83.6197, "step": 5266 }, { "compression_loss": 83.57899475097656, "epoch": 1.9, "learning_rate": 1.4745211420310807e-06, "loss": 84.1222, "step": 5267 }, { "compression_loss": 82.87932586669922, "epoch": 1.9, "learning_rate": 1.4691001084206724e-06, "loss": 83.3896, "step": 5268 }, { "compression_loss": 85.37081909179688, "epoch": 1.9, "learning_rate": 1.4636790748102638e-06, "loss": 86.0408, "step": 5269 }, { "compression_loss": 84.79130554199219, "epoch": 1.9, "learning_rate": 1.4582580411998555e-06, "loss": 85.3596, "step": 5270 }, { "compression_loss": 83.20523071289062, "epoch": 1.9, "learning_rate": 1.452837007589447e-06, "loss": 83.4883, "step": 5271 }, { "compression_loss": 83.00952911376953, "epoch": 1.91, "learning_rate": 1.4474159739790387e-06, "loss": 83.3877, "step": 5272 }, { "compression_loss": 83.73405456542969, "epoch": 1.91, "learning_rate": 1.4419949403686304e-06, "loss": 84.0794, "step": 5273 }, { "compression_loss": 84.40214538574219, "epoch": 1.91, "learning_rate": 1.436573906758222e-06, "loss": 85.009, "step": 5274 }, { "compression_loss": 84.98808288574219, "epoch": 1.91, "learning_rate": 1.4311528731478136e-06, "loss": 85.6482, "step": 5275 }, { "compression_loss": 84.0797348022461, "epoch": 1.91, "learning_rate": 1.4257318395374053e-06, "loss": 84.465, "step": 5276 }, { "compression_loss": 85.11859893798828, "epoch": 1.91, "learning_rate": 1.4203108059269968e-06, "loss": 85.4733, "step": 5277 }, { "compression_loss": 84.62060546875, "epoch": 1.91, "learning_rate": 1.4148897723165885e-06, "loss": 84.9973, "step": 5278 }, { "compression_loss": 85.84456634521484, "epoch": 1.91, "learning_rate": 1.40946873870618e-06, "loss": 86.2969, "step": 5279 }, { "compression_loss": 84.09923553466797, "epoch": 1.91, "learning_rate": 1.4040477050957716e-06, "loss": 84.8156, "step": 5280 }, { "compression_loss": 84.97386169433594, "epoch": 1.91, "learning_rate": 1.3986266714853633e-06, "loss": 85.6276, "step": 5281 }, { "compression_loss": 84.57787322998047, "epoch": 1.91, "learning_rate": 1.3932056378749548e-06, "loss": 85.0772, "step": 5282 }, { "compression_loss": 83.71826934814453, "epoch": 1.91, "learning_rate": 1.3877846042645465e-06, "loss": 84.0289, "step": 5283 }, { "compression_loss": 82.7152328491211, "epoch": 1.91, "learning_rate": 1.3823635706541382e-06, "loss": 83.1229, "step": 5284 }, { "compression_loss": 83.76123046875, "epoch": 1.91, "learning_rate": 1.3769425370437297e-06, "loss": 83.954, "step": 5285 }, { "compression_loss": 82.42694854736328, "epoch": 1.91, "learning_rate": 1.3715215034333214e-06, "loss": 82.8046, "step": 5286 }, { "compression_loss": 84.56501770019531, "epoch": 1.91, "learning_rate": 1.366100469822913e-06, "loss": 85.0981, "step": 5287 }, { "compression_loss": 83.55245971679688, "epoch": 1.91, "learning_rate": 1.3606794362125046e-06, "loss": 83.9359, "step": 5288 }, { "compression_loss": 84.51133728027344, "epoch": 1.91, "learning_rate": 1.3552584026020963e-06, "loss": 84.8345, "step": 5289 }, { "compression_loss": 83.84652709960938, "epoch": 1.91, "learning_rate": 1.3498373689916878e-06, "loss": 84.2842, "step": 5290 }, { "compression_loss": 84.58378601074219, "epoch": 1.91, "learning_rate": 1.3444163353812795e-06, "loss": 85.1731, "step": 5291 }, { "compression_loss": 82.9569091796875, "epoch": 1.91, "learning_rate": 1.3389953017708711e-06, "loss": 83.6456, "step": 5292 }, { "compression_loss": 85.48992156982422, "epoch": 1.91, "learning_rate": 1.3335742681604626e-06, "loss": 86.0742, "step": 5293 }, { "compression_loss": 85.10897827148438, "epoch": 1.91, "learning_rate": 1.3281532345500543e-06, "loss": 85.7011, "step": 5294 }, { "compression_loss": 86.05949401855469, "epoch": 1.91, "learning_rate": 1.322732200939646e-06, "loss": 86.6915, "step": 5295 }, { "compression_loss": 84.47734069824219, "epoch": 1.91, "learning_rate": 1.3173111673292375e-06, "loss": 84.7826, "step": 5296 }, { "compression_loss": 83.60833740234375, "epoch": 1.91, "learning_rate": 1.3118901337188292e-06, "loss": 84.0987, "step": 5297 }, { "compression_loss": 85.45295715332031, "epoch": 1.91, "learning_rate": 1.3064691001084207e-06, "loss": 85.8371, "step": 5298 }, { "compression_loss": 86.001708984375, "epoch": 1.92, "learning_rate": 1.3010480664980124e-06, "loss": 86.5978, "step": 5299 }, { "compression_loss": 85.23898315429688, "epoch": 1.92, "learning_rate": 1.295627032887604e-06, "loss": 85.7321, "step": 5300 }, { "compression_loss": 84.75625610351562, "epoch": 1.92, "learning_rate": 1.2902059992771956e-06, "loss": 85.3744, "step": 5301 }, { "compression_loss": 83.78059387207031, "epoch": 1.92, "learning_rate": 1.2847849656667873e-06, "loss": 84.0832, "step": 5302 }, { "compression_loss": 84.4069595336914, "epoch": 1.92, "learning_rate": 1.279363932056379e-06, "loss": 84.8152, "step": 5303 }, { "compression_loss": 84.23249053955078, "epoch": 1.92, "learning_rate": 1.2739428984459704e-06, "loss": 84.5469, "step": 5304 }, { "compression_loss": 85.24224090576172, "epoch": 1.92, "learning_rate": 1.2685218648355621e-06, "loss": 85.7957, "step": 5305 }, { "compression_loss": 84.15840911865234, "epoch": 1.92, "learning_rate": 1.2631008312251538e-06, "loss": 84.4795, "step": 5306 }, { "compression_loss": 84.31963348388672, "epoch": 1.92, "learning_rate": 1.2576797976147453e-06, "loss": 84.9418, "step": 5307 }, { "compression_loss": 82.063720703125, "epoch": 1.92, "learning_rate": 1.252258764004337e-06, "loss": 82.5331, "step": 5308 }, { "compression_loss": 84.57513427734375, "epoch": 1.92, "learning_rate": 1.2468377303939283e-06, "loss": 85.049, "step": 5309 }, { "compression_loss": 83.58917236328125, "epoch": 1.92, "learning_rate": 1.24141669678352e-06, "loss": 84.0917, "step": 5310 }, { "compression_loss": 86.46162414550781, "epoch": 1.92, "learning_rate": 1.2359956631731117e-06, "loss": 87.0747, "step": 5311 }, { "compression_loss": 84.65818786621094, "epoch": 1.92, "learning_rate": 1.2305746295627031e-06, "loss": 85.2156, "step": 5312 }, { "compression_loss": 85.07428741455078, "epoch": 1.92, "learning_rate": 1.2251535959522948e-06, "loss": 85.4294, "step": 5313 }, { "compression_loss": 84.71595764160156, "epoch": 1.92, "learning_rate": 1.2197325623418865e-06, "loss": 84.9947, "step": 5314 }, { "compression_loss": 85.11295318603516, "epoch": 1.92, "learning_rate": 1.214311528731478e-06, "loss": 85.825, "step": 5315 }, { "compression_loss": 86.14822387695312, "epoch": 1.92, "learning_rate": 1.2088904951210697e-06, "loss": 86.7582, "step": 5316 }, { "compression_loss": 83.96163940429688, "epoch": 1.92, "learning_rate": 1.2034694615106614e-06, "loss": 84.2521, "step": 5317 }, { "compression_loss": 84.64117431640625, "epoch": 1.92, "learning_rate": 1.198048427900253e-06, "loss": 84.967, "step": 5318 }, { "compression_loss": 82.80780792236328, "epoch": 1.92, "learning_rate": 1.1926273942898446e-06, "loss": 83.0979, "step": 5319 }, { "compression_loss": 82.92906188964844, "epoch": 1.92, "learning_rate": 1.187206360679436e-06, "loss": 83.1443, "step": 5320 }, { "compression_loss": 84.08602905273438, "epoch": 1.92, "learning_rate": 1.1817853270690278e-06, "loss": 84.7068, "step": 5321 }, { "compression_loss": 85.27066802978516, "epoch": 1.92, "learning_rate": 1.1763642934586195e-06, "loss": 85.7782, "step": 5322 }, { "compression_loss": 84.5802001953125, "epoch": 1.92, "learning_rate": 1.170943259848211e-06, "loss": 84.851, "step": 5323 }, { "compression_loss": 82.98612976074219, "epoch": 1.92, "learning_rate": 1.1655222262378026e-06, "loss": 83.6543, "step": 5324 }, { "compression_loss": 86.5523910522461, "epoch": 1.92, "learning_rate": 1.1601011926273943e-06, "loss": 87.0724, "step": 5325 }, { "compression_loss": 82.87325286865234, "epoch": 1.92, "learning_rate": 1.1546801590169858e-06, "loss": 83.3758, "step": 5326 }, { "compression_loss": 82.89556884765625, "epoch": 1.93, "learning_rate": 1.1492591254065775e-06, "loss": 83.1543, "step": 5327 }, { "compression_loss": 84.08406066894531, "epoch": 1.93, "learning_rate": 1.143838091796169e-06, "loss": 84.718, "step": 5328 }, { "compression_loss": 84.11192321777344, "epoch": 1.93, "learning_rate": 1.1384170581857607e-06, "loss": 84.7289, "step": 5329 }, { "compression_loss": 82.81732177734375, "epoch": 1.93, "learning_rate": 1.1329960245753524e-06, "loss": 83.3067, "step": 5330 }, { "compression_loss": 84.08668518066406, "epoch": 1.93, "learning_rate": 1.1275749909649439e-06, "loss": 84.6042, "step": 5331 }, { "compression_loss": 85.15028381347656, "epoch": 1.93, "learning_rate": 1.1221539573545356e-06, "loss": 85.7099, "step": 5332 }, { "compression_loss": 84.84949493408203, "epoch": 1.93, "learning_rate": 1.1167329237441273e-06, "loss": 85.4098, "step": 5333 }, { "compression_loss": 85.85730743408203, "epoch": 1.93, "learning_rate": 1.1113118901337188e-06, "loss": 86.3152, "step": 5334 }, { "compression_loss": 83.85482788085938, "epoch": 1.93, "learning_rate": 1.1058908565233104e-06, "loss": 84.0964, "step": 5335 }, { "compression_loss": 84.66624450683594, "epoch": 1.93, "learning_rate": 1.1004698229129021e-06, "loss": 85.0332, "step": 5336 }, { "compression_loss": 84.12513732910156, "epoch": 1.93, "learning_rate": 1.0950487893024936e-06, "loss": 85.177, "step": 5337 }, { "compression_loss": 84.95819091796875, "epoch": 1.93, "learning_rate": 1.0896277556920853e-06, "loss": 85.4491, "step": 5338 }, { "compression_loss": 85.67388916015625, "epoch": 1.93, "learning_rate": 1.0842067220816768e-06, "loss": 86.2105, "step": 5339 }, { "compression_loss": 84.20238494873047, "epoch": 1.93, "learning_rate": 1.0787856884712685e-06, "loss": 84.7057, "step": 5340 }, { "compression_loss": 85.39675903320312, "epoch": 1.93, "learning_rate": 1.0733646548608602e-06, "loss": 85.9526, "step": 5341 }, { "compression_loss": 84.04940795898438, "epoch": 1.93, "learning_rate": 1.0679436212504517e-06, "loss": 84.5866, "step": 5342 }, { "compression_loss": 83.61771392822266, "epoch": 1.93, "learning_rate": 1.0625225876400434e-06, "loss": 83.8238, "step": 5343 }, { "compression_loss": 83.11016082763672, "epoch": 1.93, "learning_rate": 1.057101554029635e-06, "loss": 83.328, "step": 5344 }, { "compression_loss": 83.69354248046875, "epoch": 1.93, "learning_rate": 1.0516805204192266e-06, "loss": 84.099, "step": 5345 }, { "compression_loss": 84.90116882324219, "epoch": 1.93, "learning_rate": 1.0462594868088182e-06, "loss": 85.315, "step": 5346 }, { "compression_loss": 85.3209228515625, "epoch": 1.93, "learning_rate": 1.04083845319841e-06, "loss": 85.906, "step": 5347 }, { "compression_loss": 85.26416015625, "epoch": 1.93, "learning_rate": 1.0354174195880014e-06, "loss": 85.8727, "step": 5348 }, { "compression_loss": 84.66349792480469, "epoch": 1.93, "learning_rate": 1.0299963859775931e-06, "loss": 85.3767, "step": 5349 }, { "compression_loss": 83.86813354492188, "epoch": 1.93, "learning_rate": 1.0245753523671846e-06, "loss": 84.3191, "step": 5350 }, { "compression_loss": 83.32635498046875, "epoch": 1.93, "learning_rate": 1.0191543187567763e-06, "loss": 83.766, "step": 5351 }, { "compression_loss": 85.22666931152344, "epoch": 1.93, "learning_rate": 1.013733285146368e-06, "loss": 85.6392, "step": 5352 }, { "compression_loss": 85.59127044677734, "epoch": 1.93, "learning_rate": 1.0083122515359595e-06, "loss": 86.2634, "step": 5353 }, { "compression_loss": 85.48540496826172, "epoch": 1.93, "learning_rate": 1.0028912179255512e-06, "loss": 85.9938, "step": 5354 }, { "compression_loss": 83.41757202148438, "epoch": 1.94, "learning_rate": 9.974701843151429e-07, "loss": 83.8619, "step": 5355 }, { "compression_loss": 83.49745178222656, "epoch": 1.94, "learning_rate": 9.920491507047344e-07, "loss": 84.2404, "step": 5356 }, { "compression_loss": 84.08087921142578, "epoch": 1.94, "learning_rate": 9.86628117094326e-07, "loss": 84.6039, "step": 5357 }, { "compression_loss": 84.44747161865234, "epoch": 1.94, "learning_rate": 9.812070834839175e-07, "loss": 84.9138, "step": 5358 }, { "compression_loss": 85.66224670410156, "epoch": 1.94, "learning_rate": 9.757860498735092e-07, "loss": 86.0796, "step": 5359 }, { "compression_loss": 85.04256439208984, "epoch": 1.94, "learning_rate": 9.70365016263101e-07, "loss": 85.3258, "step": 5360 }, { "compression_loss": 85.14508056640625, "epoch": 1.94, "learning_rate": 9.649439826526924e-07, "loss": 85.5754, "step": 5361 }, { "compression_loss": 83.01295471191406, "epoch": 1.94, "learning_rate": 9.59522949042284e-07, "loss": 83.3607, "step": 5362 }, { "compression_loss": 83.5160140991211, "epoch": 1.94, "learning_rate": 9.541019154318758e-07, "loss": 84.0127, "step": 5363 }, { "compression_loss": 82.71249389648438, "epoch": 1.94, "learning_rate": 9.486808818214674e-07, "loss": 83.019, "step": 5364 }, { "compression_loss": 83.77801513671875, "epoch": 1.94, "learning_rate": 9.43259848211059e-07, "loss": 84.1051, "step": 5365 }, { "compression_loss": 85.66883850097656, "epoch": 1.94, "learning_rate": 9.378388146006506e-07, "loss": 85.9815, "step": 5366 }, { "compression_loss": 83.656005859375, "epoch": 1.94, "learning_rate": 9.324177809902422e-07, "loss": 84.1203, "step": 5367 }, { "compression_loss": 84.76080322265625, "epoch": 1.94, "learning_rate": 9.269967473798339e-07, "loss": 85.3565, "step": 5368 }, { "compression_loss": 84.36833953857422, "epoch": 1.94, "learning_rate": 9.215757137694254e-07, "loss": 84.727, "step": 5369 }, { "compression_loss": 83.70655822753906, "epoch": 1.94, "learning_rate": 9.16154680159017e-07, "loss": 84.0102, "step": 5370 }, { "compression_loss": 85.21450805664062, "epoch": 1.94, "learning_rate": 9.107336465486086e-07, "loss": 85.5024, "step": 5371 }, { "compression_loss": 84.36210632324219, "epoch": 1.94, "learning_rate": 9.053126129382003e-07, "loss": 85.168, "step": 5372 }, { "compression_loss": 84.19454956054688, "epoch": 1.94, "learning_rate": 8.998915793277919e-07, "loss": 84.87, "step": 5373 }, { "compression_loss": 83.39297485351562, "epoch": 1.94, "learning_rate": 8.944705457173835e-07, "loss": 83.6945, "step": 5374 }, { "compression_loss": 84.33881378173828, "epoch": 1.94, "learning_rate": 8.890495121069751e-07, "loss": 85.0785, "step": 5375 }, { "compression_loss": 85.08647155761719, "epoch": 1.94, "learning_rate": 8.836284784965668e-07, "loss": 85.3589, "step": 5376 }, { "compression_loss": 85.84593200683594, "epoch": 1.94, "learning_rate": 8.782074448861584e-07, "loss": 86.3526, "step": 5377 }, { "compression_loss": 83.7976303100586, "epoch": 1.94, "learning_rate": 8.727864112757499e-07, "loss": 84.3641, "step": 5378 }, { "compression_loss": 85.05714416503906, "epoch": 1.94, "learning_rate": 8.673653776653416e-07, "loss": 85.5709, "step": 5379 }, { "compression_loss": 84.74470520019531, "epoch": 1.94, "learning_rate": 8.619443440549331e-07, "loss": 85.0575, "step": 5380 }, { "compression_loss": 83.05547332763672, "epoch": 1.94, "learning_rate": 8.565233104445247e-07, "loss": 83.7577, "step": 5381 }, { "compression_loss": 84.66844177246094, "epoch": 1.95, "learning_rate": 8.511022768341163e-07, "loss": 85.1544, "step": 5382 }, { "compression_loss": 83.93991088867188, "epoch": 1.95, "learning_rate": 8.45681243223708e-07, "loss": 84.3356, "step": 5383 }, { "compression_loss": 84.0047607421875, "epoch": 1.95, "learning_rate": 8.402602096132996e-07, "loss": 84.4069, "step": 5384 }, { "compression_loss": 84.89237976074219, "epoch": 1.95, "learning_rate": 8.348391760028912e-07, "loss": 85.537, "step": 5385 }, { "compression_loss": 84.0667953491211, "epoch": 1.95, "learning_rate": 8.294181423924828e-07, "loss": 84.6197, "step": 5386 }, { "compression_loss": 84.71391296386719, "epoch": 1.95, "learning_rate": 8.239971087820745e-07, "loss": 85.0695, "step": 5387 }, { "compression_loss": 85.150390625, "epoch": 1.95, "learning_rate": 8.185760751716661e-07, "loss": 85.6615, "step": 5388 }, { "compression_loss": 83.42566680908203, "epoch": 1.95, "learning_rate": 8.131550415612577e-07, "loss": 83.7634, "step": 5389 }, { "compression_loss": 84.61922454833984, "epoch": 1.95, "learning_rate": 8.077340079508492e-07, "loss": 85.1769, "step": 5390 }, { "compression_loss": 83.58465576171875, "epoch": 1.95, "learning_rate": 8.023129743404409e-07, "loss": 83.9361, "step": 5391 }, { "compression_loss": 83.58009338378906, "epoch": 1.95, "learning_rate": 7.968919407300325e-07, "loss": 83.9893, "step": 5392 }, { "compression_loss": 82.92304992675781, "epoch": 1.95, "learning_rate": 7.914709071196241e-07, "loss": 83.2253, "step": 5393 }, { "compression_loss": 84.18940734863281, "epoch": 1.95, "learning_rate": 7.860498735092157e-07, "loss": 84.4896, "step": 5394 }, { "compression_loss": 86.34056091308594, "epoch": 1.95, "learning_rate": 7.806288398988074e-07, "loss": 86.9697, "step": 5395 }, { "compression_loss": 84.6645736694336, "epoch": 1.95, "learning_rate": 7.75207806288399e-07, "loss": 84.8567, "step": 5396 }, { "compression_loss": 83.19134521484375, "epoch": 1.95, "learning_rate": 7.697867726779906e-07, "loss": 83.5117, "step": 5397 }, { "compression_loss": 85.01611328125, "epoch": 1.95, "learning_rate": 7.643657390675823e-07, "loss": 85.3756, "step": 5398 }, { "compression_loss": 84.05033874511719, "epoch": 1.95, "learning_rate": 7.589447054571739e-07, "loss": 84.3774, "step": 5399 }, { "compression_loss": 84.23646545410156, "epoch": 1.95, "learning_rate": 7.535236718467655e-07, "loss": 84.6425, "step": 5400 }, { "compression_loss": 86.44300842285156, "epoch": 1.95, "learning_rate": 7.48102638236357e-07, "loss": 87.1758, "step": 5401 }, { "compression_loss": 85.18597412109375, "epoch": 1.95, "learning_rate": 7.426816046259487e-07, "loss": 85.508, "step": 5402 }, { "compression_loss": 82.9363784790039, "epoch": 1.95, "learning_rate": 7.372605710155403e-07, "loss": 83.3852, "step": 5403 }, { "compression_loss": 85.68624114990234, "epoch": 1.95, "learning_rate": 7.318395374051319e-07, "loss": 86.1375, "step": 5404 }, { "compression_loss": 81.79042053222656, "epoch": 1.95, "learning_rate": 7.264185037947235e-07, "loss": 82.4828, "step": 5405 }, { "compression_loss": 85.16069030761719, "epoch": 1.95, "learning_rate": 7.209974701843152e-07, "loss": 85.5611, "step": 5406 }, { "compression_loss": 83.5752182006836, "epoch": 1.95, "learning_rate": 7.155764365739068e-07, "loss": 84.0767, "step": 5407 }, { "compression_loss": 84.23934936523438, "epoch": 1.95, "learning_rate": 7.101554029634984e-07, "loss": 84.6216, "step": 5408 }, { "compression_loss": 85.19570922851562, "epoch": 1.95, "learning_rate": 7.0473436935309e-07, "loss": 85.6213, "step": 5409 }, { "compression_loss": 84.0433349609375, "epoch": 1.96, "learning_rate": 6.993133357426817e-07, "loss": 84.4666, "step": 5410 }, { "compression_loss": 84.37173461914062, "epoch": 1.96, "learning_rate": 6.938923021322733e-07, "loss": 84.7286, "step": 5411 }, { "compression_loss": 85.09383392333984, "epoch": 1.96, "learning_rate": 6.884712685218649e-07, "loss": 85.6695, "step": 5412 }, { "compression_loss": 85.01918029785156, "epoch": 1.96, "learning_rate": 6.830502349114565e-07, "loss": 85.8465, "step": 5413 }, { "compression_loss": 82.87489318847656, "epoch": 1.96, "learning_rate": 6.776292013010481e-07, "loss": 83.3183, "step": 5414 }, { "compression_loss": 83.50031280517578, "epoch": 1.96, "learning_rate": 6.722081676906397e-07, "loss": 83.9157, "step": 5415 }, { "compression_loss": 84.97298431396484, "epoch": 1.96, "learning_rate": 6.667871340802313e-07, "loss": 85.5454, "step": 5416 }, { "compression_loss": 84.08895874023438, "epoch": 1.96, "learning_rate": 6.61366100469823e-07, "loss": 84.559, "step": 5417 }, { "compression_loss": 85.50928497314453, "epoch": 1.96, "learning_rate": 6.559450668594146e-07, "loss": 85.7462, "step": 5418 }, { "compression_loss": 84.64221954345703, "epoch": 1.96, "learning_rate": 6.505240332490062e-07, "loss": 84.9824, "step": 5419 }, { "compression_loss": 83.49061584472656, "epoch": 1.96, "learning_rate": 6.451029996385978e-07, "loss": 83.8433, "step": 5420 }, { "compression_loss": 84.73204040527344, "epoch": 1.96, "learning_rate": 6.396819660281895e-07, "loss": 85.1327, "step": 5421 }, { "compression_loss": 84.71521759033203, "epoch": 1.96, "learning_rate": 6.342609324177811e-07, "loss": 85.1703, "step": 5422 }, { "compression_loss": 83.84538269042969, "epoch": 1.96, "learning_rate": 6.288398988073727e-07, "loss": 84.0842, "step": 5423 }, { "compression_loss": 84.80558776855469, "epoch": 1.96, "learning_rate": 6.234188651969641e-07, "loss": 85.3659, "step": 5424 }, { "compression_loss": 86.27490234375, "epoch": 1.96, "learning_rate": 6.179978315865558e-07, "loss": 86.7444, "step": 5425 }, { "compression_loss": 83.63198852539062, "epoch": 1.96, "learning_rate": 6.125767979761474e-07, "loss": 83.9705, "step": 5426 }, { "compression_loss": 85.13893127441406, "epoch": 1.96, "learning_rate": 6.07155764365739e-07, "loss": 85.6186, "step": 5427 }, { "compression_loss": 85.17750549316406, "epoch": 1.96, "learning_rate": 6.017347307553307e-07, "loss": 85.8911, "step": 5428 }, { "compression_loss": 84.12359619140625, "epoch": 1.96, "learning_rate": 5.963136971449223e-07, "loss": 84.7793, "step": 5429 }, { "compression_loss": 84.89520263671875, "epoch": 1.96, "learning_rate": 5.908926635345139e-07, "loss": 85.6173, "step": 5430 }, { "compression_loss": 84.55392456054688, "epoch": 1.96, "learning_rate": 5.854716299241055e-07, "loss": 84.9272, "step": 5431 }, { "compression_loss": 84.21942138671875, "epoch": 1.96, "learning_rate": 5.800505963136972e-07, "loss": 84.6856, "step": 5432 }, { "compression_loss": 83.32350158691406, "epoch": 1.96, "learning_rate": 5.746295627032888e-07, "loss": 83.7883, "step": 5433 }, { "compression_loss": 85.04350280761719, "epoch": 1.96, "learning_rate": 5.692085290928803e-07, "loss": 85.3717, "step": 5434 }, { "compression_loss": 84.72268676757812, "epoch": 1.96, "learning_rate": 5.637874954824719e-07, "loss": 85.0333, "step": 5435 }, { "compression_loss": 83.82670593261719, "epoch": 1.96, "learning_rate": 5.583664618720636e-07, "loss": 84.2571, "step": 5436 }, { "compression_loss": 82.56675720214844, "epoch": 1.96, "learning_rate": 5.529454282616552e-07, "loss": 82.9838, "step": 5437 }, { "compression_loss": 83.76380157470703, "epoch": 1.97, "learning_rate": 5.475243946512468e-07, "loss": 84.0988, "step": 5438 }, { "compression_loss": 83.55925750732422, "epoch": 1.97, "learning_rate": 5.421033610408384e-07, "loss": 84.1128, "step": 5439 }, { "compression_loss": 83.88985443115234, "epoch": 1.97, "learning_rate": 5.366823274304301e-07, "loss": 84.2021, "step": 5440 }, { "compression_loss": 84.67516326904297, "epoch": 1.97, "learning_rate": 5.312612938200217e-07, "loss": 84.9399, "step": 5441 }, { "compression_loss": 82.48990631103516, "epoch": 1.97, "learning_rate": 5.258402602096133e-07, "loss": 82.716, "step": 5442 }, { "compression_loss": 85.75113677978516, "epoch": 1.97, "learning_rate": 5.20419226599205e-07, "loss": 86.4468, "step": 5443 }, { "compression_loss": 83.66617584228516, "epoch": 1.97, "learning_rate": 5.149981929887966e-07, "loss": 84.0539, "step": 5444 }, { "compression_loss": 84.53353118896484, "epoch": 1.97, "learning_rate": 5.095771593783882e-07, "loss": 85.2214, "step": 5445 }, { "compression_loss": 83.88858795166016, "epoch": 1.97, "learning_rate": 5.041561257679797e-07, "loss": 84.5883, "step": 5446 }, { "compression_loss": 83.14529418945312, "epoch": 1.97, "learning_rate": 4.987350921575714e-07, "loss": 83.5187, "step": 5447 }, { "compression_loss": 85.91169738769531, "epoch": 1.97, "learning_rate": 4.93314058547163e-07, "loss": 86.3426, "step": 5448 }, { "compression_loss": 84.02717590332031, "epoch": 1.97, "learning_rate": 4.878930249367546e-07, "loss": 84.3213, "step": 5449 }, { "compression_loss": 82.89852905273438, "epoch": 1.97, "learning_rate": 4.824719913263462e-07, "loss": 83.2118, "step": 5450 }, { "compression_loss": 84.4688720703125, "epoch": 1.97, "learning_rate": 4.770509577159379e-07, "loss": 84.9469, "step": 5451 }, { "compression_loss": 85.16209411621094, "epoch": 1.97, "learning_rate": 4.716299241055295e-07, "loss": 85.9014, "step": 5452 }, { "compression_loss": 83.22344970703125, "epoch": 1.97, "learning_rate": 4.662088904951211e-07, "loss": 83.7122, "step": 5453 }, { "compression_loss": 83.7003173828125, "epoch": 1.97, "learning_rate": 4.607878568847127e-07, "loss": 84.3488, "step": 5454 }, { "compression_loss": 83.76081848144531, "epoch": 1.97, "learning_rate": 4.553668232743043e-07, "loss": 84.1707, "step": 5455 }, { "compression_loss": 84.63510131835938, "epoch": 1.97, "learning_rate": 4.4994578966389595e-07, "loss": 85.2542, "step": 5456 }, { "compression_loss": 85.10285949707031, "epoch": 1.97, "learning_rate": 4.4452475605348754e-07, "loss": 85.8139, "step": 5457 }, { "compression_loss": 84.25220489501953, "epoch": 1.97, "learning_rate": 4.391037224430792e-07, "loss": 84.7465, "step": 5458 }, { "compression_loss": 82.99684143066406, "epoch": 1.97, "learning_rate": 4.336826888326708e-07, "loss": 83.3819, "step": 5459 }, { "compression_loss": 84.00447845458984, "epoch": 1.97, "learning_rate": 4.2826165522226236e-07, "loss": 84.4774, "step": 5460 }, { "compression_loss": 84.18604278564453, "epoch": 1.97, "learning_rate": 4.22840621611854e-07, "loss": 84.7395, "step": 5461 }, { "compression_loss": 84.85175323486328, "epoch": 1.97, "learning_rate": 4.174195880014456e-07, "loss": 85.2286, "step": 5462 }, { "compression_loss": 83.65498352050781, "epoch": 1.97, "learning_rate": 4.1199855439103724e-07, "loss": 83.9637, "step": 5463 }, { "compression_loss": 85.82032775878906, "epoch": 1.97, "learning_rate": 4.0657752078062883e-07, "loss": 86.1525, "step": 5464 }, { "compression_loss": 86.10971069335938, "epoch": 1.98, "learning_rate": 4.0115648717022047e-07, "loss": 86.6462, "step": 5465 }, { "compression_loss": 84.83535766601562, "epoch": 1.98, "learning_rate": 3.9573545355981206e-07, "loss": 85.4172, "step": 5466 }, { "compression_loss": 85.10771179199219, "epoch": 1.98, "learning_rate": 3.903144199494037e-07, "loss": 85.4996, "step": 5467 }, { "compression_loss": 82.93553161621094, "epoch": 1.98, "learning_rate": 3.848933863389953e-07, "loss": 83.192, "step": 5468 }, { "compression_loss": 83.2828369140625, "epoch": 1.98, "learning_rate": 3.7947235272858693e-07, "loss": 83.7255, "step": 5469 }, { "compression_loss": 83.53829193115234, "epoch": 1.98, "learning_rate": 3.740513191181785e-07, "loss": 83.9261, "step": 5470 }, { "compression_loss": 85.0662841796875, "epoch": 1.98, "learning_rate": 3.6863028550777017e-07, "loss": 85.4436, "step": 5471 }, { "compression_loss": 86.68266296386719, "epoch": 1.98, "learning_rate": 3.6320925189736176e-07, "loss": 87.453, "step": 5472 }, { "compression_loss": 84.47545623779297, "epoch": 1.98, "learning_rate": 3.577882182869534e-07, "loss": 84.8298, "step": 5473 }, { "compression_loss": 84.44767761230469, "epoch": 1.98, "learning_rate": 3.52367184676545e-07, "loss": 84.8102, "step": 5474 }, { "compression_loss": 84.22109985351562, "epoch": 1.98, "learning_rate": 3.4694615106613663e-07, "loss": 84.5191, "step": 5475 }, { "compression_loss": 84.83287048339844, "epoch": 1.98, "learning_rate": 3.4152511745572827e-07, "loss": 85.2059, "step": 5476 }, { "compression_loss": 83.66520690917969, "epoch": 1.98, "learning_rate": 3.3610408384531986e-07, "loss": 83.9838, "step": 5477 }, { "compression_loss": 84.486083984375, "epoch": 1.98, "learning_rate": 3.306830502349115e-07, "loss": 84.8949, "step": 5478 }, { "compression_loss": 85.54335021972656, "epoch": 1.98, "learning_rate": 3.252620166245031e-07, "loss": 86.1932, "step": 5479 }, { "compression_loss": 85.59603881835938, "epoch": 1.98, "learning_rate": 3.1984098301409474e-07, "loss": 85.9915, "step": 5480 }, { "compression_loss": 85.32963562011719, "epoch": 1.98, "learning_rate": 3.1441994940368633e-07, "loss": 85.6943, "step": 5481 }, { "compression_loss": 82.73797607421875, "epoch": 1.98, "learning_rate": 3.089989157932779e-07, "loss": 83.1647, "step": 5482 }, { "compression_loss": 84.78305053710938, "epoch": 1.98, "learning_rate": 3.035778821828695e-07, "loss": 85.4134, "step": 5483 }, { "compression_loss": 83.17288970947266, "epoch": 1.98, "learning_rate": 2.9815684857246115e-07, "loss": 83.8807, "step": 5484 }, { "compression_loss": 84.81248474121094, "epoch": 1.98, "learning_rate": 2.9273581496205274e-07, "loss": 85.2634, "step": 5485 }, { "compression_loss": 85.46653747558594, "epoch": 1.98, "learning_rate": 2.873147813516444e-07, "loss": 86.0312, "step": 5486 }, { "compression_loss": 85.76657104492188, "epoch": 1.98, "learning_rate": 2.8189374774123597e-07, "loss": 86.2278, "step": 5487 }, { "compression_loss": 83.86209869384766, "epoch": 1.98, "learning_rate": 2.764727141308276e-07, "loss": 84.2312, "step": 5488 }, { "compression_loss": 83.74267578125, "epoch": 1.98, "learning_rate": 2.710516805204192e-07, "loss": 84.1015, "step": 5489 }, { "compression_loss": 83.55816650390625, "epoch": 1.98, "learning_rate": 2.6563064691001084e-07, "loss": 84.2264, "step": 5490 }, { "compression_loss": 84.34867858886719, "epoch": 1.98, "learning_rate": 2.602096132996025e-07, "loss": 84.7514, "step": 5491 }, { "compression_loss": 84.45734405517578, "epoch": 1.98, "learning_rate": 2.547885796891941e-07, "loss": 84.8989, "step": 5492 }, { "compression_loss": 84.43909454345703, "epoch": 1.99, "learning_rate": 2.493675460787857e-07, "loss": 84.7447, "step": 5493 }, { "compression_loss": 84.24634552001953, "epoch": 1.99, "learning_rate": 2.439465124683773e-07, "loss": 84.7649, "step": 5494 }, { "compression_loss": 81.69123077392578, "epoch": 1.99, "learning_rate": 2.3852547885796895e-07, "loss": 81.9646, "step": 5495 }, { "compression_loss": 83.83653259277344, "epoch": 1.99, "learning_rate": 2.3310444524756054e-07, "loss": 84.3819, "step": 5496 }, { "compression_loss": 84.64825439453125, "epoch": 1.99, "learning_rate": 2.2768341163715216e-07, "loss": 85.0649, "step": 5497 }, { "compression_loss": 83.32498168945312, "epoch": 1.99, "learning_rate": 2.2226237802674377e-07, "loss": 83.6084, "step": 5498 }, { "compression_loss": 83.63139343261719, "epoch": 1.99, "learning_rate": 2.168413444163354e-07, "loss": 83.8018, "step": 5499 }, { "compression_loss": 84.41525268554688, "epoch": 1.99, "learning_rate": 2.11420310805927e-07, "loss": 85.0219, "step": 5500 }, { "epoch": 1.99, "eval_exact_match": 87.05771050141911, "eval_f1": 93.10858733407407, "step": 5500 }, { "compression_loss": 83.65408325195312, "epoch": 1.99, "learning_rate": 2.0599927719551862e-07, "loss": 84.137, "step": 5501 }, { "compression_loss": 84.80571746826172, "epoch": 1.99, "learning_rate": 2.0057824358511024e-07, "loss": 85.0704, "step": 5502 }, { "compression_loss": 84.16047668457031, "epoch": 1.99, "learning_rate": 1.9515720997470185e-07, "loss": 84.6466, "step": 5503 }, { "compression_loss": 83.14498138427734, "epoch": 1.99, "learning_rate": 1.8973617636429347e-07, "loss": 83.4423, "step": 5504 }, { "compression_loss": 84.10314178466797, "epoch": 1.99, "learning_rate": 1.8431514275388508e-07, "loss": 84.7087, "step": 5505 }, { "compression_loss": 83.357177734375, "epoch": 1.99, "learning_rate": 1.788941091434767e-07, "loss": 84.0105, "step": 5506 }, { "compression_loss": 83.79454040527344, "epoch": 1.99, "learning_rate": 1.7347307553306832e-07, "loss": 84.162, "step": 5507 }, { "compression_loss": 84.55569458007812, "epoch": 1.99, "learning_rate": 1.6805204192265993e-07, "loss": 84.9884, "step": 5508 }, { "compression_loss": 83.35787200927734, "epoch": 1.99, "learning_rate": 1.6263100831225155e-07, "loss": 83.9411, "step": 5509 }, { "compression_loss": 85.7645263671875, "epoch": 1.99, "learning_rate": 1.5720997470184316e-07, "loss": 86.3988, "step": 5510 }, { "compression_loss": 85.14952087402344, "epoch": 1.99, "learning_rate": 1.5178894109143475e-07, "loss": 85.4751, "step": 5511 }, { "compression_loss": 84.97509002685547, "epoch": 1.99, "learning_rate": 1.4636790748102637e-07, "loss": 85.8259, "step": 5512 }, { "compression_loss": 83.24527740478516, "epoch": 1.99, "learning_rate": 1.4094687387061798e-07, "loss": 83.5303, "step": 5513 }, { "compression_loss": 84.91810607910156, "epoch": 1.99, "learning_rate": 1.355258402602096e-07, "loss": 85.433, "step": 5514 }, { "compression_loss": 84.66250610351562, "epoch": 1.99, "learning_rate": 1.3010480664980124e-07, "loss": 85.0476, "step": 5515 }, { "compression_loss": 83.26564025878906, "epoch": 1.99, "learning_rate": 1.2468377303939286e-07, "loss": 83.7635, "step": 5516 }, { "compression_loss": 84.1217041015625, "epoch": 1.99, "learning_rate": 1.1926273942898448e-07, "loss": 84.4746, "step": 5517 }, { "compression_loss": 83.07740783691406, "epoch": 1.99, "learning_rate": 1.1384170581857608e-07, "loss": 83.4009, "step": 5518 }, { "compression_loss": 83.21170043945312, "epoch": 1.99, "learning_rate": 1.084206722081677e-07, "loss": 83.747, "step": 5519 }, { "compression_loss": 83.40278625488281, "epoch": 1.99, "learning_rate": 1.0299963859775931e-07, "loss": 83.6132, "step": 5520 }, { "compression_loss": 85.95453643798828, "epoch": 2.0, "learning_rate": 9.757860498735093e-08, "loss": 86.4728, "step": 5521 }, { "compression_loss": 83.88064575195312, "epoch": 2.0, "learning_rate": 9.215757137694254e-08, "loss": 84.2485, "step": 5522 }, { "compression_loss": 84.64408874511719, "epoch": 2.0, "learning_rate": 8.673653776653416e-08, "loss": 85.0533, "step": 5523 }, { "compression_loss": 82.95278930664062, "epoch": 2.0, "learning_rate": 8.131550415612577e-08, "loss": 83.5095, "step": 5524 }, { "compression_loss": 82.65150451660156, "epoch": 2.0, "learning_rate": 7.589447054571738e-08, "loss": 82.9105, "step": 5525 }, { "compression_loss": 83.65154266357422, "epoch": 2.0, "learning_rate": 7.047343693530899e-08, "loss": 84.1816, "step": 5526 }, { "compression_loss": 84.77619934082031, "epoch": 2.0, "learning_rate": 6.505240332490062e-08, "loss": 85.4092, "step": 5527 }, { "compression_loss": 85.20742797851562, "epoch": 2.0, "learning_rate": 5.963136971449224e-08, "loss": 85.4676, "step": 5528 }, { "compression_loss": 84.9150619506836, "epoch": 2.0, "learning_rate": 5.421033610408385e-08, "loss": 85.2624, "step": 5529 }, { "compression_loss": 84.0916976928711, "epoch": 2.0, "learning_rate": 4.878930249367546e-08, "loss": 84.5659, "step": 5530 }, { "compression_loss": 82.33705139160156, "epoch": 2.0, "learning_rate": 4.336826888326708e-08, "loss": 82.6642, "step": 5531 }, { "compression_loss": 84.77468872070312, "epoch": 2.0, "learning_rate": 3.794723527285869e-08, "loss": 85.1802, "step": 5532 }, { "compression_loss": 84.41414642333984, "epoch": 2.0, "learning_rate": 3.252620166245031e-08, "loss": 84.8247, "step": 5533 }, { "compression_loss": 86.73825073242188, "epoch": 2.0, "learning_rate": 2.7105168052041923e-08, "loss": 87.2081, "step": 5534 }, { "epoch": 2.0, "step": 5534, "total_flos": 2.4663914358360883e+17, "train_loss": 84.96330691090465, "train_runtime": 25554.622, "train_samples_per_second": 6.928, "train_steps_per_second": 0.217 } ], "max_steps": 5534, "num_train_epochs": 2, "total_flos": 2.4663914358360883e+17, "trial_name": null, "trial_params": null }