{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.4783180026281209, "global_step": 4500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "compression_loss": 83.32598876953125, "epoch": 0.0, "learning_rate": 3e-05, "loss": 84.3938, "step": 1 }, { "compression_loss": 86.45396423339844, "epoch": 0.0, "learning_rate": 3e-05, "loss": 88.2657, "step": 2 }, { "compression_loss": 84.32394409179688, "epoch": 0.0, "learning_rate": 2.9995072273324573e-05, "loss": 85.481, "step": 3 }, { "compression_loss": 77.87857818603516, "epoch": 0.0, "learning_rate": 2.9990144546649145e-05, "loss": 78.7099, "step": 4 }, { "compression_loss": 78.03010559082031, "epoch": 0.0, "learning_rate": 2.998521681997372e-05, "loss": 78.9405, "step": 5 }, { "compression_loss": 75.65208435058594, "epoch": 0.0, "learning_rate": 2.9980289093298292e-05, "loss": 76.3265, "step": 6 }, { "compression_loss": 79.19513702392578, "epoch": 0.0, "learning_rate": 2.9975361366622864e-05, "loss": 80.3283, "step": 7 }, { "compression_loss": 77.57250213623047, "epoch": 0.0, "learning_rate": 2.997043363994744e-05, "loss": 78.2791, "step": 8 }, { "compression_loss": 80.52154541015625, "epoch": 0.0, "learning_rate": 2.996550591327201e-05, "loss": 81.5264, "step": 9 }, { "compression_loss": 76.77779388427734, "epoch": 0.0, "learning_rate": 2.9960578186596587e-05, "loss": 77.3498, "step": 10 }, { "compression_loss": 78.97022247314453, "epoch": 0.0, "learning_rate": 2.995565045992116e-05, "loss": 79.6189, "step": 11 }, { "compression_loss": 75.529052734375, "epoch": 0.0, "learning_rate": 2.995072273324573e-05, "loss": 76.2886, "step": 12 }, { "compression_loss": 77.06999206542969, "epoch": 0.0, "learning_rate": 2.995072273324573e-05, "loss": 78.1238, "step": 13 }, { "compression_loss": 77.47019958496094, "epoch": 0.0, "learning_rate": 2.9945795006570303e-05, "loss": 78.4102, "step": 14 }, { "compression_loss": 76.24961853027344, "epoch": 0.0, "learning_rate": 2.9940867279894875e-05, "loss": 76.8266, "step": 15 }, { "compression_loss": 76.3663101196289, "epoch": 0.01, "learning_rate": 2.993593955321945e-05, "loss": 77.1729, "step": 16 }, { "compression_loss": 78.81253814697266, "epoch": 0.01, "learning_rate": 2.9931011826544022e-05, "loss": 79.9488, "step": 17 }, { "compression_loss": 79.33038330078125, "epoch": 0.01, "learning_rate": 2.9926084099868594e-05, "loss": 80.411, "step": 18 }, { "compression_loss": 76.08391571044922, "epoch": 0.01, "learning_rate": 2.9921156373193166e-05, "loss": 76.6883, "step": 19 }, { "compression_loss": 79.41412353515625, "epoch": 0.01, "learning_rate": 2.9916228646517738e-05, "loss": 80.5062, "step": 20 }, { "compression_loss": 76.67149353027344, "epoch": 0.01, "learning_rate": 2.9911300919842317e-05, "loss": 77.2514, "step": 21 }, { "compression_loss": 75.55659484863281, "epoch": 0.01, "learning_rate": 2.990637319316689e-05, "loss": 76.0355, "step": 22 }, { "compression_loss": 77.50818634033203, "epoch": 0.01, "learning_rate": 2.990144546649146e-05, "loss": 78.1316, "step": 23 }, { "compression_loss": 76.25888061523438, "epoch": 0.01, "learning_rate": 2.9896517739816032e-05, "loss": 76.9676, "step": 24 }, { "compression_loss": 77.77111053466797, "epoch": 0.01, "learning_rate": 2.9891590013140604e-05, "loss": 78.5375, "step": 25 }, { "compression_loss": 78.44304656982422, "epoch": 0.01, "learning_rate": 2.988666228646518e-05, "loss": 79.2089, "step": 26 }, { "compression_loss": 76.75584411621094, "epoch": 0.01, "learning_rate": 2.9881734559789752e-05, "loss": 77.2358, "step": 27 }, { "compression_loss": 75.962646484375, "epoch": 0.01, "learning_rate": 2.9876806833114324e-05, "loss": 76.4623, "step": 28 }, { "compression_loss": 78.2991943359375, "epoch": 0.01, "learning_rate": 2.9871879106438896e-05, "loss": 78.842, "step": 29 }, { "compression_loss": 75.40074157714844, "epoch": 0.01, "learning_rate": 2.9866951379763468e-05, "loss": 75.9163, "step": 30 }, { "compression_loss": 76.7127456665039, "epoch": 0.01, "learning_rate": 2.9862023653088043e-05, "loss": 77.68, "step": 31 }, { "compression_loss": 75.75862121582031, "epoch": 0.01, "learning_rate": 2.9857095926412615e-05, "loss": 76.1832, "step": 32 }, { "compression_loss": 75.59468078613281, "epoch": 0.01, "learning_rate": 2.985216819973719e-05, "loss": 76.2596, "step": 33 }, { "compression_loss": 75.34709167480469, "epoch": 0.01, "learning_rate": 2.9847240473061762e-05, "loss": 75.8236, "step": 34 }, { "compression_loss": 73.89505004882812, "epoch": 0.01, "learning_rate": 2.9842312746386334e-05, "loss": 74.4378, "step": 35 }, { "compression_loss": 74.7244644165039, "epoch": 0.01, "learning_rate": 2.983738501971091e-05, "loss": 75.1611, "step": 36 }, { "compression_loss": 76.5816650390625, "epoch": 0.01, "learning_rate": 2.983245729303548e-05, "loss": 77.0912, "step": 37 }, { "compression_loss": 75.05460357666016, "epoch": 0.01, "learning_rate": 2.9827529566360054e-05, "loss": 75.8121, "step": 38 }, { "compression_loss": 76.40994262695312, "epoch": 0.01, "learning_rate": 2.9822601839684626e-05, "loss": 77.2531, "step": 39 }, { "compression_loss": 77.12899780273438, "epoch": 0.01, "learning_rate": 2.9817674113009197e-05, "loss": 77.961, "step": 40 }, { "compression_loss": 76.58537292480469, "epoch": 0.01, "learning_rate": 2.9812746386333773e-05, "loss": 77.3017, "step": 41 }, { "compression_loss": 76.62289428710938, "epoch": 0.01, "learning_rate": 2.9807818659658345e-05, "loss": 77.1285, "step": 42 }, { "compression_loss": 75.92079162597656, "epoch": 0.01, "learning_rate": 2.9802890932982917e-05, "loss": 76.6256, "step": 43 }, { "compression_loss": 76.80224609375, "epoch": 0.01, "learning_rate": 2.979796320630749e-05, "loss": 77.3004, "step": 44 }, { "compression_loss": 77.24520111083984, "epoch": 0.01, "learning_rate": 2.9793035479632064e-05, "loss": 77.5333, "step": 45 }, { "compression_loss": 74.12928009033203, "epoch": 0.02, "learning_rate": 2.978810775295664e-05, "loss": 74.6589, "step": 46 }, { "compression_loss": 75.3709945678711, "epoch": 0.02, "learning_rate": 2.978318002628121e-05, "loss": 75.8864, "step": 47 }, { "compression_loss": 75.06959533691406, "epoch": 0.02, "learning_rate": 2.9778252299605783e-05, "loss": 75.4088, "step": 48 }, { "compression_loss": 75.38143157958984, "epoch": 0.02, "learning_rate": 2.9773324572930355e-05, "loss": 75.9945, "step": 49 }, { "compression_loss": 75.86177062988281, "epoch": 0.02, "learning_rate": 2.9768396846254927e-05, "loss": 76.5964, "step": 50 }, { "compression_loss": 78.66627502441406, "epoch": 0.02, "learning_rate": 2.9763469119579503e-05, "loss": 78.9309, "step": 51 }, { "compression_loss": 76.15283203125, "epoch": 0.02, "learning_rate": 2.9758541392904075e-05, "loss": 76.8048, "step": 52 }, { "compression_loss": 75.27989196777344, "epoch": 0.02, "learning_rate": 2.9753613666228647e-05, "loss": 75.6328, "step": 53 }, { "compression_loss": 77.00032043457031, "epoch": 0.02, "learning_rate": 2.974868593955322e-05, "loss": 77.3641, "step": 54 }, { "compression_loss": 74.99322509765625, "epoch": 0.02, "learning_rate": 2.974375821287779e-05, "loss": 75.3507, "step": 55 }, { "compression_loss": 75.59861755371094, "epoch": 0.02, "learning_rate": 2.9738830486202366e-05, "loss": 75.9497, "step": 56 }, { "compression_loss": 73.84629821777344, "epoch": 0.02, "learning_rate": 2.973390275952694e-05, "loss": 74.386, "step": 57 }, { "compression_loss": 76.484619140625, "epoch": 0.02, "learning_rate": 2.9728975032851513e-05, "loss": 76.8285, "step": 58 }, { "compression_loss": 74.29684448242188, "epoch": 0.02, "learning_rate": 2.9724047306176085e-05, "loss": 74.7373, "step": 59 }, { "compression_loss": 75.94050598144531, "epoch": 0.02, "learning_rate": 2.9719119579500657e-05, "loss": 76.4972, "step": 60 }, { "compression_loss": 76.47115325927734, "epoch": 0.02, "learning_rate": 2.9714191852825233e-05, "loss": 76.862, "step": 61 }, { "compression_loss": 76.67019653320312, "epoch": 0.02, "learning_rate": 2.9709264126149805e-05, "loss": 77.2467, "step": 62 }, { "compression_loss": 77.33734130859375, "epoch": 0.02, "learning_rate": 2.9704336399474377e-05, "loss": 77.6692, "step": 63 }, { "compression_loss": 77.44827270507812, "epoch": 0.02, "learning_rate": 2.969940867279895e-05, "loss": 77.9159, "step": 64 }, { "compression_loss": 74.52684020996094, "epoch": 0.02, "learning_rate": 2.969448094612352e-05, "loss": 74.8602, "step": 65 }, { "compression_loss": 75.6634521484375, "epoch": 0.02, "learning_rate": 2.9689553219448096e-05, "loss": 76.1459, "step": 66 }, { "compression_loss": 77.37470245361328, "epoch": 0.02, "learning_rate": 2.9684625492772668e-05, "loss": 77.9025, "step": 67 }, { "compression_loss": 78.55731201171875, "epoch": 0.02, "learning_rate": 2.967969776609724e-05, "loss": 79.1895, "step": 68 }, { "compression_loss": 74.72705841064453, "epoch": 0.02, "learning_rate": 2.9674770039421815e-05, "loss": 75.2304, "step": 69 }, { "compression_loss": 74.98402404785156, "epoch": 0.02, "learning_rate": 2.9669842312746387e-05, "loss": 75.3703, "step": 70 }, { "compression_loss": 77.98944091796875, "epoch": 0.02, "learning_rate": 2.9664914586070962e-05, "loss": 78.5848, "step": 71 }, { "compression_loss": 73.5915298461914, "epoch": 0.02, "learning_rate": 2.9659986859395534e-05, "loss": 74.0818, "step": 72 }, { "compression_loss": 77.10995483398438, "epoch": 0.02, "learning_rate": 2.9655059132720106e-05, "loss": 77.9447, "step": 73 }, { "compression_loss": 78.42195129394531, "epoch": 0.02, "learning_rate": 2.965013140604468e-05, "loss": 78.9486, "step": 74 }, { "compression_loss": 75.00209045410156, "epoch": 0.02, "learning_rate": 2.964520367936925e-05, "loss": 75.3491, "step": 75 }, { "compression_loss": 75.17048645019531, "epoch": 0.02, "learning_rate": 2.9640275952693826e-05, "loss": 76.0134, "step": 76 }, { "compression_loss": 73.9056167602539, "epoch": 0.03, "learning_rate": 2.9635348226018398e-05, "loss": 74.2274, "step": 77 }, { "compression_loss": 77.01741790771484, "epoch": 0.03, "learning_rate": 2.963042049934297e-05, "loss": 77.4611, "step": 78 }, { "compression_loss": 74.30995178222656, "epoch": 0.03, "learning_rate": 2.962549277266754e-05, "loss": 74.6424, "step": 79 }, { "compression_loss": 75.31745910644531, "epoch": 0.03, "learning_rate": 2.9620565045992114e-05, "loss": 75.7231, "step": 80 }, { "compression_loss": 74.95240020751953, "epoch": 0.03, "learning_rate": 2.9615637319316692e-05, "loss": 75.2796, "step": 81 }, { "compression_loss": 76.8126220703125, "epoch": 0.03, "learning_rate": 2.9610709592641264e-05, "loss": 77.2499, "step": 82 }, { "compression_loss": 78.29135131835938, "epoch": 0.03, "learning_rate": 2.9605781865965836e-05, "loss": 78.9335, "step": 83 }, { "compression_loss": 77.16011047363281, "epoch": 0.03, "learning_rate": 2.9600854139290408e-05, "loss": 77.7654, "step": 84 }, { "compression_loss": 78.00952911376953, "epoch": 0.03, "learning_rate": 2.959592641261498e-05, "loss": 78.7566, "step": 85 }, { "compression_loss": 73.58897399902344, "epoch": 0.03, "learning_rate": 2.9590998685939556e-05, "loss": 74.071, "step": 86 }, { "compression_loss": 76.68132781982422, "epoch": 0.03, "learning_rate": 2.9586070959264127e-05, "loss": 77.4953, "step": 87 }, { "compression_loss": 75.09564208984375, "epoch": 0.03, "learning_rate": 2.95811432325887e-05, "loss": 75.366, "step": 88 }, { "compression_loss": 75.58582305908203, "epoch": 0.03, "learning_rate": 2.957621550591327e-05, "loss": 76.0922, "step": 89 }, { "compression_loss": 75.13958740234375, "epoch": 0.03, "learning_rate": 2.9571287779237843e-05, "loss": 75.7384, "step": 90 }, { "compression_loss": 76.4171371459961, "epoch": 0.03, "learning_rate": 2.956636005256242e-05, "loss": 76.7594, "step": 91 }, { "compression_loss": 76.55003356933594, "epoch": 0.03, "learning_rate": 2.956143232588699e-05, "loss": 76.817, "step": 92 }, { "compression_loss": 77.27275848388672, "epoch": 0.03, "learning_rate": 2.9556504599211566e-05, "loss": 78.0263, "step": 93 }, { "compression_loss": 76.52838134765625, "epoch": 0.03, "learning_rate": 2.9551576872536138e-05, "loss": 76.9734, "step": 94 }, { "compression_loss": 73.2533187866211, "epoch": 0.03, "learning_rate": 2.954664914586071e-05, "loss": 73.5546, "step": 95 }, { "compression_loss": 76.28591918945312, "epoch": 0.03, "learning_rate": 2.9541721419185285e-05, "loss": 76.8643, "step": 96 }, { "compression_loss": 75.90972137451172, "epoch": 0.03, "learning_rate": 2.9536793692509857e-05, "loss": 76.1346, "step": 97 }, { "compression_loss": 75.44319152832031, "epoch": 0.03, "learning_rate": 2.953186596583443e-05, "loss": 75.9753, "step": 98 }, { "compression_loss": 76.22297668457031, "epoch": 0.03, "learning_rate": 2.9526938239159e-05, "loss": 76.6026, "step": 99 }, { "compression_loss": 74.00920104980469, "epoch": 0.03, "learning_rate": 2.9522010512483573e-05, "loss": 74.4485, "step": 100 }, { "compression_loss": 76.18508911132812, "epoch": 0.03, "learning_rate": 2.951708278580815e-05, "loss": 76.7362, "step": 101 }, { "compression_loss": 77.67921447753906, "epoch": 0.03, "learning_rate": 2.951215505913272e-05, "loss": 78.2017, "step": 102 }, { "compression_loss": 75.2479476928711, "epoch": 0.03, "learning_rate": 2.9507227332457293e-05, "loss": 75.7367, "step": 103 }, { "compression_loss": 77.9022445678711, "epoch": 0.03, "learning_rate": 2.9502299605781865e-05, "loss": 78.2313, "step": 104 }, { "compression_loss": 75.27488708496094, "epoch": 0.03, "learning_rate": 2.949737187910644e-05, "loss": 75.6526, "step": 105 }, { "compression_loss": 76.05467224121094, "epoch": 0.03, "learning_rate": 2.9492444152431015e-05, "loss": 76.6153, "step": 106 }, { "compression_loss": 78.19921875, "epoch": 0.04, "learning_rate": 2.9487516425755587e-05, "loss": 78.9331, "step": 107 }, { "compression_loss": 74.343017578125, "epoch": 0.04, "learning_rate": 2.948258869908016e-05, "loss": 74.8744, "step": 108 }, { "compression_loss": 76.34170532226562, "epoch": 0.04, "learning_rate": 2.947766097240473e-05, "loss": 76.9162, "step": 109 }, { "compression_loss": 77.84381103515625, "epoch": 0.04, "learning_rate": 2.9472733245729303e-05, "loss": 78.4289, "step": 110 }, { "compression_loss": 74.27777099609375, "epoch": 0.04, "learning_rate": 2.946780551905388e-05, "loss": 74.5646, "step": 111 }, { "compression_loss": 73.22571563720703, "epoch": 0.04, "learning_rate": 2.946287779237845e-05, "loss": 73.4312, "step": 112 }, { "compression_loss": 74.25314331054688, "epoch": 0.04, "learning_rate": 2.9457950065703022e-05, "loss": 74.8275, "step": 113 }, { "compression_loss": 75.49479675292969, "epoch": 0.04, "learning_rate": 2.9453022339027594e-05, "loss": 75.7403, "step": 114 }, { "compression_loss": 77.44023132324219, "epoch": 0.04, "learning_rate": 2.9448094612352166e-05, "loss": 77.7906, "step": 115 }, { "compression_loss": 76.09671020507812, "epoch": 0.04, "learning_rate": 2.944316688567674e-05, "loss": 77.0804, "step": 116 }, { "compression_loss": 74.1368179321289, "epoch": 0.04, "learning_rate": 2.9438239159001317e-05, "loss": 74.6138, "step": 117 }, { "compression_loss": 76.92250061035156, "epoch": 0.04, "learning_rate": 2.943331143232589e-05, "loss": 77.8186, "step": 118 }, { "compression_loss": 76.05874633789062, "epoch": 0.04, "learning_rate": 2.942838370565046e-05, "loss": 76.5848, "step": 119 }, { "compression_loss": 75.24482727050781, "epoch": 0.04, "learning_rate": 2.9423455978975033e-05, "loss": 75.4702, "step": 120 }, { "compression_loss": 74.17037200927734, "epoch": 0.04, "learning_rate": 2.941852825229961e-05, "loss": 74.5024, "step": 121 }, { "compression_loss": 74.9408187866211, "epoch": 0.04, "learning_rate": 2.941360052562418e-05, "loss": 75.3226, "step": 122 }, { "compression_loss": 76.81588745117188, "epoch": 0.04, "learning_rate": 2.9408672798948752e-05, "loss": 77.0867, "step": 123 }, { "compression_loss": 77.95330810546875, "epoch": 0.04, "learning_rate": 2.9403745072273324e-05, "loss": 78.7039, "step": 124 }, { "compression_loss": 77.43009185791016, "epoch": 0.04, "learning_rate": 2.9398817345597896e-05, "loss": 78.1763, "step": 125 }, { "compression_loss": 76.23763275146484, "epoch": 0.04, "learning_rate": 2.939388961892247e-05, "loss": 76.6914, "step": 126 }, { "compression_loss": 76.83124542236328, "epoch": 0.04, "learning_rate": 2.9388961892247044e-05, "loss": 77.3769, "step": 127 }, { "compression_loss": 75.59762573242188, "epoch": 0.04, "learning_rate": 2.9384034165571615e-05, "loss": 76.092, "step": 128 }, { "compression_loss": 74.75022888183594, "epoch": 0.04, "learning_rate": 2.937910643889619e-05, "loss": 75.3169, "step": 129 }, { "compression_loss": 77.38142395019531, "epoch": 0.04, "learning_rate": 2.9374178712220763e-05, "loss": 77.9114, "step": 130 }, { "compression_loss": 76.77714538574219, "epoch": 0.04, "learning_rate": 2.9369250985545338e-05, "loss": 77.1138, "step": 131 }, { "compression_loss": 75.115234375, "epoch": 0.04, "learning_rate": 2.936432325886991e-05, "loss": 75.4113, "step": 132 }, { "compression_loss": 77.13204193115234, "epoch": 0.04, "learning_rate": 2.9359395532194482e-05, "loss": 77.7157, "step": 133 }, { "compression_loss": 74.05122375488281, "epoch": 0.04, "learning_rate": 2.9354467805519054e-05, "loss": 74.4922, "step": 134 }, { "compression_loss": 72.8426742553711, "epoch": 0.04, "learning_rate": 2.9349540078843626e-05, "loss": 73.033, "step": 135 }, { "compression_loss": 76.92378997802734, "epoch": 0.04, "learning_rate": 2.93446123521682e-05, "loss": 77.4453, "step": 136 }, { "compression_loss": 74.16395568847656, "epoch": 0.05, "learning_rate": 2.9339684625492773e-05, "loss": 74.4716, "step": 137 }, { "compression_loss": 74.82159423828125, "epoch": 0.05, "learning_rate": 2.9334756898817345e-05, "loss": 75.117, "step": 138 }, { "compression_loss": 76.47021484375, "epoch": 0.05, "learning_rate": 2.9329829172141917e-05, "loss": 76.837, "step": 139 }, { "compression_loss": 77.3161849975586, "epoch": 0.05, "learning_rate": 2.9324901445466493e-05, "loss": 77.8787, "step": 140 }, { "compression_loss": 75.80555725097656, "epoch": 0.05, "learning_rate": 2.9319973718791068e-05, "loss": 76.4956, "step": 141 }, { "compression_loss": 74.9495849609375, "epoch": 0.05, "learning_rate": 2.931504599211564e-05, "loss": 75.4142, "step": 142 }, { "compression_loss": 75.82635498046875, "epoch": 0.05, "learning_rate": 2.9310118265440212e-05, "loss": 76.5205, "step": 143 }, { "compression_loss": 76.1204833984375, "epoch": 0.05, "learning_rate": 2.9305190538764784e-05, "loss": 76.6991, "step": 144 }, { "compression_loss": 78.66889190673828, "epoch": 0.05, "learning_rate": 2.9300262812089356e-05, "loss": 79.1492, "step": 145 }, { "compression_loss": 77.38497161865234, "epoch": 0.05, "learning_rate": 2.929533508541393e-05, "loss": 77.9415, "step": 146 }, { "compression_loss": 74.33131408691406, "epoch": 0.05, "learning_rate": 2.9290407358738503e-05, "loss": 74.6391, "step": 147 }, { "compression_loss": 78.04620361328125, "epoch": 0.05, "learning_rate": 2.9285479632063075e-05, "loss": 78.7915, "step": 148 }, { "compression_loss": 77.53388977050781, "epoch": 0.05, "learning_rate": 2.9280551905387647e-05, "loss": 78.1248, "step": 149 }, { "compression_loss": 75.71705627441406, "epoch": 0.05, "learning_rate": 2.927562417871222e-05, "loss": 76.3242, "step": 150 }, { "compression_loss": 75.23284912109375, "epoch": 0.05, "learning_rate": 2.9270696452036794e-05, "loss": 75.64, "step": 151 }, { "compression_loss": 74.66268920898438, "epoch": 0.05, "learning_rate": 2.926576872536137e-05, "loss": 74.9772, "step": 152 }, { "compression_loss": 74.015380859375, "epoch": 0.05, "learning_rate": 2.9260840998685942e-05, "loss": 74.485, "step": 153 }, { "compression_loss": 75.14453125, "epoch": 0.05, "learning_rate": 2.9255913272010514e-05, "loss": 75.4865, "step": 154 }, { "compression_loss": 76.34785461425781, "epoch": 0.05, "learning_rate": 2.9250985545335086e-05, "loss": 76.7353, "step": 155 }, { "compression_loss": 72.96851348876953, "epoch": 0.05, "learning_rate": 2.924605781865966e-05, "loss": 73.4389, "step": 156 }, { "compression_loss": 76.02486419677734, "epoch": 0.05, "learning_rate": 2.9241130091984233e-05, "loss": 76.3158, "step": 157 }, { "compression_loss": 76.53388214111328, "epoch": 0.05, "learning_rate": 2.9236202365308805e-05, "loss": 76.9831, "step": 158 }, { "compression_loss": 76.3155746459961, "epoch": 0.05, "learning_rate": 2.9231274638633377e-05, "loss": 76.5998, "step": 159 }, { "compression_loss": 74.29539489746094, "epoch": 0.05, "learning_rate": 2.922634691195795e-05, "loss": 74.5867, "step": 160 }, { "compression_loss": 76.72748565673828, "epoch": 0.05, "learning_rate": 2.9221419185282524e-05, "loss": 77.2575, "step": 161 }, { "compression_loss": 73.93374633789062, "epoch": 0.05, "learning_rate": 2.9216491458607096e-05, "loss": 74.3653, "step": 162 }, { "compression_loss": 77.33306884765625, "epoch": 0.05, "learning_rate": 2.9211563731931668e-05, "loss": 77.9473, "step": 163 }, { "compression_loss": 77.53567504882812, "epoch": 0.05, "learning_rate": 2.9206636005256244e-05, "loss": 78.1362, "step": 164 }, { "compression_loss": 77.04246520996094, "epoch": 0.05, "learning_rate": 2.9201708278580816e-05, "loss": 77.7045, "step": 165 }, { "compression_loss": 75.45565795898438, "epoch": 0.05, "learning_rate": 2.919678055190539e-05, "loss": 76.1339, "step": 166 }, { "compression_loss": 72.90103912353516, "epoch": 0.05, "learning_rate": 2.9191852825229963e-05, "loss": 73.2436, "step": 167 }, { "compression_loss": 76.33231353759766, "epoch": 0.06, "learning_rate": 2.9186925098554535e-05, "loss": 76.6362, "step": 168 }, { "compression_loss": 76.5862045288086, "epoch": 0.06, "learning_rate": 2.9181997371879107e-05, "loss": 77.1494, "step": 169 }, { "compression_loss": 76.74464416503906, "epoch": 0.06, "learning_rate": 2.917706964520368e-05, "loss": 77.1665, "step": 170 }, { "compression_loss": 75.16952514648438, "epoch": 0.06, "learning_rate": 2.9172141918528254e-05, "loss": 75.9335, "step": 171 }, { "compression_loss": 76.5085678100586, "epoch": 0.06, "learning_rate": 2.9167214191852826e-05, "loss": 76.9348, "step": 172 }, { "compression_loss": 75.5731430053711, "epoch": 0.06, "learning_rate": 2.9162286465177398e-05, "loss": 75.8748, "step": 173 }, { "compression_loss": 77.83612060546875, "epoch": 0.06, "learning_rate": 2.915735873850197e-05, "loss": 78.6893, "step": 174 }, { "compression_loss": 74.14400482177734, "epoch": 0.06, "learning_rate": 2.9152431011826542e-05, "loss": 74.7864, "step": 175 }, { "compression_loss": 77.03101348876953, "epoch": 0.06, "learning_rate": 2.914750328515112e-05, "loss": 77.551, "step": 176 }, { "compression_loss": 75.03648376464844, "epoch": 0.06, "learning_rate": 2.9142575558475693e-05, "loss": 75.3499, "step": 177 }, { "compression_loss": 74.00916290283203, "epoch": 0.06, "learning_rate": 2.9137647831800265e-05, "loss": 74.2702, "step": 178 }, { "compression_loss": 77.36481475830078, "epoch": 0.06, "learning_rate": 2.9132720105124837e-05, "loss": 77.6995, "step": 179 }, { "compression_loss": 76.56790924072266, "epoch": 0.06, "learning_rate": 2.912779237844941e-05, "loss": 77.0789, "step": 180 }, { "compression_loss": 72.17565155029297, "epoch": 0.06, "learning_rate": 2.9122864651773984e-05, "loss": 72.5899, "step": 181 }, { "compression_loss": 76.14957427978516, "epoch": 0.06, "learning_rate": 2.9117936925098556e-05, "loss": 76.6452, "step": 182 }, { "compression_loss": 76.39651489257812, "epoch": 0.06, "learning_rate": 2.9113009198423128e-05, "loss": 77.1679, "step": 183 }, { "compression_loss": 75.06361389160156, "epoch": 0.06, "learning_rate": 2.91080814717477e-05, "loss": 75.4611, "step": 184 }, { "compression_loss": 75.49005126953125, "epoch": 0.06, "learning_rate": 2.9103153745072272e-05, "loss": 76.1028, "step": 185 }, { "compression_loss": 77.03211975097656, "epoch": 0.06, "learning_rate": 2.9098226018396847e-05, "loss": 77.5783, "step": 186 }, { "compression_loss": 75.9092788696289, "epoch": 0.06, "learning_rate": 2.909329829172142e-05, "loss": 76.7181, "step": 187 }, { "compression_loss": 77.03887176513672, "epoch": 0.06, "learning_rate": 2.9088370565045995e-05, "loss": 77.3484, "step": 188 }, { "compression_loss": 73.23045349121094, "epoch": 0.06, "learning_rate": 2.9083442838370567e-05, "loss": 73.8298, "step": 189 }, { "compression_loss": 75.24256896972656, "epoch": 0.06, "learning_rate": 2.907851511169514e-05, "loss": 75.8112, "step": 190 }, { "compression_loss": 78.10365295410156, "epoch": 0.06, "learning_rate": 2.9073587385019714e-05, "loss": 78.8816, "step": 191 }, { "compression_loss": 75.20037841796875, "epoch": 0.06, "learning_rate": 2.9068659658344286e-05, "loss": 75.7871, "step": 192 }, { "compression_loss": 77.63656616210938, "epoch": 0.06, "learning_rate": 2.9063731931668858e-05, "loss": 78.054, "step": 193 }, { "compression_loss": 75.72503662109375, "epoch": 0.06, "learning_rate": 2.905880420499343e-05, "loss": 76.0548, "step": 194 }, { "compression_loss": 76.49850463867188, "epoch": 0.06, "learning_rate": 2.9053876478318002e-05, "loss": 77.0285, "step": 195 }, { "compression_loss": 74.20917510986328, "epoch": 0.06, "learning_rate": 2.9048948751642577e-05, "loss": 74.8247, "step": 196 }, { "compression_loss": 75.86183166503906, "epoch": 0.06, "learning_rate": 2.904402102496715e-05, "loss": 76.0683, "step": 197 }, { "compression_loss": 75.7878646850586, "epoch": 0.07, "learning_rate": 2.903909329829172e-05, "loss": 76.3288, "step": 198 }, { "compression_loss": 75.51722717285156, "epoch": 0.07, "learning_rate": 2.9034165571616293e-05, "loss": 75.8985, "step": 199 }, { "compression_loss": 73.82627868652344, "epoch": 0.07, "learning_rate": 2.902923784494087e-05, "loss": 74.2089, "step": 200 }, { "compression_loss": 76.37925720214844, "epoch": 0.07, "learning_rate": 2.9024310118265444e-05, "loss": 76.61, "step": 201 }, { "compression_loss": 74.40361022949219, "epoch": 0.07, "learning_rate": 2.9019382391590016e-05, "loss": 75.0194, "step": 202 }, { "compression_loss": 74.47918701171875, "epoch": 0.07, "learning_rate": 2.9014454664914588e-05, "loss": 74.6583, "step": 203 }, { "compression_loss": 73.78302001953125, "epoch": 0.07, "learning_rate": 2.900952693823916e-05, "loss": 74.1452, "step": 204 }, { "compression_loss": 75.33750915527344, "epoch": 0.07, "learning_rate": 2.900459921156373e-05, "loss": 75.5505, "step": 205 }, { "compression_loss": 76.87371826171875, "epoch": 0.07, "learning_rate": 2.8999671484888307e-05, "loss": 77.2306, "step": 206 }, { "compression_loss": 78.70271301269531, "epoch": 0.07, "learning_rate": 2.899474375821288e-05, "loss": 79.3709, "step": 207 }, { "compression_loss": 76.63807678222656, "epoch": 0.07, "learning_rate": 2.898981603153745e-05, "loss": 77.1524, "step": 208 }, { "compression_loss": 76.2537841796875, "epoch": 0.07, "learning_rate": 2.8984888304862023e-05, "loss": 76.7885, "step": 209 }, { "compression_loss": 76.66504669189453, "epoch": 0.07, "learning_rate": 2.8979960578186595e-05, "loss": 77.044, "step": 210 }, { "compression_loss": 77.22710418701172, "epoch": 0.07, "learning_rate": 2.897503285151117e-05, "loss": 77.7639, "step": 211 }, { "compression_loss": 76.48908996582031, "epoch": 0.07, "learning_rate": 2.8970105124835746e-05, "loss": 76.8617, "step": 212 }, { "compression_loss": 75.4661636352539, "epoch": 0.07, "learning_rate": 2.8965177398160318e-05, "loss": 75.8703, "step": 213 }, { "compression_loss": 75.65154266357422, "epoch": 0.07, "learning_rate": 2.896024967148489e-05, "loss": 75.9878, "step": 214 }, { "compression_loss": 76.90234375, "epoch": 0.07, "learning_rate": 2.895532194480946e-05, "loss": 77.2552, "step": 215 }, { "compression_loss": 75.7844009399414, "epoch": 0.07, "learning_rate": 2.8950394218134037e-05, "loss": 76.2565, "step": 216 }, { "compression_loss": 76.15606689453125, "epoch": 0.07, "learning_rate": 2.894546649145861e-05, "loss": 76.3452, "step": 217 }, { "compression_loss": 74.68834686279297, "epoch": 0.07, "learning_rate": 2.894053876478318e-05, "loss": 74.9902, "step": 218 }, { "compression_loss": 78.17327880859375, "epoch": 0.07, "learning_rate": 2.8935611038107753e-05, "loss": 78.616, "step": 219 }, { "compression_loss": 75.18898010253906, "epoch": 0.07, "learning_rate": 2.8930683311432325e-05, "loss": 75.437, "step": 220 }, { "compression_loss": 74.34846496582031, "epoch": 0.07, "learning_rate": 2.89257555847569e-05, "loss": 74.9461, "step": 221 }, { "compression_loss": 75.35258483886719, "epoch": 0.07, "learning_rate": 2.8920827858081472e-05, "loss": 76.0932, "step": 222 }, { "compression_loss": 77.31640625, "epoch": 0.07, "learning_rate": 2.8915900131406044e-05, "loss": 78.4483, "step": 223 }, { "compression_loss": 73.70903778076172, "epoch": 0.07, "learning_rate": 2.891097240473062e-05, "loss": 74.2825, "step": 224 }, { "compression_loss": 74.59931945800781, "epoch": 0.07, "learning_rate": 2.890604467805519e-05, "loss": 75.217, "step": 225 }, { "compression_loss": 74.88743591308594, "epoch": 0.07, "learning_rate": 2.8901116951379767e-05, "loss": 75.3315, "step": 226 }, { "compression_loss": 75.14473724365234, "epoch": 0.07, "learning_rate": 2.889618922470434e-05, "loss": 75.5946, "step": 227 }, { "compression_loss": 77.92741394042969, "epoch": 0.07, "learning_rate": 2.889126149802891e-05, "loss": 78.7299, "step": 228 }, { "compression_loss": 75.23294830322266, "epoch": 0.08, "learning_rate": 2.8886333771353483e-05, "loss": 75.7103, "step": 229 }, { "compression_loss": 77.17891693115234, "epoch": 0.08, "learning_rate": 2.8881406044678055e-05, "loss": 77.672, "step": 230 }, { "compression_loss": 72.88497924804688, "epoch": 0.08, "learning_rate": 2.887647831800263e-05, "loss": 73.2034, "step": 231 }, { "compression_loss": 78.50125885009766, "epoch": 0.08, "learning_rate": 2.8871550591327202e-05, "loss": 79.213, "step": 232 }, { "compression_loss": 74.96420288085938, "epoch": 0.08, "learning_rate": 2.8866622864651774e-05, "loss": 75.3866, "step": 233 }, { "compression_loss": 74.80231475830078, "epoch": 0.08, "learning_rate": 2.8861695137976346e-05, "loss": 75.4368, "step": 234 }, { "compression_loss": 75.57276916503906, "epoch": 0.08, "learning_rate": 2.8856767411300918e-05, "loss": 75.9834, "step": 235 }, { "compression_loss": 76.00504302978516, "epoch": 0.08, "learning_rate": 2.8851839684625497e-05, "loss": 76.3946, "step": 236 }, { "compression_loss": 73.33707427978516, "epoch": 0.08, "learning_rate": 2.884691195795007e-05, "loss": 73.78, "step": 237 }, { "compression_loss": 75.836181640625, "epoch": 0.08, "learning_rate": 2.884198423127464e-05, "loss": 76.3122, "step": 238 }, { "compression_loss": 77.89449310302734, "epoch": 0.08, "learning_rate": 2.8837056504599212e-05, "loss": 78.5143, "step": 239 }, { "compression_loss": 77.95869445800781, "epoch": 0.08, "learning_rate": 2.8832128777923784e-05, "loss": 78.5585, "step": 240 }, { "compression_loss": 74.84977722167969, "epoch": 0.08, "learning_rate": 2.882720105124836e-05, "loss": 75.1526, "step": 241 }, { "compression_loss": 73.61986541748047, "epoch": 0.08, "learning_rate": 2.8822273324572932e-05, "loss": 74.2395, "step": 242 }, { "compression_loss": 76.30523681640625, "epoch": 0.08, "learning_rate": 2.8817345597897504e-05, "loss": 76.8858, "step": 243 }, { "compression_loss": 74.98263549804688, "epoch": 0.08, "learning_rate": 2.8812417871222076e-05, "loss": 75.2046, "step": 244 }, { "compression_loss": 75.41921997070312, "epoch": 0.08, "learning_rate": 2.8807490144546648e-05, "loss": 75.7094, "step": 245 }, { "compression_loss": 77.78306579589844, "epoch": 0.08, "learning_rate": 2.8802562417871223e-05, "loss": 78.3286, "step": 246 }, { "compression_loss": 74.69575500488281, "epoch": 0.08, "learning_rate": 2.8797634691195795e-05, "loss": 75.1961, "step": 247 }, { "compression_loss": 74.6552734375, "epoch": 0.08, "learning_rate": 2.879270696452037e-05, "loss": 75.0338, "step": 248 }, { "compression_loss": 75.35956573486328, "epoch": 0.08, "learning_rate": 2.8787779237844942e-05, "loss": 75.7817, "step": 249 }, { "compression_loss": 75.45681762695312, "epoch": 0.08, "learning_rate": 2.8782851511169514e-05, "loss": 75.8633, "step": 250 }, { "epoch": 0.08, "eval_exact_match": 85.71428571428571, "eval_f1": 92.1895105621475, "step": 250 }, { "compression_loss": 77.50865173339844, "epoch": 0.08, "learning_rate": 2.877792378449409e-05, "loss": 77.8889, "step": 251 }, { "compression_loss": 75.48422241210938, "epoch": 0.08, "learning_rate": 2.877299605781866e-05, "loss": 76.0134, "step": 252 }, { "compression_loss": 75.33040618896484, "epoch": 0.08, "learning_rate": 2.8768068331143234e-05, "loss": 75.7734, "step": 253 }, { "compression_loss": 76.39631652832031, "epoch": 0.08, "learning_rate": 2.8763140604467806e-05, "loss": 76.9228, "step": 254 }, { "compression_loss": 78.3295669555664, "epoch": 0.08, "learning_rate": 2.8758212877792378e-05, "loss": 78.9481, "step": 255 }, { "compression_loss": 77.60240936279297, "epoch": 0.08, "learning_rate": 2.8753285151116953e-05, "loss": 77.8796, "step": 256 }, { "compression_loss": 76.11383056640625, "epoch": 0.08, "learning_rate": 2.8748357424441525e-05, "loss": 76.7309, "step": 257 }, { "compression_loss": 75.97412872314453, "epoch": 0.08, "learning_rate": 2.8743429697766097e-05, "loss": 76.1797, "step": 258 }, { "compression_loss": 76.4327163696289, "epoch": 0.09, "learning_rate": 2.873850197109067e-05, "loss": 76.9353, "step": 259 }, { "compression_loss": 78.29627990722656, "epoch": 0.09, "learning_rate": 2.8733574244415244e-05, "loss": 79.0002, "step": 260 }, { "compression_loss": 75.87815856933594, "epoch": 0.09, "learning_rate": 2.872864651773982e-05, "loss": 76.3639, "step": 261 }, { "compression_loss": 76.2979507446289, "epoch": 0.09, "learning_rate": 2.872371879106439e-05, "loss": 76.7754, "step": 262 }, { "compression_loss": 76.13604736328125, "epoch": 0.09, "learning_rate": 2.8718791064388963e-05, "loss": 76.5297, "step": 263 }, { "compression_loss": 77.0492172241211, "epoch": 0.09, "learning_rate": 2.8713863337713535e-05, "loss": 77.3113, "step": 264 }, { "compression_loss": 75.99713897705078, "epoch": 0.09, "learning_rate": 2.8708935611038107e-05, "loss": 76.375, "step": 265 }, { "compression_loss": 74.68904113769531, "epoch": 0.09, "learning_rate": 2.8708935611038107e-05, "loss": 74.9842, "step": 266 }, { "compression_loss": 79.59208679199219, "epoch": 0.09, "learning_rate": 2.8704007884362683e-05, "loss": 80.1954, "step": 267 }, { "compression_loss": 77.35648345947266, "epoch": 0.09, "learning_rate": 2.8699080157687255e-05, "loss": 77.8292, "step": 268 }, { "compression_loss": 73.66227722167969, "epoch": 0.09, "learning_rate": 2.8694152431011827e-05, "loss": 73.826, "step": 269 }, { "compression_loss": 75.49293518066406, "epoch": 0.09, "learning_rate": 2.86892247043364e-05, "loss": 76.0126, "step": 270 }, { "compression_loss": 76.47076416015625, "epoch": 0.09, "learning_rate": 2.868429697766097e-05, "loss": 76.9026, "step": 271 }, { "compression_loss": 76.11405181884766, "epoch": 0.09, "learning_rate": 2.8679369250985546e-05, "loss": 76.7562, "step": 272 }, { "compression_loss": 78.19115447998047, "epoch": 0.09, "learning_rate": 2.867444152431012e-05, "loss": 78.9335, "step": 273 }, { "compression_loss": 76.76970672607422, "epoch": 0.09, "learning_rate": 2.8669513797634693e-05, "loss": 77.2018, "step": 274 }, { "compression_loss": 74.9723892211914, "epoch": 0.09, "learning_rate": 2.8664586070959265e-05, "loss": 75.4297, "step": 275 }, { "compression_loss": 77.17861938476562, "epoch": 0.09, "learning_rate": 2.8659658344283837e-05, "loss": 77.4293, "step": 276 }, { "compression_loss": 74.9989242553711, "epoch": 0.09, "learning_rate": 2.8654730617608413e-05, "loss": 75.4433, "step": 277 }, { "compression_loss": 76.40315246582031, "epoch": 0.09, "learning_rate": 2.8649802890932985e-05, "loss": 76.7803, "step": 278 }, { "compression_loss": 75.72197723388672, "epoch": 0.09, "learning_rate": 2.8644875164257557e-05, "loss": 76.1492, "step": 279 }, { "compression_loss": 75.19316864013672, "epoch": 0.09, "learning_rate": 2.863994743758213e-05, "loss": 75.5123, "step": 280 }, { "compression_loss": 76.61006164550781, "epoch": 0.09, "learning_rate": 2.86350197109067e-05, "loss": 77.01, "step": 281 }, { "compression_loss": 74.96455383300781, "epoch": 0.09, "learning_rate": 2.8630091984231276e-05, "loss": 75.6055, "step": 282 }, { "compression_loss": 76.96006774902344, "epoch": 0.09, "learning_rate": 2.8625164257555848e-05, "loss": 77.7212, "step": 283 }, { "compression_loss": 74.46092224121094, "epoch": 0.09, "learning_rate": 2.8620236530880423e-05, "loss": 75.1615, "step": 284 }, { "compression_loss": 76.9989013671875, "epoch": 0.09, "learning_rate": 2.8615308804204995e-05, "loss": 77.4677, "step": 285 }, { "compression_loss": 75.67250061035156, "epoch": 0.09, "learning_rate": 2.8610381077529567e-05, "loss": 75.9544, "step": 286 }, { "compression_loss": 76.27839660644531, "epoch": 0.09, "learning_rate": 2.8605453350854142e-05, "loss": 76.7157, "step": 287 }, { "compression_loss": 76.05245971679688, "epoch": 0.09, "learning_rate": 2.8600525624178714e-05, "loss": 76.6766, "step": 288 }, { "compression_loss": 74.80166625976562, "epoch": 0.09, "learning_rate": 2.8595597897503286e-05, "loss": 75.4685, "step": 289 }, { "compression_loss": 74.09005737304688, "epoch": 0.1, "learning_rate": 2.859067017082786e-05, "loss": 74.6217, "step": 290 }, { "compression_loss": 74.87425231933594, "epoch": 0.1, "learning_rate": 2.859067017082786e-05, "loss": 75.078, "step": 291 }, { "compression_loss": 75.19358825683594, "epoch": 0.1, "learning_rate": 2.858574244415243e-05, "loss": 75.7171, "step": 292 }, { "compression_loss": 74.14088439941406, "epoch": 0.1, "learning_rate": 2.8580814717477006e-05, "loss": 74.3258, "step": 293 }, { "compression_loss": 75.07235717773438, "epoch": 0.1, "learning_rate": 2.8575886990801578e-05, "loss": 75.8202, "step": 294 }, { "compression_loss": 74.79788208007812, "epoch": 0.1, "learning_rate": 2.857095926412615e-05, "loss": 75.2581, "step": 295 }, { "compression_loss": 75.9828872680664, "epoch": 0.1, "learning_rate": 2.856603153745072e-05, "loss": 76.4672, "step": 296 }, { "compression_loss": 75.51312255859375, "epoch": 0.1, "learning_rate": 2.8561103810775297e-05, "loss": 75.7183, "step": 297 }, { "compression_loss": 75.61595916748047, "epoch": 0.1, "learning_rate": 2.8556176084099872e-05, "loss": 76.0361, "step": 298 }, { "compression_loss": 73.94479370117188, "epoch": 0.1, "learning_rate": 2.8551248357424444e-05, "loss": 74.422, "step": 299 }, { "compression_loss": 77.51939392089844, "epoch": 0.1, "learning_rate": 2.8546320630749016e-05, "loss": 78.0649, "step": 300 }, { "compression_loss": 75.06008911132812, "epoch": 0.1, "learning_rate": 2.8541392904073588e-05, "loss": 75.2795, "step": 301 }, { "compression_loss": 78.3770523071289, "epoch": 0.1, "learning_rate": 2.853646517739816e-05, "loss": 78.8478, "step": 302 }, { "compression_loss": 74.92798614501953, "epoch": 0.1, "learning_rate": 2.8531537450722736e-05, "loss": 75.2369, "step": 303 }, { "compression_loss": 76.89193725585938, "epoch": 0.1, "learning_rate": 2.8526609724047308e-05, "loss": 77.5608, "step": 304 }, { "compression_loss": 76.35557556152344, "epoch": 0.1, "learning_rate": 2.852168199737188e-05, "loss": 77.0678, "step": 305 }, { "compression_loss": 76.36336517333984, "epoch": 0.1, "learning_rate": 2.851675427069645e-05, "loss": 76.9399, "step": 306 }, { "compression_loss": 75.0956039428711, "epoch": 0.1, "learning_rate": 2.8511826544021023e-05, "loss": 75.4146, "step": 307 }, { "compression_loss": 74.30586242675781, "epoch": 0.1, "learning_rate": 2.85068988173456e-05, "loss": 74.5742, "step": 308 }, { "compression_loss": 74.23963928222656, "epoch": 0.1, "learning_rate": 2.8501971090670174e-05, "loss": 74.5402, "step": 309 }, { "compression_loss": 75.03922271728516, "epoch": 0.1, "learning_rate": 2.8497043363994746e-05, "loss": 75.5136, "step": 310 }, { "compression_loss": 74.86763000488281, "epoch": 0.1, "learning_rate": 2.8492115637319318e-05, "loss": 75.5294, "step": 311 }, { "compression_loss": 76.0775146484375, "epoch": 0.1, "learning_rate": 2.848718791064389e-05, "loss": 76.4303, "step": 312 }, { "compression_loss": 75.18077087402344, "epoch": 0.1, "learning_rate": 2.8482260183968465e-05, "loss": 75.8565, "step": 313 }, { "compression_loss": 76.142578125, "epoch": 0.1, "learning_rate": 2.8477332457293037e-05, "loss": 76.5314, "step": 314 }, { "compression_loss": 73.74832153320312, "epoch": 0.1, "learning_rate": 2.847240473061761e-05, "loss": 74.0029, "step": 315 }, { "compression_loss": 79.82206726074219, "epoch": 0.1, "learning_rate": 2.846747700394218e-05, "loss": 80.3686, "step": 316 }, { "compression_loss": 75.67794799804688, "epoch": 0.1, "learning_rate": 2.8462549277266753e-05, "loss": 76.2097, "step": 317 }, { "compression_loss": 74.19783020019531, "epoch": 0.1, "learning_rate": 2.845762155059133e-05, "loss": 74.6327, "step": 318 }, { "compression_loss": 75.35965728759766, "epoch": 0.1, "learning_rate": 2.84526938239159e-05, "loss": 76.1337, "step": 319 }, { "compression_loss": 74.02418518066406, "epoch": 0.11, "learning_rate": 2.8447766097240473e-05, "loss": 74.6209, "step": 320 }, { "compression_loss": 76.1018295288086, "epoch": 0.11, "learning_rate": 2.8442838370565048e-05, "loss": 76.5187, "step": 321 }, { "compression_loss": 76.47357940673828, "epoch": 0.11, "learning_rate": 2.843791064388962e-05, "loss": 77.0985, "step": 322 }, { "compression_loss": 75.61332702636719, "epoch": 0.11, "learning_rate": 2.8432982917214195e-05, "loss": 75.8774, "step": 323 }, { "compression_loss": 76.35903930664062, "epoch": 0.11, "learning_rate": 2.8428055190538767e-05, "loss": 76.7512, "step": 324 }, { "compression_loss": 77.59774780273438, "epoch": 0.11, "learning_rate": 2.842312746386334e-05, "loss": 78.018, "step": 325 }, { "compression_loss": 75.44329071044922, "epoch": 0.11, "learning_rate": 2.841819973718791e-05, "loss": 75.7616, "step": 326 }, { "compression_loss": 75.2237548828125, "epoch": 0.11, "learning_rate": 2.8413272010512483e-05, "loss": 75.6001, "step": 327 }, { "compression_loss": 75.02417755126953, "epoch": 0.11, "learning_rate": 2.840834428383706e-05, "loss": 75.4796, "step": 328 }, { "compression_loss": 76.54490661621094, "epoch": 0.11, "learning_rate": 2.840341655716163e-05, "loss": 77.122, "step": 329 }, { "compression_loss": 76.65859985351562, "epoch": 0.11, "learning_rate": 2.8398488830486202e-05, "loss": 77.0054, "step": 330 }, { "compression_loss": 75.64059448242188, "epoch": 0.11, "learning_rate": 2.8393561103810774e-05, "loss": 75.9912, "step": 331 }, { "compression_loss": 77.97250366210938, "epoch": 0.11, "learning_rate": 2.8388633377135346e-05, "loss": 79.3269, "step": 332 }, { "compression_loss": 75.92874145507812, "epoch": 0.11, "learning_rate": 2.8383705650459925e-05, "loss": 76.584, "step": 333 }, { "compression_loss": 76.76429748535156, "epoch": 0.11, "learning_rate": 2.8378777923784497e-05, "loss": 77.3049, "step": 334 }, { "compression_loss": 77.20321655273438, "epoch": 0.11, "learning_rate": 2.837385019710907e-05, "loss": 77.5, "step": 335 }, { "compression_loss": 75.21180725097656, "epoch": 0.11, "learning_rate": 2.836892247043364e-05, "loss": 75.5731, "step": 336 }, { "compression_loss": 77.2168960571289, "epoch": 0.11, "learning_rate": 2.8363994743758213e-05, "loss": 77.6389, "step": 337 }, { "compression_loss": 77.3704833984375, "epoch": 0.11, "learning_rate": 2.835906701708279e-05, "loss": 77.5832, "step": 338 }, { "compression_loss": 74.25716400146484, "epoch": 0.11, "learning_rate": 2.835413929040736e-05, "loss": 74.6668, "step": 339 }, { "compression_loss": 74.49385833740234, "epoch": 0.11, "learning_rate": 2.8349211563731932e-05, "loss": 74.8561, "step": 340 }, { "compression_loss": 78.59368896484375, "epoch": 0.11, "learning_rate": 2.8344283837056504e-05, "loss": 79.5413, "step": 341 }, { "compression_loss": 76.27571868896484, "epoch": 0.11, "learning_rate": 2.8339356110381076e-05, "loss": 76.6335, "step": 342 }, { "compression_loss": 73.20640563964844, "epoch": 0.11, "learning_rate": 2.833442838370565e-05, "loss": 73.5435, "step": 343 }, { "compression_loss": 74.97444152832031, "epoch": 0.11, "learning_rate": 2.8329500657030224e-05, "loss": 75.3985, "step": 344 }, { "compression_loss": 74.62602233886719, "epoch": 0.11, "learning_rate": 2.83245729303548e-05, "loss": 74.8896, "step": 345 }, { "compression_loss": 74.36958312988281, "epoch": 0.11, "learning_rate": 2.831964520367937e-05, "loss": 74.7326, "step": 346 }, { "compression_loss": 75.52775573730469, "epoch": 0.11, "learning_rate": 2.8314717477003943e-05, "loss": 76.3816, "step": 347 }, { "compression_loss": 74.78535461425781, "epoch": 0.11, "learning_rate": 2.8309789750328518e-05, "loss": 75.3478, "step": 348 }, { "compression_loss": 74.62637329101562, "epoch": 0.11, "learning_rate": 2.830486202365309e-05, "loss": 75.2407, "step": 349 }, { "compression_loss": 76.46245574951172, "epoch": 0.11, "learning_rate": 2.8299934296977662e-05, "loss": 76.8143, "step": 350 }, { "compression_loss": 77.4979019165039, "epoch": 0.12, "learning_rate": 2.8295006570302234e-05, "loss": 78.0427, "step": 351 }, { "compression_loss": 75.13148498535156, "epoch": 0.12, "learning_rate": 2.8290078843626806e-05, "loss": 75.5757, "step": 352 }, { "compression_loss": 76.80973815917969, "epoch": 0.12, "learning_rate": 2.828515111695138e-05, "loss": 77.4305, "step": 353 }, { "compression_loss": 72.96717834472656, "epoch": 0.12, "learning_rate": 2.8280223390275953e-05, "loss": 73.4223, "step": 354 }, { "compression_loss": 77.12020111083984, "epoch": 0.12, "learning_rate": 2.8275295663600525e-05, "loss": 77.9308, "step": 355 }, { "compression_loss": 77.11380004882812, "epoch": 0.12, "learning_rate": 2.8270367936925097e-05, "loss": 77.7013, "step": 356 }, { "compression_loss": 77.55107879638672, "epoch": 0.12, "learning_rate": 2.8265440210249673e-05, "loss": 78.5513, "step": 357 }, { "compression_loss": 75.41655731201172, "epoch": 0.12, "learning_rate": 2.8260512483574248e-05, "loss": 75.8115, "step": 358 }, { "compression_loss": 74.89767456054688, "epoch": 0.12, "learning_rate": 2.825558475689882e-05, "loss": 75.092, "step": 359 }, { "compression_loss": 75.57402038574219, "epoch": 0.12, "learning_rate": 2.8250657030223392e-05, "loss": 76.1566, "step": 360 }, { "compression_loss": 75.18212890625, "epoch": 0.12, "learning_rate": 2.8245729303547964e-05, "loss": 75.5998, "step": 361 }, { "compression_loss": 76.35773468017578, "epoch": 0.12, "learning_rate": 2.8240801576872536e-05, "loss": 76.9891, "step": 362 }, { "compression_loss": 76.31321716308594, "epoch": 0.12, "learning_rate": 2.823587385019711e-05, "loss": 76.7684, "step": 363 }, { "compression_loss": 77.01243591308594, "epoch": 0.12, "learning_rate": 2.8230946123521683e-05, "loss": 78.1152, "step": 364 }, { "compression_loss": 76.90251159667969, "epoch": 0.12, "learning_rate": 2.8226018396846255e-05, "loss": 77.3798, "step": 365 }, { "compression_loss": 75.20462036132812, "epoch": 0.12, "learning_rate": 2.8221090670170827e-05, "loss": 75.5106, "step": 366 }, { "compression_loss": 76.16737365722656, "epoch": 0.12, "learning_rate": 2.82161629434954e-05, "loss": 76.5491, "step": 367 }, { "compression_loss": 75.79910278320312, "epoch": 0.12, "learning_rate": 2.8211235216819975e-05, "loss": 76.3092, "step": 368 }, { "compression_loss": 75.85011291503906, "epoch": 0.12, "learning_rate": 2.820630749014455e-05, "loss": 76.0653, "step": 369 }, { "compression_loss": 76.33551025390625, "epoch": 0.12, "learning_rate": 2.8201379763469122e-05, "loss": 76.8052, "step": 370 }, { "compression_loss": 74.44230651855469, "epoch": 0.12, "learning_rate": 2.8196452036793694e-05, "loss": 74.9871, "step": 371 }, { "compression_loss": 73.81434631347656, "epoch": 0.12, "learning_rate": 2.8191524310118266e-05, "loss": 74.2229, "step": 372 }, { "compression_loss": 77.47750091552734, "epoch": 0.12, "learning_rate": 2.818659658344284e-05, "loss": 77.8123, "step": 373 }, { "compression_loss": 75.79161071777344, "epoch": 0.12, "learning_rate": 2.8181668856767413e-05, "loss": 76.2716, "step": 374 }, { "compression_loss": 76.20216369628906, "epoch": 0.12, "learning_rate": 2.8176741130091985e-05, "loss": 76.5844, "step": 375 }, { "compression_loss": 75.81146240234375, "epoch": 0.12, "learning_rate": 2.8171813403416557e-05, "loss": 76.4238, "step": 376 }, { "compression_loss": 75.19000244140625, "epoch": 0.12, "learning_rate": 2.816688567674113e-05, "loss": 75.6526, "step": 377 }, { "compression_loss": 76.01669311523438, "epoch": 0.12, "learning_rate": 2.8161957950065704e-05, "loss": 76.2567, "step": 378 }, { "compression_loss": 75.26980590820312, "epoch": 0.12, "learning_rate": 2.8157030223390276e-05, "loss": 75.622, "step": 379 }, { "compression_loss": 75.26224517822266, "epoch": 0.12, "learning_rate": 2.815210249671485e-05, "loss": 75.4881, "step": 380 }, { "compression_loss": 76.83143615722656, "epoch": 0.13, "learning_rate": 2.8147174770039424e-05, "loss": 77.0986, "step": 381 }, { "compression_loss": 75.34893035888672, "epoch": 0.13, "learning_rate": 2.8142247043363996e-05, "loss": 75.7328, "step": 382 }, { "compression_loss": 77.53665924072266, "epoch": 0.13, "learning_rate": 2.813731931668857e-05, "loss": 78.1777, "step": 383 }, { "compression_loss": 74.71426391601562, "epoch": 0.13, "learning_rate": 2.8132391590013143e-05, "loss": 75.1123, "step": 384 }, { "compression_loss": 73.80400085449219, "epoch": 0.13, "learning_rate": 2.8127463863337715e-05, "loss": 74.0731, "step": 385 }, { "compression_loss": 77.93877410888672, "epoch": 0.13, "learning_rate": 2.8122536136662287e-05, "loss": 78.589, "step": 386 }, { "compression_loss": 75.97171783447266, "epoch": 0.13, "learning_rate": 2.811760840998686e-05, "loss": 76.4469, "step": 387 }, { "compression_loss": 75.92143249511719, "epoch": 0.13, "learning_rate": 2.811268068331143e-05, "loss": 76.2616, "step": 388 }, { "compression_loss": 76.41030883789062, "epoch": 0.13, "learning_rate": 2.8107752956636006e-05, "loss": 76.8097, "step": 389 }, { "compression_loss": 75.85725402832031, "epoch": 0.13, "learning_rate": 2.8102825229960578e-05, "loss": 76.1612, "step": 390 }, { "compression_loss": 76.09806823730469, "epoch": 0.13, "learning_rate": 2.809789750328515e-05, "loss": 76.5047, "step": 391 }, { "compression_loss": 75.06419372558594, "epoch": 0.13, "learning_rate": 2.8092969776609722e-05, "loss": 75.6712, "step": 392 }, { "compression_loss": 78.04553985595703, "epoch": 0.13, "learning_rate": 2.8088042049934297e-05, "loss": 78.4681, "step": 393 }, { "compression_loss": 74.87519836425781, "epoch": 0.13, "learning_rate": 2.8083114323258873e-05, "loss": 75.2377, "step": 394 }, { "compression_loss": 75.23556518554688, "epoch": 0.13, "learning_rate": 2.8078186596583445e-05, "loss": 75.8558, "step": 395 }, { "compression_loss": 77.69450378417969, "epoch": 0.13, "learning_rate": 2.8073258869908017e-05, "loss": 78.1283, "step": 396 }, { "compression_loss": 75.28756713867188, "epoch": 0.13, "learning_rate": 2.806833114323259e-05, "loss": 75.6582, "step": 397 }, { "compression_loss": 74.87014770507812, "epoch": 0.13, "learning_rate": 2.806340341655716e-05, "loss": 75.1869, "step": 398 }, { "compression_loss": 75.79496002197266, "epoch": 0.13, "learning_rate": 2.8058475689881736e-05, "loss": 76.2064, "step": 399 }, { "compression_loss": 75.75946044921875, "epoch": 0.13, "learning_rate": 2.8053547963206308e-05, "loss": 76.0834, "step": 400 }, { "compression_loss": 78.20459747314453, "epoch": 0.13, "learning_rate": 2.804862023653088e-05, "loss": 78.6877, "step": 401 }, { "compression_loss": 76.69354248046875, "epoch": 0.13, "learning_rate": 2.8043692509855452e-05, "loss": 77.0423, "step": 402 }, { "compression_loss": 75.500244140625, "epoch": 0.13, "learning_rate": 2.8038764783180024e-05, "loss": 75.996, "step": 403 }, { "compression_loss": 75.691162109375, "epoch": 0.13, "learning_rate": 2.80338370565046e-05, "loss": 76.1806, "step": 404 }, { "compression_loss": 75.37199401855469, "epoch": 0.13, "learning_rate": 2.8028909329829175e-05, "loss": 75.7177, "step": 405 }, { "compression_loss": 78.19871520996094, "epoch": 0.13, "learning_rate": 2.8023981603153747e-05, "loss": 79.1633, "step": 406 }, { "compression_loss": 75.05072021484375, "epoch": 0.13, "learning_rate": 2.801905387647832e-05, "loss": 75.875, "step": 407 }, { "compression_loss": 79.16210174560547, "epoch": 0.13, "learning_rate": 2.801412614980289e-05, "loss": 79.9137, "step": 408 }, { "compression_loss": 75.24285125732422, "epoch": 0.13, "learning_rate": 2.8009198423127466e-05, "loss": 75.7783, "step": 409 }, { "compression_loss": 73.85977172851562, "epoch": 0.13, "learning_rate": 2.8004270696452038e-05, "loss": 74.1061, "step": 410 }, { "compression_loss": 76.04341125488281, "epoch": 0.14, "learning_rate": 2.799934296977661e-05, "loss": 76.4516, "step": 411 }, { "compression_loss": 75.88397216796875, "epoch": 0.14, "learning_rate": 2.7994415243101182e-05, "loss": 76.3722, "step": 412 }, { "compression_loss": 75.10623931884766, "epoch": 0.14, "learning_rate": 2.7989487516425754e-05, "loss": 75.4769, "step": 413 }, { "compression_loss": 75.89085388183594, "epoch": 0.14, "learning_rate": 2.798455978975033e-05, "loss": 76.2772, "step": 414 }, { "compression_loss": 79.3668212890625, "epoch": 0.14, "learning_rate": 2.79796320630749e-05, "loss": 80.0559, "step": 415 }, { "compression_loss": 76.00251770019531, "epoch": 0.14, "learning_rate": 2.7974704336399473e-05, "loss": 76.693, "step": 416 }, { "compression_loss": 75.89408874511719, "epoch": 0.14, "learning_rate": 2.796977660972405e-05, "loss": 76.4465, "step": 417 }, { "compression_loss": 75.64167785644531, "epoch": 0.14, "learning_rate": 2.796484888304862e-05, "loss": 76.162, "step": 418 }, { "compression_loss": 75.50347137451172, "epoch": 0.14, "learning_rate": 2.7959921156373196e-05, "loss": 76.289, "step": 419 }, { "compression_loss": 73.68327331542969, "epoch": 0.14, "learning_rate": 2.7954993429697768e-05, "loss": 74.087, "step": 420 }, { "compression_loss": 75.82661437988281, "epoch": 0.14, "learning_rate": 2.795006570302234e-05, "loss": 76.3135, "step": 421 }, { "compression_loss": 76.2021255493164, "epoch": 0.14, "learning_rate": 2.794513797634691e-05, "loss": 76.6124, "step": 422 }, { "compression_loss": 75.82408905029297, "epoch": 0.14, "learning_rate": 2.7940210249671484e-05, "loss": 76.6854, "step": 423 }, { "compression_loss": 76.21420288085938, "epoch": 0.14, "learning_rate": 2.793528252299606e-05, "loss": 76.7414, "step": 424 }, { "compression_loss": 73.57060241699219, "epoch": 0.14, "learning_rate": 2.793035479632063e-05, "loss": 73.8323, "step": 425 }, { "compression_loss": 74.60420227050781, "epoch": 0.14, "learning_rate": 2.7925427069645203e-05, "loss": 74.7916, "step": 426 }, { "compression_loss": 74.14905548095703, "epoch": 0.14, "learning_rate": 2.7920499342969775e-05, "loss": 74.6859, "step": 427 }, { "compression_loss": 74.38329315185547, "epoch": 0.14, "learning_rate": 2.791557161629435e-05, "loss": 74.6684, "step": 428 }, { "compression_loss": 76.31439208984375, "epoch": 0.14, "learning_rate": 2.7910643889618926e-05, "loss": 76.652, "step": 429 }, { "compression_loss": 73.93161010742188, "epoch": 0.14, "learning_rate": 2.7905716162943498e-05, "loss": 74.7206, "step": 430 }, { "compression_loss": 75.37800598144531, "epoch": 0.14, "learning_rate": 2.790078843626807e-05, "loss": 75.7538, "step": 431 }, { "compression_loss": 76.68069458007812, "epoch": 0.14, "learning_rate": 2.789586070959264e-05, "loss": 77.1606, "step": 432 }, { "compression_loss": 73.79312133789062, "epoch": 0.14, "learning_rate": 2.7890932982917214e-05, "loss": 74.3025, "step": 433 }, { "compression_loss": 76.89445495605469, "epoch": 0.14, "learning_rate": 2.788600525624179e-05, "loss": 77.2526, "step": 434 }, { "compression_loss": 73.53419494628906, "epoch": 0.14, "learning_rate": 2.788107752956636e-05, "loss": 73.8398, "step": 435 }, { "compression_loss": 74.4104232788086, "epoch": 0.14, "learning_rate": 2.7876149802890933e-05, "loss": 74.953, "step": 436 }, { "compression_loss": 74.31605529785156, "epoch": 0.14, "learning_rate": 2.7871222076215505e-05, "loss": 74.6081, "step": 437 }, { "compression_loss": 76.38995361328125, "epoch": 0.14, "learning_rate": 2.7866294349540077e-05, "loss": 76.5242, "step": 438 }, { "compression_loss": 76.92686462402344, "epoch": 0.14, "learning_rate": 2.7861366622864652e-05, "loss": 77.3731, "step": 439 }, { "compression_loss": 73.52784729003906, "epoch": 0.14, "learning_rate": 2.7856438896189227e-05, "loss": 73.7497, "step": 440 }, { "compression_loss": 75.53363037109375, "epoch": 0.14, "learning_rate": 2.78515111695138e-05, "loss": 75.9122, "step": 441 }, { "compression_loss": 75.63461303710938, "epoch": 0.15, "learning_rate": 2.784658344283837e-05, "loss": 76.8165, "step": 442 }, { "compression_loss": 76.34661865234375, "epoch": 0.15, "learning_rate": 2.7841655716162943e-05, "loss": 76.4572, "step": 443 }, { "compression_loss": 76.02648162841797, "epoch": 0.15, "learning_rate": 2.783672798948752e-05, "loss": 76.3532, "step": 444 }, { "compression_loss": 76.35661315917969, "epoch": 0.15, "learning_rate": 2.783180026281209e-05, "loss": 76.8314, "step": 445 }, { "compression_loss": 74.83999633789062, "epoch": 0.15, "learning_rate": 2.7826872536136663e-05, "loss": 75.3405, "step": 446 }, { "compression_loss": 76.00773620605469, "epoch": 0.15, "learning_rate": 2.7821944809461235e-05, "loss": 76.5909, "step": 447 }, { "compression_loss": 74.89166259765625, "epoch": 0.15, "learning_rate": 2.7817017082785807e-05, "loss": 75.5954, "step": 448 }, { "compression_loss": 74.87249755859375, "epoch": 0.15, "learning_rate": 2.7812089356110382e-05, "loss": 75.3042, "step": 449 }, { "compression_loss": 75.02842712402344, "epoch": 0.15, "learning_rate": 2.7807161629434954e-05, "loss": 75.3914, "step": 450 }, { "compression_loss": 74.77156066894531, "epoch": 0.15, "learning_rate": 2.7802233902759526e-05, "loss": 75.2346, "step": 451 }, { "compression_loss": 75.05839538574219, "epoch": 0.15, "learning_rate": 2.77973061760841e-05, "loss": 75.2816, "step": 452 }, { "compression_loss": 76.4306640625, "epoch": 0.15, "learning_rate": 2.7792378449408673e-05, "loss": 77.0553, "step": 453 }, { "compression_loss": 75.53495788574219, "epoch": 0.15, "learning_rate": 2.778745072273325e-05, "loss": 76.1235, "step": 454 }, { "compression_loss": 75.47538757324219, "epoch": 0.15, "learning_rate": 2.778252299605782e-05, "loss": 75.913, "step": 455 }, { "compression_loss": 76.03170013427734, "epoch": 0.15, "learning_rate": 2.7777595269382393e-05, "loss": 76.4691, "step": 456 }, { "compression_loss": 75.52935791015625, "epoch": 0.15, "learning_rate": 2.7772667542706964e-05, "loss": 75.969, "step": 457 }, { "compression_loss": 77.61917114257812, "epoch": 0.15, "learning_rate": 2.7767739816031536e-05, "loss": 78.1161, "step": 458 }, { "compression_loss": 76.44485473632812, "epoch": 0.15, "learning_rate": 2.7762812089356112e-05, "loss": 77.2068, "step": 459 }, { "compression_loss": 73.35049438476562, "epoch": 0.15, "learning_rate": 2.7757884362680684e-05, "loss": 73.51, "step": 460 }, { "compression_loss": 75.25090026855469, "epoch": 0.15, "learning_rate": 2.7752956636005256e-05, "loss": 75.7175, "step": 461 }, { "compression_loss": 73.4625473022461, "epoch": 0.15, "learning_rate": 2.7748028909329828e-05, "loss": 73.7014, "step": 462 }, { "compression_loss": 76.61968994140625, "epoch": 0.15, "learning_rate": 2.77431011826544e-05, "loss": 77.0069, "step": 463 }, { "compression_loss": 80.01747131347656, "epoch": 0.15, "learning_rate": 2.773817345597898e-05, "loss": 81.1163, "step": 464 }, { "compression_loss": 75.04171752929688, "epoch": 0.15, "learning_rate": 2.773324572930355e-05, "loss": 75.4889, "step": 465 }, { "compression_loss": 75.09938049316406, "epoch": 0.15, "learning_rate": 2.7728318002628122e-05, "loss": 75.3875, "step": 466 }, { "compression_loss": 76.67385864257812, "epoch": 0.15, "learning_rate": 2.7723390275952694e-05, "loss": 77.231, "step": 467 }, { "compression_loss": 76.23588562011719, "epoch": 0.15, "learning_rate": 2.7718462549277266e-05, "loss": 76.7477, "step": 468 }, { "compression_loss": 77.27787780761719, "epoch": 0.15, "learning_rate": 2.771353482260184e-05, "loss": 77.7479, "step": 469 }, { "compression_loss": 76.54177856445312, "epoch": 0.15, "learning_rate": 2.7708607095926414e-05, "loss": 77.1848, "step": 470 }, { "compression_loss": 74.92256164550781, "epoch": 0.15, "learning_rate": 2.7703679369250986e-05, "loss": 75.4499, "step": 471 }, { "compression_loss": 76.26750946044922, "epoch": 0.16, "learning_rate": 2.7698751642575558e-05, "loss": 76.5807, "step": 472 }, { "compression_loss": 76.13792419433594, "epoch": 0.16, "learning_rate": 2.769382391590013e-05, "loss": 76.4632, "step": 473 }, { "compression_loss": 74.9589614868164, "epoch": 0.16, "learning_rate": 2.7688896189224705e-05, "loss": 75.4841, "step": 474 }, { "compression_loss": 74.16690063476562, "epoch": 0.16, "learning_rate": 2.7683968462549277e-05, "loss": 74.5083, "step": 475 }, { "compression_loss": 78.15205383300781, "epoch": 0.16, "learning_rate": 2.7679040735873852e-05, "loss": 78.9502, "step": 476 }, { "compression_loss": 76.61897277832031, "epoch": 0.16, "learning_rate": 2.7674113009198424e-05, "loss": 76.8873, "step": 477 }, { "compression_loss": 78.07237243652344, "epoch": 0.16, "learning_rate": 2.7669185282522996e-05, "loss": 78.4409, "step": 478 }, { "compression_loss": 74.38687133789062, "epoch": 0.16, "learning_rate": 2.766425755584757e-05, "loss": 75.027, "step": 479 }, { "compression_loss": 75.3274917602539, "epoch": 0.16, "learning_rate": 2.7659329829172143e-05, "loss": 76.0962, "step": 480 }, { "compression_loss": 77.6451416015625, "epoch": 0.16, "learning_rate": 2.7654402102496715e-05, "loss": 78.9529, "step": 481 }, { "compression_loss": 75.81214904785156, "epoch": 0.16, "learning_rate": 2.7649474375821287e-05, "loss": 76.2525, "step": 482 }, { "compression_loss": 75.2957763671875, "epoch": 0.16, "learning_rate": 2.764454664914586e-05, "loss": 75.5826, "step": 483 }, { "compression_loss": 78.7001724243164, "epoch": 0.16, "learning_rate": 2.7639618922470435e-05, "loss": 79.0788, "step": 484 }, { "compression_loss": 76.38097381591797, "epoch": 0.16, "learning_rate": 2.7634691195795007e-05, "loss": 77.1575, "step": 485 }, { "compression_loss": 75.01370239257812, "epoch": 0.16, "learning_rate": 2.762976346911958e-05, "loss": 75.3615, "step": 486 }, { "compression_loss": 75.798095703125, "epoch": 0.16, "learning_rate": 2.762483574244415e-05, "loss": 76.0433, "step": 487 }, { "compression_loss": 76.85051727294922, "epoch": 0.16, "learning_rate": 2.7619908015768726e-05, "loss": 77.1938, "step": 488 }, { "compression_loss": 76.41130065917969, "epoch": 0.16, "learning_rate": 2.76149802890933e-05, "loss": 76.8143, "step": 489 }, { "compression_loss": 76.31922149658203, "epoch": 0.16, "learning_rate": 2.7610052562417873e-05, "loss": 77.0005, "step": 490 }, { "compression_loss": 76.05142974853516, "epoch": 0.16, "learning_rate": 2.7605124835742445e-05, "loss": 76.3037, "step": 491 }, { "compression_loss": 77.02668762207031, "epoch": 0.16, "learning_rate": 2.7600197109067017e-05, "loss": 77.5429, "step": 492 }, { "compression_loss": 74.64412689208984, "epoch": 0.16, "learning_rate": 2.759526938239159e-05, "loss": 75.0376, "step": 493 }, { "compression_loss": 73.28544616699219, "epoch": 0.16, "learning_rate": 2.7590341655716165e-05, "loss": 73.616, "step": 494 }, { "compression_loss": 75.7469482421875, "epoch": 0.16, "learning_rate": 2.7585413929040737e-05, "loss": 76.3474, "step": 495 }, { "compression_loss": 74.25872802734375, "epoch": 0.16, "learning_rate": 2.758048620236531e-05, "loss": 74.8209, "step": 496 }, { "compression_loss": 73.32994079589844, "epoch": 0.16, "learning_rate": 2.757555847568988e-05, "loss": 73.694, "step": 497 }, { "compression_loss": 73.65888214111328, "epoch": 0.16, "learning_rate": 2.7570630749014452e-05, "loss": 74.094, "step": 498 }, { "compression_loss": 72.74990844726562, "epoch": 0.16, "learning_rate": 2.7565703022339028e-05, "loss": 73.147, "step": 499 }, { "compression_loss": 75.53433990478516, "epoch": 0.16, "learning_rate": 2.7560775295663603e-05, "loss": 75.8766, "step": 500 }, { "epoch": 0.16, "eval_exact_match": 85.7899716177862, "eval_f1": 92.21490101898762, "step": 500 }, { "compression_loss": 76.51647186279297, "epoch": 0.16, "learning_rate": 2.7555847568988175e-05, "loss": 77.2458, "step": 501 }, { "compression_loss": 74.85919952392578, "epoch": 0.16, "learning_rate": 2.7550919842312747e-05, "loss": 75.1835, "step": 502 }, { "compression_loss": 75.40071868896484, "epoch": 0.17, "learning_rate": 2.754599211563732e-05, "loss": 75.8075, "step": 503 }, { "compression_loss": 74.68911743164062, "epoch": 0.17, "learning_rate": 2.7541064388961894e-05, "loss": 74.9951, "step": 504 }, { "compression_loss": 72.78173065185547, "epoch": 0.17, "learning_rate": 2.7536136662286466e-05, "loss": 73.0543, "step": 505 }, { "compression_loss": 74.01394653320312, "epoch": 0.17, "learning_rate": 2.753120893561104e-05, "loss": 74.4047, "step": 506 }, { "compression_loss": 73.48475646972656, "epoch": 0.17, "learning_rate": 2.752628120893561e-05, "loss": 73.9042, "step": 507 }, { "compression_loss": 74.50301361083984, "epoch": 0.17, "learning_rate": 2.7521353482260182e-05, "loss": 74.9341, "step": 508 }, { "compression_loss": 77.36134338378906, "epoch": 0.17, "learning_rate": 2.7516425755584758e-05, "loss": 77.6779, "step": 509 }, { "compression_loss": 75.77466583251953, "epoch": 0.17, "learning_rate": 2.751149802890933e-05, "loss": 76.4035, "step": 510 }, { "compression_loss": 74.36112976074219, "epoch": 0.17, "learning_rate": 2.75065703022339e-05, "loss": 74.6296, "step": 511 }, { "compression_loss": 74.2507095336914, "epoch": 0.17, "learning_rate": 2.7501642575558477e-05, "loss": 74.563, "step": 512 }, { "compression_loss": 75.5224609375, "epoch": 0.17, "learning_rate": 2.749671484888305e-05, "loss": 76.0219, "step": 513 }, { "compression_loss": 75.13352966308594, "epoch": 0.17, "learning_rate": 2.7491787122207624e-05, "loss": 75.5394, "step": 514 }, { "compression_loss": 75.13533782958984, "epoch": 0.17, "learning_rate": 2.7486859395532196e-05, "loss": 75.7753, "step": 515 }, { "compression_loss": 77.06935119628906, "epoch": 0.17, "learning_rate": 2.7481931668856768e-05, "loss": 77.6862, "step": 516 }, { "compression_loss": 75.80552673339844, "epoch": 0.17, "learning_rate": 2.747700394218134e-05, "loss": 76.1621, "step": 517 }, { "compression_loss": 75.503173828125, "epoch": 0.17, "learning_rate": 2.7472076215505912e-05, "loss": 75.7387, "step": 518 }, { "compression_loss": 77.50724792480469, "epoch": 0.17, "learning_rate": 2.7467148488830488e-05, "loss": 78.0222, "step": 519 }, { "compression_loss": 75.50270080566406, "epoch": 0.17, "learning_rate": 2.746222076215506e-05, "loss": 75.917, "step": 520 }, { "compression_loss": 73.47384643554688, "epoch": 0.17, "learning_rate": 2.745729303547963e-05, "loss": 73.8043, "step": 521 }, { "compression_loss": 76.24211120605469, "epoch": 0.17, "learning_rate": 2.7452365308804203e-05, "loss": 76.499, "step": 522 }, { "compression_loss": 77.0712890625, "epoch": 0.17, "learning_rate": 2.7447437582128775e-05, "loss": 77.7956, "step": 523 }, { "compression_loss": 72.2490234375, "epoch": 0.17, "learning_rate": 2.7442509855453354e-05, "loss": 72.747, "step": 524 }, { "compression_loss": 76.18161010742188, "epoch": 0.17, "learning_rate": 2.7437582128777926e-05, "loss": 76.8569, "step": 525 }, { "compression_loss": 74.50393676757812, "epoch": 0.17, "learning_rate": 2.7432654402102498e-05, "loss": 74.9609, "step": 526 }, { "compression_loss": 74.25588989257812, "epoch": 0.17, "learning_rate": 2.742772667542707e-05, "loss": 74.5196, "step": 527 }, { "compression_loss": 77.55440521240234, "epoch": 0.17, "learning_rate": 2.7422798948751642e-05, "loss": 78.6198, "step": 528 }, { "compression_loss": 77.597412109375, "epoch": 0.17, "learning_rate": 2.7417871222076217e-05, "loss": 78.0405, "step": 529 }, { "compression_loss": 75.96395874023438, "epoch": 0.17, "learning_rate": 2.741294349540079e-05, "loss": 76.5567, "step": 530 }, { "compression_loss": 75.84383392333984, "epoch": 0.17, "learning_rate": 2.740801576872536e-05, "loss": 76.2845, "step": 531 }, { "compression_loss": 78.63685607910156, "epoch": 0.17, "learning_rate": 2.7403088042049933e-05, "loss": 79.3222, "step": 532 }, { "compression_loss": 76.22210693359375, "epoch": 0.18, "learning_rate": 2.7398160315374505e-05, "loss": 76.6742, "step": 533 }, { "compression_loss": 76.280029296875, "epoch": 0.18, "learning_rate": 2.739323258869908e-05, "loss": 76.7468, "step": 534 }, { "compression_loss": 77.13697814941406, "epoch": 0.18, "learning_rate": 2.7388304862023653e-05, "loss": 77.6616, "step": 535 }, { "compression_loss": 74.05767059326172, "epoch": 0.18, "learning_rate": 2.7383377135348228e-05, "loss": 74.4802, "step": 536 }, { "compression_loss": 78.27745056152344, "epoch": 0.18, "learning_rate": 2.73784494086728e-05, "loss": 78.9935, "step": 537 }, { "compression_loss": 78.07682800292969, "epoch": 0.18, "learning_rate": 2.7373521681997372e-05, "loss": 78.4121, "step": 538 }, { "compression_loss": 77.05804443359375, "epoch": 0.18, "learning_rate": 2.7368593955321947e-05, "loss": 78.0842, "step": 539 }, { "compression_loss": 77.92382049560547, "epoch": 0.18, "learning_rate": 2.736366622864652e-05, "loss": 78.2235, "step": 540 }, { "compression_loss": 76.2836685180664, "epoch": 0.18, "learning_rate": 2.735873850197109e-05, "loss": 76.7702, "step": 541 }, { "compression_loss": 78.4761734008789, "epoch": 0.18, "learning_rate": 2.7353810775295663e-05, "loss": 79.1915, "step": 542 }, { "compression_loss": 74.76315307617188, "epoch": 0.18, "learning_rate": 2.7348883048620235e-05, "loss": 74.9585, "step": 543 }, { "compression_loss": 76.63992309570312, "epoch": 0.18, "learning_rate": 2.734395532194481e-05, "loss": 76.9503, "step": 544 }, { "compression_loss": 77.60247039794922, "epoch": 0.18, "learning_rate": 2.7339027595269382e-05, "loss": 78.1874, "step": 545 }, { "compression_loss": 75.07341003417969, "epoch": 0.18, "learning_rate": 2.7334099868593954e-05, "loss": 75.4752, "step": 546 }, { "compression_loss": 77.446044921875, "epoch": 0.18, "learning_rate": 2.7329172141918526e-05, "loss": 77.9623, "step": 547 }, { "compression_loss": 76.68757629394531, "epoch": 0.18, "learning_rate": 2.7324244415243102e-05, "loss": 77.2783, "step": 548 }, { "compression_loss": 75.41386413574219, "epoch": 0.18, "learning_rate": 2.7319316688567677e-05, "loss": 76.1162, "step": 549 }, { "compression_loss": 76.82261657714844, "epoch": 0.18, "learning_rate": 2.731438896189225e-05, "loss": 77.0717, "step": 550 }, { "compression_loss": 73.32501983642578, "epoch": 0.18, "learning_rate": 2.730946123521682e-05, "loss": 73.8488, "step": 551 }, { "compression_loss": 77.94982147216797, "epoch": 0.18, "learning_rate": 2.7304533508541393e-05, "loss": 78.6609, "step": 552 }, { "compression_loss": 76.47439575195312, "epoch": 0.18, "learning_rate": 2.7299605781865965e-05, "loss": 76.9373, "step": 553 }, { "compression_loss": 74.45624542236328, "epoch": 0.18, "learning_rate": 2.729467805519054e-05, "loss": 74.7848, "step": 554 }, { "compression_loss": 75.61283874511719, "epoch": 0.18, "learning_rate": 2.7289750328515112e-05, "loss": 76.2604, "step": 555 }, { "compression_loss": 76.48771667480469, "epoch": 0.18, "learning_rate": 2.7284822601839684e-05, "loss": 76.8844, "step": 556 }, { "compression_loss": 74.2625961303711, "epoch": 0.18, "learning_rate": 2.7279894875164256e-05, "loss": 74.643, "step": 557 }, { "compression_loss": 74.41314697265625, "epoch": 0.18, "learning_rate": 2.7274967148488828e-05, "loss": 74.7269, "step": 558 }, { "compression_loss": 76.11077880859375, "epoch": 0.18, "learning_rate": 2.7270039421813404e-05, "loss": 76.3789, "step": 559 }, { "compression_loss": 75.07797241210938, "epoch": 0.18, "learning_rate": 2.726511169513798e-05, "loss": 75.3963, "step": 560 }, { "compression_loss": 77.50462341308594, "epoch": 0.18, "learning_rate": 2.726018396846255e-05, "loss": 77.752, "step": 561 }, { "compression_loss": 76.00702667236328, "epoch": 0.18, "learning_rate": 2.7255256241787123e-05, "loss": 76.2875, "step": 562 }, { "compression_loss": 75.02839660644531, "epoch": 0.18, "learning_rate": 2.7250328515111695e-05, "loss": 75.7738, "step": 563 }, { "compression_loss": 76.494140625, "epoch": 0.19, "learning_rate": 2.724540078843627e-05, "loss": 76.8136, "step": 564 }, { "compression_loss": 78.44065856933594, "epoch": 0.19, "learning_rate": 2.7240473061760842e-05, "loss": 79.0148, "step": 565 }, { "compression_loss": 75.6702880859375, "epoch": 0.19, "learning_rate": 2.7235545335085414e-05, "loss": 75.9206, "step": 566 }, { "compression_loss": 75.74616241455078, "epoch": 0.19, "learning_rate": 2.7230617608409986e-05, "loss": 76.1495, "step": 567 }, { "compression_loss": 75.40595245361328, "epoch": 0.19, "learning_rate": 2.7225689881734558e-05, "loss": 75.898, "step": 568 }, { "compression_loss": 74.05290222167969, "epoch": 0.19, "learning_rate": 2.7220762155059133e-05, "loss": 74.4609, "step": 569 }, { "compression_loss": 76.64447021484375, "epoch": 0.19, "learning_rate": 2.7215834428383705e-05, "loss": 76.8277, "step": 570 }, { "compression_loss": 76.64404296875, "epoch": 0.19, "learning_rate": 2.721090670170828e-05, "loss": 77.3354, "step": 571 }, { "compression_loss": 75.10938262939453, "epoch": 0.19, "learning_rate": 2.7205978975032853e-05, "loss": 75.6476, "step": 572 }, { "compression_loss": 77.32487487792969, "epoch": 0.19, "learning_rate": 2.7201051248357425e-05, "loss": 77.8342, "step": 573 }, { "compression_loss": 75.331787109375, "epoch": 0.19, "learning_rate": 2.7196123521682e-05, "loss": 75.8464, "step": 574 }, { "compression_loss": 75.74727630615234, "epoch": 0.19, "learning_rate": 2.7191195795006572e-05, "loss": 76.1313, "step": 575 }, { "compression_loss": 78.9439926147461, "epoch": 0.19, "learning_rate": 2.7186268068331144e-05, "loss": 79.5363, "step": 576 }, { "compression_loss": 75.53726196289062, "epoch": 0.19, "learning_rate": 2.7181340341655716e-05, "loss": 75.9412, "step": 577 }, { "compression_loss": 74.64865112304688, "epoch": 0.19, "learning_rate": 2.7176412614980288e-05, "loss": 75.4367, "step": 578 }, { "compression_loss": 75.21931457519531, "epoch": 0.19, "learning_rate": 2.7171484888304863e-05, "loss": 75.7361, "step": 579 }, { "compression_loss": 75.58684539794922, "epoch": 0.19, "learning_rate": 2.7166557161629435e-05, "loss": 75.8535, "step": 580 }, { "compression_loss": 74.02142333984375, "epoch": 0.19, "learning_rate": 2.7161629434954007e-05, "loss": 74.2561, "step": 581 }, { "compression_loss": 75.13946533203125, "epoch": 0.19, "learning_rate": 2.715670170827858e-05, "loss": 75.4513, "step": 582 }, { "compression_loss": 75.48766326904297, "epoch": 0.19, "learning_rate": 2.7151773981603155e-05, "loss": 75.8068, "step": 583 }, { "compression_loss": 74.811767578125, "epoch": 0.19, "learning_rate": 2.714684625492773e-05, "loss": 75.415, "step": 584 }, { "compression_loss": 75.61771392822266, "epoch": 0.19, "learning_rate": 2.7141918528252302e-05, "loss": 76.3173, "step": 585 }, { "compression_loss": 77.91828918457031, "epoch": 0.19, "learning_rate": 2.7136990801576874e-05, "loss": 78.3992, "step": 586 }, { "compression_loss": 74.54747009277344, "epoch": 0.19, "learning_rate": 2.7132063074901446e-05, "loss": 75.0263, "step": 587 }, { "compression_loss": 74.3417739868164, "epoch": 0.19, "learning_rate": 2.7127135348226018e-05, "loss": 74.735, "step": 588 }, { "compression_loss": 75.6622543334961, "epoch": 0.19, "learning_rate": 2.7122207621550593e-05, "loss": 76.0644, "step": 589 }, { "compression_loss": 76.49593353271484, "epoch": 0.19, "learning_rate": 2.7117279894875165e-05, "loss": 76.8781, "step": 590 }, { "compression_loss": 74.07479858398438, "epoch": 0.19, "learning_rate": 2.7112352168199737e-05, "loss": 74.3966, "step": 591 }, { "compression_loss": 72.13371276855469, "epoch": 0.19, "learning_rate": 2.710742444152431e-05, "loss": 72.6106, "step": 592 }, { "compression_loss": 76.43641662597656, "epoch": 0.19, "learning_rate": 2.710249671484888e-05, "loss": 76.7801, "step": 593 }, { "compression_loss": 76.34895324707031, "epoch": 0.2, "learning_rate": 2.7097568988173456e-05, "loss": 76.9663, "step": 594 }, { "compression_loss": 76.70858001708984, "epoch": 0.2, "learning_rate": 2.7092641261498032e-05, "loss": 77.2095, "step": 595 }, { "compression_loss": 76.20101165771484, "epoch": 0.2, "learning_rate": 2.7087713534822604e-05, "loss": 76.5632, "step": 596 }, { "compression_loss": 77.85941314697266, "epoch": 0.2, "learning_rate": 2.7082785808147176e-05, "loss": 78.1375, "step": 597 }, { "compression_loss": 75.82550811767578, "epoch": 0.2, "learning_rate": 2.7077858081471748e-05, "loss": 76.6087, "step": 598 }, { "compression_loss": 74.10774993896484, "epoch": 0.2, "learning_rate": 2.7072930354796323e-05, "loss": 74.5158, "step": 599 }, { "compression_loss": 76.49610900878906, "epoch": 0.2, "learning_rate": 2.7068002628120895e-05, "loss": 76.9112, "step": 600 }, { "compression_loss": 77.18254089355469, "epoch": 0.2, "learning_rate": 2.7063074901445467e-05, "loss": 77.4576, "step": 601 }, { "compression_loss": 73.01869201660156, "epoch": 0.2, "learning_rate": 2.705814717477004e-05, "loss": 73.2622, "step": 602 }, { "compression_loss": 76.42254638671875, "epoch": 0.2, "learning_rate": 2.705321944809461e-05, "loss": 76.7434, "step": 603 }, { "compression_loss": 74.69294738769531, "epoch": 0.2, "learning_rate": 2.7048291721419186e-05, "loss": 75.0773, "step": 604 }, { "compression_loss": 75.38629150390625, "epoch": 0.2, "learning_rate": 2.7043363994743758e-05, "loss": 75.9426, "step": 605 }, { "compression_loss": 76.2300033569336, "epoch": 0.2, "learning_rate": 2.703843626806833e-05, "loss": 76.7811, "step": 606 }, { "compression_loss": 76.96832275390625, "epoch": 0.2, "learning_rate": 2.7033508541392906e-05, "loss": 77.1689, "step": 607 }, { "compression_loss": 77.31790161132812, "epoch": 0.2, "learning_rate": 2.7028580814717478e-05, "loss": 77.6463, "step": 608 }, { "compression_loss": 74.68637084960938, "epoch": 0.2, "learning_rate": 2.7023653088042053e-05, "loss": 75.0199, "step": 609 }, { "compression_loss": 74.93142700195312, "epoch": 0.2, "learning_rate": 2.7018725361366625e-05, "loss": 75.2609, "step": 610 }, { "compression_loss": 76.0052719116211, "epoch": 0.2, "learning_rate": 2.7013797634691197e-05, "loss": 76.4572, "step": 611 }, { "compression_loss": 75.82221984863281, "epoch": 0.2, "learning_rate": 2.700886990801577e-05, "loss": 76.3175, "step": 612 }, { "compression_loss": 76.46435546875, "epoch": 0.2, "learning_rate": 2.700394218134034e-05, "loss": 77.0058, "step": 613 }, { "compression_loss": 73.75770568847656, "epoch": 0.2, "learning_rate": 2.6999014454664916e-05, "loss": 73.9372, "step": 614 }, { "compression_loss": 74.49136352539062, "epoch": 0.2, "learning_rate": 2.6994086727989488e-05, "loss": 74.7271, "step": 615 }, { "compression_loss": 73.36267852783203, "epoch": 0.2, "learning_rate": 2.698915900131406e-05, "loss": 73.5956, "step": 616 }, { "compression_loss": 76.002197265625, "epoch": 0.2, "learning_rate": 2.6984231274638632e-05, "loss": 76.3234, "step": 617 }, { "compression_loss": 76.82760620117188, "epoch": 0.2, "learning_rate": 2.6979303547963204e-05, "loss": 77.2242, "step": 618 }, { "compression_loss": 75.54676818847656, "epoch": 0.2, "learning_rate": 2.6974375821287783e-05, "loss": 75.8186, "step": 619 }, { "compression_loss": 79.28472900390625, "epoch": 0.2, "learning_rate": 2.6969448094612355e-05, "loss": 79.6882, "step": 620 }, { "compression_loss": 74.4933853149414, "epoch": 0.2, "learning_rate": 2.6964520367936927e-05, "loss": 74.7844, "step": 621 }, { "compression_loss": 76.49933624267578, "epoch": 0.2, "learning_rate": 2.69595926412615e-05, "loss": 76.9591, "step": 622 }, { "compression_loss": 75.62940216064453, "epoch": 0.2, "learning_rate": 2.695466491458607e-05, "loss": 76.0595, "step": 623 }, { "compression_loss": 73.62495422363281, "epoch": 0.2, "learning_rate": 2.6949737187910646e-05, "loss": 74.1028, "step": 624 }, { "compression_loss": 74.92414855957031, "epoch": 0.21, "learning_rate": 2.6944809461235218e-05, "loss": 75.4482, "step": 625 }, { "compression_loss": 74.40018463134766, "epoch": 0.21, "learning_rate": 2.693988173455979e-05, "loss": 74.6002, "step": 626 }, { "compression_loss": 77.19481658935547, "epoch": 0.21, "learning_rate": 2.6934954007884362e-05, "loss": 77.7523, "step": 627 }, { "compression_loss": 74.8065185546875, "epoch": 0.21, "learning_rate": 2.6930026281208934e-05, "loss": 75.093, "step": 628 }, { "compression_loss": 74.70257568359375, "epoch": 0.21, "learning_rate": 2.692509855453351e-05, "loss": 74.9583, "step": 629 }, { "compression_loss": 74.93319702148438, "epoch": 0.21, "learning_rate": 2.692017082785808e-05, "loss": 75.2664, "step": 630 }, { "compression_loss": 74.12535858154297, "epoch": 0.21, "learning_rate": 2.6915243101182657e-05, "loss": 74.3676, "step": 631 }, { "compression_loss": 76.11612701416016, "epoch": 0.21, "learning_rate": 2.691031537450723e-05, "loss": 76.6432, "step": 632 }, { "compression_loss": 76.00946044921875, "epoch": 0.21, "learning_rate": 2.69053876478318e-05, "loss": 76.6081, "step": 633 }, { "compression_loss": 75.06340026855469, "epoch": 0.21, "learning_rate": 2.6900459921156376e-05, "loss": 75.4846, "step": 634 }, { "compression_loss": 80.66598510742188, "epoch": 0.21, "learning_rate": 2.6895532194480948e-05, "loss": 81.1979, "step": 635 }, { "compression_loss": 75.70448303222656, "epoch": 0.21, "learning_rate": 2.689060446780552e-05, "loss": 76.144, "step": 636 }, { "compression_loss": 76.20777893066406, "epoch": 0.21, "learning_rate": 2.6885676741130092e-05, "loss": 76.5631, "step": 637 }, { "compression_loss": 77.24947357177734, "epoch": 0.21, "learning_rate": 2.6880749014454664e-05, "loss": 77.8109, "step": 638 }, { "compression_loss": 74.68193817138672, "epoch": 0.21, "learning_rate": 2.687582128777924e-05, "loss": 75.0117, "step": 639 }, { "compression_loss": 75.48316192626953, "epoch": 0.21, "learning_rate": 2.687089356110381e-05, "loss": 76.0652, "step": 640 }, { "compression_loss": 76.63803100585938, "epoch": 0.21, "learning_rate": 2.6865965834428383e-05, "loss": 77.079, "step": 641 }, { "compression_loss": 78.05288696289062, "epoch": 0.21, "learning_rate": 2.6861038107752955e-05, "loss": 78.462, "step": 642 }, { "compression_loss": 75.5555419921875, "epoch": 0.21, "learning_rate": 2.685611038107753e-05, "loss": 75.939, "step": 643 }, { "compression_loss": 73.78616333007812, "epoch": 0.21, "learning_rate": 2.6851182654402106e-05, "loss": 73.9927, "step": 644 }, { "compression_loss": 73.503662109375, "epoch": 0.21, "learning_rate": 2.6846254927726678e-05, "loss": 74.0206, "step": 645 }, { "compression_loss": 74.46063232421875, "epoch": 0.21, "learning_rate": 2.684132720105125e-05, "loss": 74.942, "step": 646 }, { "compression_loss": 77.38908386230469, "epoch": 0.21, "learning_rate": 2.683639947437582e-05, "loss": 77.8835, "step": 647 }, { "compression_loss": 75.61263275146484, "epoch": 0.21, "learning_rate": 2.6831471747700394e-05, "loss": 76.3004, "step": 648 }, { "compression_loss": 77.11740112304688, "epoch": 0.21, "learning_rate": 2.682654402102497e-05, "loss": 77.4699, "step": 649 }, { "compression_loss": 74.64630889892578, "epoch": 0.21, "learning_rate": 2.682161629434954e-05, "loss": 74.9034, "step": 650 }, { "compression_loss": 76.1659927368164, "epoch": 0.21, "learning_rate": 2.6816688567674113e-05, "loss": 76.7, "step": 651 }, { "compression_loss": 74.78581237792969, "epoch": 0.21, "learning_rate": 2.6811760840998685e-05, "loss": 75.1627, "step": 652 }, { "compression_loss": 77.42866516113281, "epoch": 0.21, "learning_rate": 2.6806833114323257e-05, "loss": 77.8265, "step": 653 }, { "compression_loss": 78.82467651367188, "epoch": 0.21, "learning_rate": 2.6801905387647832e-05, "loss": 79.3162, "step": 654 }, { "compression_loss": 75.8014907836914, "epoch": 0.22, "learning_rate": 2.6796977660972407e-05, "loss": 76.1486, "step": 655 }, { "compression_loss": 76.11528778076172, "epoch": 0.22, "learning_rate": 2.679204993429698e-05, "loss": 76.4839, "step": 656 }, { "compression_loss": 75.1211166381836, "epoch": 0.22, "learning_rate": 2.678712220762155e-05, "loss": 75.5275, "step": 657 }, { "compression_loss": 74.93295288085938, "epoch": 0.22, "learning_rate": 2.6782194480946123e-05, "loss": 75.32, "step": 658 }, { "compression_loss": 78.52654266357422, "epoch": 0.22, "learning_rate": 2.67772667542707e-05, "loss": 78.8936, "step": 659 }, { "compression_loss": 77.23888397216797, "epoch": 0.22, "learning_rate": 2.677233902759527e-05, "loss": 77.6737, "step": 660 }, { "compression_loss": 77.94867706298828, "epoch": 0.22, "learning_rate": 2.6767411300919843e-05, "loss": 78.4594, "step": 661 }, { "compression_loss": 73.53561401367188, "epoch": 0.22, "learning_rate": 2.6762483574244415e-05, "loss": 73.7928, "step": 662 }, { "compression_loss": 75.34617614746094, "epoch": 0.22, "learning_rate": 2.6757555847568987e-05, "loss": 75.6227, "step": 663 }, { "compression_loss": 77.16145324707031, "epoch": 0.22, "learning_rate": 2.6752628120893562e-05, "loss": 77.5556, "step": 664 }, { "compression_loss": 77.6396484375, "epoch": 0.22, "learning_rate": 2.6747700394218134e-05, "loss": 78.142, "step": 665 }, { "compression_loss": 75.54558563232422, "epoch": 0.22, "learning_rate": 2.6742772667542706e-05, "loss": 76.0908, "step": 666 }, { "compression_loss": 78.2220458984375, "epoch": 0.22, "learning_rate": 2.673784494086728e-05, "loss": 78.5174, "step": 667 }, { "compression_loss": 75.26841735839844, "epoch": 0.22, "learning_rate": 2.6732917214191853e-05, "loss": 75.9697, "step": 668 }, { "compression_loss": 75.00863647460938, "epoch": 0.22, "learning_rate": 2.672798948751643e-05, "loss": 75.372, "step": 669 }, { "compression_loss": 73.90391540527344, "epoch": 0.22, "learning_rate": 2.6723061760841e-05, "loss": 74.1578, "step": 670 }, { "compression_loss": 76.83251190185547, "epoch": 0.22, "learning_rate": 2.6718134034165573e-05, "loss": 77.287, "step": 671 }, { "compression_loss": 74.66390991210938, "epoch": 0.22, "learning_rate": 2.6713206307490145e-05, "loss": 74.972, "step": 672 }, { "compression_loss": 78.04396057128906, "epoch": 0.22, "learning_rate": 2.6708278580814716e-05, "loss": 78.3457, "step": 673 }, { "compression_loss": 75.66001892089844, "epoch": 0.22, "learning_rate": 2.6703350854139292e-05, "loss": 76.303, "step": 674 }, { "compression_loss": 76.03422546386719, "epoch": 0.22, "learning_rate": 2.6698423127463864e-05, "loss": 76.4889, "step": 675 }, { "compression_loss": 76.05690002441406, "epoch": 0.22, "learning_rate": 2.6693495400788436e-05, "loss": 76.3964, "step": 676 }, { "compression_loss": 74.42251586914062, "epoch": 0.22, "learning_rate": 2.6688567674113008e-05, "loss": 74.6447, "step": 677 }, { "compression_loss": 77.09449768066406, "epoch": 0.22, "learning_rate": 2.668363994743758e-05, "loss": 77.5452, "step": 678 }, { "compression_loss": 76.06928253173828, "epoch": 0.22, "learning_rate": 2.667871222076216e-05, "loss": 76.3859, "step": 679 }, { "compression_loss": 75.71350860595703, "epoch": 0.22, "learning_rate": 2.667378449408673e-05, "loss": 76.0502, "step": 680 }, { "compression_loss": 76.81010437011719, "epoch": 0.22, "learning_rate": 2.6668856767411302e-05, "loss": 77.1956, "step": 681 }, { "compression_loss": 78.18537902832031, "epoch": 0.22, "learning_rate": 2.6663929040735874e-05, "loss": 78.6593, "step": 682 }, { "compression_loss": 75.26797485351562, "epoch": 0.22, "learning_rate": 2.6659001314060446e-05, "loss": 75.8797, "step": 683 }, { "compression_loss": 75.94635772705078, "epoch": 0.22, "learning_rate": 2.6654073587385022e-05, "loss": 76.1739, "step": 684 }, { "compression_loss": 77.46070861816406, "epoch": 0.23, "learning_rate": 2.6649145860709594e-05, "loss": 78.1448, "step": 685 }, { "compression_loss": 77.63467407226562, "epoch": 0.23, "learning_rate": 2.6644218134034166e-05, "loss": 78.2158, "step": 686 }, { "compression_loss": 76.99911499023438, "epoch": 0.23, "learning_rate": 2.6639290407358738e-05, "loss": 77.5023, "step": 687 }, { "compression_loss": 76.20722198486328, "epoch": 0.23, "learning_rate": 2.663436268068331e-05, "loss": 76.5952, "step": 688 }, { "compression_loss": 75.25719451904297, "epoch": 0.23, "learning_rate": 2.6629434954007885e-05, "loss": 75.75, "step": 689 }, { "compression_loss": 73.07713317871094, "epoch": 0.23, "learning_rate": 2.6624507227332457e-05, "loss": 73.521, "step": 690 }, { "compression_loss": 75.71110534667969, "epoch": 0.23, "learning_rate": 2.6619579500657032e-05, "loss": 76.1595, "step": 691 }, { "compression_loss": 73.72846984863281, "epoch": 0.23, "learning_rate": 2.6614651773981604e-05, "loss": 74.0837, "step": 692 }, { "compression_loss": 76.91250610351562, "epoch": 0.23, "learning_rate": 2.6609724047306176e-05, "loss": 77.2139, "step": 693 }, { "compression_loss": 74.47421264648438, "epoch": 0.23, "learning_rate": 2.660479632063075e-05, "loss": 74.7655, "step": 694 }, { "compression_loss": 78.74580383300781, "epoch": 0.23, "learning_rate": 2.6599868593955324e-05, "loss": 79.1359, "step": 695 }, { "compression_loss": 73.55398559570312, "epoch": 0.23, "learning_rate": 2.6594940867279895e-05, "loss": 73.96, "step": 696 }, { "compression_loss": 76.55179595947266, "epoch": 0.23, "learning_rate": 2.6590013140604467e-05, "loss": 76.9379, "step": 697 }, { "compression_loss": 77.14727020263672, "epoch": 0.23, "learning_rate": 2.658508541392904e-05, "loss": 77.8783, "step": 698 }, { "compression_loss": 74.79415893554688, "epoch": 0.23, "learning_rate": 2.6580157687253615e-05, "loss": 75.4158, "step": 699 }, { "compression_loss": 74.5111312866211, "epoch": 0.23, "learning_rate": 2.6575229960578187e-05, "loss": 74.917, "step": 700 }, { "compression_loss": 75.47702026367188, "epoch": 0.23, "learning_rate": 2.657030223390276e-05, "loss": 75.8499, "step": 701 }, { "compression_loss": 75.09737396240234, "epoch": 0.23, "learning_rate": 2.6565374507227334e-05, "loss": 75.5536, "step": 702 }, { "compression_loss": 74.25238037109375, "epoch": 0.23, "learning_rate": 2.6560446780551906e-05, "loss": 74.6056, "step": 703 }, { "compression_loss": 75.1063232421875, "epoch": 0.23, "learning_rate": 2.655551905387648e-05, "loss": 75.4378, "step": 704 }, { "compression_loss": 72.07884216308594, "epoch": 0.23, "learning_rate": 2.6550591327201053e-05, "loss": 72.2062, "step": 705 }, { "compression_loss": 75.93231201171875, "epoch": 0.23, "learning_rate": 2.6545663600525625e-05, "loss": 76.2356, "step": 706 }, { "compression_loss": 73.73103332519531, "epoch": 0.23, "learning_rate": 2.6540735873850197e-05, "loss": 74.1591, "step": 707 }, { "compression_loss": 73.36634826660156, "epoch": 0.23, "learning_rate": 2.653580814717477e-05, "loss": 73.8035, "step": 708 }, { "compression_loss": 74.9519271850586, "epoch": 0.23, "learning_rate": 2.6530880420499345e-05, "loss": 75.6213, "step": 709 }, { "compression_loss": 77.06005859375, "epoch": 0.23, "learning_rate": 2.6525952693823917e-05, "loss": 77.5785, "step": 710 }, { "compression_loss": 76.36036682128906, "epoch": 0.23, "learning_rate": 2.652102496714849e-05, "loss": 76.6772, "step": 711 }, { "compression_loss": 73.13965606689453, "epoch": 0.23, "learning_rate": 2.651609724047306e-05, "loss": 73.4996, "step": 712 }, { "compression_loss": 75.28207397460938, "epoch": 0.23, "learning_rate": 2.6511169513797633e-05, "loss": 75.5282, "step": 713 }, { "compression_loss": 74.10848999023438, "epoch": 0.23, "learning_rate": 2.650624178712221e-05, "loss": 74.6073, "step": 714 }, { "compression_loss": 76.76936340332031, "epoch": 0.23, "learning_rate": 2.6501314060446783e-05, "loss": 77.2421, "step": 715 }, { "compression_loss": 76.32837677001953, "epoch": 0.24, "learning_rate": 2.6496386333771355e-05, "loss": 76.7276, "step": 716 }, { "compression_loss": 75.79998779296875, "epoch": 0.24, "learning_rate": 2.6491458607095927e-05, "loss": 76.2464, "step": 717 }, { "compression_loss": 74.94847106933594, "epoch": 0.24, "learning_rate": 2.64865308804205e-05, "loss": 75.2283, "step": 718 }, { "compression_loss": 76.75939178466797, "epoch": 0.24, "learning_rate": 2.6481603153745074e-05, "loss": 77.363, "step": 719 }, { "compression_loss": 75.9052734375, "epoch": 0.24, "learning_rate": 2.6476675427069646e-05, "loss": 76.2195, "step": 720 }, { "compression_loss": 75.0926284790039, "epoch": 0.24, "learning_rate": 2.647174770039422e-05, "loss": 75.6843, "step": 721 }, { "compression_loss": 76.11614227294922, "epoch": 0.24, "learning_rate": 2.646681997371879e-05, "loss": 76.6532, "step": 722 }, { "compression_loss": 74.15274047851562, "epoch": 0.24, "learning_rate": 2.6461892247043362e-05, "loss": 74.5712, "step": 723 }, { "compression_loss": 74.41719818115234, "epoch": 0.24, "learning_rate": 2.6456964520367938e-05, "loss": 74.6353, "step": 724 }, { "compression_loss": 76.9448471069336, "epoch": 0.24, "learning_rate": 2.645203679369251e-05, "loss": 77.231, "step": 725 }, { "compression_loss": 78.06961059570312, "epoch": 0.24, "learning_rate": 2.6447109067017085e-05, "loss": 78.5914, "step": 726 }, { "compression_loss": 74.08101654052734, "epoch": 0.24, "learning_rate": 2.6442181340341657e-05, "loss": 74.7923, "step": 727 }, { "compression_loss": 75.26219177246094, "epoch": 0.24, "learning_rate": 2.643725361366623e-05, "loss": 75.4597, "step": 728 }, { "compression_loss": 75.38681030273438, "epoch": 0.24, "learning_rate": 2.6432325886990804e-05, "loss": 75.5379, "step": 729 }, { "compression_loss": 77.26354217529297, "epoch": 0.24, "learning_rate": 2.6427398160315376e-05, "loss": 77.659, "step": 730 }, { "compression_loss": 75.62438201904297, "epoch": 0.24, "learning_rate": 2.6422470433639948e-05, "loss": 76.1072, "step": 731 }, { "compression_loss": 72.71028137207031, "epoch": 0.24, "learning_rate": 2.641754270696452e-05, "loss": 73.4913, "step": 732 }, { "compression_loss": 75.67560577392578, "epoch": 0.24, "learning_rate": 2.6412614980289092e-05, "loss": 76.0652, "step": 733 }, { "compression_loss": 76.05462646484375, "epoch": 0.24, "learning_rate": 2.6407687253613668e-05, "loss": 76.5181, "step": 734 }, { "compression_loss": 75.68820190429688, "epoch": 0.24, "learning_rate": 2.640275952693824e-05, "loss": 76.2831, "step": 735 }, { "compression_loss": 75.11759948730469, "epoch": 0.24, "learning_rate": 2.639783180026281e-05, "loss": 75.5934, "step": 736 }, { "compression_loss": 76.92636108398438, "epoch": 0.24, "learning_rate": 2.6392904073587383e-05, "loss": 77.7156, "step": 737 }, { "compression_loss": 74.39114379882812, "epoch": 0.24, "learning_rate": 2.638797634691196e-05, "loss": 74.9165, "step": 738 }, { "compression_loss": 76.03325653076172, "epoch": 0.24, "learning_rate": 2.6383048620236534e-05, "loss": 76.4339, "step": 739 }, { "compression_loss": 75.50094604492188, "epoch": 0.24, "learning_rate": 2.6378120893561106e-05, "loss": 75.7677, "step": 740 }, { "compression_loss": 76.6877670288086, "epoch": 0.24, "learning_rate": 2.6373193166885678e-05, "loss": 77.1225, "step": 741 }, { "compression_loss": 78.11710357666016, "epoch": 0.24, "learning_rate": 2.636826544021025e-05, "loss": 78.6214, "step": 742 }, { "compression_loss": 76.9841079711914, "epoch": 0.24, "learning_rate": 2.6363337713534822e-05, "loss": 77.4841, "step": 743 }, { "compression_loss": 74.15127563476562, "epoch": 0.24, "learning_rate": 2.6358409986859397e-05, "loss": 74.5445, "step": 744 }, { "compression_loss": 77.04237365722656, "epoch": 0.24, "learning_rate": 2.635348226018397e-05, "loss": 77.4999, "step": 745 }, { "compression_loss": 77.35287475585938, "epoch": 0.25, "learning_rate": 2.634855453350854e-05, "loss": 77.835, "step": 746 }, { "compression_loss": 77.00286102294922, "epoch": 0.25, "learning_rate": 2.6343626806833113e-05, "loss": 77.5924, "step": 747 }, { "compression_loss": 75.41752624511719, "epoch": 0.25, "learning_rate": 2.6338699080157685e-05, "loss": 75.7813, "step": 748 }, { "compression_loss": 75.14828491210938, "epoch": 0.25, "learning_rate": 2.633377135348226e-05, "loss": 75.3917, "step": 749 }, { "compression_loss": 76.41581726074219, "epoch": 0.25, "learning_rate": 2.6328843626806836e-05, "loss": 77.1439, "step": 750 }, { "epoch": 0.25, "eval_exact_match": 85.93188268684958, "eval_f1": 92.31368099810622, "step": 750 }, { "compression_loss": 75.58445739746094, "epoch": 0.25, "learning_rate": 2.6323915900131408e-05, "loss": 75.8034, "step": 751 }, { "compression_loss": 75.5068359375, "epoch": 0.25, "learning_rate": 2.631898817345598e-05, "loss": 75.9155, "step": 752 }, { "compression_loss": 75.3727798461914, "epoch": 0.25, "learning_rate": 2.6314060446780552e-05, "loss": 75.6957, "step": 753 }, { "compression_loss": 74.59487915039062, "epoch": 0.25, "learning_rate": 2.6309132720105127e-05, "loss": 75.1558, "step": 754 }, { "compression_loss": 73.77601623535156, "epoch": 0.25, "learning_rate": 2.63042049934297e-05, "loss": 74.1129, "step": 755 }, { "compression_loss": 75.96479034423828, "epoch": 0.25, "learning_rate": 2.629927726675427e-05, "loss": 76.2823, "step": 756 }, { "compression_loss": 74.96715545654297, "epoch": 0.25, "learning_rate": 2.6294349540078843e-05, "loss": 75.6184, "step": 757 }, { "compression_loss": 73.7572021484375, "epoch": 0.25, "learning_rate": 2.6289421813403415e-05, "loss": 73.9821, "step": 758 }, { "compression_loss": 73.90113067626953, "epoch": 0.25, "learning_rate": 2.628449408672799e-05, "loss": 74.1755, "step": 759 }, { "compression_loss": 76.24119567871094, "epoch": 0.25, "learning_rate": 2.6279566360052562e-05, "loss": 76.5293, "step": 760 }, { "compression_loss": 73.76394653320312, "epoch": 0.25, "learning_rate": 2.6274638633377134e-05, "loss": 74.109, "step": 761 }, { "compression_loss": 77.39735412597656, "epoch": 0.25, "learning_rate": 2.626971090670171e-05, "loss": 77.958, "step": 762 }, { "compression_loss": 78.03610229492188, "epoch": 0.25, "learning_rate": 2.6264783180026282e-05, "loss": 78.6516, "step": 763 }, { "compression_loss": 73.53070068359375, "epoch": 0.25, "learning_rate": 2.6259855453350857e-05, "loss": 73.7568, "step": 764 }, { "compression_loss": 75.44863891601562, "epoch": 0.25, "learning_rate": 2.625492772667543e-05, "loss": 75.7487, "step": 765 }, { "compression_loss": 75.3233642578125, "epoch": 0.25, "learning_rate": 2.625e-05, "loss": 75.7048, "step": 766 }, { "compression_loss": 75.45855712890625, "epoch": 0.25, "learning_rate": 2.6245072273324573e-05, "loss": 75.693, "step": 767 }, { "compression_loss": 75.94223022460938, "epoch": 0.25, "learning_rate": 2.6240144546649145e-05, "loss": 76.3333, "step": 768 }, { "compression_loss": 74.19554138183594, "epoch": 0.25, "learning_rate": 2.623521681997372e-05, "loss": 74.7178, "step": 769 }, { "compression_loss": 75.11875915527344, "epoch": 0.25, "learning_rate": 2.6230289093298292e-05, "loss": 75.3217, "step": 770 }, { "compression_loss": 75.92781066894531, "epoch": 0.25, "learning_rate": 2.6225361366622864e-05, "loss": 76.2866, "step": 771 }, { "compression_loss": 75.80348205566406, "epoch": 0.25, "learning_rate": 2.6220433639947436e-05, "loss": 76.3123, "step": 772 }, { "compression_loss": 75.27371215820312, "epoch": 0.25, "learning_rate": 2.6215505913272008e-05, "loss": 75.8099, "step": 773 }, { "compression_loss": 77.95921325683594, "epoch": 0.25, "learning_rate": 2.6210578186596587e-05, "loss": 78.5879, "step": 774 }, { "compression_loss": 74.73515319824219, "epoch": 0.25, "learning_rate": 2.620565045992116e-05, "loss": 75.1515, "step": 775 }, { "compression_loss": 79.25530242919922, "epoch": 0.25, "learning_rate": 2.620072273324573e-05, "loss": 79.694, "step": 776 }, { "compression_loss": 75.74107360839844, "epoch": 0.26, "learning_rate": 2.6195795006570303e-05, "loss": 75.9569, "step": 777 }, { "compression_loss": 77.36885833740234, "epoch": 0.26, "learning_rate": 2.6190867279894875e-05, "loss": 77.6071, "step": 778 }, { "compression_loss": 75.63190460205078, "epoch": 0.26, "learning_rate": 2.618593955321945e-05, "loss": 75.931, "step": 779 }, { "compression_loss": 73.58209991455078, "epoch": 0.26, "learning_rate": 2.6181011826544022e-05, "loss": 73.872, "step": 780 }, { "compression_loss": 76.21342468261719, "epoch": 0.26, "learning_rate": 2.6176084099868594e-05, "loss": 76.8968, "step": 781 }, { "compression_loss": 76.76988220214844, "epoch": 0.26, "learning_rate": 2.6171156373193166e-05, "loss": 77.3328, "step": 782 }, { "compression_loss": 76.23665618896484, "epoch": 0.26, "learning_rate": 2.6166228646517738e-05, "loss": 76.7812, "step": 783 }, { "compression_loss": 74.80143737792969, "epoch": 0.26, "learning_rate": 2.6161300919842313e-05, "loss": 74.9842, "step": 784 }, { "compression_loss": 77.11146545410156, "epoch": 0.26, "learning_rate": 2.6156373193166885e-05, "loss": 77.4727, "step": 785 }, { "compression_loss": 73.97145080566406, "epoch": 0.26, "learning_rate": 2.615144546649146e-05, "loss": 74.4669, "step": 786 }, { "compression_loss": 78.91789245605469, "epoch": 0.26, "learning_rate": 2.6146517739816033e-05, "loss": 79.5558, "step": 787 }, { "compression_loss": 77.56001281738281, "epoch": 0.26, "learning_rate": 2.6141590013140605e-05, "loss": 77.8746, "step": 788 }, { "compression_loss": 77.6768798828125, "epoch": 0.26, "learning_rate": 2.613666228646518e-05, "loss": 78.6432, "step": 789 }, { "compression_loss": 75.37422180175781, "epoch": 0.26, "learning_rate": 2.6131734559789752e-05, "loss": 75.682, "step": 790 }, { "compression_loss": 74.59486389160156, "epoch": 0.26, "learning_rate": 2.6126806833114324e-05, "loss": 75.0247, "step": 791 }, { "compression_loss": 74.76690673828125, "epoch": 0.26, "learning_rate": 2.6121879106438896e-05, "loss": 74.9938, "step": 792 }, { "compression_loss": 74.40725708007812, "epoch": 0.26, "learning_rate": 2.6116951379763468e-05, "loss": 74.7838, "step": 793 }, { "compression_loss": 75.88447570800781, "epoch": 0.26, "learning_rate": 2.6112023653088043e-05, "loss": 76.3183, "step": 794 }, { "compression_loss": 77.54351806640625, "epoch": 0.26, "learning_rate": 2.6107095926412615e-05, "loss": 77.8417, "step": 795 }, { "compression_loss": 76.02684783935547, "epoch": 0.26, "learning_rate": 2.6102168199737187e-05, "loss": 76.3773, "step": 796 }, { "compression_loss": 78.34941101074219, "epoch": 0.26, "learning_rate": 2.609724047306176e-05, "loss": 78.6508, "step": 797 }, { "compression_loss": 74.92478942871094, "epoch": 0.26, "learning_rate": 2.6092312746386335e-05, "loss": 75.5277, "step": 798 }, { "compression_loss": 76.16311645507812, "epoch": 0.26, "learning_rate": 2.608738501971091e-05, "loss": 76.7416, "step": 799 }, { "compression_loss": 79.24594116210938, "epoch": 0.26, "learning_rate": 2.6082457293035482e-05, "loss": 79.758, "step": 800 }, { "compression_loss": 76.39146423339844, "epoch": 0.26, "learning_rate": 2.6077529566360054e-05, "loss": 76.7522, "step": 801 }, { "compression_loss": 74.0270004272461, "epoch": 0.26, "learning_rate": 2.6072601839684626e-05, "loss": 74.5274, "step": 802 }, { "compression_loss": 74.54388427734375, "epoch": 0.26, "learning_rate": 2.6067674113009198e-05, "loss": 75.0013, "step": 803 }, { "compression_loss": 75.07159423828125, "epoch": 0.26, "learning_rate": 2.6062746386333773e-05, "loss": 75.5615, "step": 804 }, { "compression_loss": 73.6996841430664, "epoch": 0.26, "learning_rate": 2.6057818659658345e-05, "loss": 74.2059, "step": 805 }, { "compression_loss": 76.33799743652344, "epoch": 0.26, "learning_rate": 2.6052890932982917e-05, "loss": 76.88, "step": 806 }, { "compression_loss": 76.25068664550781, "epoch": 0.27, "learning_rate": 2.604796320630749e-05, "loss": 76.5611, "step": 807 }, { "compression_loss": 78.43598937988281, "epoch": 0.27, "learning_rate": 2.604303547963206e-05, "loss": 78.9557, "step": 808 }, { "compression_loss": 74.8377685546875, "epoch": 0.27, "learning_rate": 2.6038107752956636e-05, "loss": 75.106, "step": 809 }, { "compression_loss": 75.8338851928711, "epoch": 0.27, "learning_rate": 2.6033180026281212e-05, "loss": 76.0081, "step": 810 }, { "compression_loss": 76.60589599609375, "epoch": 0.27, "learning_rate": 2.6028252299605784e-05, "loss": 76.9911, "step": 811 }, { "compression_loss": 73.56180572509766, "epoch": 0.27, "learning_rate": 2.6023324572930356e-05, "loss": 73.9668, "step": 812 }, { "compression_loss": 75.74461364746094, "epoch": 0.27, "learning_rate": 2.6018396846254928e-05, "loss": 76.1354, "step": 813 }, { "compression_loss": 74.45960998535156, "epoch": 0.27, "learning_rate": 2.6013469119579503e-05, "loss": 74.9128, "step": 814 }, { "compression_loss": 74.20196533203125, "epoch": 0.27, "learning_rate": 2.6008541392904075e-05, "loss": 74.9445, "step": 815 }, { "compression_loss": 75.10620880126953, "epoch": 0.27, "learning_rate": 2.6003613666228647e-05, "loss": 75.3999, "step": 816 }, { "compression_loss": 75.06387329101562, "epoch": 0.27, "learning_rate": 2.599868593955322e-05, "loss": 75.6888, "step": 817 }, { "compression_loss": 75.6435546875, "epoch": 0.27, "learning_rate": 2.599375821287779e-05, "loss": 76.1962, "step": 818 }, { "compression_loss": 74.92146301269531, "epoch": 0.27, "learning_rate": 2.5988830486202366e-05, "loss": 75.1593, "step": 819 }, { "compression_loss": 73.02911376953125, "epoch": 0.27, "learning_rate": 2.5983902759526938e-05, "loss": 73.5241, "step": 820 }, { "compression_loss": 76.38900756835938, "epoch": 0.27, "learning_rate": 2.597897503285151e-05, "loss": 76.686, "step": 821 }, { "compression_loss": 75.32945251464844, "epoch": 0.27, "learning_rate": 2.5974047306176086e-05, "loss": 75.4438, "step": 822 }, { "compression_loss": 75.00529479980469, "epoch": 0.27, "learning_rate": 2.5969119579500658e-05, "loss": 75.4156, "step": 823 }, { "compression_loss": 75.49118041992188, "epoch": 0.27, "learning_rate": 2.5964191852825233e-05, "loss": 76.1239, "step": 824 }, { "compression_loss": 75.0370101928711, "epoch": 0.27, "learning_rate": 2.5959264126149805e-05, "loss": 75.4078, "step": 825 }, { "compression_loss": 76.26194763183594, "epoch": 0.27, "learning_rate": 2.5954336399474377e-05, "loss": 76.8043, "step": 826 }, { "compression_loss": 79.43508911132812, "epoch": 0.27, "learning_rate": 2.594940867279895e-05, "loss": 79.9049, "step": 827 }, { "compression_loss": 74.50633239746094, "epoch": 0.27, "learning_rate": 2.594448094612352e-05, "loss": 75.2724, "step": 828 }, { "compression_loss": 74.00387573242188, "epoch": 0.27, "learning_rate": 2.5939553219448096e-05, "loss": 74.2209, "step": 829 }, { "compression_loss": 75.29681396484375, "epoch": 0.27, "learning_rate": 2.5934625492772668e-05, "loss": 75.5767, "step": 830 }, { "compression_loss": 77.79669189453125, "epoch": 0.27, "learning_rate": 2.592969776609724e-05, "loss": 78.2173, "step": 831 }, { "compression_loss": 77.80635070800781, "epoch": 0.27, "learning_rate": 2.5924770039421812e-05, "loss": 78.4125, "step": 832 }, { "compression_loss": 76.97259521484375, "epoch": 0.27, "learning_rate": 2.5919842312746387e-05, "loss": 77.3638, "step": 833 }, { "compression_loss": 76.28329467773438, "epoch": 0.27, "learning_rate": 2.5914914586070963e-05, "loss": 76.6109, "step": 834 }, { "compression_loss": 76.78678894042969, "epoch": 0.27, "learning_rate": 2.5909986859395535e-05, "loss": 77.1433, "step": 835 }, { "compression_loss": 73.34727478027344, "epoch": 0.27, "learning_rate": 2.5905059132720107e-05, "loss": 73.6514, "step": 836 }, { "compression_loss": 77.27047729492188, "epoch": 0.27, "learning_rate": 2.590013140604468e-05, "loss": 77.6071, "step": 837 }, { "compression_loss": 75.77003479003906, "epoch": 0.28, "learning_rate": 2.589520367936925e-05, "loss": 76.3923, "step": 838 }, { "compression_loss": 76.57072448730469, "epoch": 0.28, "learning_rate": 2.5890275952693826e-05, "loss": 76.8513, "step": 839 }, { "compression_loss": 74.79972076416016, "epoch": 0.28, "learning_rate": 2.5885348226018398e-05, "loss": 75.1797, "step": 840 }, { "compression_loss": 74.45264434814453, "epoch": 0.28, "learning_rate": 2.588042049934297e-05, "loss": 74.8589, "step": 841 }, { "compression_loss": 75.66279602050781, "epoch": 0.28, "learning_rate": 2.5875492772667542e-05, "loss": 76.4706, "step": 842 }, { "compression_loss": 76.35678100585938, "epoch": 0.28, "learning_rate": 2.5870565045992114e-05, "loss": 76.6609, "step": 843 }, { "compression_loss": 77.41944885253906, "epoch": 0.28, "learning_rate": 2.586563731931669e-05, "loss": 78.0894, "step": 844 }, { "compression_loss": 75.06375122070312, "epoch": 0.28, "learning_rate": 2.5860709592641265e-05, "loss": 75.3811, "step": 845 }, { "compression_loss": 76.55408477783203, "epoch": 0.28, "learning_rate": 2.5855781865965837e-05, "loss": 76.971, "step": 846 }, { "compression_loss": 76.37500762939453, "epoch": 0.28, "learning_rate": 2.585085413929041e-05, "loss": 76.9066, "step": 847 }, { "compression_loss": 71.90733337402344, "epoch": 0.28, "learning_rate": 2.584592641261498e-05, "loss": 72.3927, "step": 848 }, { "compression_loss": 76.87049865722656, "epoch": 0.28, "learning_rate": 2.5840998685939556e-05, "loss": 77.2498, "step": 849 }, { "compression_loss": 76.57371520996094, "epoch": 0.28, "learning_rate": 2.5836070959264128e-05, "loss": 76.9822, "step": 850 }, { "compression_loss": 76.70341491699219, "epoch": 0.28, "learning_rate": 2.58311432325887e-05, "loss": 77.0763, "step": 851 }, { "compression_loss": 76.482421875, "epoch": 0.28, "learning_rate": 2.5826215505913272e-05, "loss": 77.0271, "step": 852 }, { "compression_loss": 73.41143035888672, "epoch": 0.28, "learning_rate": 2.5821287779237844e-05, "loss": 73.865, "step": 853 }, { "compression_loss": 73.56269073486328, "epoch": 0.28, "learning_rate": 2.581636005256242e-05, "loss": 74.2069, "step": 854 }, { "compression_loss": 75.28087615966797, "epoch": 0.28, "learning_rate": 2.581143232588699e-05, "loss": 75.6185, "step": 855 }, { "compression_loss": 76.94354248046875, "epoch": 0.28, "learning_rate": 2.5806504599211563e-05, "loss": 77.2359, "step": 856 }, { "compression_loss": 75.34576416015625, "epoch": 0.28, "learning_rate": 2.580157687253614e-05, "loss": 75.6432, "step": 857 }, { "compression_loss": 77.7199478149414, "epoch": 0.28, "learning_rate": 2.579664914586071e-05, "loss": 78.0412, "step": 858 }, { "compression_loss": 76.60514831542969, "epoch": 0.28, "learning_rate": 2.5791721419185286e-05, "loss": 77.02, "step": 859 }, { "compression_loss": 76.01853942871094, "epoch": 0.28, "learning_rate": 2.5786793692509858e-05, "loss": 76.3289, "step": 860 }, { "compression_loss": 76.66462707519531, "epoch": 0.28, "learning_rate": 2.578186596583443e-05, "loss": 77.1775, "step": 861 }, { "compression_loss": 76.26945495605469, "epoch": 0.28, "learning_rate": 2.5776938239159e-05, "loss": 76.6362, "step": 862 }, { "compression_loss": 74.81289672851562, "epoch": 0.28, "learning_rate": 2.5772010512483574e-05, "loss": 75.3007, "step": 863 }, { "compression_loss": 77.01138305664062, "epoch": 0.28, "learning_rate": 2.576708278580815e-05, "loss": 77.5048, "step": 864 }, { "compression_loss": 75.8565444946289, "epoch": 0.28, "learning_rate": 2.576215505913272e-05, "loss": 76.274, "step": 865 }, { "compression_loss": 73.44749450683594, "epoch": 0.28, "learning_rate": 2.5757227332457293e-05, "loss": 73.7989, "step": 866 }, { "compression_loss": 75.56127166748047, "epoch": 0.28, "learning_rate": 2.5752299605781865e-05, "loss": 76.0047, "step": 867 }, { "compression_loss": 75.6731948852539, "epoch": 0.29, "learning_rate": 2.5747371879106437e-05, "loss": 75.9442, "step": 868 }, { "compression_loss": 75.8758544921875, "epoch": 0.29, "learning_rate": 2.5742444152431016e-05, "loss": 76.8466, "step": 869 }, { "compression_loss": 75.61665344238281, "epoch": 0.29, "learning_rate": 2.5737516425755588e-05, "loss": 76.0946, "step": 870 }, { "compression_loss": 77.19587707519531, "epoch": 0.29, "learning_rate": 2.573258869908016e-05, "loss": 77.584, "step": 871 }, { "compression_loss": 74.38786315917969, "epoch": 0.29, "learning_rate": 2.572766097240473e-05, "loss": 74.6485, "step": 872 }, { "compression_loss": 74.96693420410156, "epoch": 0.29, "learning_rate": 2.5722733245729303e-05, "loss": 75.4286, "step": 873 }, { "compression_loss": 75.15896606445312, "epoch": 0.29, "learning_rate": 2.571780551905388e-05, "loss": 75.7384, "step": 874 }, { "compression_loss": 74.68858337402344, "epoch": 0.29, "learning_rate": 2.571287779237845e-05, "loss": 75.063, "step": 875 }, { "compression_loss": 75.95382690429688, "epoch": 0.29, "learning_rate": 2.5707950065703023e-05, "loss": 76.46, "step": 876 }, { "compression_loss": 76.0019760131836, "epoch": 0.29, "learning_rate": 2.5703022339027595e-05, "loss": 76.4287, "step": 877 }, { "compression_loss": 74.37010955810547, "epoch": 0.29, "learning_rate": 2.5698094612352167e-05, "loss": 74.6953, "step": 878 }, { "compression_loss": 76.16937255859375, "epoch": 0.29, "learning_rate": 2.5693166885676742e-05, "loss": 76.867, "step": 879 }, { "compression_loss": 74.81929779052734, "epoch": 0.29, "learning_rate": 2.5688239159001314e-05, "loss": 75.3255, "step": 880 }, { "compression_loss": 76.07334899902344, "epoch": 0.29, "learning_rate": 2.568331143232589e-05, "loss": 76.4479, "step": 881 }, { "compression_loss": 74.70734405517578, "epoch": 0.29, "learning_rate": 2.567838370565046e-05, "loss": 75.0181, "step": 882 }, { "compression_loss": 76.0511703491211, "epoch": 0.29, "learning_rate": 2.5673455978975033e-05, "loss": 76.7639, "step": 883 }, { "compression_loss": 77.82304382324219, "epoch": 0.29, "learning_rate": 2.566852825229961e-05, "loss": 78.8241, "step": 884 }, { "compression_loss": 78.41365051269531, "epoch": 0.29, "learning_rate": 2.566360052562418e-05, "loss": 78.8368, "step": 885 }, { "compression_loss": 78.30003356933594, "epoch": 0.29, "learning_rate": 2.5658672798948753e-05, "loss": 78.6338, "step": 886 }, { "compression_loss": 73.12753295898438, "epoch": 0.29, "learning_rate": 2.5653745072273325e-05, "loss": 73.4189, "step": 887 }, { "compression_loss": 74.69491577148438, "epoch": 0.29, "learning_rate": 2.5648817345597897e-05, "loss": 75.2517, "step": 888 }, { "compression_loss": 76.80538940429688, "epoch": 0.29, "learning_rate": 2.5643889618922472e-05, "loss": 77.3831, "step": 889 }, { "compression_loss": 74.72665405273438, "epoch": 0.29, "learning_rate": 2.5638961892247044e-05, "loss": 75.3191, "step": 890 }, { "compression_loss": 78.66868591308594, "epoch": 0.29, "learning_rate": 2.5634034165571616e-05, "loss": 79.2775, "step": 891 }, { "compression_loss": 75.09654235839844, "epoch": 0.29, "learning_rate": 2.5629106438896188e-05, "loss": 75.4285, "step": 892 }, { "compression_loss": 76.47674560546875, "epoch": 0.29, "learning_rate": 2.5624178712220763e-05, "loss": 76.9309, "step": 893 }, { "compression_loss": 75.36161804199219, "epoch": 0.29, "learning_rate": 2.561925098554534e-05, "loss": 76.0176, "step": 894 }, { "compression_loss": 75.7180404663086, "epoch": 0.29, "learning_rate": 2.561432325886991e-05, "loss": 75.8841, "step": 895 }, { "compression_loss": 75.02839660644531, "epoch": 0.29, "learning_rate": 2.5609395532194482e-05, "loss": 75.3458, "step": 896 }, { "compression_loss": 76.063720703125, "epoch": 0.29, "learning_rate": 2.5604467805519054e-05, "loss": 76.364, "step": 897 }, { "compression_loss": 76.02825927734375, "epoch": 0.3, "learning_rate": 2.5599540078843626e-05, "loss": 76.312, "step": 898 }, { "compression_loss": 74.82708740234375, "epoch": 0.3, "learning_rate": 2.5594612352168202e-05, "loss": 75.1909, "step": 899 }, { "compression_loss": 77.70838165283203, "epoch": 0.3, "learning_rate": 2.5589684625492774e-05, "loss": 78.2913, "step": 900 }, { "compression_loss": 74.59722900390625, "epoch": 0.3, "learning_rate": 2.5584756898817346e-05, "loss": 75.0223, "step": 901 }, { "compression_loss": 75.14542388916016, "epoch": 0.3, "learning_rate": 2.5579829172141918e-05, "loss": 75.4019, "step": 902 }, { "compression_loss": 77.89845275878906, "epoch": 0.3, "learning_rate": 2.557490144546649e-05, "loss": 78.4955, "step": 903 }, { "compression_loss": 74.40000915527344, "epoch": 0.3, "learning_rate": 2.5569973718791065e-05, "loss": 74.8448, "step": 904 }, { "compression_loss": 73.22664642333984, "epoch": 0.3, "learning_rate": 2.556504599211564e-05, "loss": 73.7097, "step": 905 }, { "compression_loss": 76.91228485107422, "epoch": 0.3, "learning_rate": 2.5560118265440212e-05, "loss": 77.2018, "step": 906 }, { "compression_loss": 76.28392791748047, "epoch": 0.3, "learning_rate": 2.5555190538764784e-05, "loss": 76.6646, "step": 907 }, { "compression_loss": 77.94093322753906, "epoch": 0.3, "learning_rate": 2.5550262812089356e-05, "loss": 78.4984, "step": 908 }, { "compression_loss": 77.43626403808594, "epoch": 0.3, "learning_rate": 2.554533508541393e-05, "loss": 77.7924, "step": 909 }, { "compression_loss": 76.23516845703125, "epoch": 0.3, "learning_rate": 2.5540407358738504e-05, "loss": 76.6314, "step": 910 }, { "compression_loss": 73.89649200439453, "epoch": 0.3, "learning_rate": 2.5535479632063076e-05, "loss": 74.3062, "step": 911 }, { "compression_loss": 77.3267822265625, "epoch": 0.3, "learning_rate": 2.5530551905387647e-05, "loss": 77.6057, "step": 912 }, { "compression_loss": 77.6115951538086, "epoch": 0.3, "learning_rate": 2.552562417871222e-05, "loss": 78.0254, "step": 913 }, { "compression_loss": 76.74028015136719, "epoch": 0.3, "learning_rate": 2.5520696452036795e-05, "loss": 77.2203, "step": 914 }, { "compression_loss": 76.75672149658203, "epoch": 0.3, "learning_rate": 2.5515768725361367e-05, "loss": 77.448, "step": 915 }, { "compression_loss": 76.39997863769531, "epoch": 0.3, "learning_rate": 2.551084099868594e-05, "loss": 76.7879, "step": 916 }, { "compression_loss": 74.14665985107422, "epoch": 0.3, "learning_rate": 2.5505913272010514e-05, "loss": 74.6826, "step": 917 }, { "compression_loss": 75.33242797851562, "epoch": 0.3, "learning_rate": 2.5500985545335086e-05, "loss": 75.9864, "step": 918 }, { "compression_loss": 75.8909912109375, "epoch": 0.3, "learning_rate": 2.549605781865966e-05, "loss": 76.4773, "step": 919 }, { "compression_loss": 74.74566650390625, "epoch": 0.3, "learning_rate": 2.5491130091984233e-05, "loss": 75.2755, "step": 920 }, { "compression_loss": 72.99015045166016, "epoch": 0.3, "learning_rate": 2.5486202365308805e-05, "loss": 73.4308, "step": 921 }, { "compression_loss": 75.917724609375, "epoch": 0.3, "learning_rate": 2.5481274638633377e-05, "loss": 76.3896, "step": 922 }, { "compression_loss": 76.66183471679688, "epoch": 0.3, "learning_rate": 2.547634691195795e-05, "loss": 77.4059, "step": 923 }, { "compression_loss": 76.47813415527344, "epoch": 0.3, "learning_rate": 2.5471419185282525e-05, "loss": 76.6991, "step": 924 }, { "compression_loss": 76.58570861816406, "epoch": 0.3, "learning_rate": 2.5466491458607097e-05, "loss": 76.8827, "step": 925 }, { "compression_loss": 75.1261215209961, "epoch": 0.3, "learning_rate": 2.546156373193167e-05, "loss": 75.4186, "step": 926 }, { "compression_loss": 76.0444564819336, "epoch": 0.3, "learning_rate": 2.545663600525624e-05, "loss": 76.7649, "step": 927 }, { "compression_loss": 76.31124877929688, "epoch": 0.3, "learning_rate": 2.5451708278580813e-05, "loss": 76.6854, "step": 928 }, { "compression_loss": 76.12016296386719, "epoch": 0.31, "learning_rate": 2.544678055190539e-05, "loss": 76.6577, "step": 929 }, { "compression_loss": 75.23043823242188, "epoch": 0.31, "learning_rate": 2.5441852825229963e-05, "loss": 75.4293, "step": 930 }, { "compression_loss": 77.70823669433594, "epoch": 0.31, "learning_rate": 2.5436925098554535e-05, "loss": 78.4855, "step": 931 }, { "compression_loss": 76.48111724853516, "epoch": 0.31, "learning_rate": 2.5431997371879107e-05, "loss": 76.9389, "step": 932 }, { "compression_loss": 75.00910949707031, "epoch": 0.31, "learning_rate": 2.542706964520368e-05, "loss": 75.3315, "step": 933 }, { "compression_loss": 73.81132507324219, "epoch": 0.31, "learning_rate": 2.5422141918528255e-05, "loss": 74.064, "step": 934 }, { "compression_loss": 77.02596282958984, "epoch": 0.31, "learning_rate": 2.5417214191852826e-05, "loss": 78.0613, "step": 935 }, { "compression_loss": 75.37384796142578, "epoch": 0.31, "learning_rate": 2.54122864651774e-05, "loss": 76.1121, "step": 936 }, { "compression_loss": 75.36213684082031, "epoch": 0.31, "learning_rate": 2.540735873850197e-05, "loss": 76.2488, "step": 937 }, { "compression_loss": 73.36485290527344, "epoch": 0.31, "learning_rate": 2.5402431011826542e-05, "loss": 73.6884, "step": 938 }, { "compression_loss": 76.15815734863281, "epoch": 0.31, "learning_rate": 2.5397503285151118e-05, "loss": 76.4818, "step": 939 }, { "compression_loss": 75.65608215332031, "epoch": 0.31, "learning_rate": 2.539257555847569e-05, "loss": 75.9137, "step": 940 }, { "compression_loss": 75.12210083007812, "epoch": 0.31, "learning_rate": 2.5387647831800265e-05, "loss": 75.5023, "step": 941 }, { "compression_loss": 74.13884735107422, "epoch": 0.31, "learning_rate": 2.5382720105124837e-05, "loss": 74.2974, "step": 942 }, { "compression_loss": 76.58340454101562, "epoch": 0.31, "learning_rate": 2.537779237844941e-05, "loss": 77.269, "step": 943 }, { "compression_loss": 76.00332641601562, "epoch": 0.31, "learning_rate": 2.5372864651773984e-05, "loss": 76.6042, "step": 944 }, { "compression_loss": 75.32025909423828, "epoch": 0.31, "learning_rate": 2.5367936925098556e-05, "loss": 75.499, "step": 945 }, { "compression_loss": 76.68002319335938, "epoch": 0.31, "learning_rate": 2.536300919842313e-05, "loss": 76.9749, "step": 946 }, { "compression_loss": 75.43913269042969, "epoch": 0.31, "learning_rate": 2.53580814717477e-05, "loss": 75.9303, "step": 947 }, { "compression_loss": 76.60797119140625, "epoch": 0.31, "learning_rate": 2.5353153745072272e-05, "loss": 77.2885, "step": 948 }, { "compression_loss": 74.27387237548828, "epoch": 0.31, "learning_rate": 2.5348226018396848e-05, "loss": 74.7061, "step": 949 }, { "compression_loss": 75.75480651855469, "epoch": 0.31, "learning_rate": 2.534329829172142e-05, "loss": 76.3063, "step": 950 }, { "compression_loss": 75.79595184326172, "epoch": 0.31, "learning_rate": 2.533837056504599e-05, "loss": 76.5301, "step": 951 }, { "compression_loss": 73.71554565429688, "epoch": 0.31, "learning_rate": 2.5333442838370564e-05, "loss": 74.1553, "step": 952 }, { "compression_loss": 76.27751159667969, "epoch": 0.31, "learning_rate": 2.532851511169514e-05, "loss": 76.6541, "step": 953 }, { "compression_loss": 73.826904296875, "epoch": 0.31, "learning_rate": 2.5323587385019714e-05, "loss": 74.1596, "step": 954 }, { "compression_loss": 74.16844177246094, "epoch": 0.31, "learning_rate": 2.5318659658344286e-05, "loss": 74.5958, "step": 955 }, { "compression_loss": 77.71041107177734, "epoch": 0.31, "learning_rate": 2.5313731931668858e-05, "loss": 78.1776, "step": 956 }, { "compression_loss": 76.7365493774414, "epoch": 0.31, "learning_rate": 2.530880420499343e-05, "loss": 77.0908, "step": 957 }, { "compression_loss": 76.29503631591797, "epoch": 0.31, "learning_rate": 2.5303876478318002e-05, "loss": 76.6439, "step": 958 }, { "compression_loss": 75.02497863769531, "epoch": 0.32, "learning_rate": 2.5298948751642577e-05, "loss": 75.3628, "step": 959 }, { "compression_loss": 75.54483032226562, "epoch": 0.32, "learning_rate": 2.529402102496715e-05, "loss": 75.8797, "step": 960 }, { "compression_loss": 77.22688293457031, "epoch": 0.32, "learning_rate": 2.528909329829172e-05, "loss": 77.4219, "step": 961 }, { "compression_loss": 74.62763214111328, "epoch": 0.32, "learning_rate": 2.5284165571616293e-05, "loss": 75.5012, "step": 962 }, { "compression_loss": 75.76385498046875, "epoch": 0.32, "learning_rate": 2.5279237844940865e-05, "loss": 76.2867, "step": 963 }, { "compression_loss": 72.5418701171875, "epoch": 0.32, "learning_rate": 2.527431011826544e-05, "loss": 73.0562, "step": 964 }, { "compression_loss": 74.2755126953125, "epoch": 0.32, "learning_rate": 2.5269382391590016e-05, "loss": 74.5825, "step": 965 }, { "compression_loss": 76.55593872070312, "epoch": 0.32, "learning_rate": 2.5264454664914588e-05, "loss": 77.1897, "step": 966 }, { "compression_loss": 73.36320495605469, "epoch": 0.32, "learning_rate": 2.525952693823916e-05, "loss": 73.7624, "step": 967 }, { "compression_loss": 76.07953643798828, "epoch": 0.32, "learning_rate": 2.5254599211563732e-05, "loss": 76.4441, "step": 968 }, { "compression_loss": 77.37236785888672, "epoch": 0.32, "learning_rate": 2.5249671484888307e-05, "loss": 78.0153, "step": 969 }, { "compression_loss": 76.25328063964844, "epoch": 0.32, "learning_rate": 2.524474375821288e-05, "loss": 76.5889, "step": 970 }, { "compression_loss": 76.14161682128906, "epoch": 0.32, "learning_rate": 2.523981603153745e-05, "loss": 76.5364, "step": 971 }, { "compression_loss": 76.64450073242188, "epoch": 0.32, "learning_rate": 2.5234888304862023e-05, "loss": 77.0061, "step": 972 }, { "compression_loss": 74.63941192626953, "epoch": 0.32, "learning_rate": 2.5229960578186595e-05, "loss": 74.9729, "step": 973 }, { "compression_loss": 77.22631072998047, "epoch": 0.32, "learning_rate": 2.522503285151117e-05, "loss": 77.7315, "step": 974 }, { "compression_loss": 75.28347778320312, "epoch": 0.32, "learning_rate": 2.5220105124835743e-05, "loss": 75.7405, "step": 975 }, { "compression_loss": 75.90863037109375, "epoch": 0.32, "learning_rate": 2.5215177398160318e-05, "loss": 76.053, "step": 976 }, { "compression_loss": 75.17864990234375, "epoch": 0.32, "learning_rate": 2.521024967148489e-05, "loss": 75.8103, "step": 977 }, { "compression_loss": 76.82496643066406, "epoch": 0.32, "learning_rate": 2.5205321944809462e-05, "loss": 77.6322, "step": 978 }, { "compression_loss": 73.37907409667969, "epoch": 0.32, "learning_rate": 2.5200394218134037e-05, "loss": 73.7557, "step": 979 }, { "compression_loss": 77.89411926269531, "epoch": 0.32, "learning_rate": 2.519546649145861e-05, "loss": 78.3479, "step": 980 }, { "compression_loss": 74.478271484375, "epoch": 0.32, "learning_rate": 2.519053876478318e-05, "loss": 75.1043, "step": 981 }, { "compression_loss": 75.13471984863281, "epoch": 0.32, "learning_rate": 2.5185611038107753e-05, "loss": 75.4033, "step": 982 }, { "compression_loss": 75.1784896850586, "epoch": 0.32, "learning_rate": 2.5180683311432325e-05, "loss": 75.9112, "step": 983 }, { "compression_loss": 76.36214447021484, "epoch": 0.32, "learning_rate": 2.51757555847569e-05, "loss": 76.9436, "step": 984 }, { "compression_loss": 75.67510986328125, "epoch": 0.32, "learning_rate": 2.5170827858081472e-05, "loss": 76.1298, "step": 985 }, { "compression_loss": 75.68942260742188, "epoch": 0.32, "learning_rate": 2.5165900131406044e-05, "loss": 75.9523, "step": 986 }, { "compression_loss": 76.21712493896484, "epoch": 0.32, "learning_rate": 2.5160972404730616e-05, "loss": 76.613, "step": 987 }, { "compression_loss": 76.80122375488281, "epoch": 0.32, "learning_rate": 2.515604467805519e-05, "loss": 77.471, "step": 988 }, { "compression_loss": 77.91722106933594, "epoch": 0.32, "learning_rate": 2.5151116951379767e-05, "loss": 78.5956, "step": 989 }, { "compression_loss": 76.19454956054688, "epoch": 0.33, "learning_rate": 2.514618922470434e-05, "loss": 76.6235, "step": 990 }, { "compression_loss": 76.87162780761719, "epoch": 0.33, "learning_rate": 2.514126149802891e-05, "loss": 77.2695, "step": 991 }, { "compression_loss": 74.5050277709961, "epoch": 0.33, "learning_rate": 2.5136333771353483e-05, "loss": 74.917, "step": 992 }, { "compression_loss": 75.32200622558594, "epoch": 0.33, "learning_rate": 2.5131406044678055e-05, "loss": 75.6419, "step": 993 }, { "compression_loss": 76.59249877929688, "epoch": 0.33, "learning_rate": 2.512647831800263e-05, "loss": 77.1968, "step": 994 }, { "compression_loss": 75.7507553100586, "epoch": 0.33, "learning_rate": 2.5121550591327202e-05, "loss": 76.4722, "step": 995 }, { "compression_loss": 75.40194702148438, "epoch": 0.33, "learning_rate": 2.5116622864651774e-05, "loss": 75.9473, "step": 996 }, { "compression_loss": 76.52872467041016, "epoch": 0.33, "learning_rate": 2.5111695137976346e-05, "loss": 77.059, "step": 997 }, { "compression_loss": 75.58120727539062, "epoch": 0.33, "learning_rate": 2.5106767411300918e-05, "loss": 75.9144, "step": 998 }, { "compression_loss": 76.21688842773438, "epoch": 0.33, "learning_rate": 2.5101839684625494e-05, "loss": 76.705, "step": 999 }, { "compression_loss": 76.38349914550781, "epoch": 0.33, "learning_rate": 2.509691195795007e-05, "loss": 76.8573, "step": 1000 }, { "epoch": 0.33, "eval_exact_match": 85.61021759697256, "eval_f1": 92.13583512905524, "step": 1000 }, { "compression_loss": 76.03851318359375, "epoch": 0.33, "learning_rate": 2.509198423127464e-05, "loss": 76.4006, "step": 1001 }, { "compression_loss": 74.04855346679688, "epoch": 0.33, "learning_rate": 2.5087056504599213e-05, "loss": 74.5583, "step": 1002 }, { "compression_loss": 74.54679870605469, "epoch": 0.33, "learning_rate": 2.5082128777923785e-05, "loss": 74.9348, "step": 1003 }, { "compression_loss": 74.92914581298828, "epoch": 0.33, "learning_rate": 2.507720105124836e-05, "loss": 75.4178, "step": 1004 }, { "compression_loss": 74.00630187988281, "epoch": 0.33, "learning_rate": 2.5072273324572932e-05, "loss": 74.4437, "step": 1005 }, { "compression_loss": 78.17033386230469, "epoch": 0.33, "learning_rate": 2.5067345597897504e-05, "loss": 78.6142, "step": 1006 }, { "compression_loss": 75.95498657226562, "epoch": 0.33, "learning_rate": 2.5062417871222076e-05, "loss": 76.434, "step": 1007 }, { "compression_loss": 75.59843444824219, "epoch": 0.33, "learning_rate": 2.5057490144546648e-05, "loss": 75.9435, "step": 1008 }, { "compression_loss": 76.42074584960938, "epoch": 0.33, "learning_rate": 2.5052562417871223e-05, "loss": 76.7048, "step": 1009 }, { "compression_loss": 75.97257995605469, "epoch": 0.33, "learning_rate": 2.5047634691195795e-05, "loss": 76.5545, "step": 1010 }, { "compression_loss": 74.66482543945312, "epoch": 0.33, "learning_rate": 2.5042706964520367e-05, "loss": 75.0156, "step": 1011 }, { "compression_loss": 78.33641815185547, "epoch": 0.33, "learning_rate": 2.5037779237844943e-05, "loss": 78.7402, "step": 1012 }, { "compression_loss": 77.12088012695312, "epoch": 0.33, "learning_rate": 2.5032851511169515e-05, "loss": 77.6069, "step": 1013 }, { "compression_loss": 75.10132598876953, "epoch": 0.33, "learning_rate": 2.502792378449409e-05, "loss": 75.4434, "step": 1014 }, { "compression_loss": 74.09529876708984, "epoch": 0.33, "learning_rate": 2.5022996057818662e-05, "loss": 74.4985, "step": 1015 }, { "compression_loss": 73.76177215576172, "epoch": 0.33, "learning_rate": 2.5018068331143234e-05, "loss": 73.9488, "step": 1016 }, { "compression_loss": 75.87919616699219, "epoch": 0.33, "learning_rate": 2.5013140604467806e-05, "loss": 76.2607, "step": 1017 }, { "compression_loss": 75.00450134277344, "epoch": 0.33, "learning_rate": 2.5008212877792378e-05, "loss": 75.3252, "step": 1018 }, { "compression_loss": 72.81199645996094, "epoch": 0.33, "learning_rate": 2.5003285151116953e-05, "loss": 73.4373, "step": 1019 }, { "compression_loss": 74.59175109863281, "epoch": 0.34, "learning_rate": 2.4998357424441525e-05, "loss": 74.894, "step": 1020 }, { "compression_loss": 76.2168197631836, "epoch": 0.34, "learning_rate": 2.4993429697766097e-05, "loss": 76.6093, "step": 1021 }, { "compression_loss": 77.02246856689453, "epoch": 0.34, "learning_rate": 2.498850197109067e-05, "loss": 77.5811, "step": 1022 }, { "compression_loss": 73.83756256103516, "epoch": 0.34, "learning_rate": 2.498357424441524e-05, "loss": 74.0367, "step": 1023 }, { "compression_loss": 75.91946411132812, "epoch": 0.34, "learning_rate": 2.497864651773982e-05, "loss": 76.5752, "step": 1024 }, { "compression_loss": 76.04618072509766, "epoch": 0.34, "learning_rate": 2.4973718791064392e-05, "loss": 76.3987, "step": 1025 }, { "compression_loss": 75.83686828613281, "epoch": 0.34, "learning_rate": 2.4968791064388964e-05, "loss": 76.1211, "step": 1026 }, { "compression_loss": 73.58849334716797, "epoch": 0.34, "learning_rate": 2.4963863337713536e-05, "loss": 73.9405, "step": 1027 }, { "compression_loss": 75.18228149414062, "epoch": 0.34, "learning_rate": 2.4958935611038108e-05, "loss": 75.4232, "step": 1028 }, { "compression_loss": 75.02566528320312, "epoch": 0.34, "learning_rate": 2.4954007884362683e-05, "loss": 75.3454, "step": 1029 }, { "compression_loss": 75.44427490234375, "epoch": 0.34, "learning_rate": 2.4949080157687255e-05, "loss": 75.6571, "step": 1030 }, { "compression_loss": 76.50968170166016, "epoch": 0.34, "learning_rate": 2.4944152431011827e-05, "loss": 77.0269, "step": 1031 }, { "compression_loss": 76.20185852050781, "epoch": 0.34, "learning_rate": 2.49392247043364e-05, "loss": 76.4675, "step": 1032 }, { "compression_loss": 76.84738159179688, "epoch": 0.34, "learning_rate": 2.493429697766097e-05, "loss": 77.0648, "step": 1033 }, { "compression_loss": 72.97360229492188, "epoch": 0.34, "learning_rate": 2.4929369250985546e-05, "loss": 73.8046, "step": 1034 }, { "compression_loss": 75.99500274658203, "epoch": 0.34, "learning_rate": 2.4924441524310118e-05, "loss": 76.3781, "step": 1035 }, { "compression_loss": 74.08999633789062, "epoch": 0.34, "learning_rate": 2.4919513797634694e-05, "loss": 74.4048, "step": 1036 }, { "compression_loss": 77.67904663085938, "epoch": 0.34, "learning_rate": 2.4914586070959266e-05, "loss": 78.0504, "step": 1037 }, { "compression_loss": 77.17402648925781, "epoch": 0.34, "learning_rate": 2.4909658344283838e-05, "loss": 77.6071, "step": 1038 }, { "compression_loss": 76.89083862304688, "epoch": 0.34, "learning_rate": 2.4904730617608413e-05, "loss": 77.2538, "step": 1039 }, { "compression_loss": 76.04698181152344, "epoch": 0.34, "learning_rate": 2.4899802890932985e-05, "loss": 76.3586, "step": 1040 }, { "compression_loss": 76.44973754882812, "epoch": 0.34, "learning_rate": 2.4894875164257557e-05, "loss": 76.9121, "step": 1041 }, { "compression_loss": 72.33728790283203, "epoch": 0.34, "learning_rate": 2.488994743758213e-05, "loss": 72.8899, "step": 1042 }, { "compression_loss": 76.55278778076172, "epoch": 0.34, "learning_rate": 2.48850197109067e-05, "loss": 76.8444, "step": 1043 }, { "compression_loss": 77.0687255859375, "epoch": 0.34, "learning_rate": 2.4880091984231276e-05, "loss": 77.4692, "step": 1044 }, { "compression_loss": 73.23356628417969, "epoch": 0.34, "learning_rate": 2.4875164257555848e-05, "loss": 73.5954, "step": 1045 }, { "compression_loss": 74.12051391601562, "epoch": 0.34, "learning_rate": 2.487023653088042e-05, "loss": 74.366, "step": 1046 }, { "compression_loss": 76.82403564453125, "epoch": 0.34, "learning_rate": 2.4865308804204992e-05, "loss": 77.2534, "step": 1047 }, { "compression_loss": 73.64956665039062, "epoch": 0.34, "learning_rate": 2.4860381077529567e-05, "loss": 73.76, "step": 1048 }, { "compression_loss": 75.40585327148438, "epoch": 0.34, "learning_rate": 2.4855453350854143e-05, "loss": 76.2211, "step": 1049 }, { "compression_loss": 78.16334533691406, "epoch": 0.34, "learning_rate": 2.4850525624178715e-05, "loss": 78.4034, "step": 1050 }, { "compression_loss": 75.42024230957031, "epoch": 0.35, "learning_rate": 2.4845597897503287e-05, "loss": 75.7164, "step": 1051 }, { "compression_loss": 77.13371276855469, "epoch": 0.35, "learning_rate": 2.484067017082786e-05, "loss": 77.4529, "step": 1052 }, { "compression_loss": 75.47083282470703, "epoch": 0.35, "learning_rate": 2.483574244415243e-05, "loss": 75.9304, "step": 1053 }, { "compression_loss": 76.20349884033203, "epoch": 0.35, "learning_rate": 2.4830814717477006e-05, "loss": 76.6081, "step": 1054 }, { "compression_loss": 76.75390625, "epoch": 0.35, "learning_rate": 2.4825886990801578e-05, "loss": 77.3057, "step": 1055 }, { "compression_loss": 75.32998657226562, "epoch": 0.35, "learning_rate": 2.482095926412615e-05, "loss": 75.6926, "step": 1056 }, { "compression_loss": 77.0321273803711, "epoch": 0.35, "learning_rate": 2.4816031537450722e-05, "loss": 77.6216, "step": 1057 }, { "compression_loss": 75.75042724609375, "epoch": 0.35, "learning_rate": 2.4811103810775294e-05, "loss": 76.197, "step": 1058 }, { "compression_loss": 74.10926055908203, "epoch": 0.35, "learning_rate": 2.480617608409987e-05, "loss": 74.7678, "step": 1059 }, { "compression_loss": 76.62380981445312, "epoch": 0.35, "learning_rate": 2.4801248357424445e-05, "loss": 77.1888, "step": 1060 }, { "compression_loss": 75.1126708984375, "epoch": 0.35, "learning_rate": 2.4796320630749017e-05, "loss": 75.3659, "step": 1061 }, { "compression_loss": 76.73856353759766, "epoch": 0.35, "learning_rate": 2.479139290407359e-05, "loss": 77.1624, "step": 1062 }, { "compression_loss": 76.22116088867188, "epoch": 0.35, "learning_rate": 2.478646517739816e-05, "loss": 76.5849, "step": 1063 }, { "compression_loss": 76.6436767578125, "epoch": 0.35, "learning_rate": 2.4781537450722736e-05, "loss": 77.0656, "step": 1064 }, { "compression_loss": 76.92459869384766, "epoch": 0.35, "learning_rate": 2.4776609724047308e-05, "loss": 77.2873, "step": 1065 }, { "compression_loss": 74.388671875, "epoch": 0.35, "learning_rate": 2.477168199737188e-05, "loss": 74.7283, "step": 1066 }, { "compression_loss": 74.75926208496094, "epoch": 0.35, "learning_rate": 2.4766754270696452e-05, "loss": 75.1092, "step": 1067 }, { "compression_loss": 73.18669128417969, "epoch": 0.35, "learning_rate": 2.4761826544021024e-05, "loss": 73.4899, "step": 1068 }, { "compression_loss": 75.42678833007812, "epoch": 0.35, "learning_rate": 2.47568988173456e-05, "loss": 75.7059, "step": 1069 }, { "compression_loss": 78.2430648803711, "epoch": 0.35, "learning_rate": 2.475197109067017e-05, "loss": 78.5778, "step": 1070 }, { "compression_loss": 76.77153015136719, "epoch": 0.35, "learning_rate": 2.4747043363994743e-05, "loss": 77.1216, "step": 1071 }, { "compression_loss": 76.92723083496094, "epoch": 0.35, "learning_rate": 2.474211563731932e-05, "loss": 77.3731, "step": 1072 }, { "compression_loss": 78.4759750366211, "epoch": 0.35, "learning_rate": 2.473718791064389e-05, "loss": 79.2425, "step": 1073 }, { "compression_loss": 75.26060485839844, "epoch": 0.35, "learning_rate": 2.4732260183968466e-05, "loss": 75.8431, "step": 1074 }, { "compression_loss": 73.71881866455078, "epoch": 0.35, "learning_rate": 2.4727332457293038e-05, "loss": 74.0295, "step": 1075 }, { "compression_loss": 76.57136535644531, "epoch": 0.35, "learning_rate": 2.472240473061761e-05, "loss": 77.1371, "step": 1076 }, { "compression_loss": 73.070068359375, "epoch": 0.35, "learning_rate": 2.471747700394218e-05, "loss": 73.2777, "step": 1077 }, { "compression_loss": 75.34468841552734, "epoch": 0.35, "learning_rate": 2.4712549277266754e-05, "loss": 75.6342, "step": 1078 }, { "compression_loss": 74.7386245727539, "epoch": 0.35, "learning_rate": 2.470762155059133e-05, "loss": 75.4853, "step": 1079 }, { "compression_loss": 76.36392974853516, "epoch": 0.35, "learning_rate": 2.47026938239159e-05, "loss": 76.9585, "step": 1080 }, { "compression_loss": 75.46273803710938, "epoch": 0.36, "learning_rate": 2.4697766097240473e-05, "loss": 76.0457, "step": 1081 }, { "compression_loss": 78.15350341796875, "epoch": 0.36, "learning_rate": 2.4692838370565045e-05, "loss": 78.8181, "step": 1082 }, { "compression_loss": 73.212890625, "epoch": 0.36, "learning_rate": 2.4687910643889617e-05, "loss": 73.4992, "step": 1083 }, { "compression_loss": 75.41143798828125, "epoch": 0.36, "learning_rate": 2.4682982917214196e-05, "loss": 75.7232, "step": 1084 }, { "compression_loss": 73.93028259277344, "epoch": 0.36, "learning_rate": 2.4678055190538768e-05, "loss": 74.4029, "step": 1085 }, { "compression_loss": 72.45718383789062, "epoch": 0.36, "learning_rate": 2.467312746386334e-05, "loss": 72.8074, "step": 1086 }, { "compression_loss": 72.73797607421875, "epoch": 0.36, "learning_rate": 2.466819973718791e-05, "loss": 73.148, "step": 1087 }, { "compression_loss": 76.41755676269531, "epoch": 0.36, "learning_rate": 2.4663272010512483e-05, "loss": 76.7227, "step": 1088 }, { "compression_loss": 75.14752960205078, "epoch": 0.36, "learning_rate": 2.465834428383706e-05, "loss": 75.5044, "step": 1089 }, { "compression_loss": 76.76731872558594, "epoch": 0.36, "learning_rate": 2.465341655716163e-05, "loss": 77.1892, "step": 1090 }, { "compression_loss": 75.21115112304688, "epoch": 0.36, "learning_rate": 2.4648488830486203e-05, "loss": 75.8832, "step": 1091 }, { "compression_loss": 75.99147033691406, "epoch": 0.36, "learning_rate": 2.4643561103810775e-05, "loss": 76.4876, "step": 1092 }, { "compression_loss": 74.95599365234375, "epoch": 0.36, "learning_rate": 2.4638633377135347e-05, "loss": 75.3281, "step": 1093 }, { "compression_loss": 74.72752380371094, "epoch": 0.36, "learning_rate": 2.4633705650459922e-05, "loss": 75.3183, "step": 1094 }, { "compression_loss": 75.07051849365234, "epoch": 0.36, "learning_rate": 2.4628777923784494e-05, "loss": 75.4285, "step": 1095 }, { "compression_loss": 79.07938385009766, "epoch": 0.36, "learning_rate": 2.462385019710907e-05, "loss": 79.6752, "step": 1096 }, { "compression_loss": 75.40400695800781, "epoch": 0.36, "learning_rate": 2.461892247043364e-05, "loss": 75.9067, "step": 1097 }, { "compression_loss": 74.76124572753906, "epoch": 0.36, "learning_rate": 2.4613994743758213e-05, "loss": 75.1031, "step": 1098 }, { "compression_loss": 78.83189392089844, "epoch": 0.36, "learning_rate": 2.460906701708279e-05, "loss": 79.5115, "step": 1099 }, { "compression_loss": 75.96585083007812, "epoch": 0.36, "learning_rate": 2.460413929040736e-05, "loss": 76.4572, "step": 1100 }, { "compression_loss": 75.83721923828125, "epoch": 0.36, "learning_rate": 2.4599211563731933e-05, "loss": 76.5298, "step": 1101 }, { "compression_loss": 76.91724395751953, "epoch": 0.36, "learning_rate": 2.4594283837056505e-05, "loss": 77.2727, "step": 1102 }, { "compression_loss": 75.340576171875, "epoch": 0.36, "learning_rate": 2.4589356110381077e-05, "loss": 75.8467, "step": 1103 }, { "compression_loss": 74.814453125, "epoch": 0.36, "learning_rate": 2.4584428383705652e-05, "loss": 75.0292, "step": 1104 }, { "compression_loss": 76.08293151855469, "epoch": 0.36, "learning_rate": 2.4579500657030224e-05, "loss": 76.5546, "step": 1105 }, { "compression_loss": 76.87249755859375, "epoch": 0.36, "learning_rate": 2.4574572930354796e-05, "loss": 77.3148, "step": 1106 }, { "compression_loss": 77.11647033691406, "epoch": 0.36, "learning_rate": 2.4569645203679368e-05, "loss": 77.4632, "step": 1107 }, { "compression_loss": 74.80059814453125, "epoch": 0.36, "learning_rate": 2.4564717477003943e-05, "loss": 75.5332, "step": 1108 }, { "compression_loss": 75.81460571289062, "epoch": 0.36, "learning_rate": 2.455978975032852e-05, "loss": 76.6421, "step": 1109 }, { "compression_loss": 77.60086059570312, "epoch": 0.36, "learning_rate": 2.455486202365309e-05, "loss": 78.2485, "step": 1110 }, { "compression_loss": 77.7778549194336, "epoch": 0.36, "learning_rate": 2.4549934296977662e-05, "loss": 78.3398, "step": 1111 }, { "compression_loss": 77.47474670410156, "epoch": 0.37, "learning_rate": 2.4545006570302234e-05, "loss": 77.8811, "step": 1112 }, { "compression_loss": 74.4639892578125, "epoch": 0.37, "learning_rate": 2.4540078843626806e-05, "loss": 75.2037, "step": 1113 }, { "compression_loss": 75.17448425292969, "epoch": 0.37, "learning_rate": 2.4535151116951382e-05, "loss": 75.5374, "step": 1114 }, { "compression_loss": 75.25468444824219, "epoch": 0.37, "learning_rate": 2.4530223390275954e-05, "loss": 75.5737, "step": 1115 }, { "compression_loss": 76.11595153808594, "epoch": 0.37, "learning_rate": 2.4525295663600526e-05, "loss": 76.5561, "step": 1116 }, { "compression_loss": 78.10615539550781, "epoch": 0.37, "learning_rate": 2.4520367936925098e-05, "loss": 78.4929, "step": 1117 }, { "compression_loss": 73.73802185058594, "epoch": 0.37, "learning_rate": 2.451544021024967e-05, "loss": 74.1014, "step": 1118 }, { "compression_loss": 76.49676513671875, "epoch": 0.37, "learning_rate": 2.4510512483574245e-05, "loss": 76.8083, "step": 1119 }, { "compression_loss": 76.31704711914062, "epoch": 0.37, "learning_rate": 2.450558475689882e-05, "loss": 76.6757, "step": 1120 }, { "compression_loss": 75.00395202636719, "epoch": 0.37, "learning_rate": 2.4500657030223392e-05, "loss": 75.2457, "step": 1121 }, { "compression_loss": 76.94387817382812, "epoch": 0.37, "learning_rate": 2.4495729303547964e-05, "loss": 77.5301, "step": 1122 }, { "compression_loss": 77.85289001464844, "epoch": 0.37, "learning_rate": 2.4490801576872536e-05, "loss": 78.2002, "step": 1123 }, { "compression_loss": 75.62935638427734, "epoch": 0.37, "learning_rate": 2.448587385019711e-05, "loss": 76.0831, "step": 1124 }, { "compression_loss": 76.26801300048828, "epoch": 0.37, "learning_rate": 2.4480946123521684e-05, "loss": 76.5295, "step": 1125 }, { "compression_loss": 74.32608795166016, "epoch": 0.37, "learning_rate": 2.4476018396846256e-05, "loss": 74.9582, "step": 1126 }, { "compression_loss": 74.66842651367188, "epoch": 0.37, "learning_rate": 2.4471090670170828e-05, "loss": 74.9812, "step": 1127 }, { "compression_loss": 77.3962631225586, "epoch": 0.37, "learning_rate": 2.44661629434954e-05, "loss": 77.673, "step": 1128 }, { "compression_loss": 75.85955047607422, "epoch": 0.37, "learning_rate": 2.4461235216819975e-05, "loss": 76.367, "step": 1129 }, { "compression_loss": 74.74882507324219, "epoch": 0.37, "learning_rate": 2.4456307490144547e-05, "loss": 75.1988, "step": 1130 }, { "compression_loss": 74.80130004882812, "epoch": 0.37, "learning_rate": 2.4451379763469122e-05, "loss": 75.0482, "step": 1131 }, { "compression_loss": 76.01618194580078, "epoch": 0.37, "learning_rate": 2.4446452036793694e-05, "loss": 76.4141, "step": 1132 }, { "compression_loss": 74.268798828125, "epoch": 0.37, "learning_rate": 2.4441524310118266e-05, "loss": 74.5511, "step": 1133 }, { "compression_loss": 74.25418090820312, "epoch": 0.37, "learning_rate": 2.443659658344284e-05, "loss": 74.6822, "step": 1134 }, { "compression_loss": 76.42313385009766, "epoch": 0.37, "learning_rate": 2.4431668856767413e-05, "loss": 77.1264, "step": 1135 }, { "compression_loss": 77.20845031738281, "epoch": 0.37, "learning_rate": 2.4426741130091985e-05, "loss": 77.8705, "step": 1136 }, { "compression_loss": 74.46345520019531, "epoch": 0.37, "learning_rate": 2.4421813403416557e-05, "loss": 74.8277, "step": 1137 }, { "compression_loss": 75.16254425048828, "epoch": 0.37, "learning_rate": 2.441688567674113e-05, "loss": 75.6974, "step": 1138 }, { "compression_loss": 76.34503936767578, "epoch": 0.37, "learning_rate": 2.4411957950065705e-05, "loss": 76.6054, "step": 1139 }, { "compression_loss": 77.94424438476562, "epoch": 0.37, "learning_rate": 2.4407030223390277e-05, "loss": 78.3241, "step": 1140 }, { "compression_loss": 77.57725524902344, "epoch": 0.37, "learning_rate": 2.440210249671485e-05, "loss": 78.3282, "step": 1141 }, { "compression_loss": 75.81564331054688, "epoch": 0.38, "learning_rate": 2.439717477003942e-05, "loss": 76.5476, "step": 1142 }, { "compression_loss": 73.89456939697266, "epoch": 0.38, "learning_rate": 2.4392247043363996e-05, "loss": 74.2918, "step": 1143 }, { "compression_loss": 76.59613037109375, "epoch": 0.38, "learning_rate": 2.438731931668857e-05, "loss": 77.0234, "step": 1144 }, { "compression_loss": 74.7705078125, "epoch": 0.38, "learning_rate": 2.4382391590013143e-05, "loss": 75.3511, "step": 1145 }, { "compression_loss": 77.42678833007812, "epoch": 0.38, "learning_rate": 2.4377463863337715e-05, "loss": 77.8804, "step": 1146 }, { "compression_loss": 73.47016906738281, "epoch": 0.38, "learning_rate": 2.4372536136662287e-05, "loss": 73.8456, "step": 1147 }, { "compression_loss": 75.86508178710938, "epoch": 0.38, "learning_rate": 2.436760840998686e-05, "loss": 76.1137, "step": 1148 }, { "compression_loss": 74.69451904296875, "epoch": 0.38, "learning_rate": 2.436268068331143e-05, "loss": 75.098, "step": 1149 }, { "compression_loss": 76.98249816894531, "epoch": 0.38, "learning_rate": 2.4357752956636007e-05, "loss": 77.5151, "step": 1150 }, { "compression_loss": 72.70037841796875, "epoch": 0.38, "learning_rate": 2.435282522996058e-05, "loss": 73.1012, "step": 1151 }, { "compression_loss": 76.47576141357422, "epoch": 0.38, "learning_rate": 2.434789750328515e-05, "loss": 76.8131, "step": 1152 }, { "compression_loss": 74.72274780273438, "epoch": 0.38, "learning_rate": 2.4342969776609722e-05, "loss": 75.1911, "step": 1153 }, { "compression_loss": 74.86065673828125, "epoch": 0.38, "learning_rate": 2.4338042049934294e-05, "loss": 75.1731, "step": 1154 }, { "compression_loss": 75.4831771850586, "epoch": 0.38, "learning_rate": 2.4333114323258873e-05, "loss": 75.7506, "step": 1155 }, { "compression_loss": 76.16073608398438, "epoch": 0.38, "learning_rate": 2.4328186596583445e-05, "loss": 76.6313, "step": 1156 }, { "compression_loss": 76.1864013671875, "epoch": 0.38, "learning_rate": 2.4323258869908017e-05, "loss": 76.8339, "step": 1157 }, { "compression_loss": 77.32244873046875, "epoch": 0.38, "learning_rate": 2.431833114323259e-05, "loss": 77.8297, "step": 1158 }, { "compression_loss": 76.65152740478516, "epoch": 0.38, "learning_rate": 2.431340341655716e-05, "loss": 77.287, "step": 1159 }, { "compression_loss": 74.51172637939453, "epoch": 0.38, "learning_rate": 2.4308475689881736e-05, "loss": 74.8876, "step": 1160 }, { "compression_loss": 78.4332275390625, "epoch": 0.38, "learning_rate": 2.430354796320631e-05, "loss": 78.9755, "step": 1161 }, { "compression_loss": 75.19215393066406, "epoch": 0.38, "learning_rate": 2.429862023653088e-05, "loss": 75.5261, "step": 1162 }, { "compression_loss": 73.36493682861328, "epoch": 0.38, "learning_rate": 2.4293692509855452e-05, "loss": 73.8012, "step": 1163 }, { "compression_loss": 72.53424072265625, "epoch": 0.38, "learning_rate": 2.4288764783180024e-05, "loss": 73.0448, "step": 1164 }, { "compression_loss": 76.75213623046875, "epoch": 0.38, "learning_rate": 2.42838370565046e-05, "loss": 77.0759, "step": 1165 }, { "compression_loss": 76.40066528320312, "epoch": 0.38, "learning_rate": 2.427890932982917e-05, "loss": 76.8017, "step": 1166 }, { "compression_loss": 74.91970825195312, "epoch": 0.38, "learning_rate": 2.4273981603153747e-05, "loss": 75.203, "step": 1167 }, { "compression_loss": 77.69103240966797, "epoch": 0.38, "learning_rate": 2.426905387647832e-05, "loss": 78.2987, "step": 1168 }, { "compression_loss": 74.49613952636719, "epoch": 0.38, "learning_rate": 2.426412614980289e-05, "loss": 74.9049, "step": 1169 }, { "compression_loss": 77.94270324707031, "epoch": 0.38, "learning_rate": 2.4259198423127466e-05, "loss": 78.9952, "step": 1170 }, { "compression_loss": 75.01504516601562, "epoch": 0.38, "learning_rate": 2.4254270696452038e-05, "loss": 75.319, "step": 1171 }, { "compression_loss": 75.66946411132812, "epoch": 0.39, "learning_rate": 2.424934296977661e-05, "loss": 76.0121, "step": 1172 }, { "compression_loss": 76.56758880615234, "epoch": 0.39, "learning_rate": 2.4244415243101182e-05, "loss": 76.797, "step": 1173 }, { "compression_loss": 74.17927551269531, "epoch": 0.39, "learning_rate": 2.4239487516425754e-05, "loss": 74.4725, "step": 1174 }, { "compression_loss": 75.50582122802734, "epoch": 0.39, "learning_rate": 2.423455978975033e-05, "loss": 76.0843, "step": 1175 }, { "compression_loss": 76.16423034667969, "epoch": 0.39, "learning_rate": 2.42296320630749e-05, "loss": 76.5425, "step": 1176 }, { "compression_loss": 75.82735443115234, "epoch": 0.39, "learning_rate": 2.4224704336399473e-05, "loss": 76.402, "step": 1177 }, { "compression_loss": 74.04269409179688, "epoch": 0.39, "learning_rate": 2.4219776609724045e-05, "loss": 74.2536, "step": 1178 }, { "compression_loss": 75.2103271484375, "epoch": 0.39, "learning_rate": 2.421484888304862e-05, "loss": 75.4811, "step": 1179 }, { "compression_loss": 76.266357421875, "epoch": 0.39, "learning_rate": 2.4209921156373196e-05, "loss": 76.9358, "step": 1180 }, { "compression_loss": 72.88932800292969, "epoch": 0.39, "learning_rate": 2.4204993429697768e-05, "loss": 73.2592, "step": 1181 }, { "compression_loss": 77.47557830810547, "epoch": 0.39, "learning_rate": 2.420006570302234e-05, "loss": 78.1604, "step": 1182 }, { "compression_loss": 75.4631576538086, "epoch": 0.39, "learning_rate": 2.4195137976346912e-05, "loss": 75.8593, "step": 1183 }, { "compression_loss": 78.26911926269531, "epoch": 0.39, "learning_rate": 2.4190210249671484e-05, "loss": 78.8352, "step": 1184 }, { "compression_loss": 76.77720642089844, "epoch": 0.39, "learning_rate": 2.418528252299606e-05, "loss": 77.292, "step": 1185 }, { "compression_loss": 75.84379577636719, "epoch": 0.39, "learning_rate": 2.418035479632063e-05, "loss": 76.2537, "step": 1186 }, { "compression_loss": 75.53789520263672, "epoch": 0.39, "learning_rate": 2.4175427069645203e-05, "loss": 76.0709, "step": 1187 }, { "compression_loss": 77.20264434814453, "epoch": 0.39, "learning_rate": 2.4170499342969775e-05, "loss": 77.5294, "step": 1188 }, { "compression_loss": 76.88236999511719, "epoch": 0.39, "learning_rate": 2.4165571616294347e-05, "loss": 77.4466, "step": 1189 }, { "compression_loss": 76.89401245117188, "epoch": 0.39, "learning_rate": 2.4160643889618923e-05, "loss": 77.454, "step": 1190 }, { "compression_loss": 73.3030014038086, "epoch": 0.39, "learning_rate": 2.4155716162943498e-05, "loss": 73.6328, "step": 1191 }, { "compression_loss": 73.94499969482422, "epoch": 0.39, "learning_rate": 2.415078843626807e-05, "loss": 74.1955, "step": 1192 }, { "compression_loss": 74.35099792480469, "epoch": 0.39, "learning_rate": 2.4145860709592642e-05, "loss": 74.6652, "step": 1193 }, { "compression_loss": 74.83348083496094, "epoch": 0.39, "learning_rate": 2.4140932982917214e-05, "loss": 75.1486, "step": 1194 }, { "compression_loss": 78.42935180664062, "epoch": 0.39, "learning_rate": 2.413600525624179e-05, "loss": 79.0425, "step": 1195 }, { "compression_loss": 77.27269744873047, "epoch": 0.39, "learning_rate": 2.413107752956636e-05, "loss": 77.8844, "step": 1196 }, { "compression_loss": 74.28893280029297, "epoch": 0.39, "learning_rate": 2.4126149802890933e-05, "loss": 74.5658, "step": 1197 }, { "compression_loss": 76.0897216796875, "epoch": 0.39, "learning_rate": 2.4121222076215505e-05, "loss": 76.374, "step": 1198 }, { "compression_loss": 74.758056640625, "epoch": 0.39, "learning_rate": 2.4116294349540077e-05, "loss": 75.1649, "step": 1199 }, { "compression_loss": 76.7091064453125, "epoch": 0.39, "learning_rate": 2.4111366622864652e-05, "loss": 77.3055, "step": 1200 }, { "compression_loss": 77.04266357421875, "epoch": 0.39, "learning_rate": 2.4106438896189224e-05, "loss": 77.5493, "step": 1201 }, { "compression_loss": 76.14369201660156, "epoch": 0.39, "learning_rate": 2.4101511169513796e-05, "loss": 76.8208, "step": 1202 }, { "compression_loss": 76.1120834350586, "epoch": 0.4, "learning_rate": 2.4096583442838372e-05, "loss": 76.4231, "step": 1203 }, { "compression_loss": 73.53610229492188, "epoch": 0.4, "learning_rate": 2.4091655716162944e-05, "loss": 73.9409, "step": 1204 }, { "compression_loss": 76.41244506835938, "epoch": 0.4, "learning_rate": 2.408672798948752e-05, "loss": 76.9762, "step": 1205 }, { "compression_loss": 78.54864501953125, "epoch": 0.4, "learning_rate": 2.408180026281209e-05, "loss": 79.1126, "step": 1206 }, { "compression_loss": 75.00270080566406, "epoch": 0.4, "learning_rate": 2.4076872536136663e-05, "loss": 75.4678, "step": 1207 }, { "compression_loss": 76.89013671875, "epoch": 0.4, "learning_rate": 2.4071944809461235e-05, "loss": 77.643, "step": 1208 }, { "compression_loss": 77.06653594970703, "epoch": 0.4, "learning_rate": 2.4067017082785807e-05, "loss": 77.341, "step": 1209 }, { "compression_loss": 76.52151489257812, "epoch": 0.4, "learning_rate": 2.4062089356110382e-05, "loss": 77.1035, "step": 1210 }, { "compression_loss": 75.70074462890625, "epoch": 0.4, "learning_rate": 2.4057161629434954e-05, "loss": 76.0138, "step": 1211 }, { "compression_loss": 74.6846694946289, "epoch": 0.4, "learning_rate": 2.4052233902759526e-05, "loss": 74.883, "step": 1212 }, { "compression_loss": 76.29322814941406, "epoch": 0.4, "learning_rate": 2.4047306176084098e-05, "loss": 76.5657, "step": 1213 }, { "compression_loss": 75.11331176757812, "epoch": 0.4, "learning_rate": 2.404237844940867e-05, "loss": 75.551, "step": 1214 }, { "compression_loss": 77.04435729980469, "epoch": 0.4, "learning_rate": 2.403745072273325e-05, "loss": 77.341, "step": 1215 }, { "compression_loss": 75.02818298339844, "epoch": 0.4, "learning_rate": 2.403252299605782e-05, "loss": 75.2942, "step": 1216 }, { "compression_loss": 76.74701690673828, "epoch": 0.4, "learning_rate": 2.4027595269382393e-05, "loss": 77.2734, "step": 1217 }, { "compression_loss": 74.80377960205078, "epoch": 0.4, "learning_rate": 2.4022667542706965e-05, "loss": 75.3478, "step": 1218 }, { "compression_loss": 75.16809844970703, "epoch": 0.4, "learning_rate": 2.4017739816031537e-05, "loss": 75.5972, "step": 1219 }, { "compression_loss": 76.84201049804688, "epoch": 0.4, "learning_rate": 2.4012812089356112e-05, "loss": 77.429, "step": 1220 }, { "compression_loss": 74.80540466308594, "epoch": 0.4, "learning_rate": 2.4007884362680684e-05, "loss": 75.0963, "step": 1221 }, { "compression_loss": 75.812744140625, "epoch": 0.4, "learning_rate": 2.4002956636005256e-05, "loss": 76.2082, "step": 1222 }, { "compression_loss": 79.56322479248047, "epoch": 0.4, "learning_rate": 2.3998028909329828e-05, "loss": 80.1451, "step": 1223 }, { "compression_loss": 76.9559326171875, "epoch": 0.4, "learning_rate": 2.39931011826544e-05, "loss": 77.1466, "step": 1224 }, { "compression_loss": 74.9810562133789, "epoch": 0.4, "learning_rate": 2.3988173455978975e-05, "loss": 75.283, "step": 1225 }, { "compression_loss": 78.3648452758789, "epoch": 0.4, "learning_rate": 2.3983245729303547e-05, "loss": 78.9645, "step": 1226 }, { "compression_loss": 76.28429412841797, "epoch": 0.4, "learning_rate": 2.3978318002628123e-05, "loss": 76.8056, "step": 1227 }, { "compression_loss": 74.50501251220703, "epoch": 0.4, "learning_rate": 2.3973390275952695e-05, "loss": 74.8923, "step": 1228 }, { "compression_loss": 77.41087341308594, "epoch": 0.4, "learning_rate": 2.3968462549277267e-05, "loss": 77.7845, "step": 1229 }, { "compression_loss": 74.18453979492188, "epoch": 0.4, "learning_rate": 2.3963534822601842e-05, "loss": 74.7928, "step": 1230 }, { "compression_loss": 76.086669921875, "epoch": 0.4, "learning_rate": 2.3958607095926414e-05, "loss": 76.4242, "step": 1231 }, { "compression_loss": 74.80476379394531, "epoch": 0.4, "learning_rate": 2.3953679369250986e-05, "loss": 75.5628, "step": 1232 }, { "compression_loss": 74.28951263427734, "epoch": 0.41, "learning_rate": 2.3948751642575558e-05, "loss": 74.4978, "step": 1233 }, { "compression_loss": 75.153076171875, "epoch": 0.41, "learning_rate": 2.394382391590013e-05, "loss": 75.7019, "step": 1234 }, { "compression_loss": 74.73834228515625, "epoch": 0.41, "learning_rate": 2.3938896189224705e-05, "loss": 75.1843, "step": 1235 }, { "compression_loss": 75.98336029052734, "epoch": 0.41, "learning_rate": 2.3933968462549277e-05, "loss": 76.3961, "step": 1236 }, { "compression_loss": 77.09712219238281, "epoch": 0.41, "learning_rate": 2.392904073587385e-05, "loss": 77.5099, "step": 1237 }, { "compression_loss": 75.50064086914062, "epoch": 0.41, "learning_rate": 2.392411300919842e-05, "loss": 75.8754, "step": 1238 }, { "compression_loss": 75.62761688232422, "epoch": 0.41, "learning_rate": 2.3919185282522996e-05, "loss": 76.2636, "step": 1239 }, { "compression_loss": 75.64777374267578, "epoch": 0.41, "learning_rate": 2.3914257555847572e-05, "loss": 75.8758, "step": 1240 }, { "compression_loss": 72.50276947021484, "epoch": 0.41, "learning_rate": 2.3909329829172144e-05, "loss": 72.81, "step": 1241 }, { "compression_loss": 75.1602783203125, "epoch": 0.41, "learning_rate": 2.3904402102496716e-05, "loss": 75.5441, "step": 1242 }, { "compression_loss": 73.00448608398438, "epoch": 0.41, "learning_rate": 2.3899474375821288e-05, "loss": 73.3304, "step": 1243 }, { "compression_loss": 77.87008666992188, "epoch": 0.41, "learning_rate": 2.389454664914586e-05, "loss": 78.2226, "step": 1244 }, { "compression_loss": 75.91911315917969, "epoch": 0.41, "learning_rate": 2.3889618922470435e-05, "loss": 76.4513, "step": 1245 }, { "compression_loss": 76.56834411621094, "epoch": 0.41, "learning_rate": 2.3884691195795007e-05, "loss": 77.3405, "step": 1246 }, { "compression_loss": 74.89778137207031, "epoch": 0.41, "learning_rate": 2.387976346911958e-05, "loss": 75.1904, "step": 1247 }, { "compression_loss": 76.00299072265625, "epoch": 0.41, "learning_rate": 2.387483574244415e-05, "loss": 76.2809, "step": 1248 }, { "compression_loss": 77.26376342773438, "epoch": 0.41, "learning_rate": 2.3869908015768723e-05, "loss": 77.7472, "step": 1249 }, { "compression_loss": 75.65650939941406, "epoch": 0.41, "learning_rate": 2.38649802890933e-05, "loss": 76.0816, "step": 1250 }, { "epoch": 0.41, "eval_exact_match": 85.82781456953643, "eval_f1": 92.26629148036162, "step": 1250 }, { "compression_loss": 75.32132720947266, "epoch": 0.41, "learning_rate": 2.3860052562417874e-05, "loss": 75.5163, "step": 1251 }, { "compression_loss": 77.01944732666016, "epoch": 0.41, "learning_rate": 2.3855124835742446e-05, "loss": 77.382, "step": 1252 }, { "compression_loss": 77.05227661132812, "epoch": 0.41, "learning_rate": 2.3850197109067018e-05, "loss": 77.5677, "step": 1253 }, { "compression_loss": 76.5273666381836, "epoch": 0.41, "learning_rate": 2.384526938239159e-05, "loss": 76.8273, "step": 1254 }, { "compression_loss": 76.25688171386719, "epoch": 0.41, "learning_rate": 2.3840341655716165e-05, "loss": 76.5369, "step": 1255 }, { "compression_loss": 76.09782409667969, "epoch": 0.41, "learning_rate": 2.3835413929040737e-05, "loss": 76.3339, "step": 1256 }, { "compression_loss": 76.60601043701172, "epoch": 0.41, "learning_rate": 2.383048620236531e-05, "loss": 76.9459, "step": 1257 }, { "compression_loss": 76.10047149658203, "epoch": 0.41, "learning_rate": 2.382555847568988e-05, "loss": 76.6452, "step": 1258 }, { "compression_loss": 77.30000305175781, "epoch": 0.41, "learning_rate": 2.3820630749014453e-05, "loss": 77.6418, "step": 1259 }, { "compression_loss": 75.86135864257812, "epoch": 0.41, "learning_rate": 2.3815703022339028e-05, "loss": 76.1913, "step": 1260 }, { "compression_loss": 75.6583251953125, "epoch": 0.41, "learning_rate": 2.38107752956636e-05, "loss": 76.1284, "step": 1261 }, { "compression_loss": 73.19676208496094, "epoch": 0.41, "learning_rate": 2.3805847568988175e-05, "loss": 73.6416, "step": 1262 }, { "compression_loss": 75.7288818359375, "epoch": 0.41, "learning_rate": 2.3800919842312747e-05, "loss": 76.1197, "step": 1263 }, { "compression_loss": 75.73090362548828, "epoch": 0.42, "learning_rate": 2.379599211563732e-05, "loss": 76.1082, "step": 1264 }, { "compression_loss": 75.27015686035156, "epoch": 0.42, "learning_rate": 2.3791064388961895e-05, "loss": 75.6419, "step": 1265 }, { "compression_loss": 74.04745483398438, "epoch": 0.42, "learning_rate": 2.3786136662286467e-05, "loss": 74.4063, "step": 1266 }, { "compression_loss": 76.00578308105469, "epoch": 0.42, "learning_rate": 2.378120893561104e-05, "loss": 76.4733, "step": 1267 }, { "compression_loss": 75.29963684082031, "epoch": 0.42, "learning_rate": 2.377628120893561e-05, "loss": 75.8984, "step": 1268 }, { "compression_loss": 75.32962036132812, "epoch": 0.42, "learning_rate": 2.3771353482260183e-05, "loss": 75.8062, "step": 1269 }, { "compression_loss": 76.6368408203125, "epoch": 0.42, "learning_rate": 2.3766425755584758e-05, "loss": 76.8434, "step": 1270 }, { "compression_loss": 78.95269775390625, "epoch": 0.42, "learning_rate": 2.376149802890933e-05, "loss": 79.6534, "step": 1271 }, { "compression_loss": 74.112548828125, "epoch": 0.42, "learning_rate": 2.3756570302233902e-05, "loss": 74.4344, "step": 1272 }, { "compression_loss": 76.75665283203125, "epoch": 0.42, "learning_rate": 2.3751642575558474e-05, "loss": 77.0639, "step": 1273 }, { "compression_loss": 76.38213348388672, "epoch": 0.42, "learning_rate": 2.374671484888305e-05, "loss": 76.7483, "step": 1274 }, { "compression_loss": 76.9957275390625, "epoch": 0.42, "learning_rate": 2.3741787122207625e-05, "loss": 77.4434, "step": 1275 }, { "compression_loss": 74.6540756225586, "epoch": 0.42, "learning_rate": 2.3736859395532197e-05, "loss": 75.3582, "step": 1276 }, { "compression_loss": 76.43648529052734, "epoch": 0.42, "learning_rate": 2.373193166885677e-05, "loss": 76.6817, "step": 1277 }, { "compression_loss": 77.66632843017578, "epoch": 0.42, "learning_rate": 2.372700394218134e-05, "loss": 78.1564, "step": 1278 }, { "compression_loss": 76.04447174072266, "epoch": 0.42, "learning_rate": 2.3722076215505913e-05, "loss": 76.3457, "step": 1279 }, { "compression_loss": 77.41334533691406, "epoch": 0.42, "learning_rate": 2.3717148488830488e-05, "loss": 77.7691, "step": 1280 }, { "compression_loss": 73.49569702148438, "epoch": 0.42, "learning_rate": 2.371222076215506e-05, "loss": 73.9401, "step": 1281 }, { "compression_loss": 76.4646224975586, "epoch": 0.42, "learning_rate": 2.3707293035479632e-05, "loss": 77.015, "step": 1282 }, { "compression_loss": 75.05323791503906, "epoch": 0.42, "learning_rate": 2.3702365308804204e-05, "loss": 75.3181, "step": 1283 }, { "compression_loss": 76.91644287109375, "epoch": 0.42, "learning_rate": 2.3697437582128776e-05, "loss": 77.2922, "step": 1284 }, { "compression_loss": 77.74457550048828, "epoch": 0.42, "learning_rate": 2.369250985545335e-05, "loss": 78.1786, "step": 1285 }, { "compression_loss": 77.99436950683594, "epoch": 0.42, "learning_rate": 2.3687582128777926e-05, "loss": 78.3643, "step": 1286 }, { "compression_loss": 76.45126342773438, "epoch": 0.42, "learning_rate": 2.36826544021025e-05, "loss": 77.7492, "step": 1287 }, { "compression_loss": 76.57939147949219, "epoch": 0.42, "learning_rate": 2.367772667542707e-05, "loss": 77.1381, "step": 1288 }, { "compression_loss": 75.45577239990234, "epoch": 0.42, "learning_rate": 2.3672798948751642e-05, "loss": 75.9347, "step": 1289 }, { "compression_loss": 77.62718200683594, "epoch": 0.42, "learning_rate": 2.3667871222076218e-05, "loss": 78.0368, "step": 1290 }, { "compression_loss": 79.24102783203125, "epoch": 0.42, "learning_rate": 2.366294349540079e-05, "loss": 79.5884, "step": 1291 }, { "compression_loss": 74.67669677734375, "epoch": 0.42, "learning_rate": 2.365801576872536e-05, "loss": 75.0484, "step": 1292 }, { "compression_loss": 74.87474822998047, "epoch": 0.42, "learning_rate": 2.3653088042049934e-05, "loss": 75.1643, "step": 1293 }, { "compression_loss": 73.69377899169922, "epoch": 0.43, "learning_rate": 2.3648160315374506e-05, "loss": 74.0548, "step": 1294 }, { "compression_loss": 77.82450866699219, "epoch": 0.43, "learning_rate": 2.364323258869908e-05, "loss": 78.2147, "step": 1295 }, { "compression_loss": 77.78436279296875, "epoch": 0.43, "learning_rate": 2.3638304862023653e-05, "loss": 77.967, "step": 1296 }, { "compression_loss": 78.30690002441406, "epoch": 0.43, "learning_rate": 2.3633377135348225e-05, "loss": 78.6561, "step": 1297 }, { "compression_loss": 77.30936431884766, "epoch": 0.43, "learning_rate": 2.36284494086728e-05, "loss": 77.9821, "step": 1298 }, { "compression_loss": 75.92493438720703, "epoch": 0.43, "learning_rate": 2.3623521681997372e-05, "loss": 76.2134, "step": 1299 }, { "compression_loss": 74.38639831542969, "epoch": 0.43, "learning_rate": 2.3618593955321948e-05, "loss": 74.5769, "step": 1300 }, { "compression_loss": 77.52976989746094, "epoch": 0.43, "learning_rate": 2.361366622864652e-05, "loss": 77.9667, "step": 1301 }, { "compression_loss": 73.34951782226562, "epoch": 0.43, "learning_rate": 2.360873850197109e-05, "loss": 73.504, "step": 1302 }, { "compression_loss": 73.35903930664062, "epoch": 0.43, "learning_rate": 2.3603810775295663e-05, "loss": 73.9134, "step": 1303 }, { "compression_loss": 75.9433822631836, "epoch": 0.43, "learning_rate": 2.3598883048620235e-05, "loss": 76.3147, "step": 1304 }, { "compression_loss": 74.28567504882812, "epoch": 0.43, "learning_rate": 2.359395532194481e-05, "loss": 74.6251, "step": 1305 }, { "compression_loss": 74.20661926269531, "epoch": 0.43, "learning_rate": 2.3589027595269383e-05, "loss": 74.472, "step": 1306 }, { "compression_loss": 74.87081146240234, "epoch": 0.43, "learning_rate": 2.3584099868593955e-05, "loss": 75.6328, "step": 1307 }, { "compression_loss": 76.27727508544922, "epoch": 0.43, "learning_rate": 2.3579172141918527e-05, "loss": 76.6253, "step": 1308 }, { "compression_loss": 76.27743530273438, "epoch": 0.43, "learning_rate": 2.35742444152431e-05, "loss": 76.6186, "step": 1309 }, { "compression_loss": 75.53021240234375, "epoch": 0.43, "learning_rate": 2.3569316688567677e-05, "loss": 76.0703, "step": 1310 }, { "compression_loss": 77.127685546875, "epoch": 0.43, "learning_rate": 2.356438896189225e-05, "loss": 77.3851, "step": 1311 }, { "compression_loss": 74.14379119873047, "epoch": 0.43, "learning_rate": 2.355946123521682e-05, "loss": 74.546, "step": 1312 }, { "compression_loss": 75.94710540771484, "epoch": 0.43, "learning_rate": 2.3554533508541393e-05, "loss": 76.4329, "step": 1313 }, { "compression_loss": 75.07174682617188, "epoch": 0.43, "learning_rate": 2.3549605781865965e-05, "loss": 75.6049, "step": 1314 }, { "compression_loss": 75.19033813476562, "epoch": 0.43, "learning_rate": 2.354467805519054e-05, "loss": 75.6103, "step": 1315 }, { "compression_loss": 76.44731140136719, "epoch": 0.43, "learning_rate": 2.3539750328515113e-05, "loss": 77.0036, "step": 1316 }, { "compression_loss": 76.8355712890625, "epoch": 0.43, "learning_rate": 2.3534822601839685e-05, "loss": 77.1802, "step": 1317 }, { "compression_loss": 75.3316879272461, "epoch": 0.43, "learning_rate": 2.3529894875164257e-05, "loss": 75.5996, "step": 1318 }, { "compression_loss": 75.71928405761719, "epoch": 0.43, "learning_rate": 2.352496714848883e-05, "loss": 76.1653, "step": 1319 }, { "compression_loss": 77.56690979003906, "epoch": 0.43, "learning_rate": 2.3520039421813404e-05, "loss": 77.9838, "step": 1320 }, { "compression_loss": 74.21279907226562, "epoch": 0.43, "learning_rate": 2.3515111695137976e-05, "loss": 74.5116, "step": 1321 }, { "compression_loss": 77.97538757324219, "epoch": 0.43, "learning_rate": 2.351018396846255e-05, "loss": 78.5278, "step": 1322 }, { "compression_loss": 77.07666015625, "epoch": 0.43, "learning_rate": 2.3505256241787123e-05, "loss": 77.3791, "step": 1323 }, { "compression_loss": 73.07850646972656, "epoch": 0.43, "learning_rate": 2.3500328515111695e-05, "loss": 73.3714, "step": 1324 }, { "compression_loss": 75.67936706542969, "epoch": 0.44, "learning_rate": 2.349540078843627e-05, "loss": 75.9908, "step": 1325 }, { "compression_loss": 77.90711212158203, "epoch": 0.44, "learning_rate": 2.3490473061760843e-05, "loss": 78.8134, "step": 1326 }, { "compression_loss": 73.19023895263672, "epoch": 0.44, "learning_rate": 2.3485545335085414e-05, "loss": 73.4232, "step": 1327 }, { "compression_loss": 75.9398193359375, "epoch": 0.44, "learning_rate": 2.3480617608409986e-05, "loss": 76.6071, "step": 1328 }, { "compression_loss": 73.97578430175781, "epoch": 0.44, "learning_rate": 2.347568988173456e-05, "loss": 74.3054, "step": 1329 }, { "compression_loss": 76.2353515625, "epoch": 0.44, "learning_rate": 2.3470762155059134e-05, "loss": 76.5461, "step": 1330 }, { "compression_loss": 75.68681335449219, "epoch": 0.44, "learning_rate": 2.3465834428383706e-05, "loss": 76.0737, "step": 1331 }, { "compression_loss": 75.99516296386719, "epoch": 0.44, "learning_rate": 2.3460906701708278e-05, "loss": 76.2915, "step": 1332 }, { "compression_loss": 74.1313247680664, "epoch": 0.44, "learning_rate": 2.345597897503285e-05, "loss": 74.3209, "step": 1333 }, { "compression_loss": 75.03697967529297, "epoch": 0.44, "learning_rate": 2.3451051248357425e-05, "loss": 75.589, "step": 1334 }, { "compression_loss": 73.40275573730469, "epoch": 0.44, "learning_rate": 2.3446123521682e-05, "loss": 73.7279, "step": 1335 }, { "compression_loss": 75.1105728149414, "epoch": 0.44, "learning_rate": 2.3441195795006572e-05, "loss": 75.4013, "step": 1336 }, { "compression_loss": 77.11412048339844, "epoch": 0.44, "learning_rate": 2.3436268068331144e-05, "loss": 77.9855, "step": 1337 }, { "compression_loss": 75.1546859741211, "epoch": 0.44, "learning_rate": 2.3431340341655716e-05, "loss": 75.6573, "step": 1338 }, { "compression_loss": 75.70197296142578, "epoch": 0.44, "learning_rate": 2.3426412614980288e-05, "loss": 76.1339, "step": 1339 }, { "compression_loss": 73.16950988769531, "epoch": 0.44, "learning_rate": 2.3421484888304864e-05, "loss": 73.4543, "step": 1340 }, { "compression_loss": 72.71024322509766, "epoch": 0.44, "learning_rate": 2.3416557161629436e-05, "loss": 73.1069, "step": 1341 }, { "compression_loss": 74.82096862792969, "epoch": 0.44, "learning_rate": 2.3411629434954008e-05, "loss": 75.1379, "step": 1342 }, { "compression_loss": 76.00465393066406, "epoch": 0.44, "learning_rate": 2.340670170827858e-05, "loss": 76.6195, "step": 1343 }, { "compression_loss": 74.39222717285156, "epoch": 0.44, "learning_rate": 2.340177398160315e-05, "loss": 74.7864, "step": 1344 }, { "compression_loss": 74.6722412109375, "epoch": 0.44, "learning_rate": 2.3396846254927727e-05, "loss": 74.9163, "step": 1345 }, { "compression_loss": 76.03089904785156, "epoch": 0.44, "learning_rate": 2.3391918528252302e-05, "loss": 76.2043, "step": 1346 }, { "compression_loss": 78.78286743164062, "epoch": 0.44, "learning_rate": 2.3386990801576874e-05, "loss": 79.2323, "step": 1347 }, { "compression_loss": 75.25013732910156, "epoch": 0.44, "learning_rate": 2.3382063074901446e-05, "loss": 75.5456, "step": 1348 }, { "compression_loss": 77.21611022949219, "epoch": 0.44, "learning_rate": 2.3377135348226018e-05, "loss": 77.6567, "step": 1349 }, { "compression_loss": 76.79515838623047, "epoch": 0.44, "learning_rate": 2.3372207621550593e-05, "loss": 77.0487, "step": 1350 }, { "compression_loss": 74.76878356933594, "epoch": 0.44, "learning_rate": 2.3367279894875165e-05, "loss": 75.3196, "step": 1351 }, { "compression_loss": 74.00032806396484, "epoch": 0.44, "learning_rate": 2.3362352168199737e-05, "loss": 74.3192, "step": 1352 }, { "compression_loss": 75.28224182128906, "epoch": 0.44, "learning_rate": 2.335742444152431e-05, "loss": 75.7978, "step": 1353 }, { "compression_loss": 76.8321533203125, "epoch": 0.44, "learning_rate": 2.335249671484888e-05, "loss": 77.2195, "step": 1354 }, { "compression_loss": 75.74684143066406, "epoch": 0.45, "learning_rate": 2.3347568988173457e-05, "loss": 75.9517, "step": 1355 }, { "compression_loss": 76.79127502441406, "epoch": 0.45, "learning_rate": 2.334264126149803e-05, "loss": 77.3393, "step": 1356 }, { "compression_loss": 75.05281829833984, "epoch": 0.45, "learning_rate": 2.33377135348226e-05, "loss": 75.3805, "step": 1357 }, { "compression_loss": 75.39938354492188, "epoch": 0.45, "learning_rate": 2.3332785808147176e-05, "loss": 75.6819, "step": 1358 }, { "compression_loss": 77.38658142089844, "epoch": 0.45, "learning_rate": 2.3327858081471748e-05, "loss": 77.7417, "step": 1359 }, { "compression_loss": 74.43121337890625, "epoch": 0.45, "learning_rate": 2.3322930354796323e-05, "loss": 74.6968, "step": 1360 }, { "compression_loss": 76.27394104003906, "epoch": 0.45, "learning_rate": 2.3318002628120895e-05, "loss": 76.9247, "step": 1361 }, { "compression_loss": 76.2271728515625, "epoch": 0.45, "learning_rate": 2.3313074901445467e-05, "loss": 76.7093, "step": 1362 }, { "compression_loss": 75.58567810058594, "epoch": 0.45, "learning_rate": 2.330814717477004e-05, "loss": 76.0798, "step": 1363 }, { "compression_loss": 76.11273956298828, "epoch": 0.45, "learning_rate": 2.330321944809461e-05, "loss": 76.6434, "step": 1364 }, { "compression_loss": 74.27301025390625, "epoch": 0.45, "learning_rate": 2.3298291721419187e-05, "loss": 74.6253, "step": 1365 }, { "compression_loss": 74.197998046875, "epoch": 0.45, "learning_rate": 2.329336399474376e-05, "loss": 74.6188, "step": 1366 }, { "compression_loss": 77.47794342041016, "epoch": 0.45, "learning_rate": 2.328843626806833e-05, "loss": 78.0327, "step": 1367 }, { "compression_loss": 77.42997741699219, "epoch": 0.45, "learning_rate": 2.3283508541392902e-05, "loss": 78.07, "step": 1368 }, { "compression_loss": 78.35997772216797, "epoch": 0.45, "learning_rate": 2.3278580814717474e-05, "loss": 78.7374, "step": 1369 }, { "compression_loss": 76.87786865234375, "epoch": 0.45, "learning_rate": 2.3273653088042053e-05, "loss": 77.5424, "step": 1370 }, { "compression_loss": 74.93963623046875, "epoch": 0.45, "learning_rate": 2.3268725361366625e-05, "loss": 75.1753, "step": 1371 }, { "compression_loss": 74.8924560546875, "epoch": 0.45, "learning_rate": 2.3263797634691197e-05, "loss": 75.247, "step": 1372 }, { "compression_loss": 75.4969253540039, "epoch": 0.45, "learning_rate": 2.325886990801577e-05, "loss": 75.6708, "step": 1373 }, { "compression_loss": 78.32154846191406, "epoch": 0.45, "learning_rate": 2.325394218134034e-05, "loss": 78.8534, "step": 1374 }, { "compression_loss": 74.11418151855469, "epoch": 0.45, "learning_rate": 2.3249014454664916e-05, "loss": 74.5744, "step": 1375 }, { "compression_loss": 75.26136779785156, "epoch": 0.45, "learning_rate": 2.324408672798949e-05, "loss": 75.4078, "step": 1376 }, { "compression_loss": 73.47821044921875, "epoch": 0.45, "learning_rate": 2.323915900131406e-05, "loss": 73.882, "step": 1377 }, { "compression_loss": 75.7191390991211, "epoch": 0.45, "learning_rate": 2.3234231274638632e-05, "loss": 76.3658, "step": 1378 }, { "compression_loss": 74.15553283691406, "epoch": 0.45, "learning_rate": 2.3229303547963204e-05, "loss": 74.5746, "step": 1379 }, { "compression_loss": 77.67323303222656, "epoch": 0.45, "learning_rate": 2.322437582128778e-05, "loss": 77.9514, "step": 1380 }, { "compression_loss": 74.19922637939453, "epoch": 0.45, "learning_rate": 2.321944809461235e-05, "loss": 74.6996, "step": 1381 }, { "compression_loss": 75.20943450927734, "epoch": 0.45, "learning_rate": 2.3214520367936927e-05, "loss": 75.5712, "step": 1382 }, { "compression_loss": 73.03390502929688, "epoch": 0.45, "learning_rate": 2.32095926412615e-05, "loss": 73.1993, "step": 1383 }, { "compression_loss": 76.34945678710938, "epoch": 0.45, "learning_rate": 2.320466491458607e-05, "loss": 76.7027, "step": 1384 }, { "compression_loss": 75.86277770996094, "epoch": 0.45, "learning_rate": 2.3199737187910646e-05, "loss": 76.0876, "step": 1385 }, { "compression_loss": 78.45631408691406, "epoch": 0.46, "learning_rate": 2.3194809461235218e-05, "loss": 78.739, "step": 1386 }, { "compression_loss": 76.81365966796875, "epoch": 0.46, "learning_rate": 2.318988173455979e-05, "loss": 77.2167, "step": 1387 }, { "compression_loss": 76.02232360839844, "epoch": 0.46, "learning_rate": 2.3184954007884362e-05, "loss": 76.4812, "step": 1388 }, { "compression_loss": 75.63827514648438, "epoch": 0.46, "learning_rate": 2.3180026281208934e-05, "loss": 76.2256, "step": 1389 }, { "compression_loss": 76.69847869873047, "epoch": 0.46, "learning_rate": 2.317509855453351e-05, "loss": 77.0372, "step": 1390 }, { "compression_loss": 73.11077880859375, "epoch": 0.46, "learning_rate": 2.317017082785808e-05, "loss": 73.4077, "step": 1391 }, { "compression_loss": 74.92182922363281, "epoch": 0.46, "learning_rate": 2.3165243101182653e-05, "loss": 75.2759, "step": 1392 }, { "compression_loss": 77.65512084960938, "epoch": 0.46, "learning_rate": 2.316031537450723e-05, "loss": 77.9936, "step": 1393 }, { "compression_loss": 74.50338745117188, "epoch": 0.46, "learning_rate": 2.31553876478318e-05, "loss": 74.9079, "step": 1394 }, { "compression_loss": 76.5574951171875, "epoch": 0.46, "learning_rate": 2.3150459921156376e-05, "loss": 76.9001, "step": 1395 }, { "compression_loss": 76.1092758178711, "epoch": 0.46, "learning_rate": 2.3145532194480948e-05, "loss": 76.623, "step": 1396 }, { "compression_loss": 77.41769409179688, "epoch": 0.46, "learning_rate": 2.314060446780552e-05, "loss": 78.0993, "step": 1397 }, { "compression_loss": 74.54932403564453, "epoch": 0.46, "learning_rate": 2.3135676741130092e-05, "loss": 74.7874, "step": 1398 }, { "compression_loss": 77.1164321899414, "epoch": 0.46, "learning_rate": 2.3130749014454664e-05, "loss": 77.5011, "step": 1399 }, { "compression_loss": 75.63487243652344, "epoch": 0.46, "learning_rate": 2.312582128777924e-05, "loss": 75.8918, "step": 1400 }, { "compression_loss": 74.73231506347656, "epoch": 0.46, "learning_rate": 2.312089356110381e-05, "loss": 75.0596, "step": 1401 }, { "compression_loss": 72.86199951171875, "epoch": 0.46, "learning_rate": 2.3115965834428383e-05, "loss": 73.1015, "step": 1402 }, { "compression_loss": 76.06163787841797, "epoch": 0.46, "learning_rate": 2.3111038107752955e-05, "loss": 76.4265, "step": 1403 }, { "compression_loss": 75.19062805175781, "epoch": 0.46, "learning_rate": 2.3106110381077527e-05, "loss": 75.7198, "step": 1404 }, { "compression_loss": 77.83827209472656, "epoch": 0.46, "learning_rate": 2.3101182654402106e-05, "loss": 78.2146, "step": 1405 }, { "compression_loss": 74.4255142211914, "epoch": 0.46, "learning_rate": 2.3096254927726678e-05, "loss": 74.9004, "step": 1406 }, { "compression_loss": 76.06036376953125, "epoch": 0.46, "learning_rate": 2.309132720105125e-05, "loss": 76.4294, "step": 1407 }, { "compression_loss": 73.78916931152344, "epoch": 0.46, "learning_rate": 2.3086399474375822e-05, "loss": 74.0833, "step": 1408 }, { "compression_loss": 75.25413513183594, "epoch": 0.46, "learning_rate": 2.3081471747700394e-05, "loss": 75.5432, "step": 1409 }, { "compression_loss": 77.47044372558594, "epoch": 0.46, "learning_rate": 2.307654402102497e-05, "loss": 78.0847, "step": 1410 }, { "compression_loss": 76.37332153320312, "epoch": 0.46, "learning_rate": 2.307161629434954e-05, "loss": 76.8029, "step": 1411 }, { "compression_loss": 73.34280395507812, "epoch": 0.46, "learning_rate": 2.3066688567674113e-05, "loss": 73.9275, "step": 1412 }, { "compression_loss": 74.04435729980469, "epoch": 0.46, "learning_rate": 2.3061760840998685e-05, "loss": 74.4218, "step": 1413 }, { "compression_loss": 76.59390258789062, "epoch": 0.46, "learning_rate": 2.3056833114323257e-05, "loss": 77.0698, "step": 1414 }, { "compression_loss": 77.27462768554688, "epoch": 0.46, "learning_rate": 2.3051905387647832e-05, "loss": 77.6571, "step": 1415 }, { "compression_loss": 76.78672790527344, "epoch": 0.47, "learning_rate": 2.3046977660972404e-05, "loss": 77.2243, "step": 1416 }, { "compression_loss": 74.34864044189453, "epoch": 0.47, "learning_rate": 2.304204993429698e-05, "loss": 74.7589, "step": 1417 }, { "compression_loss": 76.71854400634766, "epoch": 0.47, "learning_rate": 2.3037122207621552e-05, "loss": 77.6796, "step": 1418 }, { "compression_loss": 76.77104949951172, "epoch": 0.47, "learning_rate": 2.3032194480946124e-05, "loss": 77.3183, "step": 1419 }, { "compression_loss": 78.61924743652344, "epoch": 0.47, "learning_rate": 2.30272667542707e-05, "loss": 79.1969, "step": 1420 }, { "compression_loss": 78.7808609008789, "epoch": 0.47, "learning_rate": 2.302233902759527e-05, "loss": 79.9014, "step": 1421 }, { "compression_loss": 73.82730102539062, "epoch": 0.47, "learning_rate": 2.3017411300919843e-05, "loss": 74.3933, "step": 1422 }, { "compression_loss": 76.59474182128906, "epoch": 0.47, "learning_rate": 2.3012483574244415e-05, "loss": 76.9212, "step": 1423 }, { "compression_loss": 75.03352355957031, "epoch": 0.47, "learning_rate": 2.3007555847568987e-05, "loss": 75.6118, "step": 1424 }, { "compression_loss": 74.70626831054688, "epoch": 0.47, "learning_rate": 2.3002628120893562e-05, "loss": 74.9244, "step": 1425 }, { "compression_loss": 74.75141906738281, "epoch": 0.47, "learning_rate": 2.2997700394218134e-05, "loss": 75.1487, "step": 1426 }, { "compression_loss": 72.31967163085938, "epoch": 0.47, "learning_rate": 2.2992772667542706e-05, "loss": 72.5986, "step": 1427 }, { "compression_loss": 77.75454711914062, "epoch": 0.47, "learning_rate": 2.2987844940867278e-05, "loss": 78.1598, "step": 1428 }, { "compression_loss": 74.4293212890625, "epoch": 0.47, "learning_rate": 2.2982917214191854e-05, "loss": 75.0304, "step": 1429 }, { "compression_loss": 74.5172348022461, "epoch": 0.47, "learning_rate": 2.297798948751643e-05, "loss": 74.786, "step": 1430 }, { "compression_loss": 76.03694152832031, "epoch": 0.47, "learning_rate": 2.2973061760841e-05, "loss": 76.2838, "step": 1431 }, { "compression_loss": 75.65663146972656, "epoch": 0.47, "learning_rate": 2.2968134034165573e-05, "loss": 75.8911, "step": 1432 }, { "compression_loss": 74.88063049316406, "epoch": 0.47, "learning_rate": 2.2963206307490145e-05, "loss": 75.11, "step": 1433 }, { "compression_loss": 74.00909423828125, "epoch": 0.47, "learning_rate": 2.2958278580814717e-05, "loss": 74.3174, "step": 1434 }, { "compression_loss": 75.52208709716797, "epoch": 0.47, "learning_rate": 2.2953350854139292e-05, "loss": 75.9888, "step": 1435 }, { "compression_loss": 76.11994934082031, "epoch": 0.47, "learning_rate": 2.2948423127463864e-05, "loss": 76.9097, "step": 1436 }, { "compression_loss": 77.14031982421875, "epoch": 0.47, "learning_rate": 2.2943495400788436e-05, "loss": 77.4874, "step": 1437 }, { "compression_loss": 75.84729766845703, "epoch": 0.47, "learning_rate": 2.2938567674113008e-05, "loss": 76.5746, "step": 1438 }, { "compression_loss": 74.16987609863281, "epoch": 0.47, "learning_rate": 2.293363994743758e-05, "loss": 74.5921, "step": 1439 }, { "compression_loss": 76.22505187988281, "epoch": 0.47, "learning_rate": 2.2928712220762155e-05, "loss": 76.6267, "step": 1440 }, { "compression_loss": 77.10429382324219, "epoch": 0.47, "learning_rate": 2.292378449408673e-05, "loss": 77.5776, "step": 1441 }, { "compression_loss": 75.8865737915039, "epoch": 0.47, "learning_rate": 2.2918856767411303e-05, "loss": 76.3482, "step": 1442 }, { "compression_loss": 73.1197509765625, "epoch": 0.47, "learning_rate": 2.2913929040735875e-05, "loss": 73.5364, "step": 1443 }, { "compression_loss": 73.91925811767578, "epoch": 0.47, "learning_rate": 2.2909001314060447e-05, "loss": 74.187, "step": 1444 }, { "compression_loss": 73.55792999267578, "epoch": 0.47, "learning_rate": 2.2904073587385022e-05, "loss": 73.8044, "step": 1445 }, { "compression_loss": 74.25358581542969, "epoch": 0.48, "learning_rate": 2.2899145860709594e-05, "loss": 74.6999, "step": 1446 }, { "compression_loss": 76.01113891601562, "epoch": 0.48, "learning_rate": 2.2894218134034166e-05, "loss": 76.511, "step": 1447 }, { "compression_loss": 75.73887634277344, "epoch": 0.48, "learning_rate": 2.2889290407358738e-05, "loss": 76.0412, "step": 1448 }, { "compression_loss": 74.99154663085938, "epoch": 0.48, "learning_rate": 2.288436268068331e-05, "loss": 75.3042, "step": 1449 }, { "compression_loss": 75.73924255371094, "epoch": 0.48, "learning_rate": 2.2879434954007885e-05, "loss": 76.1024, "step": 1450 }, { "compression_loss": 73.35899353027344, "epoch": 0.48, "learning_rate": 2.2874507227332457e-05, "loss": 73.5574, "step": 1451 }, { "compression_loss": 77.49580383300781, "epoch": 0.48, "learning_rate": 2.286957950065703e-05, "loss": 77.8263, "step": 1452 }, { "compression_loss": 75.36337280273438, "epoch": 0.48, "learning_rate": 2.2864651773981605e-05, "loss": 75.9082, "step": 1453 }, { "compression_loss": 75.62959289550781, "epoch": 0.48, "learning_rate": 2.2859724047306177e-05, "loss": 76.473, "step": 1454 }, { "compression_loss": 74.9559326171875, "epoch": 0.48, "learning_rate": 2.2854796320630752e-05, "loss": 75.7459, "step": 1455 }, { "compression_loss": 78.20927429199219, "epoch": 0.48, "learning_rate": 2.2849868593955324e-05, "loss": 78.9737, "step": 1456 }, { "compression_loss": 76.07830810546875, "epoch": 0.48, "learning_rate": 2.2844940867279896e-05, "loss": 76.3755, "step": 1457 }, { "compression_loss": 74.61356353759766, "epoch": 0.48, "learning_rate": 2.2840013140604468e-05, "loss": 74.8448, "step": 1458 }, { "compression_loss": 74.03121948242188, "epoch": 0.48, "learning_rate": 2.283508541392904e-05, "loss": 74.2364, "step": 1459 }, { "compression_loss": 76.7228775024414, "epoch": 0.48, "learning_rate": 2.2830157687253615e-05, "loss": 77.1581, "step": 1460 }, { "compression_loss": 77.21524047851562, "epoch": 0.48, "learning_rate": 2.2825229960578187e-05, "loss": 77.9769, "step": 1461 }, { "compression_loss": 75.84562683105469, "epoch": 0.48, "learning_rate": 2.282030223390276e-05, "loss": 76.2366, "step": 1462 }, { "compression_loss": 74.13874816894531, "epoch": 0.48, "learning_rate": 2.281537450722733e-05, "loss": 74.61, "step": 1463 }, { "compression_loss": 74.76532745361328, "epoch": 0.48, "learning_rate": 2.2810446780551903e-05, "loss": 75.1103, "step": 1464 }, { "compression_loss": 73.4688491821289, "epoch": 0.48, "learning_rate": 2.2805519053876482e-05, "loss": 73.6194, "step": 1465 }, { "compression_loss": 75.50507354736328, "epoch": 0.48, "learning_rate": 2.2800591327201054e-05, "loss": 75.9461, "step": 1466 }, { "compression_loss": 74.59496307373047, "epoch": 0.48, "learning_rate": 2.2795663600525626e-05, "loss": 74.9784, "step": 1467 }, { "compression_loss": 77.25350952148438, "epoch": 0.48, "learning_rate": 2.2790735873850198e-05, "loss": 77.9319, "step": 1468 }, { "compression_loss": 77.61578369140625, "epoch": 0.48, "learning_rate": 2.278580814717477e-05, "loss": 77.9186, "step": 1469 }, { "compression_loss": 77.1017837524414, "epoch": 0.48, "learning_rate": 2.2780880420499345e-05, "loss": 77.276, "step": 1470 }, { "compression_loss": 74.91259765625, "epoch": 0.48, "learning_rate": 2.2775952693823917e-05, "loss": 75.4191, "step": 1471 }, { "compression_loss": 76.0722427368164, "epoch": 0.48, "learning_rate": 2.277102496714849e-05, "loss": 76.4247, "step": 1472 }, { "compression_loss": 75.62474060058594, "epoch": 0.48, "learning_rate": 2.276609724047306e-05, "loss": 76.0971, "step": 1473 }, { "compression_loss": 74.66835021972656, "epoch": 0.48, "learning_rate": 2.2761169513797633e-05, "loss": 75.1031, "step": 1474 }, { "compression_loss": 75.64799499511719, "epoch": 0.48, "learning_rate": 2.2756241787122208e-05, "loss": 75.9856, "step": 1475 }, { "compression_loss": 78.06112670898438, "epoch": 0.48, "learning_rate": 2.275131406044678e-05, "loss": 78.9643, "step": 1476 }, { "compression_loss": 75.90830993652344, "epoch": 0.49, "learning_rate": 2.2746386333771356e-05, "loss": 76.1708, "step": 1477 }, { "compression_loss": 73.92487335205078, "epoch": 0.49, "learning_rate": 2.2741458607095927e-05, "loss": 74.1698, "step": 1478 }, { "compression_loss": 74.90689086914062, "epoch": 0.49, "learning_rate": 2.27365308804205e-05, "loss": 75.1694, "step": 1479 }, { "compression_loss": 75.18128204345703, "epoch": 0.49, "learning_rate": 2.2731603153745075e-05, "loss": 75.7151, "step": 1480 }, { "compression_loss": 75.29216766357422, "epoch": 0.49, "learning_rate": 2.2726675427069647e-05, "loss": 75.7246, "step": 1481 }, { "compression_loss": 75.42068481445312, "epoch": 0.49, "learning_rate": 2.272174770039422e-05, "loss": 75.5963, "step": 1482 }, { "compression_loss": 75.69802856445312, "epoch": 0.49, "learning_rate": 2.271681997371879e-05, "loss": 75.9426, "step": 1483 }, { "compression_loss": 75.79456329345703, "epoch": 0.49, "learning_rate": 2.2711892247043363e-05, "loss": 76.3929, "step": 1484 }, { "compression_loss": 76.33423614501953, "epoch": 0.49, "learning_rate": 2.2706964520367938e-05, "loss": 76.6647, "step": 1485 }, { "compression_loss": 77.28173828125, "epoch": 0.49, "learning_rate": 2.270203679369251e-05, "loss": 77.7282, "step": 1486 }, { "compression_loss": 77.72386169433594, "epoch": 0.49, "learning_rate": 2.2697109067017082e-05, "loss": 78.0956, "step": 1487 }, { "compression_loss": 75.06908416748047, "epoch": 0.49, "learning_rate": 2.2692181340341654e-05, "loss": 75.4118, "step": 1488 }, { "compression_loss": 74.88236999511719, "epoch": 0.49, "learning_rate": 2.268725361366623e-05, "loss": 75.2438, "step": 1489 }, { "compression_loss": 76.257568359375, "epoch": 0.49, "learning_rate": 2.2682325886990805e-05, "loss": 76.6219, "step": 1490 }, { "compression_loss": 75.04757690429688, "epoch": 0.49, "learning_rate": 2.2677398160315377e-05, "loss": 75.6048, "step": 1491 }, { "compression_loss": 78.40178680419922, "epoch": 0.49, "learning_rate": 2.267247043363995e-05, "loss": 78.8448, "step": 1492 }, { "compression_loss": 76.20527648925781, "epoch": 0.49, "learning_rate": 2.266754270696452e-05, "loss": 76.5363, "step": 1493 }, { "compression_loss": 78.11514282226562, "epoch": 0.49, "learning_rate": 2.2662614980289093e-05, "loss": 78.5986, "step": 1494 }, { "compression_loss": 76.8272705078125, "epoch": 0.49, "learning_rate": 2.2657687253613668e-05, "loss": 77.2376, "step": 1495 }, { "compression_loss": 75.77752685546875, "epoch": 0.49, "learning_rate": 2.265275952693824e-05, "loss": 76.0852, "step": 1496 }, { "compression_loss": 75.48158264160156, "epoch": 0.49, "learning_rate": 2.2647831800262812e-05, "loss": 75.7959, "step": 1497 }, { "compression_loss": 76.14924621582031, "epoch": 0.49, "learning_rate": 2.2642904073587384e-05, "loss": 76.3284, "step": 1498 }, { "compression_loss": 75.69233703613281, "epoch": 0.49, "learning_rate": 2.2637976346911956e-05, "loss": 75.9268, "step": 1499 }, { "compression_loss": 74.18671417236328, "epoch": 0.49, "learning_rate": 2.263304862023653e-05, "loss": 74.5421, "step": 1500 }, { "epoch": 0.49, "eval_exact_match": 85.69536423841059, "eval_f1": 92.20476831452969, "step": 1500 }, { "compression_loss": 74.82463073730469, "epoch": 0.49, "learning_rate": 2.2628120893561107e-05, "loss": 75.1268, "step": 1501 }, { "compression_loss": 74.3650131225586, "epoch": 0.49, "learning_rate": 2.262319316688568e-05, "loss": 74.7008, "step": 1502 }, { "compression_loss": 74.7654800415039, "epoch": 0.49, "learning_rate": 2.261826544021025e-05, "loss": 75.1695, "step": 1503 }, { "compression_loss": 76.7178955078125, "epoch": 0.49, "learning_rate": 2.2613337713534822e-05, "loss": 77.3963, "step": 1504 }, { "compression_loss": 72.98741149902344, "epoch": 0.49, "learning_rate": 2.2608409986859398e-05, "loss": 73.6405, "step": 1505 }, { "compression_loss": 73.26692199707031, "epoch": 0.49, "learning_rate": 2.260348226018397e-05, "loss": 73.5802, "step": 1506 }, { "compression_loss": 74.745361328125, "epoch": 0.5, "learning_rate": 2.2598554533508542e-05, "loss": 75.1683, "step": 1507 }, { "compression_loss": 75.79586791992188, "epoch": 0.5, "learning_rate": 2.2593626806833114e-05, "loss": 76.2799, "step": 1508 }, { "compression_loss": 74.79562377929688, "epoch": 0.5, "learning_rate": 2.2588699080157686e-05, "loss": 75.0207, "step": 1509 }, { "compression_loss": 77.38924407958984, "epoch": 0.5, "learning_rate": 2.258377135348226e-05, "loss": 77.8548, "step": 1510 }, { "compression_loss": 74.10713195800781, "epoch": 0.5, "learning_rate": 2.2578843626806833e-05, "loss": 74.577, "step": 1511 }, { "compression_loss": 76.59857177734375, "epoch": 0.5, "learning_rate": 2.2573915900131405e-05, "loss": 77.1544, "step": 1512 }, { "compression_loss": 74.34640502929688, "epoch": 0.5, "learning_rate": 2.256898817345598e-05, "loss": 74.6922, "step": 1513 }, { "compression_loss": 73.85005187988281, "epoch": 0.5, "learning_rate": 2.2564060446780552e-05, "loss": 74.2307, "step": 1514 }, { "compression_loss": 73.97698211669922, "epoch": 0.5, "learning_rate": 2.2559132720105128e-05, "loss": 74.6271, "step": 1515 }, { "compression_loss": 77.12825775146484, "epoch": 0.5, "learning_rate": 2.25542049934297e-05, "loss": 77.6616, "step": 1516 }, { "compression_loss": 74.79940795898438, "epoch": 0.5, "learning_rate": 2.254927726675427e-05, "loss": 75.1191, "step": 1517 }, { "compression_loss": 74.85298919677734, "epoch": 0.5, "learning_rate": 2.2544349540078844e-05, "loss": 75.1988, "step": 1518 }, { "compression_loss": 78.08265686035156, "epoch": 0.5, "learning_rate": 2.2539421813403415e-05, "loss": 78.5403, "step": 1519 }, { "compression_loss": 76.03762817382812, "epoch": 0.5, "learning_rate": 2.253449408672799e-05, "loss": 76.4031, "step": 1520 }, { "compression_loss": 76.24079895019531, "epoch": 0.5, "learning_rate": 2.2529566360052563e-05, "loss": 76.7988, "step": 1521 }, { "compression_loss": 75.44092559814453, "epoch": 0.5, "learning_rate": 2.2524638633377135e-05, "loss": 75.7522, "step": 1522 }, { "compression_loss": 75.96868896484375, "epoch": 0.5, "learning_rate": 2.2519710906701707e-05, "loss": 76.3596, "step": 1523 }, { "compression_loss": 73.9196548461914, "epoch": 0.5, "learning_rate": 2.2514783180026282e-05, "loss": 74.3839, "step": 1524 }, { "compression_loss": 73.52731323242188, "epoch": 0.5, "learning_rate": 2.2509855453350857e-05, "loss": 73.8993, "step": 1525 }, { "compression_loss": 75.21199035644531, "epoch": 0.5, "learning_rate": 2.250492772667543e-05, "loss": 75.5223, "step": 1526 }, { "compression_loss": 74.75788116455078, "epoch": 0.5, "learning_rate": 2.25e-05, "loss": 75.2563, "step": 1527 }, { "compression_loss": 74.13339233398438, "epoch": 0.5, "learning_rate": 2.2495072273324573e-05, "loss": 74.8231, "step": 1528 }, { "compression_loss": 77.35670471191406, "epoch": 0.5, "learning_rate": 2.2490144546649145e-05, "loss": 77.8982, "step": 1529 }, { "compression_loss": 77.35313415527344, "epoch": 0.5, "learning_rate": 2.248521681997372e-05, "loss": 77.6376, "step": 1530 }, { "compression_loss": 72.92851257324219, "epoch": 0.5, "learning_rate": 2.2480289093298293e-05, "loss": 73.2344, "step": 1531 }, { "compression_loss": 76.31100463867188, "epoch": 0.5, "learning_rate": 2.2475361366622865e-05, "loss": 76.8904, "step": 1532 }, { "compression_loss": 76.53548431396484, "epoch": 0.5, "learning_rate": 2.2470433639947437e-05, "loss": 76.916, "step": 1533 }, { "compression_loss": 74.84725952148438, "epoch": 0.5, "learning_rate": 2.246550591327201e-05, "loss": 75.3646, "step": 1534 }, { "compression_loss": 76.67975616455078, "epoch": 0.5, "learning_rate": 2.2460578186596584e-05, "loss": 77.0284, "step": 1535 }, { "compression_loss": 77.00389099121094, "epoch": 0.5, "learning_rate": 2.245565045992116e-05, "loss": 77.4363, "step": 1536 }, { "compression_loss": 76.19036865234375, "epoch": 0.5, "learning_rate": 2.245072273324573e-05, "loss": 76.5905, "step": 1537 }, { "compression_loss": 74.67306518554688, "epoch": 0.51, "learning_rate": 2.2445795006570303e-05, "loss": 75.2935, "step": 1538 }, { "compression_loss": 76.53231811523438, "epoch": 0.51, "learning_rate": 2.2440867279894875e-05, "loss": 76.9522, "step": 1539 }, { "compression_loss": 75.31729125976562, "epoch": 0.51, "learning_rate": 2.243593955321945e-05, "loss": 75.6536, "step": 1540 }, { "compression_loss": 77.37890625, "epoch": 0.51, "learning_rate": 2.2431011826544023e-05, "loss": 78.1619, "step": 1541 }, { "compression_loss": 75.5369644165039, "epoch": 0.51, "learning_rate": 2.2426084099868595e-05, "loss": 75.8638, "step": 1542 }, { "compression_loss": 75.78627014160156, "epoch": 0.51, "learning_rate": 2.2421156373193166e-05, "loss": 76.2728, "step": 1543 }, { "compression_loss": 74.33973693847656, "epoch": 0.51, "learning_rate": 2.241622864651774e-05, "loss": 74.6518, "step": 1544 }, { "compression_loss": 75.39083862304688, "epoch": 0.51, "learning_rate": 2.2411300919842314e-05, "loss": 75.6087, "step": 1545 }, { "compression_loss": 75.51518249511719, "epoch": 0.51, "learning_rate": 2.2406373193166886e-05, "loss": 76.136, "step": 1546 }, { "compression_loss": 76.2033462524414, "epoch": 0.51, "learning_rate": 2.2401445466491458e-05, "loss": 76.6535, "step": 1547 }, { "compression_loss": 76.55091857910156, "epoch": 0.51, "learning_rate": 2.2396517739816033e-05, "loss": 76.8193, "step": 1548 }, { "compression_loss": 74.09190368652344, "epoch": 0.51, "learning_rate": 2.2391590013140605e-05, "loss": 74.4244, "step": 1549 }, { "compression_loss": 74.58661651611328, "epoch": 0.51, "learning_rate": 2.238666228646518e-05, "loss": 74.918, "step": 1550 }, { "compression_loss": 76.26412963867188, "epoch": 0.51, "learning_rate": 2.2381734559789752e-05, "loss": 76.6613, "step": 1551 }, { "compression_loss": 75.63663482666016, "epoch": 0.51, "learning_rate": 2.2376806833114324e-05, "loss": 76.1536, "step": 1552 }, { "compression_loss": 74.90176391601562, "epoch": 0.51, "learning_rate": 2.2371879106438896e-05, "loss": 75.2459, "step": 1553 }, { "compression_loss": 75.72588348388672, "epoch": 0.51, "learning_rate": 2.2366951379763468e-05, "loss": 76.462, "step": 1554 }, { "compression_loss": 77.413818359375, "epoch": 0.51, "learning_rate": 2.2362023653088044e-05, "loss": 77.914, "step": 1555 }, { "compression_loss": 77.048583984375, "epoch": 0.51, "learning_rate": 2.2357095926412616e-05, "loss": 77.379, "step": 1556 }, { "compression_loss": 75.9238052368164, "epoch": 0.51, "learning_rate": 2.2352168199737188e-05, "loss": 76.4753, "step": 1557 }, { "compression_loss": 78.92996215820312, "epoch": 0.51, "learning_rate": 2.234724047306176e-05, "loss": 79.3184, "step": 1558 }, { "compression_loss": 75.5709457397461, "epoch": 0.51, "learning_rate": 2.234231274638633e-05, "loss": 76.0488, "step": 1559 }, { "compression_loss": 76.28514099121094, "epoch": 0.51, "learning_rate": 2.233738501971091e-05, "loss": 77.3382, "step": 1560 }, { "compression_loss": 74.32967376708984, "epoch": 0.51, "learning_rate": 2.2332457293035482e-05, "loss": 75.0041, "step": 1561 }, { "compression_loss": 75.07188415527344, "epoch": 0.51, "learning_rate": 2.2327529566360054e-05, "loss": 75.2721, "step": 1562 }, { "compression_loss": 74.62023162841797, "epoch": 0.51, "learning_rate": 2.2322601839684626e-05, "loss": 74.9487, "step": 1563 }, { "compression_loss": 77.71825408935547, "epoch": 0.51, "learning_rate": 2.2317674113009198e-05, "loss": 78.1023, "step": 1564 }, { "compression_loss": 74.74739074707031, "epoch": 0.51, "learning_rate": 2.2312746386333774e-05, "loss": 75.2844, "step": 1565 }, { "compression_loss": 77.84782409667969, "epoch": 0.51, "learning_rate": 2.2307818659658345e-05, "loss": 78.4608, "step": 1566 }, { "compression_loss": 77.78390502929688, "epoch": 0.51, "learning_rate": 2.2302890932982917e-05, "loss": 78.3583, "step": 1567 }, { "compression_loss": 75.12432861328125, "epoch": 0.52, "learning_rate": 2.229796320630749e-05, "loss": 75.5906, "step": 1568 }, { "compression_loss": 75.33584594726562, "epoch": 0.52, "learning_rate": 2.229303547963206e-05, "loss": 75.7053, "step": 1569 }, { "compression_loss": 75.3369140625, "epoch": 0.52, "learning_rate": 2.2288107752956637e-05, "loss": 75.6302, "step": 1570 }, { "compression_loss": 74.00288391113281, "epoch": 0.52, "learning_rate": 2.228318002628121e-05, "loss": 74.4455, "step": 1571 }, { "compression_loss": 76.93495178222656, "epoch": 0.52, "learning_rate": 2.2278252299605784e-05, "loss": 77.3181, "step": 1572 }, { "compression_loss": 75.40631103515625, "epoch": 0.52, "learning_rate": 2.2273324572930356e-05, "loss": 75.8337, "step": 1573 }, { "compression_loss": 75.44346618652344, "epoch": 0.52, "learning_rate": 2.2268396846254928e-05, "loss": 75.846, "step": 1574 }, { "compression_loss": 74.78324890136719, "epoch": 0.52, "learning_rate": 2.2263469119579503e-05, "loss": 75.1341, "step": 1575 }, { "compression_loss": 72.6159439086914, "epoch": 0.52, "learning_rate": 2.2258541392904075e-05, "loss": 72.81, "step": 1576 }, { "compression_loss": 74.70463562011719, "epoch": 0.52, "learning_rate": 2.2253613666228647e-05, "loss": 74.9436, "step": 1577 }, { "compression_loss": 75.37132263183594, "epoch": 0.52, "learning_rate": 2.224868593955322e-05, "loss": 75.8825, "step": 1578 }, { "compression_loss": 77.59294128417969, "epoch": 0.52, "learning_rate": 2.224375821287779e-05, "loss": 78.1096, "step": 1579 }, { "compression_loss": 77.49850463867188, "epoch": 0.52, "learning_rate": 2.2238830486202367e-05, "loss": 77.9023, "step": 1580 }, { "compression_loss": 76.00870513916016, "epoch": 0.52, "learning_rate": 2.223390275952694e-05, "loss": 76.4437, "step": 1581 }, { "compression_loss": 77.2686767578125, "epoch": 0.52, "learning_rate": 2.222897503285151e-05, "loss": 77.7521, "step": 1582 }, { "compression_loss": 74.63278198242188, "epoch": 0.52, "learning_rate": 2.2224047306176083e-05, "loss": 75.1666, "step": 1583 }, { "compression_loss": 76.47116088867188, "epoch": 0.52, "learning_rate": 2.2219119579500658e-05, "loss": 76.7564, "step": 1584 }, { "compression_loss": 78.09589385986328, "epoch": 0.52, "learning_rate": 2.2214191852825233e-05, "loss": 78.4521, "step": 1585 }, { "compression_loss": 76.8382568359375, "epoch": 0.52, "learning_rate": 2.2209264126149805e-05, "loss": 77.2918, "step": 1586 }, { "compression_loss": 74.43258666992188, "epoch": 0.52, "learning_rate": 2.2204336399474377e-05, "loss": 74.7398, "step": 1587 }, { "compression_loss": 75.85079956054688, "epoch": 0.52, "learning_rate": 2.219940867279895e-05, "loss": 76.2348, "step": 1588 }, { "compression_loss": 76.23812866210938, "epoch": 0.52, "learning_rate": 2.219448094612352e-05, "loss": 76.715, "step": 1589 }, { "compression_loss": 75.33490753173828, "epoch": 0.52, "learning_rate": 2.2189553219448096e-05, "loss": 75.8962, "step": 1590 }, { "compression_loss": 76.266357421875, "epoch": 0.52, "learning_rate": 2.218462549277267e-05, "loss": 76.9231, "step": 1591 }, { "compression_loss": 77.5110092163086, "epoch": 0.52, "learning_rate": 2.217969776609724e-05, "loss": 77.9057, "step": 1592 }, { "compression_loss": 74.84557342529297, "epoch": 0.52, "learning_rate": 2.2174770039421812e-05, "loss": 75.1581, "step": 1593 }, { "compression_loss": 75.42716979980469, "epoch": 0.52, "learning_rate": 2.2169842312746384e-05, "loss": 76.0282, "step": 1594 }, { "compression_loss": 72.96583557128906, "epoch": 0.52, "learning_rate": 2.216491458607096e-05, "loss": 73.1849, "step": 1595 }, { "compression_loss": 74.19113159179688, "epoch": 0.52, "learning_rate": 2.2159986859395535e-05, "loss": 74.4909, "step": 1596 }, { "compression_loss": 76.03644561767578, "epoch": 0.52, "learning_rate": 2.2155059132720107e-05, "loss": 76.5644, "step": 1597 }, { "compression_loss": 73.8751220703125, "epoch": 0.52, "learning_rate": 2.215013140604468e-05, "loss": 74.2541, "step": 1598 }, { "compression_loss": 77.36744689941406, "epoch": 0.53, "learning_rate": 2.214520367936925e-05, "loss": 77.8007, "step": 1599 }, { "compression_loss": 76.17439270019531, "epoch": 0.53, "learning_rate": 2.2140275952693826e-05, "loss": 76.6119, "step": 1600 }, { "compression_loss": 78.45246887207031, "epoch": 0.53, "learning_rate": 2.2135348226018398e-05, "loss": 79.1823, "step": 1601 }, { "compression_loss": 75.57612609863281, "epoch": 0.53, "learning_rate": 2.213042049934297e-05, "loss": 76.2032, "step": 1602 }, { "compression_loss": 76.35967254638672, "epoch": 0.53, "learning_rate": 2.2125492772667542e-05, "loss": 76.9726, "step": 1603 }, { "compression_loss": 75.47576904296875, "epoch": 0.53, "learning_rate": 2.2120565045992114e-05, "loss": 75.8594, "step": 1604 }, { "compression_loss": 75.29034423828125, "epoch": 0.53, "learning_rate": 2.211563731931669e-05, "loss": 75.5614, "step": 1605 }, { "compression_loss": 76.66232299804688, "epoch": 0.53, "learning_rate": 2.211070959264126e-05, "loss": 77.2833, "step": 1606 }, { "compression_loss": 77.79546356201172, "epoch": 0.53, "learning_rate": 2.2105781865965833e-05, "loss": 78.127, "step": 1607 }, { "compression_loss": 77.14787292480469, "epoch": 0.53, "learning_rate": 2.210085413929041e-05, "loss": 77.3554, "step": 1608 }, { "compression_loss": 76.05047607421875, "epoch": 0.53, "learning_rate": 2.209592641261498e-05, "loss": 76.4445, "step": 1609 }, { "compression_loss": 75.36672973632812, "epoch": 0.53, "learning_rate": 2.2090998685939556e-05, "loss": 75.5523, "step": 1610 }, { "compression_loss": 73.62781524658203, "epoch": 0.53, "learning_rate": 2.2086070959264128e-05, "loss": 73.9848, "step": 1611 }, { "compression_loss": 75.74346160888672, "epoch": 0.53, "learning_rate": 2.20811432325887e-05, "loss": 76.2991, "step": 1612 }, { "compression_loss": 74.9449234008789, "epoch": 0.53, "learning_rate": 2.2076215505913272e-05, "loss": 75.4707, "step": 1613 }, { "compression_loss": 75.66929626464844, "epoch": 0.53, "learning_rate": 2.2071287779237844e-05, "loss": 75.9334, "step": 1614 }, { "compression_loss": 74.10460662841797, "epoch": 0.53, "learning_rate": 2.206636005256242e-05, "loss": 74.6504, "step": 1615 }, { "compression_loss": 76.85565185546875, "epoch": 0.53, "learning_rate": 2.206143232588699e-05, "loss": 77.5332, "step": 1616 }, { "compression_loss": 74.22091674804688, "epoch": 0.53, "learning_rate": 2.2056504599211563e-05, "loss": 74.4714, "step": 1617 }, { "compression_loss": 74.78052520751953, "epoch": 0.53, "learning_rate": 2.2051576872536135e-05, "loss": 75.2622, "step": 1618 }, { "compression_loss": 77.12612915039062, "epoch": 0.53, "learning_rate": 2.2046649145860707e-05, "loss": 77.4641, "step": 1619 }, { "compression_loss": 75.35136413574219, "epoch": 0.53, "learning_rate": 2.2041721419185286e-05, "loss": 75.6079, "step": 1620 }, { "compression_loss": 74.50651550292969, "epoch": 0.53, "learning_rate": 2.2036793692509858e-05, "loss": 74.6972, "step": 1621 }, { "compression_loss": 77.3287124633789, "epoch": 0.53, "learning_rate": 2.203186596583443e-05, "loss": 77.8788, "step": 1622 }, { "compression_loss": 74.35106658935547, "epoch": 0.53, "learning_rate": 2.2026938239159002e-05, "loss": 74.4665, "step": 1623 }, { "compression_loss": 76.95722961425781, "epoch": 0.53, "learning_rate": 2.2022010512483574e-05, "loss": 77.2555, "step": 1624 }, { "compression_loss": 74.18766021728516, "epoch": 0.53, "learning_rate": 2.201708278580815e-05, "loss": 74.6684, "step": 1625 }, { "compression_loss": 76.84503936767578, "epoch": 0.53, "learning_rate": 2.201215505913272e-05, "loss": 77.4681, "step": 1626 }, { "compression_loss": 74.0609359741211, "epoch": 0.53, "learning_rate": 2.2007227332457293e-05, "loss": 74.3062, "step": 1627 }, { "compression_loss": 73.86238861083984, "epoch": 0.53, "learning_rate": 2.2002299605781865e-05, "loss": 74.1431, "step": 1628 }, { "compression_loss": 75.80247497558594, "epoch": 0.54, "learning_rate": 2.1997371879106437e-05, "loss": 76.339, "step": 1629 }, { "compression_loss": 76.14568328857422, "epoch": 0.54, "learning_rate": 2.1992444152431012e-05, "loss": 76.4888, "step": 1630 }, { "compression_loss": 74.09754180908203, "epoch": 0.54, "learning_rate": 2.1987516425755584e-05, "loss": 74.5641, "step": 1631 }, { "compression_loss": 75.26226806640625, "epoch": 0.54, "learning_rate": 2.198258869908016e-05, "loss": 75.6447, "step": 1632 }, { "compression_loss": 75.77291107177734, "epoch": 0.54, "learning_rate": 2.1977660972404732e-05, "loss": 76.0569, "step": 1633 }, { "compression_loss": 78.64692687988281, "epoch": 0.54, "learning_rate": 2.1972733245729304e-05, "loss": 79.2355, "step": 1634 }, { "compression_loss": 77.52544403076172, "epoch": 0.54, "learning_rate": 2.196780551905388e-05, "loss": 78.0914, "step": 1635 }, { "compression_loss": 76.16845703125, "epoch": 0.54, "learning_rate": 2.196287779237845e-05, "loss": 76.5386, "step": 1636 }, { "compression_loss": 75.45692443847656, "epoch": 0.54, "learning_rate": 2.1957950065703023e-05, "loss": 75.7227, "step": 1637 }, { "compression_loss": 72.33480072021484, "epoch": 0.54, "learning_rate": 2.1953022339027595e-05, "loss": 72.7854, "step": 1638 }, { "compression_loss": 74.78141021728516, "epoch": 0.54, "learning_rate": 2.1948094612352167e-05, "loss": 75.021, "step": 1639 }, { "compression_loss": 76.10465240478516, "epoch": 0.54, "learning_rate": 2.1943166885676742e-05, "loss": 76.5438, "step": 1640 }, { "compression_loss": 73.37660217285156, "epoch": 0.54, "learning_rate": 2.1938239159001314e-05, "loss": 73.6975, "step": 1641 }, { "compression_loss": 74.19585418701172, "epoch": 0.54, "learning_rate": 2.1933311432325886e-05, "loss": 74.4901, "step": 1642 }, { "compression_loss": 75.32743835449219, "epoch": 0.54, "learning_rate": 2.1928383705650458e-05, "loss": 75.6331, "step": 1643 }, { "compression_loss": 75.54734802246094, "epoch": 0.54, "learning_rate": 2.1923455978975034e-05, "loss": 76.0331, "step": 1644 }, { "compression_loss": 76.10176849365234, "epoch": 0.54, "learning_rate": 2.191852825229961e-05, "loss": 76.5563, "step": 1645 }, { "compression_loss": 77.74795532226562, "epoch": 0.54, "learning_rate": 2.191360052562418e-05, "loss": 78.2018, "step": 1646 }, { "compression_loss": 76.14421844482422, "epoch": 0.54, "learning_rate": 2.1908672798948753e-05, "loss": 76.4518, "step": 1647 }, { "compression_loss": 75.29953002929688, "epoch": 0.54, "learning_rate": 2.1903745072273325e-05, "loss": 75.8493, "step": 1648 }, { "compression_loss": 76.06184387207031, "epoch": 0.54, "learning_rate": 2.1898817345597897e-05, "loss": 76.4497, "step": 1649 }, { "compression_loss": 74.43330383300781, "epoch": 0.54, "learning_rate": 2.1893889618922472e-05, "loss": 74.8298, "step": 1650 }, { "compression_loss": 75.48634338378906, "epoch": 0.54, "learning_rate": 2.1888961892247044e-05, "loss": 76.0654, "step": 1651 }, { "compression_loss": 76.55927276611328, "epoch": 0.54, "learning_rate": 2.1884034165571616e-05, "loss": 77.3435, "step": 1652 }, { "compression_loss": 75.8421630859375, "epoch": 0.54, "learning_rate": 2.1879106438896188e-05, "loss": 76.4821, "step": 1653 }, { "compression_loss": 78.22576904296875, "epoch": 0.54, "learning_rate": 2.187417871222076e-05, "loss": 78.6373, "step": 1654 }, { "compression_loss": 75.45285034179688, "epoch": 0.54, "learning_rate": 2.1869250985545335e-05, "loss": 75.7415, "step": 1655 }, { "compression_loss": 73.47624206542969, "epoch": 0.54, "learning_rate": 2.186432325886991e-05, "loss": 73.7072, "step": 1656 }, { "compression_loss": 75.51042175292969, "epoch": 0.54, "learning_rate": 2.1859395532194483e-05, "loss": 75.863, "step": 1657 }, { "compression_loss": 71.85987854003906, "epoch": 0.54, "learning_rate": 2.1854467805519055e-05, "loss": 72.0147, "step": 1658 }, { "compression_loss": 77.72590637207031, "epoch": 0.55, "learning_rate": 2.1849540078843627e-05, "loss": 78.3806, "step": 1659 }, { "compression_loss": 74.85035705566406, "epoch": 0.55, "learning_rate": 2.1844612352168202e-05, "loss": 75.2515, "step": 1660 }, { "compression_loss": 76.31066131591797, "epoch": 0.55, "learning_rate": 2.1839684625492774e-05, "loss": 76.6259, "step": 1661 }, { "compression_loss": 76.01081848144531, "epoch": 0.55, "learning_rate": 2.1834756898817346e-05, "loss": 76.3995, "step": 1662 }, { "compression_loss": 74.33633422851562, "epoch": 0.55, "learning_rate": 2.1829829172141918e-05, "loss": 74.5487, "step": 1663 }, { "compression_loss": 74.0689926147461, "epoch": 0.55, "learning_rate": 2.182490144546649e-05, "loss": 74.4463, "step": 1664 }, { "compression_loss": 75.90422058105469, "epoch": 0.55, "learning_rate": 2.1819973718791065e-05, "loss": 76.2658, "step": 1665 }, { "compression_loss": 75.6801528930664, "epoch": 0.55, "learning_rate": 2.1815045992115637e-05, "loss": 75.9494, "step": 1666 }, { "compression_loss": 77.11851501464844, "epoch": 0.55, "learning_rate": 2.1810118265440213e-05, "loss": 77.5065, "step": 1667 }, { "compression_loss": 74.73053741455078, "epoch": 0.55, "learning_rate": 2.1805190538764785e-05, "loss": 75.0441, "step": 1668 }, { "compression_loss": 75.3892593383789, "epoch": 0.55, "learning_rate": 2.1800262812089357e-05, "loss": 75.689, "step": 1669 }, { "compression_loss": 76.35330200195312, "epoch": 0.55, "learning_rate": 2.1795335085413932e-05, "loss": 76.5944, "step": 1670 }, { "compression_loss": 75.60545349121094, "epoch": 0.55, "learning_rate": 2.1790407358738504e-05, "loss": 75.9784, "step": 1671 }, { "compression_loss": 76.79185485839844, "epoch": 0.55, "learning_rate": 2.1785479632063076e-05, "loss": 77.4394, "step": 1672 }, { "compression_loss": 78.06690979003906, "epoch": 0.55, "learning_rate": 2.1780551905387648e-05, "loss": 78.6819, "step": 1673 }, { "compression_loss": 74.51656341552734, "epoch": 0.55, "learning_rate": 2.177562417871222e-05, "loss": 74.8573, "step": 1674 }, { "compression_loss": 76.09962463378906, "epoch": 0.55, "learning_rate": 2.1770696452036795e-05, "loss": 76.4226, "step": 1675 }, { "compression_loss": 74.87515258789062, "epoch": 0.55, "learning_rate": 2.1765768725361367e-05, "loss": 75.146, "step": 1676 }, { "compression_loss": 76.57431030273438, "epoch": 0.55, "learning_rate": 2.176084099868594e-05, "loss": 76.9389, "step": 1677 }, { "compression_loss": 73.61866760253906, "epoch": 0.55, "learning_rate": 2.175591327201051e-05, "loss": 74.0309, "step": 1678 }, { "compression_loss": 75.50120544433594, "epoch": 0.55, "learning_rate": 2.1750985545335086e-05, "loss": 75.9476, "step": 1679 }, { "compression_loss": 73.08696746826172, "epoch": 0.55, "learning_rate": 2.1746057818659662e-05, "loss": 73.4419, "step": 1680 }, { "compression_loss": 75.17456817626953, "epoch": 0.55, "learning_rate": 2.1741130091984234e-05, "loss": 75.9466, "step": 1681 }, { "compression_loss": 74.41161346435547, "epoch": 0.55, "learning_rate": 2.1736202365308806e-05, "loss": 74.9419, "step": 1682 }, { "compression_loss": 75.52928924560547, "epoch": 0.55, "learning_rate": 2.1731274638633378e-05, "loss": 75.9511, "step": 1683 }, { "compression_loss": 76.029541015625, "epoch": 0.55, "learning_rate": 2.172634691195795e-05, "loss": 76.5748, "step": 1684 }, { "compression_loss": 76.47083282470703, "epoch": 0.55, "learning_rate": 2.1721419185282525e-05, "loss": 76.9468, "step": 1685 }, { "compression_loss": 75.5156478881836, "epoch": 0.55, "learning_rate": 2.1716491458607097e-05, "loss": 76.2057, "step": 1686 }, { "compression_loss": 78.34613037109375, "epoch": 0.55, "learning_rate": 2.171156373193167e-05, "loss": 78.8033, "step": 1687 }, { "compression_loss": 76.67054748535156, "epoch": 0.55, "learning_rate": 2.170663600525624e-05, "loss": 76.9106, "step": 1688 }, { "compression_loss": 74.94766998291016, "epoch": 0.55, "learning_rate": 2.1701708278580813e-05, "loss": 75.581, "step": 1689 }, { "compression_loss": 75.85467529296875, "epoch": 0.56, "learning_rate": 2.1696780551905388e-05, "loss": 76.0756, "step": 1690 }, { "compression_loss": 73.63671875, "epoch": 0.56, "learning_rate": 2.1691852825229964e-05, "loss": 74.168, "step": 1691 }, { "compression_loss": 77.94276428222656, "epoch": 0.56, "learning_rate": 2.1686925098554536e-05, "loss": 78.3611, "step": 1692 }, { "compression_loss": 74.9266586303711, "epoch": 0.56, "learning_rate": 2.1681997371879108e-05, "loss": 75.3872, "step": 1693 }, { "compression_loss": 75.93891906738281, "epoch": 0.56, "learning_rate": 2.167706964520368e-05, "loss": 76.457, "step": 1694 }, { "compression_loss": 74.73518371582031, "epoch": 0.56, "learning_rate": 2.1672141918528255e-05, "loss": 75.0248, "step": 1695 }, { "compression_loss": 74.90202331542969, "epoch": 0.56, "learning_rate": 2.1667214191852827e-05, "loss": 75.4429, "step": 1696 }, { "compression_loss": 76.19255065917969, "epoch": 0.56, "learning_rate": 2.16622864651774e-05, "loss": 76.5761, "step": 1697 }, { "compression_loss": 74.53523254394531, "epoch": 0.56, "learning_rate": 2.165735873850197e-05, "loss": 74.7338, "step": 1698 }, { "compression_loss": 75.1122055053711, "epoch": 0.56, "learning_rate": 2.1652431011826543e-05, "loss": 75.639, "step": 1699 }, { "compression_loss": 76.48912811279297, "epoch": 0.56, "learning_rate": 2.1647503285151118e-05, "loss": 77.0083, "step": 1700 }, { "compression_loss": 75.46299743652344, "epoch": 0.56, "learning_rate": 2.164257555847569e-05, "loss": 75.726, "step": 1701 }, { "compression_loss": 76.31478118896484, "epoch": 0.56, "learning_rate": 2.1637647831800262e-05, "loss": 76.715, "step": 1702 }, { "compression_loss": 74.9323501586914, "epoch": 0.56, "learning_rate": 2.1632720105124837e-05, "loss": 75.389, "step": 1703 }, { "compression_loss": 76.81776428222656, "epoch": 0.56, "learning_rate": 2.162779237844941e-05, "loss": 77.203, "step": 1704 }, { "compression_loss": 76.54707336425781, "epoch": 0.56, "learning_rate": 2.1622864651773985e-05, "loss": 76.8013, "step": 1705 }, { "compression_loss": 74.25006103515625, "epoch": 0.56, "learning_rate": 2.1617936925098557e-05, "loss": 74.6547, "step": 1706 }, { "compression_loss": 78.1575698852539, "epoch": 0.56, "learning_rate": 2.161300919842313e-05, "loss": 78.6964, "step": 1707 }, { "compression_loss": 77.02886199951172, "epoch": 0.56, "learning_rate": 2.16080814717477e-05, "loss": 77.535, "step": 1708 }, { "compression_loss": 76.31272888183594, "epoch": 0.56, "learning_rate": 2.1603153745072273e-05, "loss": 76.6074, "step": 1709 }, { "compression_loss": 75.03732299804688, "epoch": 0.56, "learning_rate": 2.1598226018396848e-05, "loss": 75.5153, "step": 1710 }, { "compression_loss": 76.498046875, "epoch": 0.56, "learning_rate": 2.159329829172142e-05, "loss": 76.8825, "step": 1711 }, { "compression_loss": 77.38994598388672, "epoch": 0.56, "learning_rate": 2.1588370565045992e-05, "loss": 77.6681, "step": 1712 }, { "compression_loss": 74.8598403930664, "epoch": 0.56, "learning_rate": 2.1583442838370564e-05, "loss": 75.1296, "step": 1713 }, { "compression_loss": 75.2442626953125, "epoch": 0.56, "learning_rate": 2.1578515111695136e-05, "loss": 75.5241, "step": 1714 }, { "compression_loss": 74.49652099609375, "epoch": 0.56, "learning_rate": 2.1573587385019715e-05, "loss": 74.8993, "step": 1715 }, { "compression_loss": 75.77052307128906, "epoch": 0.56, "learning_rate": 2.1568659658344287e-05, "loss": 76.2934, "step": 1716 }, { "compression_loss": 74.3677978515625, "epoch": 0.56, "learning_rate": 2.156373193166886e-05, "loss": 74.8359, "step": 1717 }, { "compression_loss": 75.07919311523438, "epoch": 0.56, "learning_rate": 2.155880420499343e-05, "loss": 75.5857, "step": 1718 }, { "compression_loss": 75.68045043945312, "epoch": 0.56, "learning_rate": 2.1553876478318002e-05, "loss": 75.9414, "step": 1719 }, { "compression_loss": 77.64389038085938, "epoch": 0.57, "learning_rate": 2.1548948751642578e-05, "loss": 78.1089, "step": 1720 }, { "compression_loss": 75.94056701660156, "epoch": 0.57, "learning_rate": 2.154402102496715e-05, "loss": 76.3168, "step": 1721 }, { "compression_loss": 73.29087829589844, "epoch": 0.57, "learning_rate": 2.1539093298291722e-05, "loss": 73.5832, "step": 1722 }, { "compression_loss": 74.85301208496094, "epoch": 0.57, "learning_rate": 2.1534165571616294e-05, "loss": 75.2558, "step": 1723 }, { "compression_loss": 77.1651611328125, "epoch": 0.57, "learning_rate": 2.1529237844940866e-05, "loss": 77.3767, "step": 1724 }, { "compression_loss": 76.70068359375, "epoch": 0.57, "learning_rate": 2.152431011826544e-05, "loss": 77.2174, "step": 1725 }, { "compression_loss": 77.39287567138672, "epoch": 0.57, "learning_rate": 2.1519382391590013e-05, "loss": 77.8054, "step": 1726 }, { "compression_loss": 74.07107543945312, "epoch": 0.57, "learning_rate": 2.151445466491459e-05, "loss": 74.6103, "step": 1727 }, { "compression_loss": 76.43960571289062, "epoch": 0.57, "learning_rate": 2.150952693823916e-05, "loss": 76.9291, "step": 1728 }, { "compression_loss": 75.70806884765625, "epoch": 0.57, "learning_rate": 2.1504599211563732e-05, "loss": 76.3936, "step": 1729 }, { "compression_loss": 75.39608764648438, "epoch": 0.57, "learning_rate": 2.1499671484888308e-05, "loss": 75.7672, "step": 1730 }, { "compression_loss": 77.82893371582031, "epoch": 0.57, "learning_rate": 2.149474375821288e-05, "loss": 78.3338, "step": 1731 }, { "compression_loss": 76.215576171875, "epoch": 0.57, "learning_rate": 2.148981603153745e-05, "loss": 76.5582, "step": 1732 }, { "compression_loss": 73.82400512695312, "epoch": 0.57, "learning_rate": 2.1484888304862024e-05, "loss": 74.2851, "step": 1733 }, { "compression_loss": 74.1493148803711, "epoch": 0.57, "learning_rate": 2.1479960578186596e-05, "loss": 74.4796, "step": 1734 }, { "compression_loss": 74.87773132324219, "epoch": 0.57, "learning_rate": 2.147503285151117e-05, "loss": 75.1581, "step": 1735 }, { "compression_loss": 74.01294708251953, "epoch": 0.57, "learning_rate": 2.1470105124835743e-05, "loss": 74.2848, "step": 1736 }, { "compression_loss": 75.56625366210938, "epoch": 0.57, "learning_rate": 2.1465177398160315e-05, "loss": 76.0461, "step": 1737 }, { "compression_loss": 76.63775634765625, "epoch": 0.57, "learning_rate": 2.1460249671484887e-05, "loss": 77.084, "step": 1738 }, { "compression_loss": 73.97274780273438, "epoch": 0.57, "learning_rate": 2.1455321944809462e-05, "loss": 74.2143, "step": 1739 }, { "compression_loss": 75.36927032470703, "epoch": 0.57, "learning_rate": 2.1450394218134038e-05, "loss": 75.8911, "step": 1740 }, { "compression_loss": 76.6137466430664, "epoch": 0.57, "learning_rate": 2.144546649145861e-05, "loss": 76.7866, "step": 1741 }, { "compression_loss": 76.52096557617188, "epoch": 0.57, "learning_rate": 2.144053876478318e-05, "loss": 76.7343, "step": 1742 }, { "compression_loss": 76.729736328125, "epoch": 0.57, "learning_rate": 2.1435611038107753e-05, "loss": 77.31, "step": 1743 }, { "compression_loss": 75.89350891113281, "epoch": 0.57, "learning_rate": 2.1430683311432325e-05, "loss": 76.5777, "step": 1744 }, { "compression_loss": 75.92019653320312, "epoch": 0.57, "learning_rate": 2.14257555847569e-05, "loss": 76.2453, "step": 1745 }, { "compression_loss": 77.16370391845703, "epoch": 0.57, "learning_rate": 2.1420827858081473e-05, "loss": 77.4759, "step": 1746 }, { "compression_loss": 77.45657348632812, "epoch": 0.57, "learning_rate": 2.1415900131406045e-05, "loss": 78.1508, "step": 1747 }, { "compression_loss": 76.43241882324219, "epoch": 0.57, "learning_rate": 2.1410972404730617e-05, "loss": 76.7954, "step": 1748 }, { "compression_loss": 77.37158203125, "epoch": 0.57, "learning_rate": 2.140604467805519e-05, "loss": 77.8878, "step": 1749 }, { "compression_loss": 77.16315460205078, "epoch": 0.57, "learning_rate": 2.1401116951379764e-05, "loss": 77.6117, "step": 1750 }, { "epoch": 0.57, "eval_exact_match": 85.78051087984863, "eval_f1": 92.40663943141142, "step": 1750 }, { "compression_loss": 75.1058349609375, "epoch": 0.58, "learning_rate": 2.139618922470434e-05, "loss": 75.4987, "step": 1751 }, { "compression_loss": 74.15979766845703, "epoch": 0.58, "learning_rate": 2.139126149802891e-05, "loss": 74.4533, "step": 1752 }, { "compression_loss": 75.80075073242188, "epoch": 0.58, "learning_rate": 2.1386333771353483e-05, "loss": 76.3598, "step": 1753 }, { "compression_loss": 79.77607727050781, "epoch": 0.58, "learning_rate": 2.1381406044678055e-05, "loss": 80.0308, "step": 1754 }, { "compression_loss": 75.84557342529297, "epoch": 0.58, "learning_rate": 2.137647831800263e-05, "loss": 76.272, "step": 1755 }, { "compression_loss": 77.1348876953125, "epoch": 0.58, "learning_rate": 2.1371550591327203e-05, "loss": 77.5227, "step": 1756 }, { "compression_loss": 72.8821029663086, "epoch": 0.58, "learning_rate": 2.1366622864651775e-05, "loss": 73.2182, "step": 1757 }, { "compression_loss": 76.59471130371094, "epoch": 0.58, "learning_rate": 2.1361695137976347e-05, "loss": 77.2743, "step": 1758 }, { "compression_loss": 73.9395751953125, "epoch": 0.58, "learning_rate": 2.135676741130092e-05, "loss": 74.2914, "step": 1759 }, { "compression_loss": 76.0220718383789, "epoch": 0.58, "learning_rate": 2.1351839684625494e-05, "loss": 76.7189, "step": 1760 }, { "compression_loss": 75.98711395263672, "epoch": 0.58, "learning_rate": 2.1346911957950066e-05, "loss": 76.4402, "step": 1761 }, { "compression_loss": 76.9683837890625, "epoch": 0.58, "learning_rate": 2.1341984231274638e-05, "loss": 77.5465, "step": 1762 }, { "compression_loss": 77.5102767944336, "epoch": 0.58, "learning_rate": 2.1337056504599213e-05, "loss": 77.9317, "step": 1763 }, { "compression_loss": 73.73435974121094, "epoch": 0.58, "learning_rate": 2.1332128777923785e-05, "loss": 74.0311, "step": 1764 }, { "compression_loss": 76.94489288330078, "epoch": 0.58, "learning_rate": 2.132720105124836e-05, "loss": 77.5534, "step": 1765 }, { "compression_loss": 75.0605239868164, "epoch": 0.58, "learning_rate": 2.1322273324572932e-05, "loss": 75.4262, "step": 1766 }, { "compression_loss": 78.0181884765625, "epoch": 0.58, "learning_rate": 2.1317345597897504e-05, "loss": 78.3867, "step": 1767 }, { "compression_loss": 77.30625915527344, "epoch": 0.58, "learning_rate": 2.1312417871222076e-05, "loss": 77.7709, "step": 1768 }, { "compression_loss": 78.32052612304688, "epoch": 0.58, "learning_rate": 2.130749014454665e-05, "loss": 78.8251, "step": 1769 }, { "compression_loss": 75.61058044433594, "epoch": 0.58, "learning_rate": 2.1302562417871224e-05, "loss": 76.1247, "step": 1770 }, { "compression_loss": 77.84649658203125, "epoch": 0.58, "learning_rate": 2.1297634691195796e-05, "loss": 78.6654, "step": 1771 }, { "compression_loss": 77.69164276123047, "epoch": 0.58, "learning_rate": 2.1292706964520368e-05, "loss": 78.3619, "step": 1772 }, { "compression_loss": 72.50984954833984, "epoch": 0.58, "learning_rate": 2.128777923784494e-05, "loss": 72.7951, "step": 1773 }, { "compression_loss": 79.13247680664062, "epoch": 0.58, "learning_rate": 2.128285151116951e-05, "loss": 79.5922, "step": 1774 }, { "compression_loss": 75.20292663574219, "epoch": 0.58, "learning_rate": 2.127792378449409e-05, "loss": 76.013, "step": 1775 }, { "compression_loss": 74.45907592773438, "epoch": 0.58, "learning_rate": 2.1272996057818662e-05, "loss": 75.1884, "step": 1776 }, { "compression_loss": 75.99690246582031, "epoch": 0.58, "learning_rate": 2.1268068331143234e-05, "loss": 76.4014, "step": 1777 }, { "compression_loss": 74.70825958251953, "epoch": 0.58, "learning_rate": 2.1263140604467806e-05, "loss": 75.0266, "step": 1778 }, { "compression_loss": 73.35723876953125, "epoch": 0.58, "learning_rate": 2.1258212877792378e-05, "loss": 73.6799, "step": 1779 }, { "compression_loss": 73.90086364746094, "epoch": 0.58, "learning_rate": 2.1253285151116954e-05, "loss": 74.3935, "step": 1780 }, { "compression_loss": 72.91390991210938, "epoch": 0.59, "learning_rate": 2.1248357424441526e-05, "loss": 73.251, "step": 1781 }, { "compression_loss": 75.55140686035156, "epoch": 0.59, "learning_rate": 2.1243429697766097e-05, "loss": 76.3006, "step": 1782 }, { "compression_loss": 76.16502380371094, "epoch": 0.59, "learning_rate": 2.123850197109067e-05, "loss": 76.6347, "step": 1783 }, { "compression_loss": 76.30262756347656, "epoch": 0.59, "learning_rate": 2.123357424441524e-05, "loss": 76.6934, "step": 1784 }, { "compression_loss": 77.23271179199219, "epoch": 0.59, "learning_rate": 2.1228646517739817e-05, "loss": 77.4854, "step": 1785 }, { "compression_loss": 74.5569076538086, "epoch": 0.59, "learning_rate": 2.122371879106439e-05, "loss": 74.8776, "step": 1786 }, { "compression_loss": 73.10721588134766, "epoch": 0.59, "learning_rate": 2.1218791064388964e-05, "loss": 73.4108, "step": 1787 }, { "compression_loss": 76.8255615234375, "epoch": 0.59, "learning_rate": 2.1213863337713536e-05, "loss": 77.2217, "step": 1788 }, { "compression_loss": 75.44747161865234, "epoch": 0.59, "learning_rate": 2.1208935611038108e-05, "loss": 75.8701, "step": 1789 }, { "compression_loss": 77.0290298461914, "epoch": 0.59, "learning_rate": 2.1204007884362683e-05, "loss": 77.6348, "step": 1790 }, { "compression_loss": 75.25994873046875, "epoch": 0.59, "learning_rate": 2.1199080157687255e-05, "loss": 75.4955, "step": 1791 }, { "compression_loss": 75.16159057617188, "epoch": 0.59, "learning_rate": 2.1194152431011827e-05, "loss": 75.4988, "step": 1792 }, { "compression_loss": 76.6692886352539, "epoch": 0.59, "learning_rate": 2.11892247043364e-05, "loss": 76.9572, "step": 1793 }, { "compression_loss": 77.20706939697266, "epoch": 0.59, "learning_rate": 2.118429697766097e-05, "loss": 78.2426, "step": 1794 }, { "compression_loss": 76.54414367675781, "epoch": 0.59, "learning_rate": 2.1179369250985547e-05, "loss": 76.9723, "step": 1795 }, { "compression_loss": 76.38867950439453, "epoch": 0.59, "learning_rate": 2.117444152431012e-05, "loss": 76.8373, "step": 1796 }, { "compression_loss": 74.68475341796875, "epoch": 0.59, "learning_rate": 2.116951379763469e-05, "loss": 74.9317, "step": 1797 }, { "compression_loss": 76.34695434570312, "epoch": 0.59, "learning_rate": 2.1164586070959263e-05, "loss": 76.7572, "step": 1798 }, { "compression_loss": 74.51426696777344, "epoch": 0.59, "learning_rate": 2.1159658344283838e-05, "loss": 74.7695, "step": 1799 }, { "compression_loss": 75.22850036621094, "epoch": 0.59, "learning_rate": 2.1154730617608413e-05, "loss": 75.5562, "step": 1800 }, { "compression_loss": 74.02621459960938, "epoch": 0.59, "learning_rate": 2.1149802890932985e-05, "loss": 74.3551, "step": 1801 }, { "compression_loss": 77.53412628173828, "epoch": 0.59, "learning_rate": 2.1144875164257557e-05, "loss": 77.9179, "step": 1802 }, { "compression_loss": 76.84015655517578, "epoch": 0.59, "learning_rate": 2.113994743758213e-05, "loss": 77.2256, "step": 1803 }, { "compression_loss": 73.70983123779297, "epoch": 0.59, "learning_rate": 2.11350197109067e-05, "loss": 74.0624, "step": 1804 }, { "compression_loss": 76.64524841308594, "epoch": 0.59, "learning_rate": 2.1130091984231276e-05, "loss": 77.0701, "step": 1805 }, { "compression_loss": 74.12098693847656, "epoch": 0.59, "learning_rate": 2.112516425755585e-05, "loss": 74.4644, "step": 1806 }, { "compression_loss": 75.74234008789062, "epoch": 0.59, "learning_rate": 2.112023653088042e-05, "loss": 76.1291, "step": 1807 }, { "compression_loss": 73.84033966064453, "epoch": 0.59, "learning_rate": 2.1115308804204992e-05, "loss": 74.1195, "step": 1808 }, { "compression_loss": 76.98944091796875, "epoch": 0.59, "learning_rate": 2.1110381077529564e-05, "loss": 77.2781, "step": 1809 }, { "compression_loss": 76.545654296875, "epoch": 0.59, "learning_rate": 2.110545335085414e-05, "loss": 76.813, "step": 1810 }, { "compression_loss": 76.3826904296875, "epoch": 0.59, "learning_rate": 2.1100525624178715e-05, "loss": 76.6426, "step": 1811 }, { "compression_loss": 76.85687255859375, "epoch": 0.6, "learning_rate": 2.1095597897503287e-05, "loss": 77.3027, "step": 1812 }, { "compression_loss": 77.52239990234375, "epoch": 0.6, "learning_rate": 2.109067017082786e-05, "loss": 78.001, "step": 1813 }, { "compression_loss": 75.46049499511719, "epoch": 0.6, "learning_rate": 2.108574244415243e-05, "loss": 75.9427, "step": 1814 }, { "compression_loss": 74.62218475341797, "epoch": 0.6, "learning_rate": 2.1080814717477006e-05, "loss": 74.938, "step": 1815 }, { "compression_loss": 76.7823486328125, "epoch": 0.6, "learning_rate": 2.107588699080158e-05, "loss": 77.4753, "step": 1816 }, { "compression_loss": 78.22354125976562, "epoch": 0.6, "learning_rate": 2.107095926412615e-05, "loss": 78.4719, "step": 1817 }, { "compression_loss": 79.1426010131836, "epoch": 0.6, "learning_rate": 2.1066031537450722e-05, "loss": 79.5648, "step": 1818 }, { "compression_loss": 73.92790985107422, "epoch": 0.6, "learning_rate": 2.1061103810775294e-05, "loss": 74.227, "step": 1819 }, { "compression_loss": 76.51050567626953, "epoch": 0.6, "learning_rate": 2.105617608409987e-05, "loss": 76.9584, "step": 1820 }, { "compression_loss": 77.57745361328125, "epoch": 0.6, "learning_rate": 2.105124835742444e-05, "loss": 78.111, "step": 1821 }, { "compression_loss": 74.0067138671875, "epoch": 0.6, "learning_rate": 2.1046320630749017e-05, "loss": 74.5001, "step": 1822 }, { "compression_loss": 77.77896118164062, "epoch": 0.6, "learning_rate": 2.104139290407359e-05, "loss": 78.2295, "step": 1823 }, { "compression_loss": 76.17839050292969, "epoch": 0.6, "learning_rate": 2.103646517739816e-05, "loss": 76.6324, "step": 1824 }, { "compression_loss": 72.79295349121094, "epoch": 0.6, "learning_rate": 2.1031537450722736e-05, "loss": 73.0672, "step": 1825 }, { "compression_loss": 75.30856323242188, "epoch": 0.6, "learning_rate": 2.1026609724047308e-05, "loss": 75.7325, "step": 1826 }, { "compression_loss": 76.21416473388672, "epoch": 0.6, "learning_rate": 2.102168199737188e-05, "loss": 76.4922, "step": 1827 }, { "compression_loss": 77.50270080566406, "epoch": 0.6, "learning_rate": 2.1016754270696452e-05, "loss": 78.2441, "step": 1828 }, { "compression_loss": 73.41825103759766, "epoch": 0.6, "learning_rate": 2.1011826544021024e-05, "loss": 73.7463, "step": 1829 }, { "compression_loss": 74.90390014648438, "epoch": 0.6, "learning_rate": 2.10068988173456e-05, "loss": 75.2344, "step": 1830 }, { "compression_loss": 74.37946319580078, "epoch": 0.6, "learning_rate": 2.100197109067017e-05, "loss": 74.7857, "step": 1831 }, { "compression_loss": 76.0509033203125, "epoch": 0.6, "learning_rate": 2.0997043363994743e-05, "loss": 76.3866, "step": 1832 }, { "compression_loss": 74.9376220703125, "epoch": 0.6, "learning_rate": 2.0992115637319315e-05, "loss": 75.4387, "step": 1833 }, { "compression_loss": 72.72439575195312, "epoch": 0.6, "learning_rate": 2.098718791064389e-05, "loss": 73.2218, "step": 1834 }, { "compression_loss": 77.85264587402344, "epoch": 0.6, "learning_rate": 2.0982260183968466e-05, "loss": 78.2042, "step": 1835 }, { "compression_loss": 78.20791625976562, "epoch": 0.6, "learning_rate": 2.0977332457293038e-05, "loss": 78.8759, "step": 1836 }, { "compression_loss": 77.18181610107422, "epoch": 0.6, "learning_rate": 2.097240473061761e-05, "loss": 77.6697, "step": 1837 }, { "compression_loss": 76.99442291259766, "epoch": 0.6, "learning_rate": 2.0967477003942182e-05, "loss": 77.7001, "step": 1838 }, { "compression_loss": 75.73155975341797, "epoch": 0.6, "learning_rate": 2.0962549277266754e-05, "loss": 76.0554, "step": 1839 }, { "compression_loss": 75.726806640625, "epoch": 0.6, "learning_rate": 2.095762155059133e-05, "loss": 76.2772, "step": 1840 }, { "compression_loss": 77.38151550292969, "epoch": 0.6, "learning_rate": 2.09526938239159e-05, "loss": 77.7018, "step": 1841 }, { "compression_loss": 76.44279479980469, "epoch": 0.61, "learning_rate": 2.0947766097240473e-05, "loss": 76.9236, "step": 1842 }, { "compression_loss": 76.65831756591797, "epoch": 0.61, "learning_rate": 2.0942838370565045e-05, "loss": 76.9453, "step": 1843 }, { "compression_loss": 75.30210876464844, "epoch": 0.61, "learning_rate": 2.0937910643889617e-05, "loss": 75.661, "step": 1844 }, { "compression_loss": 77.79039001464844, "epoch": 0.61, "learning_rate": 2.0932982917214193e-05, "loss": 78.1275, "step": 1845 }, { "compression_loss": 74.16572570800781, "epoch": 0.61, "learning_rate": 2.0928055190538768e-05, "loss": 74.8199, "step": 1846 }, { "compression_loss": 73.9644775390625, "epoch": 0.61, "learning_rate": 2.092312746386334e-05, "loss": 74.3121, "step": 1847 }, { "compression_loss": 74.27696228027344, "epoch": 0.61, "learning_rate": 2.0918199737187912e-05, "loss": 75.1842, "step": 1848 }, { "compression_loss": 75.39554595947266, "epoch": 0.61, "learning_rate": 2.0913272010512484e-05, "loss": 75.9712, "step": 1849 }, { "compression_loss": 74.69192504882812, "epoch": 0.61, "learning_rate": 2.090834428383706e-05, "loss": 75.0056, "step": 1850 }, { "compression_loss": 73.78857421875, "epoch": 0.61, "learning_rate": 2.090341655716163e-05, "loss": 74.2738, "step": 1851 }, { "compression_loss": 76.44668579101562, "epoch": 0.61, "learning_rate": 2.0898488830486203e-05, "loss": 77.1318, "step": 1852 }, { "compression_loss": 76.90381622314453, "epoch": 0.61, "learning_rate": 2.0893561103810775e-05, "loss": 77.4038, "step": 1853 }, { "compression_loss": 77.57475280761719, "epoch": 0.61, "learning_rate": 2.0888633377135347e-05, "loss": 77.9794, "step": 1854 }, { "compression_loss": 75.4014663696289, "epoch": 0.61, "learning_rate": 2.0883705650459922e-05, "loss": 75.8677, "step": 1855 }, { "compression_loss": 75.3317642211914, "epoch": 0.61, "learning_rate": 2.0878777923784494e-05, "loss": 75.6355, "step": 1856 }, { "compression_loss": 77.40216827392578, "epoch": 0.61, "learning_rate": 2.0873850197109066e-05, "loss": 77.7487, "step": 1857 }, { "compression_loss": 74.0274658203125, "epoch": 0.61, "learning_rate": 2.086892247043364e-05, "loss": 74.3163, "step": 1858 }, { "compression_loss": 75.55270385742188, "epoch": 0.61, "learning_rate": 2.0863994743758214e-05, "loss": 76.2197, "step": 1859 }, { "compression_loss": 76.59015655517578, "epoch": 0.61, "learning_rate": 2.085906701708279e-05, "loss": 76.8753, "step": 1860 }, { "compression_loss": 74.94790649414062, "epoch": 0.61, "learning_rate": 2.085413929040736e-05, "loss": 75.4071, "step": 1861 }, { "compression_loss": 75.38705444335938, "epoch": 0.61, "learning_rate": 2.0849211563731933e-05, "loss": 75.6602, "step": 1862 }, { "compression_loss": 76.67802429199219, "epoch": 0.61, "learning_rate": 2.0844283837056505e-05, "loss": 77.0065, "step": 1863 }, { "compression_loss": 77.45801544189453, "epoch": 0.61, "learning_rate": 2.0839356110381077e-05, "loss": 78.1143, "step": 1864 }, { "compression_loss": 75.41703796386719, "epoch": 0.61, "learning_rate": 2.0834428383705652e-05, "loss": 76.0878, "step": 1865 }, { "compression_loss": 77.2567138671875, "epoch": 0.61, "learning_rate": 2.0829500657030224e-05, "loss": 77.6408, "step": 1866 }, { "compression_loss": 77.28756713867188, "epoch": 0.61, "learning_rate": 2.0824572930354796e-05, "loss": 77.9372, "step": 1867 }, { "compression_loss": 74.67106628417969, "epoch": 0.61, "learning_rate": 2.0819645203679368e-05, "loss": 74.9781, "step": 1868 }, { "compression_loss": 75.0561752319336, "epoch": 0.61, "learning_rate": 2.081471747700394e-05, "loss": 75.126, "step": 1869 }, { "compression_loss": 75.21934509277344, "epoch": 0.61, "learning_rate": 2.080978975032852e-05, "loss": 75.7236, "step": 1870 }, { "compression_loss": 73.86114501953125, "epoch": 0.61, "learning_rate": 2.080486202365309e-05, "loss": 74.2796, "step": 1871 }, { "compression_loss": 75.35702514648438, "epoch": 0.61, "learning_rate": 2.0799934296977663e-05, "loss": 75.7399, "step": 1872 }, { "compression_loss": 77.22467041015625, "epoch": 0.62, "learning_rate": 2.0795006570302235e-05, "loss": 77.649, "step": 1873 }, { "compression_loss": 75.38542175292969, "epoch": 0.62, "learning_rate": 2.0790078843626807e-05, "loss": 75.5826, "step": 1874 }, { "compression_loss": 75.5643310546875, "epoch": 0.62, "learning_rate": 2.0785151116951382e-05, "loss": 76.0775, "step": 1875 }, { "compression_loss": 76.47431945800781, "epoch": 0.62, "learning_rate": 2.0780223390275954e-05, "loss": 76.9, "step": 1876 }, { "compression_loss": 74.10923767089844, "epoch": 0.62, "learning_rate": 2.0775295663600526e-05, "loss": 74.4956, "step": 1877 }, { "compression_loss": 73.8182373046875, "epoch": 0.62, "learning_rate": 2.0770367936925098e-05, "loss": 74.2353, "step": 1878 }, { "compression_loss": 74.93711853027344, "epoch": 0.62, "learning_rate": 2.076544021024967e-05, "loss": 75.3717, "step": 1879 }, { "compression_loss": 74.16873168945312, "epoch": 0.62, "learning_rate": 2.0760512483574245e-05, "loss": 74.7781, "step": 1880 }, { "compression_loss": 72.40654754638672, "epoch": 0.62, "learning_rate": 2.0755584756898817e-05, "loss": 72.6198, "step": 1881 }, { "compression_loss": 76.39607238769531, "epoch": 0.62, "learning_rate": 2.0750657030223393e-05, "loss": 76.786, "step": 1882 }, { "compression_loss": 74.58499145507812, "epoch": 0.62, "learning_rate": 2.0745729303547965e-05, "loss": 74.831, "step": 1883 }, { "compression_loss": 73.63877868652344, "epoch": 0.62, "learning_rate": 2.0740801576872537e-05, "loss": 73.8341, "step": 1884 }, { "compression_loss": 75.99854278564453, "epoch": 0.62, "learning_rate": 2.0735873850197112e-05, "loss": 76.2625, "step": 1885 }, { "compression_loss": 76.01081085205078, "epoch": 0.62, "learning_rate": 2.0730946123521684e-05, "loss": 76.2687, "step": 1886 }, { "compression_loss": 76.14561462402344, "epoch": 0.62, "learning_rate": 2.0726018396846256e-05, "loss": 76.5801, "step": 1887 }, { "compression_loss": 77.10679626464844, "epoch": 0.62, "learning_rate": 2.0721090670170828e-05, "loss": 77.7093, "step": 1888 }, { "compression_loss": 78.89474487304688, "epoch": 0.62, "learning_rate": 2.07161629434954e-05, "loss": 79.5207, "step": 1889 }, { "compression_loss": 74.91426086425781, "epoch": 0.62, "learning_rate": 2.0711235216819975e-05, "loss": 75.5363, "step": 1890 }, { "compression_loss": 76.4323501586914, "epoch": 0.62, "learning_rate": 2.0706307490144547e-05, "loss": 76.6433, "step": 1891 }, { "compression_loss": 75.19380187988281, "epoch": 0.62, "learning_rate": 2.070137976346912e-05, "loss": 75.5548, "step": 1892 }, { "compression_loss": 75.41204833984375, "epoch": 0.62, "learning_rate": 2.069645203679369e-05, "loss": 75.7155, "step": 1893 }, { "compression_loss": 76.32234954833984, "epoch": 0.62, "learning_rate": 2.0691524310118266e-05, "loss": 76.8016, "step": 1894 }, { "compression_loss": 76.27059936523438, "epoch": 0.62, "learning_rate": 2.0686596583442842e-05, "loss": 76.594, "step": 1895 }, { "compression_loss": 76.51066589355469, "epoch": 0.62, "learning_rate": 2.0681668856767414e-05, "loss": 76.9923, "step": 1896 }, { "compression_loss": 73.62735748291016, "epoch": 0.62, "learning_rate": 2.0676741130091986e-05, "loss": 73.9916, "step": 1897 }, { "compression_loss": 76.2170639038086, "epoch": 0.62, "learning_rate": 2.0671813403416558e-05, "loss": 76.6505, "step": 1898 }, { "compression_loss": 75.1943359375, "epoch": 0.62, "learning_rate": 2.066688567674113e-05, "loss": 75.4562, "step": 1899 }, { "compression_loss": 75.94800567626953, "epoch": 0.62, "learning_rate": 2.0661957950065705e-05, "loss": 76.239, "step": 1900 }, { "compression_loss": 76.47882843017578, "epoch": 0.62, "learning_rate": 2.0657030223390277e-05, "loss": 77.1074, "step": 1901 }, { "compression_loss": 74.15398406982422, "epoch": 0.62, "learning_rate": 2.065210249671485e-05, "loss": 74.5342, "step": 1902 }, { "compression_loss": 75.65422058105469, "epoch": 0.63, "learning_rate": 2.064717477003942e-05, "loss": 76.0653, "step": 1903 }, { "compression_loss": 76.29209899902344, "epoch": 0.63, "learning_rate": 2.0642247043363993e-05, "loss": 76.9584, "step": 1904 }, { "compression_loss": 71.93624114990234, "epoch": 0.63, "learning_rate": 2.0637319316688568e-05, "loss": 72.2119, "step": 1905 }, { "compression_loss": 76.66273498535156, "epoch": 0.63, "learning_rate": 2.0632391590013144e-05, "loss": 77.0999, "step": 1906 }, { "compression_loss": 76.46534729003906, "epoch": 0.63, "learning_rate": 2.0627463863337716e-05, "loss": 76.9812, "step": 1907 }, { "compression_loss": 74.52398681640625, "epoch": 0.63, "learning_rate": 2.0622536136662288e-05, "loss": 74.7726, "step": 1908 }, { "compression_loss": 76.32947540283203, "epoch": 0.63, "learning_rate": 2.061760840998686e-05, "loss": 76.6181, "step": 1909 }, { "compression_loss": 76.61486053466797, "epoch": 0.63, "learning_rate": 2.061268068331143e-05, "loss": 77.0807, "step": 1910 }, { "compression_loss": 75.603759765625, "epoch": 0.63, "learning_rate": 2.0607752956636007e-05, "loss": 75.9596, "step": 1911 }, { "compression_loss": 76.23245239257812, "epoch": 0.63, "learning_rate": 2.060282522996058e-05, "loss": 76.6002, "step": 1912 }, { "compression_loss": 75.05587768554688, "epoch": 0.63, "learning_rate": 2.059789750328515e-05, "loss": 75.3607, "step": 1913 }, { "compression_loss": 76.81143951416016, "epoch": 0.63, "learning_rate": 2.0592969776609723e-05, "loss": 77.3464, "step": 1914 }, { "compression_loss": 77.45478057861328, "epoch": 0.63, "learning_rate": 2.0588042049934295e-05, "loss": 77.9434, "step": 1915 }, { "compression_loss": 75.23643493652344, "epoch": 0.63, "learning_rate": 2.058311432325887e-05, "loss": 75.7597, "step": 1916 }, { "compression_loss": 77.6742935180664, "epoch": 0.63, "learning_rate": 2.0578186596583442e-05, "loss": 78.1668, "step": 1917 }, { "compression_loss": 76.71304321289062, "epoch": 0.63, "learning_rate": 2.0573258869908017e-05, "loss": 77.2383, "step": 1918 }, { "compression_loss": 75.35628509521484, "epoch": 0.63, "learning_rate": 2.056833114323259e-05, "loss": 75.7687, "step": 1919 }, { "compression_loss": 75.86167907714844, "epoch": 0.63, "learning_rate": 2.056340341655716e-05, "loss": 76.3259, "step": 1920 }, { "compression_loss": 76.97002410888672, "epoch": 0.63, "learning_rate": 2.0558475689881737e-05, "loss": 77.2364, "step": 1921 }, { "compression_loss": 74.05045318603516, "epoch": 0.63, "learning_rate": 2.055354796320631e-05, "loss": 74.4232, "step": 1922 }, { "compression_loss": 76.18231201171875, "epoch": 0.63, "learning_rate": 2.054862023653088e-05, "loss": 76.604, "step": 1923 }, { "compression_loss": 76.1355209350586, "epoch": 0.63, "learning_rate": 2.0543692509855453e-05, "loss": 76.5205, "step": 1924 }, { "compression_loss": 78.6526870727539, "epoch": 0.63, "learning_rate": 2.0538764783180025e-05, "loss": 79.2021, "step": 1925 }, { "compression_loss": 76.79264068603516, "epoch": 0.63, "learning_rate": 2.05338370565046e-05, "loss": 77.2963, "step": 1926 }, { "compression_loss": 75.76703643798828, "epoch": 0.63, "learning_rate": 2.0528909329829172e-05, "loss": 76.1822, "step": 1927 }, { "compression_loss": 76.42189025878906, "epoch": 0.63, "learning_rate": 2.0523981603153744e-05, "loss": 76.8534, "step": 1928 }, { "compression_loss": 75.21685791015625, "epoch": 0.63, "learning_rate": 2.0519053876478316e-05, "loss": 75.4847, "step": 1929 }, { "compression_loss": 76.04531860351562, "epoch": 0.63, "learning_rate": 2.051412614980289e-05, "loss": 76.4372, "step": 1930 }, { "compression_loss": 76.70574951171875, "epoch": 0.63, "learning_rate": 2.0509198423127467e-05, "loss": 77.1504, "step": 1931 }, { "compression_loss": 74.9390869140625, "epoch": 0.63, "learning_rate": 2.050427069645204e-05, "loss": 75.4801, "step": 1932 }, { "compression_loss": 76.49578857421875, "epoch": 0.64, "learning_rate": 2.049934296977661e-05, "loss": 76.9498, "step": 1933 }, { "compression_loss": 76.02902221679688, "epoch": 0.64, "learning_rate": 2.0494415243101182e-05, "loss": 76.4085, "step": 1934 }, { "compression_loss": 73.71177673339844, "epoch": 0.64, "learning_rate": 2.0489487516425754e-05, "loss": 74.0074, "step": 1935 }, { "compression_loss": 75.7916030883789, "epoch": 0.64, "learning_rate": 2.048455978975033e-05, "loss": 76.5524, "step": 1936 }, { "compression_loss": 76.06027221679688, "epoch": 0.64, "learning_rate": 2.0479632063074902e-05, "loss": 76.4131, "step": 1937 }, { "compression_loss": 75.09332275390625, "epoch": 0.64, "learning_rate": 2.0474704336399474e-05, "loss": 75.6163, "step": 1938 }, { "compression_loss": 78.04136657714844, "epoch": 0.64, "learning_rate": 2.0469776609724046e-05, "loss": 78.3408, "step": 1939 }, { "compression_loss": 74.80207824707031, "epoch": 0.64, "learning_rate": 2.0464848883048618e-05, "loss": 75.2315, "step": 1940 }, { "compression_loss": 74.2066650390625, "epoch": 0.64, "learning_rate": 2.0459921156373193e-05, "loss": 74.639, "step": 1941 }, { "compression_loss": 75.52996063232422, "epoch": 0.64, "learning_rate": 2.045499342969777e-05, "loss": 76.0711, "step": 1942 }, { "compression_loss": 76.72209167480469, "epoch": 0.64, "learning_rate": 2.045006570302234e-05, "loss": 77.079, "step": 1943 }, { "compression_loss": 73.78787231445312, "epoch": 0.64, "learning_rate": 2.0445137976346912e-05, "loss": 74.0365, "step": 1944 }, { "compression_loss": 75.0322265625, "epoch": 0.64, "learning_rate": 2.0440210249671484e-05, "loss": 75.2566, "step": 1945 }, { "compression_loss": 75.57068634033203, "epoch": 0.64, "learning_rate": 2.043528252299606e-05, "loss": 75.9247, "step": 1946 }, { "compression_loss": 74.65745544433594, "epoch": 0.64, "learning_rate": 2.043035479632063e-05, "loss": 74.8278, "step": 1947 }, { "compression_loss": 75.95213317871094, "epoch": 0.64, "learning_rate": 2.0425427069645204e-05, "loss": 76.4279, "step": 1948 }, { "compression_loss": 78.23578643798828, "epoch": 0.64, "learning_rate": 2.0420499342969776e-05, "loss": 78.6247, "step": 1949 }, { "compression_loss": 73.65308380126953, "epoch": 0.64, "learning_rate": 2.0415571616294348e-05, "loss": 73.9369, "step": 1950 }, { "compression_loss": 74.93989562988281, "epoch": 0.64, "learning_rate": 2.0410643889618923e-05, "loss": 75.1821, "step": 1951 }, { "compression_loss": 75.53411865234375, "epoch": 0.64, "learning_rate": 2.0405716162943495e-05, "loss": 75.8901, "step": 1952 }, { "compression_loss": 74.23777770996094, "epoch": 0.64, "learning_rate": 2.040078843626807e-05, "loss": 74.752, "step": 1953 }, { "compression_loss": 74.91738891601562, "epoch": 0.64, "learning_rate": 2.0395860709592642e-05, "loss": 75.1774, "step": 1954 }, { "compression_loss": 75.0574951171875, "epoch": 0.64, "learning_rate": 2.0390932982917214e-05, "loss": 75.4693, "step": 1955 }, { "compression_loss": 75.79580688476562, "epoch": 0.64, "learning_rate": 2.038600525624179e-05, "loss": 76.0644, "step": 1956 }, { "compression_loss": 78.12519073486328, "epoch": 0.64, "learning_rate": 2.038107752956636e-05, "loss": 78.6968, "step": 1957 }, { "compression_loss": 76.43771362304688, "epoch": 0.64, "learning_rate": 2.0376149802890933e-05, "loss": 76.8658, "step": 1958 }, { "compression_loss": 74.72689819335938, "epoch": 0.64, "learning_rate": 2.0371222076215505e-05, "loss": 75.1758, "step": 1959 }, { "compression_loss": 77.24043273925781, "epoch": 0.64, "learning_rate": 2.0366294349540077e-05, "loss": 78.0854, "step": 1960 }, { "compression_loss": 77.3773193359375, "epoch": 0.64, "learning_rate": 2.0361366622864653e-05, "loss": 77.953, "step": 1961 }, { "compression_loss": 77.59123229980469, "epoch": 0.64, "learning_rate": 2.0356438896189225e-05, "loss": 78.2415, "step": 1962 }, { "compression_loss": 77.4551010131836, "epoch": 0.64, "learning_rate": 2.0351511169513797e-05, "loss": 78.081, "step": 1963 }, { "compression_loss": 74.33566284179688, "epoch": 0.65, "learning_rate": 2.034658344283837e-05, "loss": 74.9258, "step": 1964 }, { "compression_loss": 75.29393768310547, "epoch": 0.65, "learning_rate": 2.0341655716162944e-05, "loss": 75.5608, "step": 1965 }, { "compression_loss": 77.88878631591797, "epoch": 0.65, "learning_rate": 2.033672798948752e-05, "loss": 78.4103, "step": 1966 }, { "compression_loss": 74.51048278808594, "epoch": 0.65, "learning_rate": 2.033180026281209e-05, "loss": 75.0265, "step": 1967 }, { "compression_loss": 78.38330841064453, "epoch": 0.65, "learning_rate": 2.0326872536136663e-05, "loss": 78.6959, "step": 1968 }, { "compression_loss": 74.57865142822266, "epoch": 0.65, "learning_rate": 2.0321944809461235e-05, "loss": 75.0404, "step": 1969 }, { "compression_loss": 73.20004272460938, "epoch": 0.65, "learning_rate": 2.0317017082785807e-05, "loss": 73.4094, "step": 1970 }, { "compression_loss": 79.10769653320312, "epoch": 0.65, "learning_rate": 2.0312089356110383e-05, "loss": 79.6029, "step": 1971 }, { "compression_loss": 74.43273162841797, "epoch": 0.65, "learning_rate": 2.0307161629434955e-05, "loss": 74.7147, "step": 1972 }, { "compression_loss": 76.96382141113281, "epoch": 0.65, "learning_rate": 2.0302233902759527e-05, "loss": 77.463, "step": 1973 }, { "compression_loss": 76.25477600097656, "epoch": 0.65, "learning_rate": 2.02973061760841e-05, "loss": 76.939, "step": 1974 }, { "compression_loss": 77.70050048828125, "epoch": 0.65, "learning_rate": 2.029237844940867e-05, "loss": 78.1534, "step": 1975 }, { "compression_loss": 76.36967468261719, "epoch": 0.65, "learning_rate": 2.0287450722733246e-05, "loss": 76.7776, "step": 1976 }, { "compression_loss": 77.4141845703125, "epoch": 0.65, "learning_rate": 2.028252299605782e-05, "loss": 77.8463, "step": 1977 }, { "compression_loss": 74.15812683105469, "epoch": 0.65, "learning_rate": 2.0277595269382393e-05, "loss": 74.4432, "step": 1978 }, { "compression_loss": 74.74708557128906, "epoch": 0.65, "learning_rate": 2.0272667542706965e-05, "loss": 75.1037, "step": 1979 }, { "compression_loss": 75.35286712646484, "epoch": 0.65, "learning_rate": 2.0267739816031537e-05, "loss": 75.8674, "step": 1980 }, { "compression_loss": 76.20592498779297, "epoch": 0.65, "learning_rate": 2.0262812089356112e-05, "loss": 76.7934, "step": 1981 }, { "compression_loss": 71.54249572753906, "epoch": 0.65, "learning_rate": 2.0257884362680684e-05, "loss": 71.8689, "step": 1982 }, { "compression_loss": 75.23481750488281, "epoch": 0.65, "learning_rate": 2.0252956636005256e-05, "loss": 75.6394, "step": 1983 }, { "compression_loss": 77.55015563964844, "epoch": 0.65, "learning_rate": 2.024802890932983e-05, "loss": 78.1863, "step": 1984 }, { "compression_loss": 75.71626281738281, "epoch": 0.65, "learning_rate": 2.02431011826544e-05, "loss": 76.0563, "step": 1985 }, { "compression_loss": 77.3556900024414, "epoch": 0.65, "learning_rate": 2.0238173455978976e-05, "loss": 77.7272, "step": 1986 }, { "compression_loss": 77.83049774169922, "epoch": 0.65, "learning_rate": 2.0233245729303548e-05, "loss": 78.3969, "step": 1987 }, { "compression_loss": 74.97361755371094, "epoch": 0.65, "learning_rate": 2.022831800262812e-05, "loss": 75.3137, "step": 1988 }, { "compression_loss": 76.38961791992188, "epoch": 0.65, "learning_rate": 2.0223390275952695e-05, "loss": 77.0667, "step": 1989 }, { "compression_loss": 73.9745864868164, "epoch": 0.65, "learning_rate": 2.0218462549277267e-05, "loss": 74.3508, "step": 1990 }, { "compression_loss": 76.00891876220703, "epoch": 0.65, "learning_rate": 2.0213534822601842e-05, "loss": 76.194, "step": 1991 }, { "compression_loss": 73.70856475830078, "epoch": 0.65, "learning_rate": 2.0208607095926414e-05, "loss": 74.0971, "step": 1992 }, { "compression_loss": 77.45050811767578, "epoch": 0.65, "learning_rate": 2.0203679369250986e-05, "loss": 77.8867, "step": 1993 }, { "compression_loss": 73.5611801147461, "epoch": 0.66, "learning_rate": 2.0198751642575558e-05, "loss": 73.8204, "step": 1994 }, { "compression_loss": 73.89698028564453, "epoch": 0.66, "learning_rate": 2.019382391590013e-05, "loss": 74.3582, "step": 1995 }, { "compression_loss": 76.04844665527344, "epoch": 0.66, "learning_rate": 2.0188896189224706e-05, "loss": 76.3701, "step": 1996 }, { "compression_loss": 75.65998840332031, "epoch": 0.66, "learning_rate": 2.0183968462549278e-05, "loss": 76.1293, "step": 1997 }, { "compression_loss": 76.32345581054688, "epoch": 0.66, "learning_rate": 2.017904073587385e-05, "loss": 76.6471, "step": 1998 }, { "compression_loss": 75.03227233886719, "epoch": 0.66, "learning_rate": 2.017411300919842e-05, "loss": 75.2614, "step": 1999 }, { "compression_loss": 76.73362731933594, "epoch": 0.66, "learning_rate": 2.0169185282522993e-05, "loss": 77.3208, "step": 2000 }, { "epoch": 0.66, "eval_exact_match": 86.29139072847683, "eval_f1": 92.50538373819413, "step": 2000 }, { "compression_loss": 74.88578796386719, "epoch": 0.66, "learning_rate": 2.0164257555847572e-05, "loss": 75.4185, "step": 2001 }, { "compression_loss": 75.21112060546875, "epoch": 0.66, "learning_rate": 2.0159329829172144e-05, "loss": 75.5445, "step": 2002 }, { "compression_loss": 74.72452545166016, "epoch": 0.66, "learning_rate": 2.0154402102496716e-05, "loss": 75.0682, "step": 2003 }, { "compression_loss": 75.505615234375, "epoch": 0.66, "learning_rate": 2.0149474375821288e-05, "loss": 75.8486, "step": 2004 }, { "compression_loss": 77.25303649902344, "epoch": 0.66, "learning_rate": 2.014454664914586e-05, "loss": 77.5906, "step": 2005 }, { "compression_loss": 74.56629943847656, "epoch": 0.66, "learning_rate": 2.0139618922470435e-05, "loss": 74.8639, "step": 2006 }, { "compression_loss": 77.17404174804688, "epoch": 0.66, "learning_rate": 2.0134691195795007e-05, "loss": 77.4957, "step": 2007 }, { "compression_loss": 75.78129577636719, "epoch": 0.66, "learning_rate": 2.012976346911958e-05, "loss": 76.2462, "step": 2008 }, { "compression_loss": 75.49705505371094, "epoch": 0.66, "learning_rate": 2.012483574244415e-05, "loss": 76.0023, "step": 2009 }, { "compression_loss": 76.09326934814453, "epoch": 0.66, "learning_rate": 2.0119908015768723e-05, "loss": 76.3201, "step": 2010 }, { "compression_loss": 77.51109313964844, "epoch": 0.66, "learning_rate": 2.01149802890933e-05, "loss": 78.2467, "step": 2011 }, { "compression_loss": 75.12763214111328, "epoch": 0.66, "learning_rate": 2.011005256241787e-05, "loss": 75.5509, "step": 2012 }, { "compression_loss": 74.49169921875, "epoch": 0.66, "learning_rate": 2.0105124835742446e-05, "loss": 74.8953, "step": 2013 }, { "compression_loss": 74.69638061523438, "epoch": 0.66, "learning_rate": 2.0100197109067018e-05, "loss": 75.1195, "step": 2014 }, { "compression_loss": 78.59827423095703, "epoch": 0.66, "learning_rate": 2.009526938239159e-05, "loss": 79.318, "step": 2015 }, { "compression_loss": 74.77079772949219, "epoch": 0.66, "learning_rate": 2.0090341655716165e-05, "loss": 75.2041, "step": 2016 }, { "compression_loss": 76.36074829101562, "epoch": 0.66, "learning_rate": 2.0085413929040737e-05, "loss": 76.7817, "step": 2017 }, { "compression_loss": 72.81227111816406, "epoch": 0.66, "learning_rate": 2.008048620236531e-05, "loss": 73.2562, "step": 2018 }, { "compression_loss": 76.36967468261719, "epoch": 0.66, "learning_rate": 2.007555847568988e-05, "loss": 76.7308, "step": 2019 }, { "compression_loss": 74.45895385742188, "epoch": 0.66, "learning_rate": 2.0070630749014453e-05, "loss": 74.7981, "step": 2020 }, { "compression_loss": 75.85382843017578, "epoch": 0.66, "learning_rate": 2.006570302233903e-05, "loss": 76.1989, "step": 2021 }, { "compression_loss": 75.58341217041016, "epoch": 0.66, "learning_rate": 2.00607752956636e-05, "loss": 76.0433, "step": 2022 }, { "compression_loss": 75.27838134765625, "epoch": 0.66, "learning_rate": 2.0055847568988172e-05, "loss": 75.7866, "step": 2023 }, { "compression_loss": 76.66521453857422, "epoch": 0.66, "learning_rate": 2.0050919842312744e-05, "loss": 77.1346, "step": 2024 }, { "compression_loss": 74.34486389160156, "epoch": 0.67, "learning_rate": 2.004599211563732e-05, "loss": 74.5669, "step": 2025 }, { "compression_loss": 77.38711547851562, "epoch": 0.67, "learning_rate": 2.0041064388961895e-05, "loss": 77.6706, "step": 2026 }, { "compression_loss": 76.06629943847656, "epoch": 0.67, "learning_rate": 2.0036136662286467e-05, "loss": 76.6026, "step": 2027 }, { "compression_loss": 74.96553039550781, "epoch": 0.67, "learning_rate": 2.003120893561104e-05, "loss": 75.2139, "step": 2028 }, { "compression_loss": 76.89330291748047, "epoch": 0.67, "learning_rate": 2.002628120893561e-05, "loss": 77.2719, "step": 2029 }, { "compression_loss": 75.20922088623047, "epoch": 0.67, "learning_rate": 2.0021353482260183e-05, "loss": 75.499, "step": 2030 }, { "compression_loss": 73.86817932128906, "epoch": 0.67, "learning_rate": 2.001642575558476e-05, "loss": 74.4393, "step": 2031 }, { "compression_loss": 76.39979553222656, "epoch": 0.67, "learning_rate": 2.001149802890933e-05, "loss": 76.632, "step": 2032 }, { "compression_loss": 76.25411987304688, "epoch": 0.67, "learning_rate": 2.0006570302233902e-05, "loss": 76.6613, "step": 2033 }, { "compression_loss": 76.45719909667969, "epoch": 0.67, "learning_rate": 2.0001642575558474e-05, "loss": 77.0217, "step": 2034 }, { "compression_loss": 76.50393676757812, "epoch": 0.67, "learning_rate": 1.9996714848883046e-05, "loss": 76.6548, "step": 2035 }, { "compression_loss": 77.11212158203125, "epoch": 0.67, "learning_rate": 1.999178712220762e-05, "loss": 77.382, "step": 2036 }, { "compression_loss": 77.09336853027344, "epoch": 0.67, "learning_rate": 1.9986859395532197e-05, "loss": 77.5796, "step": 2037 }, { "compression_loss": 74.14299011230469, "epoch": 0.67, "learning_rate": 1.998193166885677e-05, "loss": 74.2348, "step": 2038 }, { "compression_loss": 73.78987121582031, "epoch": 0.67, "learning_rate": 1.997700394218134e-05, "loss": 74.4363, "step": 2039 }, { "compression_loss": 75.61375427246094, "epoch": 0.67, "learning_rate": 1.9972076215505913e-05, "loss": 76.0303, "step": 2040 }, { "compression_loss": 77.27782440185547, "epoch": 0.67, "learning_rate": 1.9967148488830488e-05, "loss": 77.6947, "step": 2041 }, { "compression_loss": 73.60147857666016, "epoch": 0.67, "learning_rate": 1.996222076215506e-05, "loss": 74.251, "step": 2042 }, { "compression_loss": 73.5968017578125, "epoch": 0.67, "learning_rate": 1.9957293035479632e-05, "loss": 73.8089, "step": 2043 }, { "compression_loss": 78.25401306152344, "epoch": 0.67, "learning_rate": 1.9952365308804204e-05, "loss": 78.8099, "step": 2044 }, { "compression_loss": 77.7777099609375, "epoch": 0.67, "learning_rate": 1.9947437582128776e-05, "loss": 78.2474, "step": 2045 }, { "compression_loss": 76.65934753417969, "epoch": 0.67, "learning_rate": 1.994250985545335e-05, "loss": 77.3099, "step": 2046 }, { "compression_loss": 75.94932556152344, "epoch": 0.67, "learning_rate": 1.9937582128777923e-05, "loss": 76.0831, "step": 2047 }, { "compression_loss": 76.37418365478516, "epoch": 0.67, "learning_rate": 1.9932654402102495e-05, "loss": 76.7053, "step": 2048 }, { "compression_loss": 75.32997131347656, "epoch": 0.67, "learning_rate": 1.992772667542707e-05, "loss": 75.5991, "step": 2049 }, { "compression_loss": 76.4509506225586, "epoch": 0.67, "learning_rate": 1.9922798948751643e-05, "loss": 76.7221, "step": 2050 }, { "compression_loss": 75.01995849609375, "epoch": 0.67, "learning_rate": 1.9917871222076218e-05, "loss": 75.2782, "step": 2051 }, { "compression_loss": 74.90457153320312, "epoch": 0.67, "learning_rate": 1.991294349540079e-05, "loss": 75.1711, "step": 2052 }, { "compression_loss": 74.69131469726562, "epoch": 0.67, "learning_rate": 1.9908015768725362e-05, "loss": 74.9941, "step": 2053 }, { "compression_loss": 73.89149475097656, "epoch": 0.67, "learning_rate": 1.9903088042049934e-05, "loss": 74.3172, "step": 2054 }, { "compression_loss": 74.89460754394531, "epoch": 0.68, "learning_rate": 1.9898160315374506e-05, "loss": 75.4104, "step": 2055 }, { "compression_loss": 72.69834899902344, "epoch": 0.68, "learning_rate": 1.989323258869908e-05, "loss": 73.1544, "step": 2056 }, { "compression_loss": 73.78819274902344, "epoch": 0.68, "learning_rate": 1.9888304862023653e-05, "loss": 74.0984, "step": 2057 }, { "compression_loss": 73.6429672241211, "epoch": 0.68, "learning_rate": 1.9883377135348225e-05, "loss": 74.3351, "step": 2058 }, { "compression_loss": 75.89401245117188, "epoch": 0.68, "learning_rate": 1.9878449408672797e-05, "loss": 76.4067, "step": 2059 }, { "compression_loss": 73.92329406738281, "epoch": 0.68, "learning_rate": 1.987352168199737e-05, "loss": 74.4718, "step": 2060 }, { "compression_loss": 74.67826080322266, "epoch": 0.68, "learning_rate": 1.9868593955321948e-05, "loss": 74.9339, "step": 2061 }, { "compression_loss": 75.9259262084961, "epoch": 0.68, "learning_rate": 1.986366622864652e-05, "loss": 76.3204, "step": 2062 }, { "compression_loss": 76.1596908569336, "epoch": 0.68, "learning_rate": 1.9858738501971092e-05, "loss": 76.5336, "step": 2063 }, { "compression_loss": 77.18843078613281, "epoch": 0.68, "learning_rate": 1.9853810775295664e-05, "loss": 77.6811, "step": 2064 }, { "compression_loss": 72.50721740722656, "epoch": 0.68, "learning_rate": 1.9848883048620236e-05, "loss": 72.6487, "step": 2065 }, { "compression_loss": 77.026123046875, "epoch": 0.68, "learning_rate": 1.984395532194481e-05, "loss": 77.6883, "step": 2066 }, { "compression_loss": 76.67431640625, "epoch": 0.68, "learning_rate": 1.9839027595269383e-05, "loss": 76.9711, "step": 2067 }, { "compression_loss": 76.13371276855469, "epoch": 0.68, "learning_rate": 1.9834099868593955e-05, "loss": 76.6284, "step": 2068 }, { "compression_loss": 77.02962493896484, "epoch": 0.68, "learning_rate": 1.9829172141918527e-05, "loss": 77.911, "step": 2069 }, { "compression_loss": 78.06982421875, "epoch": 0.68, "learning_rate": 1.98242444152431e-05, "loss": 78.8086, "step": 2070 }, { "compression_loss": 75.21781158447266, "epoch": 0.68, "learning_rate": 1.9819316688567674e-05, "loss": 75.6799, "step": 2071 }, { "compression_loss": 74.237548828125, "epoch": 0.68, "learning_rate": 1.9814388961892246e-05, "loss": 74.8212, "step": 2072 }, { "compression_loss": 76.18905639648438, "epoch": 0.68, "learning_rate": 1.9809461235216822e-05, "loss": 76.6855, "step": 2073 }, { "compression_loss": 75.49847412109375, "epoch": 0.68, "learning_rate": 1.9804533508541394e-05, "loss": 75.9197, "step": 2074 }, { "compression_loss": 75.54096984863281, "epoch": 0.68, "learning_rate": 1.9799605781865966e-05, "loss": 75.9798, "step": 2075 }, { "compression_loss": 75.09483337402344, "epoch": 0.68, "learning_rate": 1.979467805519054e-05, "loss": 75.4026, "step": 2076 }, { "compression_loss": 75.6442642211914, "epoch": 0.68, "learning_rate": 1.9789750328515113e-05, "loss": 75.9576, "step": 2077 }, { "compression_loss": 77.10155487060547, "epoch": 0.68, "learning_rate": 1.9784822601839685e-05, "loss": 77.8171, "step": 2078 }, { "compression_loss": 75.96160888671875, "epoch": 0.68, "learning_rate": 1.9779894875164257e-05, "loss": 76.2866, "step": 2079 }, { "compression_loss": 75.18626403808594, "epoch": 0.68, "learning_rate": 1.977496714848883e-05, "loss": 75.4984, "step": 2080 }, { "compression_loss": 76.02721405029297, "epoch": 0.68, "learning_rate": 1.9770039421813404e-05, "loss": 76.6407, "step": 2081 }, { "compression_loss": 76.0082015991211, "epoch": 0.68, "learning_rate": 1.9765111695137976e-05, "loss": 76.4184, "step": 2082 }, { "compression_loss": 76.4130630493164, "epoch": 0.68, "learning_rate": 1.9760183968462548e-05, "loss": 77.2012, "step": 2083 }, { "compression_loss": 76.28345489501953, "epoch": 0.68, "learning_rate": 1.9755256241787124e-05, "loss": 76.7215, "step": 2084 }, { "compression_loss": 75.67054748535156, "epoch": 0.68, "learning_rate": 1.9750328515111696e-05, "loss": 76.1668, "step": 2085 }, { "compression_loss": 73.70074462890625, "epoch": 0.69, "learning_rate": 1.974540078843627e-05, "loss": 74.0153, "step": 2086 }, { "compression_loss": 78.14967346191406, "epoch": 0.69, "learning_rate": 1.9740473061760843e-05, "loss": 78.5192, "step": 2087 }, { "compression_loss": 76.69244384765625, "epoch": 0.69, "learning_rate": 1.9735545335085415e-05, "loss": 77.2346, "step": 2088 }, { "compression_loss": 76.56510162353516, "epoch": 0.69, "learning_rate": 1.9730617608409987e-05, "loss": 77.4274, "step": 2089 }, { "compression_loss": 77.25199890136719, "epoch": 0.69, "learning_rate": 1.972568988173456e-05, "loss": 77.7728, "step": 2090 }, { "compression_loss": 76.89029693603516, "epoch": 0.69, "learning_rate": 1.9720762155059134e-05, "loss": 77.2272, "step": 2091 }, { "compression_loss": 72.3401107788086, "epoch": 0.69, "learning_rate": 1.9715834428383706e-05, "loss": 72.8092, "step": 2092 }, { "compression_loss": 73.89888000488281, "epoch": 0.69, "learning_rate": 1.9710906701708278e-05, "loss": 74.4648, "step": 2093 }, { "compression_loss": 75.57884216308594, "epoch": 0.69, "learning_rate": 1.970597897503285e-05, "loss": 76.1513, "step": 2094 }, { "compression_loss": 75.61319732666016, "epoch": 0.69, "learning_rate": 1.9701051248357422e-05, "loss": 76.0072, "step": 2095 }, { "compression_loss": 77.35038757324219, "epoch": 0.69, "learning_rate": 1.9696123521682e-05, "loss": 77.6213, "step": 2096 }, { "compression_loss": 75.5692138671875, "epoch": 0.69, "learning_rate": 1.9691195795006573e-05, "loss": 75.9297, "step": 2097 }, { "compression_loss": 73.60284423828125, "epoch": 0.69, "learning_rate": 1.9686268068331145e-05, "loss": 73.9328, "step": 2098 }, { "compression_loss": 75.42228698730469, "epoch": 0.69, "learning_rate": 1.9681340341655717e-05, "loss": 75.7213, "step": 2099 }, { "compression_loss": 76.36540222167969, "epoch": 0.69, "learning_rate": 1.967641261498029e-05, "loss": 76.804, "step": 2100 }, { "compression_loss": 74.7614517211914, "epoch": 0.69, "learning_rate": 1.9671484888304864e-05, "loss": 74.9176, "step": 2101 }, { "compression_loss": 74.04485321044922, "epoch": 0.69, "learning_rate": 1.9666557161629436e-05, "loss": 74.3691, "step": 2102 }, { "compression_loss": 76.11412811279297, "epoch": 0.69, "learning_rate": 1.9661629434954008e-05, "loss": 76.6195, "step": 2103 }, { "compression_loss": 75.87068939208984, "epoch": 0.69, "learning_rate": 1.965670170827858e-05, "loss": 76.3417, "step": 2104 }, { "compression_loss": 76.83189392089844, "epoch": 0.69, "learning_rate": 1.9651773981603152e-05, "loss": 77.5745, "step": 2105 }, { "compression_loss": 75.04769897460938, "epoch": 0.69, "learning_rate": 1.9646846254927727e-05, "loss": 75.3809, "step": 2106 }, { "compression_loss": 73.56455993652344, "epoch": 0.69, "learning_rate": 1.96419185282523e-05, "loss": 73.9137, "step": 2107 }, { "compression_loss": 74.88691711425781, "epoch": 0.69, "learning_rate": 1.9636990801576875e-05, "loss": 75.3632, "step": 2108 }, { "compression_loss": 77.180419921875, "epoch": 0.69, "learning_rate": 1.9632063074901446e-05, "loss": 77.7441, "step": 2109 }, { "compression_loss": 75.62344360351562, "epoch": 0.69, "learning_rate": 1.962713534822602e-05, "loss": 76.036, "step": 2110 }, { "compression_loss": 74.58877563476562, "epoch": 0.69, "learning_rate": 1.9622207621550594e-05, "loss": 75.0455, "step": 2111 }, { "compression_loss": 76.58628845214844, "epoch": 0.69, "learning_rate": 1.9617279894875166e-05, "loss": 76.9055, "step": 2112 }, { "compression_loss": 75.83294677734375, "epoch": 0.69, "learning_rate": 1.9612352168199738e-05, "loss": 76.2823, "step": 2113 }, { "compression_loss": 76.10084533691406, "epoch": 0.69, "learning_rate": 1.960742444152431e-05, "loss": 76.5398, "step": 2114 }, { "compression_loss": 75.36026000976562, "epoch": 0.69, "learning_rate": 1.960249671484888e-05, "loss": 75.8219, "step": 2115 }, { "compression_loss": 75.07198333740234, "epoch": 0.7, "learning_rate": 1.9597568988173457e-05, "loss": 75.6694, "step": 2116 }, { "compression_loss": 73.72782897949219, "epoch": 0.7, "learning_rate": 1.959264126149803e-05, "loss": 73.9481, "step": 2117 }, { "compression_loss": 75.8790283203125, "epoch": 0.7, "learning_rate": 1.95877135348226e-05, "loss": 76.2775, "step": 2118 }, { "compression_loss": 74.41297912597656, "epoch": 0.7, "learning_rate": 1.9582785808147173e-05, "loss": 74.8676, "step": 2119 }, { "compression_loss": 76.50502014160156, "epoch": 0.7, "learning_rate": 1.9577858081471748e-05, "loss": 76.8019, "step": 2120 }, { "compression_loss": 75.2585678100586, "epoch": 0.7, "learning_rate": 1.9572930354796324e-05, "loss": 75.7647, "step": 2121 }, { "compression_loss": 76.25411987304688, "epoch": 0.7, "learning_rate": 1.9568002628120896e-05, "loss": 76.8663, "step": 2122 }, { "compression_loss": 74.40669250488281, "epoch": 0.7, "learning_rate": 1.9563074901445468e-05, "loss": 74.8956, "step": 2123 }, { "compression_loss": 72.43794250488281, "epoch": 0.7, "learning_rate": 1.955814717477004e-05, "loss": 72.7328, "step": 2124 }, { "compression_loss": 74.23696899414062, "epoch": 0.7, "learning_rate": 1.955321944809461e-05, "loss": 74.5253, "step": 2125 }, { "compression_loss": 76.14041900634766, "epoch": 0.7, "learning_rate": 1.9548291721419187e-05, "loss": 76.51, "step": 2126 }, { "compression_loss": 75.77576446533203, "epoch": 0.7, "learning_rate": 1.954336399474376e-05, "loss": 76.1984, "step": 2127 }, { "compression_loss": 74.5972900390625, "epoch": 0.7, "learning_rate": 1.953843626806833e-05, "loss": 74.9366, "step": 2128 }, { "compression_loss": 75.36349487304688, "epoch": 0.7, "learning_rate": 1.9533508541392903e-05, "loss": 75.9183, "step": 2129 }, { "compression_loss": 74.92723083496094, "epoch": 0.7, "learning_rate": 1.9528580814717475e-05, "loss": 75.3223, "step": 2130 }, { "compression_loss": 75.33657836914062, "epoch": 0.7, "learning_rate": 1.952365308804205e-05, "loss": 75.6342, "step": 2131 }, { "compression_loss": 74.00248718261719, "epoch": 0.7, "learning_rate": 1.9518725361366625e-05, "loss": 74.453, "step": 2132 }, { "compression_loss": 75.07777404785156, "epoch": 0.7, "learning_rate": 1.9513797634691197e-05, "loss": 75.3861, "step": 2133 }, { "compression_loss": 75.96086120605469, "epoch": 0.7, "learning_rate": 1.950886990801577e-05, "loss": 76.2954, "step": 2134 }, { "compression_loss": 77.27047729492188, "epoch": 0.7, "learning_rate": 1.950394218134034e-05, "loss": 77.8017, "step": 2135 }, { "compression_loss": 76.16975402832031, "epoch": 0.7, "learning_rate": 1.9499014454664917e-05, "loss": 76.6331, "step": 2136 }, { "compression_loss": 75.6307601928711, "epoch": 0.7, "learning_rate": 1.949408672798949e-05, "loss": 75.9755, "step": 2137 }, { "compression_loss": 75.05058288574219, "epoch": 0.7, "learning_rate": 1.948915900131406e-05, "loss": 75.6341, "step": 2138 }, { "compression_loss": 71.72218322753906, "epoch": 0.7, "learning_rate": 1.9484231274638633e-05, "loss": 71.91, "step": 2139 }, { "compression_loss": 76.47039794921875, "epoch": 0.7, "learning_rate": 1.9479303547963205e-05, "loss": 76.6527, "step": 2140 }, { "compression_loss": 78.38096618652344, "epoch": 0.7, "learning_rate": 1.947437582128778e-05, "loss": 78.7084, "step": 2141 }, { "compression_loss": 76.18441772460938, "epoch": 0.7, "learning_rate": 1.9469448094612352e-05, "loss": 76.5628, "step": 2142 }, { "compression_loss": 77.1583251953125, "epoch": 0.7, "learning_rate": 1.9464520367936924e-05, "loss": 77.8318, "step": 2143 }, { "compression_loss": 74.53269958496094, "epoch": 0.7, "learning_rate": 1.94595926412615e-05, "loss": 75.1155, "step": 2144 }, { "compression_loss": 75.32971954345703, "epoch": 0.7, "learning_rate": 1.945466491458607e-05, "loss": 75.6302, "step": 2145 }, { "compression_loss": 78.35045623779297, "epoch": 0.7, "learning_rate": 1.9449737187910647e-05, "loss": 78.6964, "step": 2146 }, { "compression_loss": 75.0120849609375, "epoch": 0.71, "learning_rate": 1.944480946123522e-05, "loss": 75.2961, "step": 2147 }, { "compression_loss": 75.16754913330078, "epoch": 0.71, "learning_rate": 1.943988173455979e-05, "loss": 75.5066, "step": 2148 }, { "compression_loss": 76.4659652709961, "epoch": 0.71, "learning_rate": 1.9434954007884363e-05, "loss": 77.0243, "step": 2149 }, { "compression_loss": 76.04177856445312, "epoch": 0.71, "learning_rate": 1.9430026281208934e-05, "loss": 76.7655, "step": 2150 }, { "compression_loss": 75.82689666748047, "epoch": 0.71, "learning_rate": 1.942509855453351e-05, "loss": 76.6443, "step": 2151 }, { "compression_loss": 75.77635192871094, "epoch": 0.71, "learning_rate": 1.9420170827858082e-05, "loss": 76.403, "step": 2152 }, { "compression_loss": 75.08858489990234, "epoch": 0.71, "learning_rate": 1.9415243101182654e-05, "loss": 75.3626, "step": 2153 }, { "compression_loss": 75.7988510131836, "epoch": 0.71, "learning_rate": 1.9410315374507226e-05, "loss": 76.1727, "step": 2154 }, { "compression_loss": 74.6719970703125, "epoch": 0.71, "learning_rate": 1.9405387647831798e-05, "loss": 75.092, "step": 2155 }, { "compression_loss": 76.24285888671875, "epoch": 0.71, "learning_rate": 1.9400459921156376e-05, "loss": 76.472, "step": 2156 }, { "compression_loss": 76.67402648925781, "epoch": 0.71, "learning_rate": 1.939553219448095e-05, "loss": 77.1364, "step": 2157 }, { "compression_loss": 74.26083374023438, "epoch": 0.71, "learning_rate": 1.939060446780552e-05, "loss": 74.9443, "step": 2158 }, { "compression_loss": 75.52238464355469, "epoch": 0.71, "learning_rate": 1.9385676741130092e-05, "loss": 76.0895, "step": 2159 }, { "compression_loss": 75.42813110351562, "epoch": 0.71, "learning_rate": 1.9380749014454664e-05, "loss": 75.6282, "step": 2160 }, { "compression_loss": 76.36378479003906, "epoch": 0.71, "learning_rate": 1.937582128777924e-05, "loss": 76.8145, "step": 2161 }, { "compression_loss": 77.132080078125, "epoch": 0.71, "learning_rate": 1.937089356110381e-05, "loss": 77.7737, "step": 2162 }, { "compression_loss": 71.85513305664062, "epoch": 0.71, "learning_rate": 1.9365965834428384e-05, "loss": 72.159, "step": 2163 }, { "compression_loss": 75.19483947753906, "epoch": 0.71, "learning_rate": 1.9361038107752956e-05, "loss": 75.6988, "step": 2164 }, { "compression_loss": 75.57731628417969, "epoch": 0.71, "learning_rate": 1.9356110381077528e-05, "loss": 76.1968, "step": 2165 }, { "compression_loss": 77.88420104980469, "epoch": 0.71, "learning_rate": 1.9351182654402103e-05, "loss": 78.3437, "step": 2166 }, { "compression_loss": 76.89801788330078, "epoch": 0.71, "learning_rate": 1.9346254927726675e-05, "loss": 77.1915, "step": 2167 }, { "compression_loss": 73.1663818359375, "epoch": 0.71, "learning_rate": 1.934132720105125e-05, "loss": 73.5096, "step": 2168 }, { "compression_loss": 75.9994125366211, "epoch": 0.71, "learning_rate": 1.9336399474375822e-05, "loss": 76.4635, "step": 2169 }, { "compression_loss": 75.12471008300781, "epoch": 0.71, "learning_rate": 1.9331471747700394e-05, "loss": 75.4242, "step": 2170 }, { "compression_loss": 78.99937438964844, "epoch": 0.71, "learning_rate": 1.932654402102497e-05, "loss": 79.4844, "step": 2171 }, { "compression_loss": 76.78019714355469, "epoch": 0.71, "learning_rate": 1.932161629434954e-05, "loss": 77.1797, "step": 2172 }, { "compression_loss": 76.74423217773438, "epoch": 0.71, "learning_rate": 1.9316688567674113e-05, "loss": 77.1293, "step": 2173 }, { "compression_loss": 73.95800018310547, "epoch": 0.71, "learning_rate": 1.9311760840998685e-05, "loss": 74.1218, "step": 2174 }, { "compression_loss": 77.6907958984375, "epoch": 0.71, "learning_rate": 1.9306833114323257e-05, "loss": 77.9132, "step": 2175 }, { "compression_loss": 77.3897705078125, "epoch": 0.71, "learning_rate": 1.9301905387647833e-05, "loss": 78.2414, "step": 2176 }, { "compression_loss": 77.30064392089844, "epoch": 0.72, "learning_rate": 1.9296977660972405e-05, "loss": 77.9686, "step": 2177 }, { "compression_loss": 75.1585693359375, "epoch": 0.72, "learning_rate": 1.9292049934296977e-05, "loss": 75.6487, "step": 2178 }, { "compression_loss": 74.61688995361328, "epoch": 0.72, "learning_rate": 1.928712220762155e-05, "loss": 75.4686, "step": 2179 }, { "compression_loss": 74.64148712158203, "epoch": 0.72, "learning_rate": 1.9282194480946124e-05, "loss": 75.0581, "step": 2180 }, { "compression_loss": 76.18013000488281, "epoch": 0.72, "learning_rate": 1.92772667542707e-05, "loss": 76.6296, "step": 2181 }, { "compression_loss": 76.30931854248047, "epoch": 0.72, "learning_rate": 1.927233902759527e-05, "loss": 76.8101, "step": 2182 }, { "compression_loss": 75.89542388916016, "epoch": 0.72, "learning_rate": 1.9267411300919843e-05, "loss": 76.2529, "step": 2183 }, { "compression_loss": 76.95632934570312, "epoch": 0.72, "learning_rate": 1.9262483574244415e-05, "loss": 77.4426, "step": 2184 }, { "compression_loss": 77.43508911132812, "epoch": 0.72, "learning_rate": 1.9257555847568987e-05, "loss": 77.6318, "step": 2185 }, { "compression_loss": 78.6867904663086, "epoch": 0.72, "learning_rate": 1.9252628120893563e-05, "loss": 79.3264, "step": 2186 }, { "compression_loss": 76.3311767578125, "epoch": 0.72, "learning_rate": 1.9247700394218135e-05, "loss": 76.9789, "step": 2187 }, { "compression_loss": 76.24758911132812, "epoch": 0.72, "learning_rate": 1.9242772667542707e-05, "loss": 76.4631, "step": 2188 }, { "compression_loss": 74.955810546875, "epoch": 0.72, "learning_rate": 1.923784494086728e-05, "loss": 75.3407, "step": 2189 }, { "compression_loss": 75.0875015258789, "epoch": 0.72, "learning_rate": 1.923291721419185e-05, "loss": 75.6793, "step": 2190 }, { "compression_loss": 76.41549682617188, "epoch": 0.72, "learning_rate": 1.9227989487516426e-05, "loss": 76.9295, "step": 2191 }, { "compression_loss": 74.00315856933594, "epoch": 0.72, "learning_rate": 1.9223061760841e-05, "loss": 74.558, "step": 2192 }, { "compression_loss": 74.672607421875, "epoch": 0.72, "learning_rate": 1.9218134034165573e-05, "loss": 75.0816, "step": 2193 }, { "compression_loss": 75.62214660644531, "epoch": 0.72, "learning_rate": 1.9213206307490145e-05, "loss": 76.0901, "step": 2194 }, { "compression_loss": 77.34127807617188, "epoch": 0.72, "learning_rate": 1.9208278580814717e-05, "loss": 77.9833, "step": 2195 }, { "compression_loss": 74.34689331054688, "epoch": 0.72, "learning_rate": 1.9203350854139292e-05, "loss": 74.7188, "step": 2196 }, { "compression_loss": 78.66734313964844, "epoch": 0.72, "learning_rate": 1.9198423127463864e-05, "loss": 79.1176, "step": 2197 }, { "compression_loss": 74.76720428466797, "epoch": 0.72, "learning_rate": 1.9193495400788436e-05, "loss": 75.0893, "step": 2198 }, { "compression_loss": 77.1739273071289, "epoch": 0.72, "learning_rate": 1.918856767411301e-05, "loss": 77.8604, "step": 2199 }, { "compression_loss": 76.10232543945312, "epoch": 0.72, "learning_rate": 1.918363994743758e-05, "loss": 76.4611, "step": 2200 }, { "compression_loss": 77.92611694335938, "epoch": 0.72, "learning_rate": 1.9178712220762156e-05, "loss": 78.4515, "step": 2201 }, { "compression_loss": 76.54701232910156, "epoch": 0.72, "learning_rate": 1.9173784494086728e-05, "loss": 76.9201, "step": 2202 }, { "compression_loss": 76.01858520507812, "epoch": 0.72, "learning_rate": 1.91688567674113e-05, "loss": 76.5611, "step": 2203 }, { "compression_loss": 76.17073059082031, "epoch": 0.72, "learning_rate": 1.9163929040735875e-05, "loss": 76.7453, "step": 2204 }, { "compression_loss": 75.02337646484375, "epoch": 0.72, "learning_rate": 1.9159001314060447e-05, "loss": 75.4122, "step": 2205 }, { "compression_loss": 75.76466369628906, "epoch": 0.72, "learning_rate": 1.9154073587385022e-05, "loss": 76.0064, "step": 2206 }, { "compression_loss": 76.77720642089844, "epoch": 0.73, "learning_rate": 1.9149145860709594e-05, "loss": 77.0933, "step": 2207 }, { "compression_loss": 75.30123138427734, "epoch": 0.73, "learning_rate": 1.9144218134034166e-05, "loss": 75.5011, "step": 2208 }, { "compression_loss": 74.99249267578125, "epoch": 0.73, "learning_rate": 1.9139290407358738e-05, "loss": 75.2851, "step": 2209 }, { "compression_loss": 76.20291900634766, "epoch": 0.73, "learning_rate": 1.913436268068331e-05, "loss": 76.4953, "step": 2210 }, { "compression_loss": 76.38334655761719, "epoch": 0.73, "learning_rate": 1.9129434954007886e-05, "loss": 77.0968, "step": 2211 }, { "compression_loss": 75.66239929199219, "epoch": 0.73, "learning_rate": 1.9124507227332458e-05, "loss": 76.3229, "step": 2212 }, { "compression_loss": 76.04246520996094, "epoch": 0.73, "learning_rate": 1.911957950065703e-05, "loss": 76.423, "step": 2213 }, { "compression_loss": 73.76502990722656, "epoch": 0.73, "learning_rate": 1.91146517739816e-05, "loss": 73.9962, "step": 2214 }, { "compression_loss": 75.33004760742188, "epoch": 0.73, "learning_rate": 1.9109724047306177e-05, "loss": 75.7384, "step": 2215 }, { "compression_loss": 75.39249420166016, "epoch": 0.73, "learning_rate": 1.9104796320630752e-05, "loss": 76.065, "step": 2216 }, { "compression_loss": 75.65975952148438, "epoch": 0.73, "learning_rate": 1.9099868593955324e-05, "loss": 76.4523, "step": 2217 }, { "compression_loss": 77.2260971069336, "epoch": 0.73, "learning_rate": 1.9094940867279896e-05, "loss": 77.4487, "step": 2218 }, { "compression_loss": 77.84190368652344, "epoch": 0.73, "learning_rate": 1.9090013140604468e-05, "loss": 78.2435, "step": 2219 }, { "compression_loss": 76.65557861328125, "epoch": 0.73, "learning_rate": 1.908508541392904e-05, "loss": 76.9464, "step": 2220 }, { "compression_loss": 74.60871887207031, "epoch": 0.73, "learning_rate": 1.9080157687253615e-05, "loss": 74.9646, "step": 2221 }, { "compression_loss": 74.7629165649414, "epoch": 0.73, "learning_rate": 1.9075229960578187e-05, "loss": 75.2669, "step": 2222 }, { "compression_loss": 73.33627319335938, "epoch": 0.73, "learning_rate": 1.907030223390276e-05, "loss": 73.7446, "step": 2223 }, { "compression_loss": 75.53285217285156, "epoch": 0.73, "learning_rate": 1.906537450722733e-05, "loss": 75.9138, "step": 2224 }, { "compression_loss": 75.72639465332031, "epoch": 0.73, "learning_rate": 1.9060446780551903e-05, "loss": 75.9847, "step": 2225 }, { "compression_loss": 74.48011779785156, "epoch": 0.73, "learning_rate": 1.905551905387648e-05, "loss": 74.7151, "step": 2226 }, { "compression_loss": 76.37004852294922, "epoch": 0.73, "learning_rate": 1.9050591327201054e-05, "loss": 77.3756, "step": 2227 }, { "compression_loss": 74.28805541992188, "epoch": 0.73, "learning_rate": 1.9045663600525626e-05, "loss": 74.8571, "step": 2228 }, { "compression_loss": 75.96858215332031, "epoch": 0.73, "learning_rate": 1.9040735873850198e-05, "loss": 76.202, "step": 2229 }, { "compression_loss": 75.9226303100586, "epoch": 0.73, "learning_rate": 1.903580814717477e-05, "loss": 76.3584, "step": 2230 }, { "compression_loss": 77.44743347167969, "epoch": 0.73, "learning_rate": 1.9030880420499345e-05, "loss": 77.6162, "step": 2231 }, { "compression_loss": 75.25776672363281, "epoch": 0.73, "learning_rate": 1.9025952693823917e-05, "loss": 75.8427, "step": 2232 }, { "compression_loss": 75.33706665039062, "epoch": 0.73, "learning_rate": 1.902102496714849e-05, "loss": 75.5424, "step": 2233 }, { "compression_loss": 73.90570068359375, "epoch": 0.73, "learning_rate": 1.901609724047306e-05, "loss": 74.2153, "step": 2234 }, { "compression_loss": 75.22080993652344, "epoch": 0.73, "learning_rate": 1.9011169513797633e-05, "loss": 75.6115, "step": 2235 }, { "compression_loss": 76.07656860351562, "epoch": 0.73, "learning_rate": 1.900624178712221e-05, "loss": 76.6433, "step": 2236 }, { "compression_loss": 73.89393615722656, "epoch": 0.73, "learning_rate": 1.900131406044678e-05, "loss": 74.1037, "step": 2237 }, { "compression_loss": 75.13722229003906, "epoch": 0.74, "learning_rate": 1.8996386333771352e-05, "loss": 75.5153, "step": 2238 }, { "compression_loss": 74.8584976196289, "epoch": 0.74, "learning_rate": 1.8991458607095928e-05, "loss": 75.4412, "step": 2239 }, { "compression_loss": 74.50373840332031, "epoch": 0.74, "learning_rate": 1.89865308804205e-05, "loss": 74.816, "step": 2240 }, { "compression_loss": 75.90113067626953, "epoch": 0.74, "learning_rate": 1.8981603153745075e-05, "loss": 76.5152, "step": 2241 }, { "compression_loss": 75.6273193359375, "epoch": 0.74, "learning_rate": 1.8976675427069647e-05, "loss": 75.8633, "step": 2242 }, { "compression_loss": 76.38385009765625, "epoch": 0.74, "learning_rate": 1.897174770039422e-05, "loss": 76.7589, "step": 2243 }, { "compression_loss": 72.92920684814453, "epoch": 0.74, "learning_rate": 1.896681997371879e-05, "loss": 73.3539, "step": 2244 }, { "compression_loss": 75.37184143066406, "epoch": 0.74, "learning_rate": 1.8961892247043363e-05, "loss": 75.593, "step": 2245 }, { "compression_loss": 75.61030578613281, "epoch": 0.74, "learning_rate": 1.895696452036794e-05, "loss": 75.8062, "step": 2246 }, { "compression_loss": 74.31312561035156, "epoch": 0.74, "learning_rate": 1.895203679369251e-05, "loss": 74.682, "step": 2247 }, { "compression_loss": 78.15211486816406, "epoch": 0.74, "learning_rate": 1.8947109067017082e-05, "loss": 78.431, "step": 2248 }, { "compression_loss": 74.08544921875, "epoch": 0.74, "learning_rate": 1.8942181340341654e-05, "loss": 74.7999, "step": 2249 }, { "compression_loss": 74.49113464355469, "epoch": 0.74, "learning_rate": 1.8937253613666226e-05, "loss": 74.7863, "step": 2250 }, { "epoch": 0.74, "eval_exact_match": 86.46168401135289, "eval_f1": 92.81092628229678, "step": 2250 }, { "compression_loss": 76.9566650390625, "epoch": 0.74, "learning_rate": 1.8932325886990805e-05, "loss": 77.704, "step": 2251 }, { "compression_loss": 78.2602310180664, "epoch": 0.74, "learning_rate": 1.8927398160315377e-05, "loss": 78.6383, "step": 2252 }, { "compression_loss": 75.89555358886719, "epoch": 0.74, "learning_rate": 1.892247043363995e-05, "loss": 76.3061, "step": 2253 }, { "compression_loss": 75.62835693359375, "epoch": 0.74, "learning_rate": 1.891754270696452e-05, "loss": 76.2111, "step": 2254 }, { "compression_loss": 76.63946533203125, "epoch": 0.74, "learning_rate": 1.8912614980289093e-05, "loss": 77.0681, "step": 2255 }, { "compression_loss": 76.01246643066406, "epoch": 0.74, "learning_rate": 1.8907687253613668e-05, "loss": 76.4364, "step": 2256 }, { "compression_loss": 75.73919677734375, "epoch": 0.74, "learning_rate": 1.890275952693824e-05, "loss": 75.9833, "step": 2257 }, { "compression_loss": 77.70226287841797, "epoch": 0.74, "learning_rate": 1.8897831800262812e-05, "loss": 78.2669, "step": 2258 }, { "compression_loss": 74.27574920654297, "epoch": 0.74, "learning_rate": 1.8892904073587384e-05, "loss": 74.4907, "step": 2259 }, { "compression_loss": 75.69286346435547, "epoch": 0.74, "learning_rate": 1.8887976346911956e-05, "loss": 76.1218, "step": 2260 }, { "compression_loss": 77.19658660888672, "epoch": 0.74, "learning_rate": 1.888304862023653e-05, "loss": 77.5444, "step": 2261 }, { "compression_loss": 78.84701538085938, "epoch": 0.74, "learning_rate": 1.8878120893561103e-05, "loss": 79.384, "step": 2262 }, { "compression_loss": 77.494384765625, "epoch": 0.74, "learning_rate": 1.887319316688568e-05, "loss": 77.7789, "step": 2263 }, { "compression_loss": 77.23129272460938, "epoch": 0.74, "learning_rate": 1.886826544021025e-05, "loss": 77.7947, "step": 2264 }, { "compression_loss": 75.64675903320312, "epoch": 0.74, "learning_rate": 1.8863337713534823e-05, "loss": 76.0333, "step": 2265 }, { "compression_loss": 75.34769439697266, "epoch": 0.74, "learning_rate": 1.8858409986859398e-05, "loss": 76.0059, "step": 2266 }, { "compression_loss": 75.07366180419922, "epoch": 0.74, "learning_rate": 1.885348226018397e-05, "loss": 75.5236, "step": 2267 }, { "compression_loss": 75.41958618164062, "epoch": 0.75, "learning_rate": 1.8848554533508542e-05, "loss": 75.6263, "step": 2268 }, { "compression_loss": 75.44096374511719, "epoch": 0.75, "learning_rate": 1.8843626806833114e-05, "loss": 75.9244, "step": 2269 }, { "compression_loss": 73.68809509277344, "epoch": 0.75, "learning_rate": 1.8838699080157686e-05, "loss": 73.9444, "step": 2270 }, { "compression_loss": 75.25993347167969, "epoch": 0.75, "learning_rate": 1.883377135348226e-05, "loss": 75.6329, "step": 2271 }, { "compression_loss": 76.19720458984375, "epoch": 0.75, "learning_rate": 1.8828843626806833e-05, "loss": 76.7324, "step": 2272 }, { "compression_loss": 77.12109375, "epoch": 0.75, "learning_rate": 1.8823915900131405e-05, "loss": 77.4455, "step": 2273 }, { "compression_loss": 75.1524658203125, "epoch": 0.75, "learning_rate": 1.8818988173455977e-05, "loss": 75.5977, "step": 2274 }, { "compression_loss": 79.58419799804688, "epoch": 0.75, "learning_rate": 1.8814060446780553e-05, "loss": 79.8388, "step": 2275 }, { "compression_loss": 76.25747680664062, "epoch": 0.75, "learning_rate": 1.8809132720105128e-05, "loss": 76.706, "step": 2276 }, { "compression_loss": 74.70982360839844, "epoch": 0.75, "learning_rate": 1.88042049934297e-05, "loss": 75.2069, "step": 2277 }, { "compression_loss": 74.0767593383789, "epoch": 0.75, "learning_rate": 1.8799277266754272e-05, "loss": 74.2954, "step": 2278 }, { "compression_loss": 77.54193878173828, "epoch": 0.75, "learning_rate": 1.8794349540078844e-05, "loss": 77.8473, "step": 2279 }, { "compression_loss": 75.3125, "epoch": 0.75, "learning_rate": 1.8789421813403416e-05, "loss": 75.6687, "step": 2280 }, { "compression_loss": 76.88805389404297, "epoch": 0.75, "learning_rate": 1.878449408672799e-05, "loss": 77.3746, "step": 2281 }, { "compression_loss": 76.89601135253906, "epoch": 0.75, "learning_rate": 1.8779566360052563e-05, "loss": 77.6048, "step": 2282 }, { "compression_loss": 73.48692321777344, "epoch": 0.75, "learning_rate": 1.8774638633377135e-05, "loss": 73.8189, "step": 2283 }, { "compression_loss": 77.59615325927734, "epoch": 0.75, "learning_rate": 1.8769710906701707e-05, "loss": 78.019, "step": 2284 }, { "compression_loss": 73.80783081054688, "epoch": 0.75, "learning_rate": 1.876478318002628e-05, "loss": 74.0087, "step": 2285 }, { "compression_loss": 75.97425079345703, "epoch": 0.75, "learning_rate": 1.8759855453350854e-05, "loss": 76.5472, "step": 2286 }, { "compression_loss": 77.97329711914062, "epoch": 0.75, "learning_rate": 1.875492772667543e-05, "loss": 78.4091, "step": 2287 }, { "compression_loss": 78.88800048828125, "epoch": 0.75, "learning_rate": 1.8750000000000002e-05, "loss": 79.6193, "step": 2288 }, { "compression_loss": 76.0050048828125, "epoch": 0.75, "learning_rate": 1.8745072273324574e-05, "loss": 76.4544, "step": 2289 }, { "compression_loss": 75.96923065185547, "epoch": 0.75, "learning_rate": 1.8740144546649146e-05, "loss": 76.6246, "step": 2290 }, { "compression_loss": 75.27249908447266, "epoch": 0.75, "learning_rate": 1.873521681997372e-05, "loss": 75.6375, "step": 2291 }, { "compression_loss": 75.98779296875, "epoch": 0.75, "learning_rate": 1.8730289093298293e-05, "loss": 76.3859, "step": 2292 }, { "compression_loss": 73.49317932128906, "epoch": 0.75, "learning_rate": 1.8725361366622865e-05, "loss": 73.9156, "step": 2293 }, { "compression_loss": 75.3675537109375, "epoch": 0.75, "learning_rate": 1.8720433639947437e-05, "loss": 75.8268, "step": 2294 }, { "compression_loss": 70.45028686523438, "epoch": 0.75, "learning_rate": 1.871550591327201e-05, "loss": 70.5867, "step": 2295 }, { "compression_loss": 75.11604309082031, "epoch": 0.75, "learning_rate": 1.8710578186596584e-05, "loss": 75.4777, "step": 2296 }, { "compression_loss": 74.49234771728516, "epoch": 0.75, "learning_rate": 1.8705650459921156e-05, "loss": 74.86, "step": 2297 }, { "compression_loss": 76.06854248046875, "epoch": 0.75, "learning_rate": 1.8700722733245728e-05, "loss": 76.7012, "step": 2298 }, { "compression_loss": 76.05692291259766, "epoch": 0.76, "learning_rate": 1.8695795006570304e-05, "loss": 76.4791, "step": 2299 }, { "compression_loss": 75.32432556152344, "epoch": 0.76, "learning_rate": 1.8690867279894876e-05, "loss": 75.5038, "step": 2300 }, { "compression_loss": 73.41126251220703, "epoch": 0.76, "learning_rate": 1.868593955321945e-05, "loss": 73.7448, "step": 2301 }, { "compression_loss": 73.80810546875, "epoch": 0.76, "learning_rate": 1.8681011826544023e-05, "loss": 73.9426, "step": 2302 }, { "compression_loss": 76.25218963623047, "epoch": 0.76, "learning_rate": 1.8676084099868595e-05, "loss": 76.6581, "step": 2303 }, { "compression_loss": 76.28610229492188, "epoch": 0.76, "learning_rate": 1.8671156373193167e-05, "loss": 76.5731, "step": 2304 }, { "compression_loss": 72.84844970703125, "epoch": 0.76, "learning_rate": 1.866622864651774e-05, "loss": 73.28, "step": 2305 }, { "compression_loss": 75.34741973876953, "epoch": 0.76, "learning_rate": 1.8661300919842314e-05, "loss": 75.9351, "step": 2306 }, { "compression_loss": 73.93913269042969, "epoch": 0.76, "learning_rate": 1.8656373193166886e-05, "loss": 74.1085, "step": 2307 }, { "compression_loss": 76.4912109375, "epoch": 0.76, "learning_rate": 1.8651445466491458e-05, "loss": 76.8989, "step": 2308 }, { "compression_loss": 76.07601928710938, "epoch": 0.76, "learning_rate": 1.864651773981603e-05, "loss": 76.4861, "step": 2309 }, { "compression_loss": 75.58454895019531, "epoch": 0.76, "learning_rate": 1.8641590013140602e-05, "loss": 76.0584, "step": 2310 }, { "compression_loss": 76.8675765991211, "epoch": 0.76, "learning_rate": 1.863666228646518e-05, "loss": 77.2155, "step": 2311 }, { "compression_loss": 75.26138305664062, "epoch": 0.76, "learning_rate": 1.8631734559789753e-05, "loss": 75.6405, "step": 2312 }, { "compression_loss": 76.11579895019531, "epoch": 0.76, "learning_rate": 1.8626806833114325e-05, "loss": 76.2646, "step": 2313 }, { "compression_loss": 76.62288665771484, "epoch": 0.76, "learning_rate": 1.8621879106438897e-05, "loss": 77.3049, "step": 2314 }, { "compression_loss": 73.09176635742188, "epoch": 0.76, "learning_rate": 1.861695137976347e-05, "loss": 73.3417, "step": 2315 }, { "compression_loss": 79.61555480957031, "epoch": 0.76, "learning_rate": 1.8612023653088044e-05, "loss": 80.1773, "step": 2316 }, { "compression_loss": 77.50171661376953, "epoch": 0.76, "learning_rate": 1.8607095926412616e-05, "loss": 78.0183, "step": 2317 }, { "compression_loss": 75.56660461425781, "epoch": 0.76, "learning_rate": 1.8602168199737188e-05, "loss": 75.9609, "step": 2318 }, { "compression_loss": 75.72505187988281, "epoch": 0.76, "learning_rate": 1.859724047306176e-05, "loss": 76.0406, "step": 2319 }, { "compression_loss": 76.91133880615234, "epoch": 0.76, "learning_rate": 1.8592312746386332e-05, "loss": 77.2296, "step": 2320 }, { "compression_loss": 76.31227111816406, "epoch": 0.76, "learning_rate": 1.8587385019710907e-05, "loss": 76.9004, "step": 2321 }, { "compression_loss": 74.99430847167969, "epoch": 0.76, "learning_rate": 1.858245729303548e-05, "loss": 75.3818, "step": 2322 }, { "compression_loss": 77.2065200805664, "epoch": 0.76, "learning_rate": 1.8577529566360055e-05, "loss": 77.7892, "step": 2323 }, { "compression_loss": 76.03108215332031, "epoch": 0.76, "learning_rate": 1.8572601839684627e-05, "loss": 76.342, "step": 2324 }, { "compression_loss": 76.18260955810547, "epoch": 0.76, "learning_rate": 1.85676741130092e-05, "loss": 76.6396, "step": 2325 }, { "compression_loss": 77.49589538574219, "epoch": 0.76, "learning_rate": 1.8562746386333774e-05, "loss": 77.7842, "step": 2326 }, { "compression_loss": 74.28379821777344, "epoch": 0.76, "learning_rate": 1.8557818659658346e-05, "loss": 74.7126, "step": 2327 }, { "compression_loss": 74.91835021972656, "epoch": 0.76, "learning_rate": 1.8552890932982918e-05, "loss": 75.2304, "step": 2328 }, { "compression_loss": 75.48748779296875, "epoch": 0.77, "learning_rate": 1.854796320630749e-05, "loss": 75.8056, "step": 2329 }, { "compression_loss": 77.10371398925781, "epoch": 0.77, "learning_rate": 1.8543035479632062e-05, "loss": 77.6597, "step": 2330 }, { "compression_loss": 74.73632049560547, "epoch": 0.77, "learning_rate": 1.8538107752956637e-05, "loss": 75.0855, "step": 2331 }, { "compression_loss": 75.44346618652344, "epoch": 0.77, "learning_rate": 1.853318002628121e-05, "loss": 75.775, "step": 2332 }, { "compression_loss": 74.490234375, "epoch": 0.77, "learning_rate": 1.852825229960578e-05, "loss": 74.8138, "step": 2333 }, { "compression_loss": 75.78173065185547, "epoch": 0.77, "learning_rate": 1.8523324572930353e-05, "loss": 76.1915, "step": 2334 }, { "compression_loss": 76.26628112792969, "epoch": 0.77, "learning_rate": 1.851839684625493e-05, "loss": 76.8819, "step": 2335 }, { "compression_loss": 76.17095947265625, "epoch": 0.77, "learning_rate": 1.8513469119579504e-05, "loss": 76.5681, "step": 2336 }, { "compression_loss": 76.98857879638672, "epoch": 0.77, "learning_rate": 1.8508541392904076e-05, "loss": 77.2595, "step": 2337 }, { "compression_loss": 76.43620300292969, "epoch": 0.77, "learning_rate": 1.8503613666228648e-05, "loss": 76.7809, "step": 2338 }, { "compression_loss": 76.4462661743164, "epoch": 0.77, "learning_rate": 1.849868593955322e-05, "loss": 76.8147, "step": 2339 }, { "compression_loss": 76.10411834716797, "epoch": 0.77, "learning_rate": 1.849375821287779e-05, "loss": 76.4071, "step": 2340 }, { "compression_loss": 75.66239929199219, "epoch": 0.77, "learning_rate": 1.8488830486202367e-05, "loss": 76.0719, "step": 2341 }, { "compression_loss": 73.88452911376953, "epoch": 0.77, "learning_rate": 1.848390275952694e-05, "loss": 74.2097, "step": 2342 }, { "compression_loss": 77.56043243408203, "epoch": 0.77, "learning_rate": 1.847897503285151e-05, "loss": 77.9208, "step": 2343 }, { "compression_loss": 74.9137954711914, "epoch": 0.77, "learning_rate": 1.8474047306176083e-05, "loss": 75.4296, "step": 2344 }, { "compression_loss": 72.41001892089844, "epoch": 0.77, "learning_rate": 1.8469119579500655e-05, "loss": 72.6282, "step": 2345 }, { "compression_loss": 75.74185180664062, "epoch": 0.77, "learning_rate": 1.846419185282523e-05, "loss": 76.2409, "step": 2346 }, { "compression_loss": 75.17280578613281, "epoch": 0.77, "learning_rate": 1.8459264126149806e-05, "loss": 75.6742, "step": 2347 }, { "compression_loss": 76.62889862060547, "epoch": 0.77, "learning_rate": 1.8454336399474377e-05, "loss": 76.8757, "step": 2348 }, { "compression_loss": 75.82928466796875, "epoch": 0.77, "learning_rate": 1.844940867279895e-05, "loss": 76.2648, "step": 2349 }, { "compression_loss": 75.80769348144531, "epoch": 0.77, "learning_rate": 1.844448094612352e-05, "loss": 76.1049, "step": 2350 }, { "compression_loss": 76.86407470703125, "epoch": 0.77, "learning_rate": 1.8439553219448097e-05, "loss": 77.3245, "step": 2351 }, { "compression_loss": 74.17231750488281, "epoch": 0.77, "learning_rate": 1.843462549277267e-05, "loss": 74.5327, "step": 2352 }, { "compression_loss": 74.07296752929688, "epoch": 0.77, "learning_rate": 1.842969776609724e-05, "loss": 74.3198, "step": 2353 }, { "compression_loss": 76.9600601196289, "epoch": 0.77, "learning_rate": 1.8424770039421813e-05, "loss": 77.4251, "step": 2354 }, { "compression_loss": 76.7874755859375, "epoch": 0.77, "learning_rate": 1.8419842312746385e-05, "loss": 77.4771, "step": 2355 }, { "compression_loss": 75.66314697265625, "epoch": 0.77, "learning_rate": 1.841491458607096e-05, "loss": 76.0619, "step": 2356 }, { "compression_loss": 77.253662109375, "epoch": 0.77, "learning_rate": 1.8409986859395532e-05, "loss": 77.6047, "step": 2357 }, { "compression_loss": 76.47171783447266, "epoch": 0.77, "learning_rate": 1.8405059132720107e-05, "loss": 76.7969, "step": 2358 }, { "compression_loss": 77.18482971191406, "epoch": 0.77, "learning_rate": 1.840013140604468e-05, "loss": 77.4327, "step": 2359 }, { "compression_loss": 76.27526092529297, "epoch": 0.78, "learning_rate": 1.839520367936925e-05, "loss": 76.8837, "step": 2360 }, { "compression_loss": 74.1195068359375, "epoch": 0.78, "learning_rate": 1.8390275952693827e-05, "loss": 74.4598, "step": 2361 }, { "compression_loss": 75.0107650756836, "epoch": 0.78, "learning_rate": 1.83853482260184e-05, "loss": 75.3047, "step": 2362 }, { "compression_loss": 75.41819763183594, "epoch": 0.78, "learning_rate": 1.838042049934297e-05, "loss": 75.8063, "step": 2363 }, { "compression_loss": 74.20466613769531, "epoch": 0.78, "learning_rate": 1.8375492772667543e-05, "loss": 74.5205, "step": 2364 }, { "compression_loss": 72.89991760253906, "epoch": 0.78, "learning_rate": 1.8370565045992115e-05, "loss": 73.2893, "step": 2365 }, { "compression_loss": 76.57487487792969, "epoch": 0.78, "learning_rate": 1.836563731931669e-05, "loss": 77.2107, "step": 2366 }, { "compression_loss": 74.22149658203125, "epoch": 0.78, "learning_rate": 1.8360709592641262e-05, "loss": 74.5113, "step": 2367 }, { "compression_loss": 75.27651977539062, "epoch": 0.78, "learning_rate": 1.8355781865965834e-05, "loss": 75.6045, "step": 2368 }, { "compression_loss": 76.93250274658203, "epoch": 0.78, "learning_rate": 1.8350854139290406e-05, "loss": 77.0946, "step": 2369 }, { "compression_loss": 74.72750854492188, "epoch": 0.78, "learning_rate": 1.834592641261498e-05, "loss": 75.0021, "step": 2370 }, { "compression_loss": 76.72217559814453, "epoch": 0.78, "learning_rate": 1.8340998685939556e-05, "loss": 77.3212, "step": 2371 }, { "compression_loss": 77.17826080322266, "epoch": 0.78, "learning_rate": 1.833607095926413e-05, "loss": 77.488, "step": 2372 }, { "compression_loss": 76.35343933105469, "epoch": 0.78, "learning_rate": 1.83311432325887e-05, "loss": 76.6657, "step": 2373 }, { "compression_loss": 75.34455871582031, "epoch": 0.78, "learning_rate": 1.8326215505913272e-05, "loss": 75.8668, "step": 2374 }, { "compression_loss": 75.48442840576172, "epoch": 0.78, "learning_rate": 1.8321287779237844e-05, "loss": 75.795, "step": 2375 }, { "compression_loss": 75.48858642578125, "epoch": 0.78, "learning_rate": 1.831636005256242e-05, "loss": 75.961, "step": 2376 }, { "compression_loss": 75.68707275390625, "epoch": 0.78, "learning_rate": 1.8311432325886992e-05, "loss": 76.1876, "step": 2377 }, { "compression_loss": 73.80717468261719, "epoch": 0.78, "learning_rate": 1.8306504599211564e-05, "loss": 74.0111, "step": 2378 }, { "compression_loss": 74.47877502441406, "epoch": 0.78, "learning_rate": 1.8301576872536136e-05, "loss": 75.061, "step": 2379 }, { "compression_loss": 73.72080993652344, "epoch": 0.78, "learning_rate": 1.8296649145860708e-05, "loss": 74.0924, "step": 2380 }, { "compression_loss": 75.52752685546875, "epoch": 0.78, "learning_rate": 1.8291721419185283e-05, "loss": 75.8461, "step": 2381 }, { "compression_loss": 77.02229309082031, "epoch": 0.78, "learning_rate": 1.828679369250986e-05, "loss": 77.4637, "step": 2382 }, { "compression_loss": 75.14462280273438, "epoch": 0.78, "learning_rate": 1.828186596583443e-05, "loss": 75.6397, "step": 2383 }, { "compression_loss": 74.68359375, "epoch": 0.78, "learning_rate": 1.8276938239159002e-05, "loss": 75.028, "step": 2384 }, { "compression_loss": 75.19347381591797, "epoch": 0.78, "learning_rate": 1.8272010512483574e-05, "loss": 75.6012, "step": 2385 }, { "compression_loss": 76.99790954589844, "epoch": 0.78, "learning_rate": 1.826708278580815e-05, "loss": 77.5821, "step": 2386 }, { "compression_loss": 73.93174743652344, "epoch": 0.78, "learning_rate": 1.826215505913272e-05, "loss": 74.3866, "step": 2387 }, { "compression_loss": 75.23095703125, "epoch": 0.78, "learning_rate": 1.8257227332457294e-05, "loss": 75.5275, "step": 2388 }, { "compression_loss": 75.92945098876953, "epoch": 0.78, "learning_rate": 1.8252299605781865e-05, "loss": 76.2312, "step": 2389 }, { "compression_loss": 75.96501922607422, "epoch": 0.79, "learning_rate": 1.8247371879106437e-05, "loss": 76.5677, "step": 2390 }, { "compression_loss": 75.49539947509766, "epoch": 0.79, "learning_rate": 1.8242444152431013e-05, "loss": 75.8987, "step": 2391 }, { "compression_loss": 73.05293273925781, "epoch": 0.79, "learning_rate": 1.8237516425755585e-05, "loss": 73.4094, "step": 2392 }, { "compression_loss": 78.65193176269531, "epoch": 0.79, "learning_rate": 1.8232588699080157e-05, "loss": 78.9593, "step": 2393 }, { "compression_loss": 74.65565490722656, "epoch": 0.79, "learning_rate": 1.8227660972404732e-05, "loss": 74.9395, "step": 2394 }, { "compression_loss": 74.89580535888672, "epoch": 0.79, "learning_rate": 1.8222733245729304e-05, "loss": 75.1884, "step": 2395 }, { "compression_loss": 73.23831939697266, "epoch": 0.79, "learning_rate": 1.821780551905388e-05, "loss": 73.641, "step": 2396 }, { "compression_loss": 73.88139343261719, "epoch": 0.79, "learning_rate": 1.821287779237845e-05, "loss": 74.1615, "step": 2397 }, { "compression_loss": 71.09992980957031, "epoch": 0.79, "learning_rate": 1.8207950065703023e-05, "loss": 71.4683, "step": 2398 }, { "compression_loss": 75.76509094238281, "epoch": 0.79, "learning_rate": 1.8203022339027595e-05, "loss": 76.4514, "step": 2399 }, { "compression_loss": 74.3895263671875, "epoch": 0.79, "learning_rate": 1.8198094612352167e-05, "loss": 74.5989, "step": 2400 }, { "compression_loss": 76.7343978881836, "epoch": 0.79, "learning_rate": 1.8193166885676743e-05, "loss": 77.1823, "step": 2401 }, { "compression_loss": 76.00634765625, "epoch": 0.79, "learning_rate": 1.8188239159001315e-05, "loss": 76.3237, "step": 2402 }, { "compression_loss": 75.36526489257812, "epoch": 0.79, "learning_rate": 1.8183311432325887e-05, "loss": 75.8193, "step": 2403 }, { "compression_loss": 76.22848510742188, "epoch": 0.79, "learning_rate": 1.817838370565046e-05, "loss": 76.6068, "step": 2404 }, { "compression_loss": 75.76763916015625, "epoch": 0.79, "learning_rate": 1.817345597897503e-05, "loss": 76.2791, "step": 2405 }, { "compression_loss": 76.53150939941406, "epoch": 0.79, "learning_rate": 1.816852825229961e-05, "loss": 77.0483, "step": 2406 }, { "compression_loss": 75.92098999023438, "epoch": 0.79, "learning_rate": 1.816360052562418e-05, "loss": 76.1971, "step": 2407 }, { "compression_loss": 75.66797637939453, "epoch": 0.79, "learning_rate": 1.8158672798948753e-05, "loss": 75.9089, "step": 2408 }, { "compression_loss": 77.46792602539062, "epoch": 0.79, "learning_rate": 1.8153745072273325e-05, "loss": 77.9628, "step": 2409 }, { "compression_loss": 76.41897583007812, "epoch": 0.79, "learning_rate": 1.8148817345597897e-05, "loss": 76.9227, "step": 2410 }, { "compression_loss": 77.5321273803711, "epoch": 0.79, "learning_rate": 1.8143889618922473e-05, "loss": 77.9423, "step": 2411 }, { "compression_loss": 75.10570526123047, "epoch": 0.79, "learning_rate": 1.8138961892247044e-05, "loss": 75.3782, "step": 2412 }, { "compression_loss": 75.40132141113281, "epoch": 0.79, "learning_rate": 1.8134034165571616e-05, "loss": 75.7698, "step": 2413 }, { "compression_loss": 73.01881408691406, "epoch": 0.79, "learning_rate": 1.812910643889619e-05, "loss": 73.1969, "step": 2414 }, { "compression_loss": 76.85418701171875, "epoch": 0.79, "learning_rate": 1.812417871222076e-05, "loss": 77.2119, "step": 2415 }, { "compression_loss": 74.65492248535156, "epoch": 0.79, "learning_rate": 1.8119250985545336e-05, "loss": 75.1018, "step": 2416 }, { "compression_loss": 72.50923156738281, "epoch": 0.79, "learning_rate": 1.8114323258869908e-05, "loss": 72.8306, "step": 2417 }, { "compression_loss": 73.36712646484375, "epoch": 0.79, "learning_rate": 1.8109395532194483e-05, "loss": 73.6376, "step": 2418 }, { "compression_loss": 75.09535217285156, "epoch": 0.79, "learning_rate": 1.8104467805519055e-05, "loss": 75.6819, "step": 2419 }, { "compression_loss": 74.89114379882812, "epoch": 0.8, "learning_rate": 1.8099540078843627e-05, "loss": 75.6152, "step": 2420 }, { "compression_loss": 76.67716217041016, "epoch": 0.8, "learning_rate": 1.8094612352168202e-05, "loss": 77.1396, "step": 2421 }, { "compression_loss": 75.11766815185547, "epoch": 0.8, "learning_rate": 1.8089684625492774e-05, "loss": 75.4011, "step": 2422 }, { "compression_loss": 75.4896240234375, "epoch": 0.8, "learning_rate": 1.8084756898817346e-05, "loss": 75.7837, "step": 2423 }, { "compression_loss": 76.54850769042969, "epoch": 0.8, "learning_rate": 1.8079829172141918e-05, "loss": 77.1935, "step": 2424 }, { "compression_loss": 74.08755493164062, "epoch": 0.8, "learning_rate": 1.807490144546649e-05, "loss": 74.3418, "step": 2425 }, { "compression_loss": 76.1146240234375, "epoch": 0.8, "learning_rate": 1.8069973718791066e-05, "loss": 76.546, "step": 2426 }, { "compression_loss": 74.55171203613281, "epoch": 0.8, "learning_rate": 1.8065045992115638e-05, "loss": 74.8128, "step": 2427 }, { "compression_loss": 76.29200744628906, "epoch": 0.8, "learning_rate": 1.806011826544021e-05, "loss": 76.6251, "step": 2428 }, { "compression_loss": 75.46343231201172, "epoch": 0.8, "learning_rate": 1.805519053876478e-05, "loss": 75.8246, "step": 2429 }, { "compression_loss": 74.70275115966797, "epoch": 0.8, "learning_rate": 1.8050262812089357e-05, "loss": 75.0935, "step": 2430 }, { "compression_loss": 77.75375366210938, "epoch": 0.8, "learning_rate": 1.8045335085413932e-05, "loss": 77.9555, "step": 2431 }, { "compression_loss": 75.71324157714844, "epoch": 0.8, "learning_rate": 1.8040407358738504e-05, "loss": 76.3524, "step": 2432 }, { "compression_loss": 77.05426025390625, "epoch": 0.8, "learning_rate": 1.8035479632063076e-05, "loss": 77.3199, "step": 2433 }, { "compression_loss": 77.07321166992188, "epoch": 0.8, "learning_rate": 1.8030551905387648e-05, "loss": 77.4528, "step": 2434 }, { "compression_loss": 76.60398864746094, "epoch": 0.8, "learning_rate": 1.802562417871222e-05, "loss": 77.0283, "step": 2435 }, { "compression_loss": 75.52938842773438, "epoch": 0.8, "learning_rate": 1.8020696452036795e-05, "loss": 76.0829, "step": 2436 }, { "compression_loss": 75.31002044677734, "epoch": 0.8, "learning_rate": 1.8015768725361367e-05, "loss": 75.5045, "step": 2437 }, { "compression_loss": 76.7418212890625, "epoch": 0.8, "learning_rate": 1.801084099868594e-05, "loss": 77.0425, "step": 2438 }, { "compression_loss": 78.14012908935547, "epoch": 0.8, "learning_rate": 1.800591327201051e-05, "loss": 78.4102, "step": 2439 }, { "compression_loss": 74.966552734375, "epoch": 0.8, "learning_rate": 1.8000985545335083e-05, "loss": 75.6414, "step": 2440 }, { "compression_loss": 74.38496398925781, "epoch": 0.8, "learning_rate": 1.799605781865966e-05, "loss": 74.7369, "step": 2441 }, { "compression_loss": 76.95275115966797, "epoch": 0.8, "learning_rate": 1.7991130091984234e-05, "loss": 77.402, "step": 2442 }, { "compression_loss": 78.38703918457031, "epoch": 0.8, "learning_rate": 1.7986202365308806e-05, "loss": 79.3272, "step": 2443 }, { "compression_loss": 75.25812530517578, "epoch": 0.8, "learning_rate": 1.7981274638633378e-05, "loss": 75.449, "step": 2444 }, { "compression_loss": 76.81609344482422, "epoch": 0.8, "learning_rate": 1.797634691195795e-05, "loss": 77.2387, "step": 2445 }, { "compression_loss": 76.17505645751953, "epoch": 0.8, "learning_rate": 1.7971419185282525e-05, "loss": 76.5428, "step": 2446 }, { "compression_loss": 75.70095825195312, "epoch": 0.8, "learning_rate": 1.7966491458607097e-05, "loss": 76.1314, "step": 2447 }, { "compression_loss": 75.25900268554688, "epoch": 0.8, "learning_rate": 1.796156373193167e-05, "loss": 75.7098, "step": 2448 }, { "compression_loss": 74.78191375732422, "epoch": 0.8, "learning_rate": 1.795663600525624e-05, "loss": 75.1726, "step": 2449 }, { "compression_loss": 76.20235443115234, "epoch": 0.8, "learning_rate": 1.7951708278580813e-05, "loss": 76.663, "step": 2450 }, { "compression_loss": 75.48097229003906, "epoch": 0.81, "learning_rate": 1.794678055190539e-05, "loss": 75.7763, "step": 2451 }, { "compression_loss": 77.45220947265625, "epoch": 0.81, "learning_rate": 1.794185282522996e-05, "loss": 78.0359, "step": 2452 }, { "compression_loss": 74.55848693847656, "epoch": 0.81, "learning_rate": 1.7936925098554533e-05, "loss": 74.8581, "step": 2453 }, { "compression_loss": 77.36883544921875, "epoch": 0.81, "learning_rate": 1.7931997371879108e-05, "loss": 77.8312, "step": 2454 }, { "compression_loss": 77.27251434326172, "epoch": 0.81, "learning_rate": 1.792706964520368e-05, "loss": 77.685, "step": 2455 }, { "compression_loss": 75.05424499511719, "epoch": 0.81, "learning_rate": 1.7922141918528255e-05, "loss": 75.381, "step": 2456 }, { "compression_loss": 76.02384948730469, "epoch": 0.81, "learning_rate": 1.7917214191852827e-05, "loss": 76.5401, "step": 2457 }, { "compression_loss": 75.84040832519531, "epoch": 0.81, "learning_rate": 1.79122864651774e-05, "loss": 76.0424, "step": 2458 }, { "compression_loss": 75.89981079101562, "epoch": 0.81, "learning_rate": 1.790735873850197e-05, "loss": 76.2767, "step": 2459 }, { "compression_loss": 74.41572570800781, "epoch": 0.81, "learning_rate": 1.7902431011826543e-05, "loss": 74.8805, "step": 2460 }, { "compression_loss": 74.61323547363281, "epoch": 0.81, "learning_rate": 1.789750328515112e-05, "loss": 74.9113, "step": 2461 }, { "compression_loss": 75.65321350097656, "epoch": 0.81, "learning_rate": 1.789257555847569e-05, "loss": 75.9601, "step": 2462 }, { "compression_loss": 74.12808227539062, "epoch": 0.81, "learning_rate": 1.7887647831800262e-05, "loss": 74.4138, "step": 2463 }, { "compression_loss": 76.73883056640625, "epoch": 0.81, "learning_rate": 1.7882720105124834e-05, "loss": 77.4607, "step": 2464 }, { "compression_loss": 75.24176025390625, "epoch": 0.81, "learning_rate": 1.7877792378449406e-05, "loss": 75.7762, "step": 2465 }, { "compression_loss": 75.65363311767578, "epoch": 0.81, "learning_rate": 1.7872864651773985e-05, "loss": 76.3055, "step": 2466 }, { "compression_loss": 73.96023559570312, "epoch": 0.81, "learning_rate": 1.7867936925098557e-05, "loss": 74.3061, "step": 2467 }, { "compression_loss": 75.82307434082031, "epoch": 0.81, "learning_rate": 1.786300919842313e-05, "loss": 76.2014, "step": 2468 }, { "compression_loss": 74.46887969970703, "epoch": 0.81, "learning_rate": 1.78580814717477e-05, "loss": 74.8727, "step": 2469 }, { "compression_loss": 76.44223022460938, "epoch": 0.81, "learning_rate": 1.7853153745072273e-05, "loss": 76.9751, "step": 2470 }, { "compression_loss": 78.45756530761719, "epoch": 0.81, "learning_rate": 1.7848226018396848e-05, "loss": 79.2481, "step": 2471 }, { "compression_loss": 75.44839477539062, "epoch": 0.81, "learning_rate": 1.784329829172142e-05, "loss": 75.7837, "step": 2472 }, { "compression_loss": 73.54247283935547, "epoch": 0.81, "learning_rate": 1.7838370565045992e-05, "loss": 74.0813, "step": 2473 }, { "compression_loss": 74.71624755859375, "epoch": 0.81, "learning_rate": 1.7833442838370564e-05, "loss": 75.0157, "step": 2474 }, { "compression_loss": 78.23133850097656, "epoch": 0.81, "learning_rate": 1.7828515111695136e-05, "loss": 78.5816, "step": 2475 }, { "compression_loss": 73.59690856933594, "epoch": 0.81, "learning_rate": 1.782358738501971e-05, "loss": 73.854, "step": 2476 }, { "compression_loss": 75.68357849121094, "epoch": 0.81, "learning_rate": 1.7818659658344283e-05, "loss": 75.9818, "step": 2477 }, { "compression_loss": 76.17045593261719, "epoch": 0.81, "learning_rate": 1.781373193166886e-05, "loss": 76.4759, "step": 2478 }, { "compression_loss": 74.24368286132812, "epoch": 0.81, "learning_rate": 1.780880420499343e-05, "loss": 74.6471, "step": 2479 }, { "compression_loss": 74.63346862792969, "epoch": 0.81, "learning_rate": 1.7803876478318003e-05, "loss": 75.051, "step": 2480 }, { "compression_loss": 76.6799545288086, "epoch": 0.82, "learning_rate": 1.7798948751642578e-05, "loss": 77.4968, "step": 2481 }, { "compression_loss": 74.65585327148438, "epoch": 0.82, "learning_rate": 1.779402102496715e-05, "loss": 75.2452, "step": 2482 }, { "compression_loss": 76.20934295654297, "epoch": 0.82, "learning_rate": 1.7789093298291722e-05, "loss": 76.7584, "step": 2483 }, { "compression_loss": 76.80070495605469, "epoch": 0.82, "learning_rate": 1.7784165571616294e-05, "loss": 77.1239, "step": 2484 }, { "compression_loss": 76.33409118652344, "epoch": 0.82, "learning_rate": 1.7779237844940866e-05, "loss": 76.8388, "step": 2485 }, { "compression_loss": 74.92146301269531, "epoch": 0.82, "learning_rate": 1.777431011826544e-05, "loss": 75.3109, "step": 2486 }, { "compression_loss": 75.4592514038086, "epoch": 0.82, "learning_rate": 1.7769382391590013e-05, "loss": 76.041, "step": 2487 }, { "compression_loss": 78.592529296875, "epoch": 0.82, "learning_rate": 1.7764454664914585e-05, "loss": 79.1012, "step": 2488 }, { "compression_loss": 76.42533874511719, "epoch": 0.82, "learning_rate": 1.7759526938239157e-05, "loss": 76.7991, "step": 2489 }, { "compression_loss": 75.29940795898438, "epoch": 0.82, "learning_rate": 1.7754599211563733e-05, "loss": 75.7555, "step": 2490 }, { "compression_loss": 74.43756103515625, "epoch": 0.82, "learning_rate": 1.7749671484888308e-05, "loss": 74.9274, "step": 2491 }, { "compression_loss": 75.21072387695312, "epoch": 0.82, "learning_rate": 1.774474375821288e-05, "loss": 75.8142, "step": 2492 }, { "compression_loss": 73.41800689697266, "epoch": 0.82, "learning_rate": 1.7739816031537452e-05, "loss": 73.7204, "step": 2493 }, { "compression_loss": 76.22215270996094, "epoch": 0.82, "learning_rate": 1.7734888304862024e-05, "loss": 76.619, "step": 2494 }, { "compression_loss": 75.99593353271484, "epoch": 0.82, "learning_rate": 1.7729960578186596e-05, "loss": 76.3617, "step": 2495 }, { "compression_loss": 76.29755401611328, "epoch": 0.82, "learning_rate": 1.772503285151117e-05, "loss": 76.9872, "step": 2496 }, { "compression_loss": 76.189208984375, "epoch": 0.82, "learning_rate": 1.7720105124835743e-05, "loss": 76.5629, "step": 2497 }, { "compression_loss": 74.51951599121094, "epoch": 0.82, "learning_rate": 1.7715177398160315e-05, "loss": 74.7063, "step": 2498 }, { "compression_loss": 75.39103698730469, "epoch": 0.82, "learning_rate": 1.7710249671484887e-05, "loss": 75.6943, "step": 2499 }, { "compression_loss": 75.93753051757812, "epoch": 0.82, "learning_rate": 1.770532194480946e-05, "loss": 76.378, "step": 2500 }, { "epoch": 0.82, "eval_exact_match": 86.30085146641439, "eval_f1": 92.5902417360272, "step": 2500 }, { "compression_loss": 74.27301788330078, "epoch": 0.82, "learning_rate": 1.7700394218134034e-05, "loss": 74.6883, "step": 2501 }, { "compression_loss": 76.56202697753906, "epoch": 0.82, "learning_rate": 1.769546649145861e-05, "loss": 76.9661, "step": 2502 }, { "compression_loss": 75.91564178466797, "epoch": 0.82, "learning_rate": 1.7690538764783182e-05, "loss": 76.6683, "step": 2503 }, { "compression_loss": 75.78331756591797, "epoch": 0.82, "learning_rate": 1.7685611038107754e-05, "loss": 76.4376, "step": 2504 }, { "compression_loss": 76.55966186523438, "epoch": 0.82, "learning_rate": 1.7680683311432326e-05, "loss": 77.0554, "step": 2505 }, { "compression_loss": 78.23638153076172, "epoch": 0.82, "learning_rate": 1.76757555847569e-05, "loss": 78.772, "step": 2506 }, { "compression_loss": 77.5640869140625, "epoch": 0.82, "learning_rate": 1.7670827858081473e-05, "loss": 78.0652, "step": 2507 }, { "compression_loss": 75.98818969726562, "epoch": 0.82, "learning_rate": 1.7665900131406045e-05, "loss": 76.735, "step": 2508 }, { "compression_loss": 74.28897094726562, "epoch": 0.82, "learning_rate": 1.7660972404730617e-05, "loss": 74.699, "step": 2509 }, { "compression_loss": 74.37225341796875, "epoch": 0.82, "learning_rate": 1.765604467805519e-05, "loss": 74.7321, "step": 2510 }, { "compression_loss": 78.7568130493164, "epoch": 0.82, "learning_rate": 1.7651116951379764e-05, "loss": 79.5947, "step": 2511 }, { "compression_loss": 77.08164978027344, "epoch": 0.83, "learning_rate": 1.7646189224704336e-05, "loss": 77.3956, "step": 2512 }, { "compression_loss": 75.90776824951172, "epoch": 0.83, "learning_rate": 1.764126149802891e-05, "loss": 76.446, "step": 2513 }, { "compression_loss": 75.16230010986328, "epoch": 0.83, "learning_rate": 1.7636333771353484e-05, "loss": 75.6194, "step": 2514 }, { "compression_loss": 76.25325012207031, "epoch": 0.83, "learning_rate": 1.7631406044678056e-05, "loss": 76.4443, "step": 2515 }, { "compression_loss": 75.42631530761719, "epoch": 0.83, "learning_rate": 1.762647831800263e-05, "loss": 75.7357, "step": 2516 }, { "compression_loss": 74.67412567138672, "epoch": 0.83, "learning_rate": 1.7621550591327203e-05, "loss": 74.9938, "step": 2517 }, { "compression_loss": 75.81114196777344, "epoch": 0.83, "learning_rate": 1.7616622864651775e-05, "loss": 76.361, "step": 2518 }, { "compression_loss": 75.60613250732422, "epoch": 0.83, "learning_rate": 1.7611695137976347e-05, "loss": 76.1172, "step": 2519 }, { "compression_loss": 75.26362609863281, "epoch": 0.83, "learning_rate": 1.760676741130092e-05, "loss": 75.4472, "step": 2520 }, { "compression_loss": 76.20588684082031, "epoch": 0.83, "learning_rate": 1.7601839684625494e-05, "loss": 76.472, "step": 2521 }, { "compression_loss": 74.6200180053711, "epoch": 0.83, "learning_rate": 1.7596911957950066e-05, "loss": 74.9362, "step": 2522 }, { "compression_loss": 75.02464294433594, "epoch": 0.83, "learning_rate": 1.7591984231274638e-05, "loss": 75.3374, "step": 2523 }, { "compression_loss": 74.86224365234375, "epoch": 0.83, "learning_rate": 1.758705650459921e-05, "loss": 75.3997, "step": 2524 }, { "compression_loss": 76.08817291259766, "epoch": 0.83, "learning_rate": 1.7582128777923785e-05, "loss": 76.7186, "step": 2525 }, { "compression_loss": 74.9935302734375, "epoch": 0.83, "learning_rate": 1.757720105124836e-05, "loss": 75.5938, "step": 2526 }, { "compression_loss": 74.59942626953125, "epoch": 0.83, "learning_rate": 1.7572273324572933e-05, "loss": 75.2316, "step": 2527 }, { "compression_loss": 79.48331451416016, "epoch": 0.83, "learning_rate": 1.7567345597897505e-05, "loss": 80.0699, "step": 2528 }, { "compression_loss": 74.49589538574219, "epoch": 0.83, "learning_rate": 1.7562417871222077e-05, "loss": 74.8296, "step": 2529 }, { "compression_loss": 74.46454620361328, "epoch": 0.83, "learning_rate": 1.755749014454665e-05, "loss": 74.9625, "step": 2530 }, { "compression_loss": 75.2376937866211, "epoch": 0.83, "learning_rate": 1.7552562417871224e-05, "loss": 75.6154, "step": 2531 }, { "compression_loss": 76.28339385986328, "epoch": 0.83, "learning_rate": 1.7547634691195796e-05, "loss": 76.7354, "step": 2532 }, { "compression_loss": 75.24842834472656, "epoch": 0.83, "learning_rate": 1.7542706964520368e-05, "loss": 75.6716, "step": 2533 }, { "compression_loss": 75.64498901367188, "epoch": 0.83, "learning_rate": 1.753777923784494e-05, "loss": 75.9881, "step": 2534 }, { "compression_loss": 76.09675598144531, "epoch": 0.83, "learning_rate": 1.7532851511169512e-05, "loss": 76.2954, "step": 2535 }, { "compression_loss": 74.77780151367188, "epoch": 0.83, "learning_rate": 1.7527923784494087e-05, "loss": 75.0216, "step": 2536 }, { "compression_loss": 75.894775390625, "epoch": 0.83, "learning_rate": 1.7522996057818663e-05, "loss": 76.5791, "step": 2537 }, { "compression_loss": 75.80210876464844, "epoch": 0.83, "learning_rate": 1.7518068331143235e-05, "loss": 76.2485, "step": 2538 }, { "compression_loss": 75.10977935791016, "epoch": 0.83, "learning_rate": 1.7513140604467807e-05, "loss": 75.4444, "step": 2539 }, { "compression_loss": 73.9869155883789, "epoch": 0.83, "learning_rate": 1.750821287779238e-05, "loss": 74.5604, "step": 2540 }, { "compression_loss": 75.9797134399414, "epoch": 0.83, "learning_rate": 1.7503285151116954e-05, "loss": 76.5716, "step": 2541 }, { "compression_loss": 76.36837768554688, "epoch": 0.84, "learning_rate": 1.7498357424441526e-05, "loss": 76.9224, "step": 2542 }, { "compression_loss": 73.68243408203125, "epoch": 0.84, "learning_rate": 1.7493429697766098e-05, "loss": 74.0918, "step": 2543 }, { "compression_loss": 76.91227722167969, "epoch": 0.84, "learning_rate": 1.748850197109067e-05, "loss": 77.5252, "step": 2544 }, { "compression_loss": 75.99566650390625, "epoch": 0.84, "learning_rate": 1.7483574244415242e-05, "loss": 76.4076, "step": 2545 }, { "compression_loss": 76.47382354736328, "epoch": 0.84, "learning_rate": 1.7478646517739817e-05, "loss": 76.9876, "step": 2546 }, { "compression_loss": 75.70161437988281, "epoch": 0.84, "learning_rate": 1.747371879106439e-05, "loss": 76.0685, "step": 2547 }, { "compression_loss": 74.13417053222656, "epoch": 0.84, "learning_rate": 1.746879106438896e-05, "loss": 74.8114, "step": 2548 }, { "compression_loss": 74.31497955322266, "epoch": 0.84, "learning_rate": 1.7463863337713536e-05, "loss": 74.5676, "step": 2549 }, { "compression_loss": 74.30189514160156, "epoch": 0.84, "learning_rate": 1.745893561103811e-05, "loss": 74.7699, "step": 2550 }, { "compression_loss": 74.82379150390625, "epoch": 0.84, "learning_rate": 1.7454007884362684e-05, "loss": 75.3482, "step": 2551 }, { "compression_loss": 75.01121520996094, "epoch": 0.84, "learning_rate": 1.7449080157687256e-05, "loss": 75.4769, "step": 2552 }, { "compression_loss": 76.00126647949219, "epoch": 0.84, "learning_rate": 1.7444152431011828e-05, "loss": 76.4974, "step": 2553 }, { "compression_loss": 76.24882507324219, "epoch": 0.84, "learning_rate": 1.74392247043364e-05, "loss": 76.6571, "step": 2554 }, { "compression_loss": 72.58364868164062, "epoch": 0.84, "learning_rate": 1.743429697766097e-05, "loss": 73.1532, "step": 2555 }, { "compression_loss": 75.4610595703125, "epoch": 0.84, "learning_rate": 1.7429369250985547e-05, "loss": 75.834, "step": 2556 }, { "compression_loss": 74.46580505371094, "epoch": 0.84, "learning_rate": 1.742444152431012e-05, "loss": 74.959, "step": 2557 }, { "compression_loss": 73.42239379882812, "epoch": 0.84, "learning_rate": 1.741951379763469e-05, "loss": 73.8017, "step": 2558 }, { "compression_loss": 73.47238159179688, "epoch": 0.84, "learning_rate": 1.7414586070959263e-05, "loss": 73.7514, "step": 2559 }, { "compression_loss": 73.38613891601562, "epoch": 0.84, "learning_rate": 1.7409658344283835e-05, "loss": 74.1188, "step": 2560 }, { "compression_loss": 73.62715148925781, "epoch": 0.84, "learning_rate": 1.7404730617608414e-05, "loss": 74.2695, "step": 2561 }, { "compression_loss": 74.91703033447266, "epoch": 0.84, "learning_rate": 1.7399802890932986e-05, "loss": 75.0952, "step": 2562 }, { "compression_loss": 76.07322692871094, "epoch": 0.84, "learning_rate": 1.7394875164257558e-05, "loss": 76.8198, "step": 2563 }, { "compression_loss": 74.9301528930664, "epoch": 0.84, "learning_rate": 1.738994743758213e-05, "loss": 75.2669, "step": 2564 }, { "compression_loss": 74.02899932861328, "epoch": 0.84, "learning_rate": 1.73850197109067e-05, "loss": 74.2936, "step": 2565 }, { "compression_loss": 76.79856872558594, "epoch": 0.84, "learning_rate": 1.7380091984231277e-05, "loss": 77.0404, "step": 2566 }, { "compression_loss": 75.43368530273438, "epoch": 0.84, "learning_rate": 1.737516425755585e-05, "loss": 75.858, "step": 2567 }, { "compression_loss": 76.3055419921875, "epoch": 0.84, "learning_rate": 1.737023653088042e-05, "loss": 76.6065, "step": 2568 }, { "compression_loss": 73.62326049804688, "epoch": 0.84, "learning_rate": 1.7365308804204993e-05, "loss": 73.8809, "step": 2569 }, { "compression_loss": 73.1448745727539, "epoch": 0.84, "learning_rate": 1.7360381077529565e-05, "loss": 73.4515, "step": 2570 }, { "compression_loss": 75.39981079101562, "epoch": 0.84, "learning_rate": 1.735545335085414e-05, "loss": 75.875, "step": 2571 }, { "compression_loss": 74.3123550415039, "epoch": 0.84, "learning_rate": 1.7350525624178712e-05, "loss": 74.6218, "step": 2572 }, { "compression_loss": 76.55575561523438, "epoch": 0.85, "learning_rate": 1.7345597897503287e-05, "loss": 77.0, "step": 2573 }, { "compression_loss": 73.90672302246094, "epoch": 0.85, "learning_rate": 1.734067017082786e-05, "loss": 74.273, "step": 2574 }, { "compression_loss": 75.32843017578125, "epoch": 0.85, "learning_rate": 1.733574244415243e-05, "loss": 75.6953, "step": 2575 }, { "compression_loss": 77.28150939941406, "epoch": 0.85, "learning_rate": 1.7330814717477007e-05, "loss": 77.7826, "step": 2576 }, { "compression_loss": 75.06260681152344, "epoch": 0.85, "learning_rate": 1.732588699080158e-05, "loss": 75.4598, "step": 2577 }, { "compression_loss": 74.86605834960938, "epoch": 0.85, "learning_rate": 1.732095926412615e-05, "loss": 75.333, "step": 2578 }, { "compression_loss": 77.1517333984375, "epoch": 0.85, "learning_rate": 1.7316031537450723e-05, "loss": 77.5076, "step": 2579 }, { "compression_loss": 74.63313293457031, "epoch": 0.85, "learning_rate": 1.7311103810775295e-05, "loss": 75.0446, "step": 2580 }, { "compression_loss": 74.96528625488281, "epoch": 0.85, "learning_rate": 1.730617608409987e-05, "loss": 75.2318, "step": 2581 }, { "compression_loss": 73.7007827758789, "epoch": 0.85, "learning_rate": 1.7301248357424442e-05, "loss": 73.9565, "step": 2582 }, { "compression_loss": 78.154052734375, "epoch": 0.85, "learning_rate": 1.7296320630749014e-05, "loss": 78.673, "step": 2583 }, { "compression_loss": 75.78921508789062, "epoch": 0.85, "learning_rate": 1.7291392904073586e-05, "loss": 76.3095, "step": 2584 }, { "compression_loss": 76.1656494140625, "epoch": 0.85, "learning_rate": 1.728646517739816e-05, "loss": 76.7638, "step": 2585 }, { "compression_loss": 77.997802734375, "epoch": 0.85, "learning_rate": 1.7281537450722737e-05, "loss": 78.4475, "step": 2586 }, { "compression_loss": 76.0586929321289, "epoch": 0.85, "learning_rate": 1.727660972404731e-05, "loss": 76.2691, "step": 2587 }, { "compression_loss": 75.67284393310547, "epoch": 0.85, "learning_rate": 1.727168199737188e-05, "loss": 75.9596, "step": 2588 }, { "compression_loss": 75.1490249633789, "epoch": 0.85, "learning_rate": 1.7266754270696452e-05, "loss": 75.6209, "step": 2589 }, { "compression_loss": 75.89755249023438, "epoch": 0.85, "learning_rate": 1.7261826544021024e-05, "loss": 76.3486, "step": 2590 }, { "compression_loss": 78.93061828613281, "epoch": 0.85, "learning_rate": 1.72568988173456e-05, "loss": 79.3538, "step": 2591 }, { "compression_loss": 74.26747131347656, "epoch": 0.85, "learning_rate": 1.7251971090670172e-05, "loss": 75.0575, "step": 2592 }, { "compression_loss": 74.94882202148438, "epoch": 0.85, "learning_rate": 1.7247043363994744e-05, "loss": 75.1866, "step": 2593 }, { "compression_loss": 75.67452239990234, "epoch": 0.85, "learning_rate": 1.7242115637319316e-05, "loss": 76.5609, "step": 2594 }, { "compression_loss": 75.33865356445312, "epoch": 0.85, "learning_rate": 1.7237187910643888e-05, "loss": 75.7478, "step": 2595 }, { "compression_loss": 74.64332580566406, "epoch": 0.85, "learning_rate": 1.7232260183968463e-05, "loss": 74.9699, "step": 2596 }, { "compression_loss": 77.43710327148438, "epoch": 0.85, "learning_rate": 1.722733245729304e-05, "loss": 77.9115, "step": 2597 }, { "compression_loss": 76.14246368408203, "epoch": 0.85, "learning_rate": 1.722240473061761e-05, "loss": 76.3849, "step": 2598 }, { "compression_loss": 75.53916931152344, "epoch": 0.85, "learning_rate": 1.7217477003942182e-05, "loss": 76.0013, "step": 2599 }, { "compression_loss": 77.31758880615234, "epoch": 0.85, "learning_rate": 1.7212549277266754e-05, "loss": 78.0111, "step": 2600 }, { "compression_loss": 75.21601867675781, "epoch": 0.85, "learning_rate": 1.720762155059133e-05, "loss": 75.4021, "step": 2601 }, { "compression_loss": 74.66822052001953, "epoch": 0.85, "learning_rate": 1.72026938239159e-05, "loss": 74.9661, "step": 2602 }, { "compression_loss": 75.94776916503906, "epoch": 0.86, "learning_rate": 1.7197766097240474e-05, "loss": 76.6406, "step": 2603 }, { "compression_loss": 76.37574768066406, "epoch": 0.86, "learning_rate": 1.7192838370565046e-05, "loss": 76.9253, "step": 2604 }, { "compression_loss": 74.05667114257812, "epoch": 0.86, "learning_rate": 1.7187910643889617e-05, "loss": 74.5994, "step": 2605 }, { "compression_loss": 76.19343566894531, "epoch": 0.86, "learning_rate": 1.7182982917214193e-05, "loss": 76.688, "step": 2606 }, { "compression_loss": 75.33972930908203, "epoch": 0.86, "learning_rate": 1.7178055190538765e-05, "loss": 75.7934, "step": 2607 }, { "compression_loss": 73.72200775146484, "epoch": 0.86, "learning_rate": 1.7173127463863337e-05, "loss": 74.0128, "step": 2608 }, { "compression_loss": 75.4324951171875, "epoch": 0.86, "learning_rate": 1.7168199737187912e-05, "loss": 76.093, "step": 2609 }, { "compression_loss": 77.30621337890625, "epoch": 0.86, "learning_rate": 1.7163272010512484e-05, "loss": 77.9502, "step": 2610 }, { "compression_loss": 74.66928100585938, "epoch": 0.86, "learning_rate": 1.715834428383706e-05, "loss": 75.1871, "step": 2611 }, { "compression_loss": 77.73945617675781, "epoch": 0.86, "learning_rate": 1.715341655716163e-05, "loss": 78.391, "step": 2612 }, { "compression_loss": 78.86802673339844, "epoch": 0.86, "learning_rate": 1.7148488830486203e-05, "loss": 79.3851, "step": 2613 }, { "compression_loss": 77.02227783203125, "epoch": 0.86, "learning_rate": 1.7143561103810775e-05, "loss": 77.6241, "step": 2614 }, { "compression_loss": 76.33735656738281, "epoch": 0.86, "learning_rate": 1.7138633377135347e-05, "loss": 76.7575, "step": 2615 }, { "compression_loss": 77.01591491699219, "epoch": 0.86, "learning_rate": 1.7133705650459923e-05, "loss": 77.629, "step": 2616 }, { "compression_loss": 77.83220672607422, "epoch": 0.86, "learning_rate": 1.7128777923784495e-05, "loss": 78.3, "step": 2617 }, { "compression_loss": 77.04531860351562, "epoch": 0.86, "learning_rate": 1.7123850197109067e-05, "loss": 77.2861, "step": 2618 }, { "compression_loss": 75.72196960449219, "epoch": 0.86, "learning_rate": 1.711892247043364e-05, "loss": 76.287, "step": 2619 }, { "compression_loss": 77.93954467773438, "epoch": 0.86, "learning_rate": 1.711399474375821e-05, "loss": 78.1567, "step": 2620 }, { "compression_loss": 75.93013000488281, "epoch": 0.86, "learning_rate": 1.710906701708279e-05, "loss": 76.3268, "step": 2621 }, { "compression_loss": 75.63520050048828, "epoch": 0.86, "learning_rate": 1.710413929040736e-05, "loss": 76.0451, "step": 2622 }, { "compression_loss": 75.53463745117188, "epoch": 0.86, "learning_rate": 1.7099211563731933e-05, "loss": 75.8224, "step": 2623 }, { "compression_loss": 76.83306884765625, "epoch": 0.86, "learning_rate": 1.7094283837056505e-05, "loss": 77.0551, "step": 2624 }, { "compression_loss": 73.35404205322266, "epoch": 0.86, "learning_rate": 1.7089356110381077e-05, "loss": 73.6154, "step": 2625 }, { "compression_loss": 78.3812255859375, "epoch": 0.86, "learning_rate": 1.7084428383705653e-05, "loss": 78.8806, "step": 2626 }, { "compression_loss": 72.62727355957031, "epoch": 0.86, "learning_rate": 1.7079500657030225e-05, "loss": 72.9949, "step": 2627 }, { "compression_loss": 75.89850616455078, "epoch": 0.86, "learning_rate": 1.7074572930354797e-05, "loss": 76.2889, "step": 2628 }, { "compression_loss": 76.45746612548828, "epoch": 0.86, "learning_rate": 1.706964520367937e-05, "loss": 76.8999, "step": 2629 }, { "compression_loss": 76.40148162841797, "epoch": 0.86, "learning_rate": 1.706471747700394e-05, "loss": 77.007, "step": 2630 }, { "compression_loss": 73.66362762451172, "epoch": 0.86, "learning_rate": 1.7059789750328516e-05, "loss": 74.1895, "step": 2631 }, { "compression_loss": 77.2634506225586, "epoch": 0.86, "learning_rate": 1.7054862023653088e-05, "loss": 77.6145, "step": 2632 }, { "compression_loss": 77.01783752441406, "epoch": 0.86, "learning_rate": 1.7049934296977663e-05, "loss": 77.6644, "step": 2633 }, { "compression_loss": 76.27117919921875, "epoch": 0.87, "learning_rate": 1.7045006570302235e-05, "loss": 76.5557, "step": 2634 }, { "compression_loss": 73.965576171875, "epoch": 0.87, "learning_rate": 1.7040078843626807e-05, "loss": 74.2166, "step": 2635 }, { "compression_loss": 76.67503356933594, "epoch": 0.87, "learning_rate": 1.7035151116951382e-05, "loss": 76.7768, "step": 2636 }, { "compression_loss": 77.44540405273438, "epoch": 0.87, "learning_rate": 1.7030223390275954e-05, "loss": 77.7544, "step": 2637 }, { "compression_loss": 73.61602783203125, "epoch": 0.87, "learning_rate": 1.7025295663600526e-05, "loss": 73.96, "step": 2638 }, { "compression_loss": 74.91342163085938, "epoch": 0.87, "learning_rate": 1.70203679369251e-05, "loss": 75.4819, "step": 2639 }, { "compression_loss": 74.94927215576172, "epoch": 0.87, "learning_rate": 1.701544021024967e-05, "loss": 75.3772, "step": 2640 }, { "compression_loss": 73.89060974121094, "epoch": 0.87, "learning_rate": 1.7010512483574246e-05, "loss": 74.1902, "step": 2641 }, { "compression_loss": 75.70258331298828, "epoch": 0.87, "learning_rate": 1.7005584756898818e-05, "loss": 76.2424, "step": 2642 }, { "compression_loss": 76.05802154541016, "epoch": 0.87, "learning_rate": 1.700065703022339e-05, "loss": 76.3722, "step": 2643 }, { "compression_loss": 74.80067443847656, "epoch": 0.87, "learning_rate": 1.6995729303547965e-05, "loss": 75.6013, "step": 2644 }, { "compression_loss": 74.748779296875, "epoch": 0.87, "learning_rate": 1.6990801576872537e-05, "loss": 75.2734, "step": 2645 }, { "compression_loss": 73.72561645507812, "epoch": 0.87, "learning_rate": 1.6985873850197112e-05, "loss": 74.2024, "step": 2646 }, { "compression_loss": 76.1856918334961, "epoch": 0.87, "learning_rate": 1.6980946123521684e-05, "loss": 76.5818, "step": 2647 }, { "compression_loss": 75.84588623046875, "epoch": 0.87, "learning_rate": 1.6976018396846256e-05, "loss": 76.1599, "step": 2648 }, { "compression_loss": 75.04933166503906, "epoch": 0.87, "learning_rate": 1.6971090670170828e-05, "loss": 75.3994, "step": 2649 }, { "compression_loss": 76.533935546875, "epoch": 0.87, "learning_rate": 1.69661629434954e-05, "loss": 77.1626, "step": 2650 }, { "compression_loss": 76.15292358398438, "epoch": 0.87, "learning_rate": 1.6961235216819976e-05, "loss": 76.8268, "step": 2651 }, { "compression_loss": 76.34736633300781, "epoch": 0.87, "learning_rate": 1.6956307490144547e-05, "loss": 76.767, "step": 2652 }, { "compression_loss": 78.66326141357422, "epoch": 0.87, "learning_rate": 1.695137976346912e-05, "loss": 78.9193, "step": 2653 }, { "compression_loss": 76.94514465332031, "epoch": 0.87, "learning_rate": 1.694645203679369e-05, "loss": 77.449, "step": 2654 }, { "compression_loss": 74.88442993164062, "epoch": 0.87, "learning_rate": 1.6941524310118263e-05, "loss": 75.2045, "step": 2655 }, { "compression_loss": 74.81629943847656, "epoch": 0.87, "learning_rate": 1.6936596583442842e-05, "loss": 75.1652, "step": 2656 }, { "compression_loss": 75.64319610595703, "epoch": 0.87, "learning_rate": 1.6931668856767414e-05, "loss": 76.3032, "step": 2657 }, { "compression_loss": 75.96405029296875, "epoch": 0.87, "learning_rate": 1.6926741130091986e-05, "loss": 76.0729, "step": 2658 }, { "compression_loss": 74.88522338867188, "epoch": 0.87, "learning_rate": 1.6921813403416558e-05, "loss": 75.4145, "step": 2659 }, { "compression_loss": 74.7957763671875, "epoch": 0.87, "learning_rate": 1.691688567674113e-05, "loss": 75.1512, "step": 2660 }, { "compression_loss": 74.0363998413086, "epoch": 0.87, "learning_rate": 1.6911957950065705e-05, "loss": 74.5325, "step": 2661 }, { "compression_loss": 79.59412384033203, "epoch": 0.87, "learning_rate": 1.6907030223390277e-05, "loss": 80.2978, "step": 2662 }, { "compression_loss": 75.8028564453125, "epoch": 0.87, "learning_rate": 1.690210249671485e-05, "loss": 76.2148, "step": 2663 }, { "compression_loss": 76.12033081054688, "epoch": 0.88, "learning_rate": 1.689717477003942e-05, "loss": 76.6672, "step": 2664 }, { "compression_loss": 74.00347137451172, "epoch": 0.88, "learning_rate": 1.6892247043363993e-05, "loss": 74.3585, "step": 2665 }, { "compression_loss": 74.60504150390625, "epoch": 0.88, "learning_rate": 1.688731931668857e-05, "loss": 75.1145, "step": 2666 }, { "compression_loss": 75.63481903076172, "epoch": 0.88, "learning_rate": 1.688239159001314e-05, "loss": 75.8575, "step": 2667 }, { "compression_loss": 75.3844985961914, "epoch": 0.88, "learning_rate": 1.6877463863337716e-05, "loss": 75.9125, "step": 2668 }, { "compression_loss": 74.10002136230469, "epoch": 0.88, "learning_rate": 1.6872536136662288e-05, "loss": 74.3545, "step": 2669 }, { "compression_loss": 73.24994659423828, "epoch": 0.88, "learning_rate": 1.686760840998686e-05, "loss": 73.4585, "step": 2670 }, { "compression_loss": 76.52091217041016, "epoch": 0.88, "learning_rate": 1.6862680683311432e-05, "loss": 77.0493, "step": 2671 }, { "compression_loss": 76.83303833007812, "epoch": 0.88, "learning_rate": 1.6857752956636007e-05, "loss": 77.1218, "step": 2672 }, { "compression_loss": 74.48619079589844, "epoch": 0.88, "learning_rate": 1.685282522996058e-05, "loss": 75.1824, "step": 2673 }, { "compression_loss": 76.10357666015625, "epoch": 0.88, "learning_rate": 1.684789750328515e-05, "loss": 76.531, "step": 2674 }, { "compression_loss": 73.78019714355469, "epoch": 0.88, "learning_rate": 1.6842969776609723e-05, "loss": 74.0042, "step": 2675 }, { "compression_loss": 77.02656555175781, "epoch": 0.88, "learning_rate": 1.6838042049934295e-05, "loss": 77.9382, "step": 2676 }, { "compression_loss": 76.18803405761719, "epoch": 0.88, "learning_rate": 1.683311432325887e-05, "loss": 76.6671, "step": 2677 }, { "compression_loss": 74.66392517089844, "epoch": 0.88, "learning_rate": 1.6828186596583442e-05, "loss": 74.936, "step": 2678 }, { "compression_loss": 75.32819366455078, "epoch": 0.88, "learning_rate": 1.6823258869908014e-05, "loss": 75.7151, "step": 2679 }, { "compression_loss": 75.52596282958984, "epoch": 0.88, "learning_rate": 1.681833114323259e-05, "loss": 75.9517, "step": 2680 }, { "compression_loss": 76.40274047851562, "epoch": 0.88, "learning_rate": 1.681340341655716e-05, "loss": 77.3226, "step": 2681 }, { "compression_loss": 75.44647216796875, "epoch": 0.88, "learning_rate": 1.6808475689881737e-05, "loss": 76.0072, "step": 2682 }, { "compression_loss": 77.98319244384766, "epoch": 0.88, "learning_rate": 1.680354796320631e-05, "loss": 78.3894, "step": 2683 }, { "compression_loss": 75.82080841064453, "epoch": 0.88, "learning_rate": 1.679862023653088e-05, "loss": 76.1685, "step": 2684 }, { "compression_loss": 73.72894287109375, "epoch": 0.88, "learning_rate": 1.6793692509855453e-05, "loss": 74.018, "step": 2685 }, { "compression_loss": 75.31831359863281, "epoch": 0.88, "learning_rate": 1.6788764783180025e-05, "loss": 75.7769, "step": 2686 }, { "compression_loss": 74.74555969238281, "epoch": 0.88, "learning_rate": 1.67838370565046e-05, "loss": 74.9344, "step": 2687 }, { "compression_loss": 77.42142486572266, "epoch": 0.88, "learning_rate": 1.6778909329829172e-05, "loss": 77.7597, "step": 2688 }, { "compression_loss": 76.69819641113281, "epoch": 0.88, "learning_rate": 1.6773981603153744e-05, "loss": 76.9419, "step": 2689 }, { "compression_loss": 76.248291015625, "epoch": 0.88, "learning_rate": 1.6769053876478316e-05, "loss": 76.5823, "step": 2690 }, { "compression_loss": 75.51921081542969, "epoch": 0.88, "learning_rate": 1.6764126149802888e-05, "loss": 75.921, "step": 2691 }, { "compression_loss": 76.23255920410156, "epoch": 0.88, "learning_rate": 1.6759198423127467e-05, "loss": 76.7049, "step": 2692 }, { "compression_loss": 75.59103393554688, "epoch": 0.88, "learning_rate": 1.675427069645204e-05, "loss": 76.1121, "step": 2693 }, { "compression_loss": 77.37593078613281, "epoch": 0.89, "learning_rate": 1.674934296977661e-05, "loss": 77.7674, "step": 2694 }, { "compression_loss": 73.95298767089844, "epoch": 0.89, "learning_rate": 1.6744415243101183e-05, "loss": 74.3027, "step": 2695 }, { "compression_loss": 76.10799407958984, "epoch": 0.89, "learning_rate": 1.6739487516425755e-05, "loss": 76.5845, "step": 2696 }, { "compression_loss": 75.01485443115234, "epoch": 0.89, "learning_rate": 1.673455978975033e-05, "loss": 75.3743, "step": 2697 }, { "compression_loss": 74.8719482421875, "epoch": 0.89, "learning_rate": 1.6729632063074902e-05, "loss": 75.3112, "step": 2698 }, { "compression_loss": 75.67581176757812, "epoch": 0.89, "learning_rate": 1.6724704336399474e-05, "loss": 76.0762, "step": 2699 }, { "compression_loss": 73.22663116455078, "epoch": 0.89, "learning_rate": 1.6719776609724046e-05, "loss": 73.5271, "step": 2700 }, { "compression_loss": 73.95487213134766, "epoch": 0.89, "learning_rate": 1.6714848883048618e-05, "loss": 74.2848, "step": 2701 }, { "compression_loss": 76.78900146484375, "epoch": 0.89, "learning_rate": 1.6709921156373193e-05, "loss": 77.1142, "step": 2702 }, { "compression_loss": 76.2796630859375, "epoch": 0.89, "learning_rate": 1.6704993429697765e-05, "loss": 76.6512, "step": 2703 }, { "compression_loss": 74.87823486328125, "epoch": 0.89, "learning_rate": 1.670006570302234e-05, "loss": 75.2036, "step": 2704 }, { "compression_loss": 75.34434509277344, "epoch": 0.89, "learning_rate": 1.6695137976346913e-05, "loss": 75.7415, "step": 2705 }, { "compression_loss": 73.93787384033203, "epoch": 0.89, "learning_rate": 1.6690210249671485e-05, "loss": 74.3125, "step": 2706 }, { "compression_loss": 75.28138732910156, "epoch": 0.89, "learning_rate": 1.668528252299606e-05, "loss": 75.6874, "step": 2707 }, { "compression_loss": 75.808837890625, "epoch": 0.89, "learning_rate": 1.6680354796320632e-05, "loss": 76.3365, "step": 2708 }, { "compression_loss": 75.1055908203125, "epoch": 0.89, "learning_rate": 1.6675427069645204e-05, "loss": 75.5401, "step": 2709 }, { "compression_loss": 76.86092376708984, "epoch": 0.89, "learning_rate": 1.6670499342969776e-05, "loss": 77.3306, "step": 2710 }, { "compression_loss": 77.64093017578125, "epoch": 0.89, "learning_rate": 1.6665571616294348e-05, "loss": 77.8359, "step": 2711 }, { "compression_loss": 76.35111236572266, "epoch": 0.89, "learning_rate": 1.6660643889618923e-05, "loss": 76.9591, "step": 2712 }, { "compression_loss": 75.5794448852539, "epoch": 0.89, "learning_rate": 1.6655716162943495e-05, "loss": 75.9793, "step": 2713 }, { "compression_loss": 74.8402099609375, "epoch": 0.89, "learning_rate": 1.6650788436268067e-05, "loss": 75.1423, "step": 2714 }, { "compression_loss": 77.65736389160156, "epoch": 0.89, "learning_rate": 1.664586070959264e-05, "loss": 78.1129, "step": 2715 }, { "compression_loss": 75.26974487304688, "epoch": 0.89, "learning_rate": 1.6640932982917214e-05, "loss": 75.9189, "step": 2716 }, { "compression_loss": 77.12493896484375, "epoch": 0.89, "learning_rate": 1.663600525624179e-05, "loss": 77.6473, "step": 2717 }, { "compression_loss": 74.55355834960938, "epoch": 0.89, "learning_rate": 1.6631077529566362e-05, "loss": 75.2992, "step": 2718 }, { "compression_loss": 77.31744384765625, "epoch": 0.89, "learning_rate": 1.6626149802890934e-05, "loss": 78.2159, "step": 2719 }, { "compression_loss": 76.46258544921875, "epoch": 0.89, "learning_rate": 1.6621222076215506e-05, "loss": 76.7679, "step": 2720 }, { "compression_loss": 74.25106811523438, "epoch": 0.89, "learning_rate": 1.6616294349540078e-05, "loss": 74.7117, "step": 2721 }, { "compression_loss": 75.46815490722656, "epoch": 0.89, "learning_rate": 1.6611366622864653e-05, "loss": 76.0639, "step": 2722 }, { "compression_loss": 76.41470336914062, "epoch": 0.89, "learning_rate": 1.6606438896189225e-05, "loss": 76.9084, "step": 2723 }, { "compression_loss": 74.73402404785156, "epoch": 0.89, "learning_rate": 1.6601511169513797e-05, "loss": 75.1228, "step": 2724 }, { "compression_loss": 75.77354431152344, "epoch": 0.9, "learning_rate": 1.659658344283837e-05, "loss": 75.9846, "step": 2725 }, { "compression_loss": 76.42417907714844, "epoch": 0.9, "learning_rate": 1.659165571616294e-05, "loss": 76.9523, "step": 2726 }, { "compression_loss": 75.34868621826172, "epoch": 0.9, "learning_rate": 1.6586727989487516e-05, "loss": 75.6201, "step": 2727 }, { "compression_loss": 75.49468994140625, "epoch": 0.9, "learning_rate": 1.658180026281209e-05, "loss": 75.796, "step": 2728 }, { "compression_loss": 74.8768310546875, "epoch": 0.9, "learning_rate": 1.6576872536136664e-05, "loss": 75.16, "step": 2729 }, { "compression_loss": 75.84322357177734, "epoch": 0.9, "learning_rate": 1.6571944809461236e-05, "loss": 76.4065, "step": 2730 }, { "compression_loss": 76.64994049072266, "epoch": 0.9, "learning_rate": 1.6567017082785808e-05, "loss": 76.881, "step": 2731 }, { "compression_loss": 76.65289306640625, "epoch": 0.9, "learning_rate": 1.6562089356110383e-05, "loss": 77.155, "step": 2732 }, { "compression_loss": 76.24757385253906, "epoch": 0.9, "learning_rate": 1.6557161629434955e-05, "loss": 76.6738, "step": 2733 }, { "compression_loss": 76.82662963867188, "epoch": 0.9, "learning_rate": 1.6552233902759527e-05, "loss": 77.5133, "step": 2734 }, { "compression_loss": 73.23135375976562, "epoch": 0.9, "learning_rate": 1.65473061760841e-05, "loss": 73.4893, "step": 2735 }, { "compression_loss": 78.30955505371094, "epoch": 0.9, "learning_rate": 1.654237844940867e-05, "loss": 78.627, "step": 2736 }, { "compression_loss": 75.49173736572266, "epoch": 0.9, "learning_rate": 1.6537450722733246e-05, "loss": 76.294, "step": 2737 }, { "compression_loss": 76.9596939086914, "epoch": 0.9, "learning_rate": 1.6532522996057818e-05, "loss": 77.5699, "step": 2738 }, { "compression_loss": 75.52782440185547, "epoch": 0.9, "learning_rate": 1.652759526938239e-05, "loss": 75.7182, "step": 2739 }, { "compression_loss": 74.82439422607422, "epoch": 0.9, "learning_rate": 1.6522667542706965e-05, "loss": 75.2573, "step": 2740 }, { "compression_loss": 74.6290283203125, "epoch": 0.9, "learning_rate": 1.6517739816031537e-05, "loss": 75.0914, "step": 2741 }, { "compression_loss": 75.09817504882812, "epoch": 0.9, "learning_rate": 1.6512812089356113e-05, "loss": 75.3487, "step": 2742 }, { "compression_loss": 75.22460174560547, "epoch": 0.9, "learning_rate": 1.6507884362680685e-05, "loss": 75.5578, "step": 2743 }, { "compression_loss": 75.88468170166016, "epoch": 0.9, "learning_rate": 1.6502956636005257e-05, "loss": 76.4964, "step": 2744 }, { "compression_loss": 76.43295288085938, "epoch": 0.9, "learning_rate": 1.649802890932983e-05, "loss": 76.7245, "step": 2745 }, { "compression_loss": 74.84296417236328, "epoch": 0.9, "learning_rate": 1.64931011826544e-05, "loss": 75.3739, "step": 2746 }, { "compression_loss": 75.88294219970703, "epoch": 0.9, "learning_rate": 1.6488173455978976e-05, "loss": 76.5545, "step": 2747 }, { "compression_loss": 75.87491607666016, "epoch": 0.9, "learning_rate": 1.6483245729303548e-05, "loss": 76.354, "step": 2748 }, { "compression_loss": 76.47418975830078, "epoch": 0.9, "learning_rate": 1.647831800262812e-05, "loss": 76.9329, "step": 2749 }, { "compression_loss": 73.50695037841797, "epoch": 0.9, "learning_rate": 1.6473390275952692e-05, "loss": 73.8932, "step": 2750 }, { "epoch": 0.9, "eval_exact_match": 86.42384105960265, "eval_f1": 92.72944810031328, "step": 2750 }, { "compression_loss": 75.38846588134766, "epoch": 0.9, "learning_rate": 1.6468462549277264e-05, "loss": 75.81, "step": 2751 }, { "compression_loss": 72.66111755371094, "epoch": 0.9, "learning_rate": 1.6463534822601843e-05, "loss": 73.2067, "step": 2752 }, { "compression_loss": 71.57412719726562, "epoch": 0.9, "learning_rate": 1.6458607095926415e-05, "loss": 71.8363, "step": 2753 }, { "compression_loss": 75.8919677734375, "epoch": 0.9, "learning_rate": 1.6453679369250987e-05, "loss": 76.5058, "step": 2754 }, { "compression_loss": 74.20104217529297, "epoch": 0.91, "learning_rate": 1.644875164257556e-05, "loss": 74.5534, "step": 2755 }, { "compression_loss": 74.30094909667969, "epoch": 0.91, "learning_rate": 1.644382391590013e-05, "loss": 74.5133, "step": 2756 }, { "compression_loss": 75.99610137939453, "epoch": 0.91, "learning_rate": 1.6438896189224706e-05, "loss": 76.2971, "step": 2757 }, { "compression_loss": 75.40231323242188, "epoch": 0.91, "learning_rate": 1.6433968462549278e-05, "loss": 75.623, "step": 2758 }, { "compression_loss": 74.767333984375, "epoch": 0.91, "learning_rate": 1.642904073587385e-05, "loss": 75.3482, "step": 2759 }, { "compression_loss": 73.70635986328125, "epoch": 0.91, "learning_rate": 1.6424113009198422e-05, "loss": 73.9548, "step": 2760 }, { "compression_loss": 75.75703430175781, "epoch": 0.91, "learning_rate": 1.6419185282522994e-05, "loss": 76.4066, "step": 2761 }, { "compression_loss": 77.05740356445312, "epoch": 0.91, "learning_rate": 1.641425755584757e-05, "loss": 77.4587, "step": 2762 }, { "compression_loss": 76.75566101074219, "epoch": 0.91, "learning_rate": 1.640932982917214e-05, "loss": 77.1565, "step": 2763 }, { "compression_loss": 75.29566192626953, "epoch": 0.91, "learning_rate": 1.6404402102496716e-05, "loss": 75.8172, "step": 2764 }, { "compression_loss": 73.14503479003906, "epoch": 0.91, "learning_rate": 1.639947437582129e-05, "loss": 73.4549, "step": 2765 }, { "compression_loss": 77.88409423828125, "epoch": 0.91, "learning_rate": 1.639454664914586e-05, "loss": 78.5548, "step": 2766 }, { "compression_loss": 75.43622589111328, "epoch": 0.91, "learning_rate": 1.6389618922470436e-05, "loss": 75.6733, "step": 2767 }, { "compression_loss": 75.91902923583984, "epoch": 0.91, "learning_rate": 1.6384691195795008e-05, "loss": 76.2183, "step": 2768 }, { "compression_loss": 75.66574096679688, "epoch": 0.91, "learning_rate": 1.637976346911958e-05, "loss": 76.1576, "step": 2769 }, { "compression_loss": 77.70738220214844, "epoch": 0.91, "learning_rate": 1.637483574244415e-05, "loss": 78.2531, "step": 2770 }, { "compression_loss": 76.6207275390625, "epoch": 0.91, "learning_rate": 1.6369908015768724e-05, "loss": 77.4095, "step": 2771 }, { "compression_loss": 75.03999328613281, "epoch": 0.91, "learning_rate": 1.63649802890933e-05, "loss": 75.4532, "step": 2772 }, { "compression_loss": 77.19767761230469, "epoch": 0.91, "learning_rate": 1.636005256241787e-05, "loss": 77.5869, "step": 2773 }, { "compression_loss": 78.10542297363281, "epoch": 0.91, "learning_rate": 1.6355124835742443e-05, "loss": 78.5593, "step": 2774 }, { "compression_loss": 78.74911499023438, "epoch": 0.91, "learning_rate": 1.6350197109067018e-05, "loss": 79.2193, "step": 2775 }, { "compression_loss": 79.45819854736328, "epoch": 0.91, "learning_rate": 1.634526938239159e-05, "loss": 80.3213, "step": 2776 }, { "compression_loss": 75.65404510498047, "epoch": 0.91, "learning_rate": 1.6340341655716166e-05, "loss": 75.8874, "step": 2777 }, { "compression_loss": 75.66634368896484, "epoch": 0.91, "learning_rate": 1.6335413929040738e-05, "loss": 76.2625, "step": 2778 }, { "compression_loss": 76.332275390625, "epoch": 0.91, "learning_rate": 1.633048620236531e-05, "loss": 76.8233, "step": 2779 }, { "compression_loss": 73.75948333740234, "epoch": 0.91, "learning_rate": 1.632555847568988e-05, "loss": 74.1621, "step": 2780 }, { "compression_loss": 76.16539001464844, "epoch": 0.91, "learning_rate": 1.6320630749014453e-05, "loss": 76.8835, "step": 2781 }, { "compression_loss": 74.21723937988281, "epoch": 0.91, "learning_rate": 1.631570302233903e-05, "loss": 74.6415, "step": 2782 }, { "compression_loss": 77.17110443115234, "epoch": 0.91, "learning_rate": 1.63107752956636e-05, "loss": 77.5226, "step": 2783 }, { "compression_loss": 77.67329406738281, "epoch": 0.91, "learning_rate": 1.6305847568988173e-05, "loss": 78.1408, "step": 2784 }, { "compression_loss": 76.86851501464844, "epoch": 0.91, "learning_rate": 1.6300919842312745e-05, "loss": 77.1606, "step": 2785 }, { "compression_loss": 77.29196166992188, "epoch": 0.92, "learning_rate": 1.6295992115637317e-05, "loss": 78.036, "step": 2786 }, { "compression_loss": 75.11542510986328, "epoch": 0.92, "learning_rate": 1.6291064388961895e-05, "loss": 75.2734, "step": 2787 }, { "compression_loss": 76.1707763671875, "epoch": 0.92, "learning_rate": 1.6286136662286467e-05, "loss": 76.5331, "step": 2788 }, { "compression_loss": 76.539794921875, "epoch": 0.92, "learning_rate": 1.628120893561104e-05, "loss": 76.8912, "step": 2789 }, { "compression_loss": 77.07003021240234, "epoch": 0.92, "learning_rate": 1.627628120893561e-05, "loss": 77.4171, "step": 2790 }, { "compression_loss": 75.85684204101562, "epoch": 0.92, "learning_rate": 1.6271353482260183e-05, "loss": 76.063, "step": 2791 }, { "compression_loss": 76.2483139038086, "epoch": 0.92, "learning_rate": 1.626642575558476e-05, "loss": 76.4803, "step": 2792 }, { "compression_loss": 73.08747863769531, "epoch": 0.92, "learning_rate": 1.626149802890933e-05, "loss": 73.2773, "step": 2793 }, { "compression_loss": 75.87271118164062, "epoch": 0.92, "learning_rate": 1.6256570302233903e-05, "loss": 76.1206, "step": 2794 }, { "compression_loss": 74.59609985351562, "epoch": 0.92, "learning_rate": 1.6251642575558475e-05, "loss": 75.3824, "step": 2795 }, { "compression_loss": 75.75624084472656, "epoch": 0.92, "learning_rate": 1.6246714848883047e-05, "loss": 76.1486, "step": 2796 }, { "compression_loss": 76.00820922851562, "epoch": 0.92, "learning_rate": 1.6241787122207622e-05, "loss": 76.4974, "step": 2797 }, { "compression_loss": 75.25619506835938, "epoch": 0.92, "learning_rate": 1.6236859395532194e-05, "loss": 75.5808, "step": 2798 }, { "compression_loss": 78.10275268554688, "epoch": 0.92, "learning_rate": 1.623193166885677e-05, "loss": 78.6028, "step": 2799 }, { "compression_loss": 76.85847473144531, "epoch": 0.92, "learning_rate": 1.622700394218134e-05, "loss": 77.2076, "step": 2800 }, { "compression_loss": 77.18424987792969, "epoch": 0.92, "learning_rate": 1.6222076215505913e-05, "loss": 77.6153, "step": 2801 }, { "compression_loss": 73.40463256835938, "epoch": 0.92, "learning_rate": 1.621714848883049e-05, "loss": 73.7319, "step": 2802 }, { "compression_loss": 73.74888610839844, "epoch": 0.92, "learning_rate": 1.621222076215506e-05, "loss": 73.9898, "step": 2803 }, { "compression_loss": 74.06068420410156, "epoch": 0.92, "learning_rate": 1.6207293035479632e-05, "loss": 74.5513, "step": 2804 }, { "compression_loss": 75.56552124023438, "epoch": 0.92, "learning_rate": 1.6202365308804204e-05, "loss": 75.7959, "step": 2805 }, { "compression_loss": 75.3897933959961, "epoch": 0.92, "learning_rate": 1.6197437582128776e-05, "loss": 75.7111, "step": 2806 }, { "compression_loss": 75.55399322509766, "epoch": 0.92, "learning_rate": 1.6192509855453352e-05, "loss": 75.8074, "step": 2807 }, { "compression_loss": 76.78154754638672, "epoch": 0.92, "learning_rate": 1.6187582128777924e-05, "loss": 77.4863, "step": 2808 }, { "compression_loss": 74.78269958496094, "epoch": 0.92, "learning_rate": 1.6182654402102496e-05, "loss": 75.2871, "step": 2809 }, { "compression_loss": 74.17192840576172, "epoch": 0.92, "learning_rate": 1.6177726675427068e-05, "loss": 74.381, "step": 2810 }, { "compression_loss": 75.12493896484375, "epoch": 0.92, "learning_rate": 1.6172798948751643e-05, "loss": 75.4406, "step": 2811 }, { "compression_loss": 76.540771484375, "epoch": 0.92, "learning_rate": 1.616787122207622e-05, "loss": 76.9547, "step": 2812 }, { "compression_loss": 74.81394958496094, "epoch": 0.92, "learning_rate": 1.616294349540079e-05, "loss": 75.2975, "step": 2813 }, { "compression_loss": 78.11334991455078, "epoch": 0.92, "learning_rate": 1.6158015768725362e-05, "loss": 78.7338, "step": 2814 }, { "compression_loss": 76.31611633300781, "epoch": 0.92, "learning_rate": 1.6153088042049934e-05, "loss": 76.7168, "step": 2815 }, { "compression_loss": 75.22541046142578, "epoch": 0.93, "learning_rate": 1.6148160315374506e-05, "loss": 75.5921, "step": 2816 }, { "compression_loss": 77.6578369140625, "epoch": 0.93, "learning_rate": 1.614323258869908e-05, "loss": 78.1623, "step": 2817 }, { "compression_loss": 75.12239837646484, "epoch": 0.93, "learning_rate": 1.6138304862023654e-05, "loss": 75.3732, "step": 2818 }, { "compression_loss": 75.99229431152344, "epoch": 0.93, "learning_rate": 1.6133377135348226e-05, "loss": 76.3615, "step": 2819 }, { "compression_loss": 75.64952087402344, "epoch": 0.93, "learning_rate": 1.6128449408672798e-05, "loss": 76.0685, "step": 2820 }, { "compression_loss": 74.73223114013672, "epoch": 0.93, "learning_rate": 1.612352168199737e-05, "loss": 75.5284, "step": 2821 }, { "compression_loss": 73.45634460449219, "epoch": 0.93, "learning_rate": 1.6118593955321945e-05, "loss": 73.7774, "step": 2822 }, { "compression_loss": 79.47265625, "epoch": 0.93, "learning_rate": 1.611366622864652e-05, "loss": 79.8627, "step": 2823 }, { "compression_loss": 75.85954284667969, "epoch": 0.93, "learning_rate": 1.6108738501971092e-05, "loss": 76.1339, "step": 2824 }, { "compression_loss": 76.42218780517578, "epoch": 0.93, "learning_rate": 1.6103810775295664e-05, "loss": 76.8581, "step": 2825 }, { "compression_loss": 75.26994323730469, "epoch": 0.93, "learning_rate": 1.6098883048620236e-05, "loss": 75.6011, "step": 2826 }, { "compression_loss": 75.80720520019531, "epoch": 0.93, "learning_rate": 1.609395532194481e-05, "loss": 76.2137, "step": 2827 }, { "compression_loss": 74.09929656982422, "epoch": 0.93, "learning_rate": 1.6089027595269383e-05, "loss": 74.3941, "step": 2828 }, { "compression_loss": 73.72296905517578, "epoch": 0.93, "learning_rate": 1.6084099868593955e-05, "loss": 74.0647, "step": 2829 }, { "compression_loss": 75.03746795654297, "epoch": 0.93, "learning_rate": 1.6079172141918527e-05, "loss": 75.1803, "step": 2830 }, { "compression_loss": 75.95042419433594, "epoch": 0.93, "learning_rate": 1.60742444152431e-05, "loss": 76.6415, "step": 2831 }, { "compression_loss": 74.8966064453125, "epoch": 0.93, "learning_rate": 1.6069316688567675e-05, "loss": 75.3333, "step": 2832 }, { "compression_loss": 76.359375, "epoch": 0.93, "learning_rate": 1.6064388961892247e-05, "loss": 76.6916, "step": 2833 }, { "compression_loss": 77.1133041381836, "epoch": 0.93, "learning_rate": 1.605946123521682e-05, "loss": 77.6074, "step": 2834 }, { "compression_loss": 76.5877685546875, "epoch": 0.93, "learning_rate": 1.6054533508541394e-05, "loss": 77.0767, "step": 2835 }, { "compression_loss": 77.20748901367188, "epoch": 0.93, "learning_rate": 1.6049605781865966e-05, "loss": 78.2522, "step": 2836 }, { "compression_loss": 74.77839660644531, "epoch": 0.93, "learning_rate": 1.604467805519054e-05, "loss": 75.1864, "step": 2837 }, { "compression_loss": 74.11029052734375, "epoch": 0.93, "learning_rate": 1.6039750328515113e-05, "loss": 74.6581, "step": 2838 }, { "compression_loss": 73.95234680175781, "epoch": 0.93, "learning_rate": 1.6034822601839685e-05, "loss": 74.1985, "step": 2839 }, { "compression_loss": 75.769287109375, "epoch": 0.93, "learning_rate": 1.6029894875164257e-05, "loss": 76.7296, "step": 2840 }, { "compression_loss": 72.33897399902344, "epoch": 0.93, "learning_rate": 1.602496714848883e-05, "loss": 72.7932, "step": 2841 }, { "compression_loss": 76.33570098876953, "epoch": 0.93, "learning_rate": 1.6020039421813405e-05, "loss": 77.1793, "step": 2842 }, { "compression_loss": 76.99617004394531, "epoch": 0.93, "learning_rate": 1.6015111695137977e-05, "loss": 77.3253, "step": 2843 }, { "compression_loss": 76.25654602050781, "epoch": 0.93, "learning_rate": 1.601018396846255e-05, "loss": 76.5808, "step": 2844 }, { "compression_loss": 76.09689331054688, "epoch": 0.93, "learning_rate": 1.600525624178712e-05, "loss": 76.4555, "step": 2845 }, { "compression_loss": 75.16133117675781, "epoch": 0.93, "learning_rate": 1.6000328515111692e-05, "loss": 75.3939, "step": 2846 }, { "compression_loss": 77.29948425292969, "epoch": 0.94, "learning_rate": 1.599540078843627e-05, "loss": 77.7669, "step": 2847 }, { "compression_loss": 76.06826782226562, "epoch": 0.94, "learning_rate": 1.5990473061760843e-05, "loss": 76.2738, "step": 2848 }, { "compression_loss": 76.69061279296875, "epoch": 0.94, "learning_rate": 1.5985545335085415e-05, "loss": 76.969, "step": 2849 }, { "compression_loss": 78.89090728759766, "epoch": 0.94, "learning_rate": 1.5980617608409987e-05, "loss": 79.7048, "step": 2850 }, { "compression_loss": 73.91506958007812, "epoch": 0.94, "learning_rate": 1.597568988173456e-05, "loss": 74.1624, "step": 2851 }, { "compression_loss": 75.47048950195312, "epoch": 0.94, "learning_rate": 1.5970762155059134e-05, "loss": 76.0088, "step": 2852 }, { "compression_loss": 77.03762817382812, "epoch": 0.94, "learning_rate": 1.5965834428383706e-05, "loss": 77.6602, "step": 2853 }, { "compression_loss": 76.8829574584961, "epoch": 0.94, "learning_rate": 1.596090670170828e-05, "loss": 77.1605, "step": 2854 }, { "compression_loss": 77.33480072021484, "epoch": 0.94, "learning_rate": 1.595597897503285e-05, "loss": 78.1061, "step": 2855 }, { "compression_loss": 75.4820556640625, "epoch": 0.94, "learning_rate": 1.5951051248357422e-05, "loss": 76.0382, "step": 2856 }, { "compression_loss": 76.17256164550781, "epoch": 0.94, "learning_rate": 1.5946123521681998e-05, "loss": 76.9223, "step": 2857 }, { "compression_loss": 77.19800567626953, "epoch": 0.94, "learning_rate": 1.594119579500657e-05, "loss": 77.3942, "step": 2858 }, { "compression_loss": 75.23497772216797, "epoch": 0.94, "learning_rate": 1.5936268068331145e-05, "loss": 75.5373, "step": 2859 }, { "compression_loss": 76.64328002929688, "epoch": 0.94, "learning_rate": 1.5931340341655717e-05, "loss": 77.1549, "step": 2860 }, { "compression_loss": 74.61561584472656, "epoch": 0.94, "learning_rate": 1.592641261498029e-05, "loss": 75.128, "step": 2861 }, { "compression_loss": 71.46611022949219, "epoch": 0.94, "learning_rate": 1.5921484888304864e-05, "loss": 71.7783, "step": 2862 }, { "compression_loss": 75.93324279785156, "epoch": 0.94, "learning_rate": 1.5916557161629436e-05, "loss": 76.2202, "step": 2863 }, { "compression_loss": 75.6153564453125, "epoch": 0.94, "learning_rate": 1.5911629434954008e-05, "loss": 75.9932, "step": 2864 }, { "compression_loss": 73.80425262451172, "epoch": 0.94, "learning_rate": 1.590670170827858e-05, "loss": 74.1264, "step": 2865 }, { "compression_loss": 77.75212097167969, "epoch": 0.94, "learning_rate": 1.5901773981603152e-05, "loss": 78.2752, "step": 2866 }, { "compression_loss": 74.7979736328125, "epoch": 0.94, "learning_rate": 1.5896846254927728e-05, "loss": 75.1929, "step": 2867 }, { "compression_loss": 75.78331756591797, "epoch": 0.94, "learning_rate": 1.58919185282523e-05, "loss": 76.2727, "step": 2868 }, { "compression_loss": 75.56732940673828, "epoch": 0.94, "learning_rate": 1.588699080157687e-05, "loss": 76.0914, "step": 2869 }, { "compression_loss": 74.87611389160156, "epoch": 0.94, "learning_rate": 1.5882063074901443e-05, "loss": 75.1691, "step": 2870 }, { "compression_loss": 75.671142578125, "epoch": 0.94, "learning_rate": 1.587713534822602e-05, "loss": 75.9667, "step": 2871 }, { "compression_loss": 76.5256576538086, "epoch": 0.94, "learning_rate": 1.5872207621550594e-05, "loss": 77.0582, "step": 2872 }, { "compression_loss": 76.39361572265625, "epoch": 0.94, "learning_rate": 1.5867279894875166e-05, "loss": 76.8122, "step": 2873 }, { "compression_loss": 76.80158996582031, "epoch": 0.94, "learning_rate": 1.5862352168199738e-05, "loss": 77.375, "step": 2874 }, { "compression_loss": 76.30084991455078, "epoch": 0.94, "learning_rate": 1.585742444152431e-05, "loss": 76.8887, "step": 2875 }, { "compression_loss": 77.49400329589844, "epoch": 0.94, "learning_rate": 1.5852496714848882e-05, "loss": 77.965, "step": 2876 }, { "compression_loss": 78.53592681884766, "epoch": 0.95, "learning_rate": 1.5847568988173457e-05, "loss": 78.8495, "step": 2877 }, { "compression_loss": 79.43556213378906, "epoch": 0.95, "learning_rate": 1.584264126149803e-05, "loss": 79.9665, "step": 2878 }, { "compression_loss": 75.11679077148438, "epoch": 0.95, "learning_rate": 1.58377135348226e-05, "loss": 75.3881, "step": 2879 }, { "compression_loss": 74.49748229980469, "epoch": 0.95, "learning_rate": 1.5832785808147173e-05, "loss": 74.8611, "step": 2880 }, { "compression_loss": 75.61842346191406, "epoch": 0.95, "learning_rate": 1.5827858081471745e-05, "loss": 76.3863, "step": 2881 }, { "compression_loss": 73.18769073486328, "epoch": 0.95, "learning_rate": 1.582293035479632e-05, "loss": 73.5473, "step": 2882 }, { "compression_loss": 76.49197387695312, "epoch": 0.95, "learning_rate": 1.5818002628120896e-05, "loss": 76.7972, "step": 2883 }, { "compression_loss": 76.94786834716797, "epoch": 0.95, "learning_rate": 1.5813074901445468e-05, "loss": 77.4338, "step": 2884 }, { "compression_loss": 75.76525115966797, "epoch": 0.95, "learning_rate": 1.580814717477004e-05, "loss": 76.0985, "step": 2885 }, { "compression_loss": 76.13407897949219, "epoch": 0.95, "learning_rate": 1.5803219448094612e-05, "loss": 76.5338, "step": 2886 }, { "compression_loss": 77.20637512207031, "epoch": 0.95, "learning_rate": 1.5798291721419187e-05, "loss": 77.4606, "step": 2887 }, { "compression_loss": 76.23958587646484, "epoch": 0.95, "learning_rate": 1.579336399474376e-05, "loss": 77.0721, "step": 2888 }, { "compression_loss": 76.64481353759766, "epoch": 0.95, "learning_rate": 1.578843626806833e-05, "loss": 77.2331, "step": 2889 }, { "compression_loss": 74.85755157470703, "epoch": 0.95, "learning_rate": 1.5783508541392903e-05, "loss": 75.1302, "step": 2890 }, { "compression_loss": 76.11885070800781, "epoch": 0.95, "learning_rate": 1.5778580814717475e-05, "loss": 76.5062, "step": 2891 }, { "compression_loss": 77.59136962890625, "epoch": 0.95, "learning_rate": 1.577365308804205e-05, "loss": 78.1686, "step": 2892 }, { "compression_loss": 74.86663055419922, "epoch": 0.95, "learning_rate": 1.5768725361366622e-05, "loss": 75.2911, "step": 2893 }, { "compression_loss": 76.73597717285156, "epoch": 0.95, "learning_rate": 1.5763797634691194e-05, "loss": 77.143, "step": 2894 }, { "compression_loss": 76.95468139648438, "epoch": 0.95, "learning_rate": 1.575886990801577e-05, "loss": 77.4852, "step": 2895 }, { "compression_loss": 76.15068817138672, "epoch": 0.95, "learning_rate": 1.5753942181340342e-05, "loss": 76.4186, "step": 2896 }, { "compression_loss": 77.69955444335938, "epoch": 0.95, "learning_rate": 1.5749014454664917e-05, "loss": 77.9698, "step": 2897 }, { "compression_loss": 77.4146957397461, "epoch": 0.95, "learning_rate": 1.574408672798949e-05, "loss": 78.1064, "step": 2898 }, { "compression_loss": 76.71971893310547, "epoch": 0.95, "learning_rate": 1.573915900131406e-05, "loss": 77.3456, "step": 2899 }, { "compression_loss": 74.78284454345703, "epoch": 0.95, "learning_rate": 1.5734231274638633e-05, "loss": 75.0791, "step": 2900 }, { "compression_loss": 75.8049087524414, "epoch": 0.95, "learning_rate": 1.5729303547963205e-05, "loss": 76.114, "step": 2901 }, { "compression_loss": 74.79496002197266, "epoch": 0.95, "learning_rate": 1.572437582128778e-05, "loss": 75.0531, "step": 2902 }, { "compression_loss": 77.18086242675781, "epoch": 0.95, "learning_rate": 1.5719448094612352e-05, "loss": 77.9183, "step": 2903 }, { "compression_loss": 74.45010375976562, "epoch": 0.95, "learning_rate": 1.5714520367936924e-05, "loss": 74.7637, "step": 2904 }, { "compression_loss": 78.49347686767578, "epoch": 0.95, "learning_rate": 1.5709592641261496e-05, "loss": 78.9739, "step": 2905 }, { "compression_loss": 74.64936828613281, "epoch": 0.95, "learning_rate": 1.570466491458607e-05, "loss": 74.8326, "step": 2906 }, { "compression_loss": 76.12796783447266, "epoch": 0.95, "learning_rate": 1.5699737187910647e-05, "loss": 76.5893, "step": 2907 }, { "compression_loss": 73.47093200683594, "epoch": 0.96, "learning_rate": 1.569480946123522e-05, "loss": 73.6534, "step": 2908 }, { "compression_loss": 74.20967102050781, "epoch": 0.96, "learning_rate": 1.568988173455979e-05, "loss": 74.4458, "step": 2909 }, { "compression_loss": 76.40391540527344, "epoch": 0.96, "learning_rate": 1.5684954007884363e-05, "loss": 76.8742, "step": 2910 }, { "compression_loss": 74.90056610107422, "epoch": 0.96, "learning_rate": 1.5680026281208935e-05, "loss": 75.116, "step": 2911 }, { "compression_loss": 77.99530792236328, "epoch": 0.96, "learning_rate": 1.567509855453351e-05, "loss": 78.3063, "step": 2912 }, { "compression_loss": 75.0551986694336, "epoch": 0.96, "learning_rate": 1.5670170827858082e-05, "loss": 75.7915, "step": 2913 }, { "compression_loss": 77.7935562133789, "epoch": 0.96, "learning_rate": 1.5665243101182654e-05, "loss": 78.1453, "step": 2914 }, { "compression_loss": 74.5939712524414, "epoch": 0.96, "learning_rate": 1.5660315374507226e-05, "loss": 75.035, "step": 2915 }, { "compression_loss": 75.6129150390625, "epoch": 0.96, "learning_rate": 1.5655387647831798e-05, "loss": 76.0102, "step": 2916 }, { "compression_loss": 78.4631576538086, "epoch": 0.96, "learning_rate": 1.5650459921156373e-05, "loss": 79.0828, "step": 2917 }, { "compression_loss": 75.3790512084961, "epoch": 0.96, "learning_rate": 1.564553219448095e-05, "loss": 75.6269, "step": 2918 }, { "compression_loss": 76.22132110595703, "epoch": 0.96, "learning_rate": 1.564060446780552e-05, "loss": 76.863, "step": 2919 }, { "compression_loss": 76.42599487304688, "epoch": 0.96, "learning_rate": 1.5635676741130093e-05, "loss": 76.618, "step": 2920 }, { "compression_loss": 74.97193145751953, "epoch": 0.96, "learning_rate": 1.5630749014454665e-05, "loss": 75.5445, "step": 2921 }, { "compression_loss": 72.78303527832031, "epoch": 0.96, "learning_rate": 1.562582128777924e-05, "loss": 73.1107, "step": 2922 }, { "compression_loss": 76.57919311523438, "epoch": 0.96, "learning_rate": 1.5620893561103812e-05, "loss": 76.9933, "step": 2923 }, { "compression_loss": 75.94287109375, "epoch": 0.96, "learning_rate": 1.5615965834428384e-05, "loss": 76.3686, "step": 2924 }, { "compression_loss": 76.6729507446289, "epoch": 0.96, "learning_rate": 1.5611038107752956e-05, "loss": 77.1012, "step": 2925 }, { "compression_loss": 74.59309387207031, "epoch": 0.96, "learning_rate": 1.5606110381077528e-05, "loss": 75.0667, "step": 2926 }, { "compression_loss": 75.52742004394531, "epoch": 0.96, "learning_rate": 1.5601182654402103e-05, "loss": 75.8483, "step": 2927 }, { "compression_loss": 77.63777160644531, "epoch": 0.96, "learning_rate": 1.5596254927726675e-05, "loss": 78.1241, "step": 2928 }, { "compression_loss": 75.50177764892578, "epoch": 0.96, "learning_rate": 1.5591327201051247e-05, "loss": 75.9519, "step": 2929 }, { "compression_loss": 78.14065551757812, "epoch": 0.96, "learning_rate": 1.5586399474375823e-05, "loss": 78.7609, "step": 2930 }, { "compression_loss": 75.50395965576172, "epoch": 0.96, "learning_rate": 1.5581471747700395e-05, "loss": 75.8666, "step": 2931 }, { "compression_loss": 75.00633239746094, "epoch": 0.96, "learning_rate": 1.557654402102497e-05, "loss": 75.4359, "step": 2932 }, { "compression_loss": 76.38973999023438, "epoch": 0.96, "learning_rate": 1.5571616294349542e-05, "loss": 76.7139, "step": 2933 }, { "compression_loss": 75.96217346191406, "epoch": 0.96, "learning_rate": 1.5566688567674114e-05, "loss": 76.2148, "step": 2934 }, { "compression_loss": 75.34477996826172, "epoch": 0.96, "learning_rate": 1.5561760840998686e-05, "loss": 75.6793, "step": 2935 }, { "compression_loss": 75.03329467773438, "epoch": 0.96, "learning_rate": 1.5556833114323258e-05, "loss": 75.5146, "step": 2936 }, { "compression_loss": 74.64742279052734, "epoch": 0.96, "learning_rate": 1.5551905387647833e-05, "loss": 75.1328, "step": 2937 }, { "compression_loss": 77.30865478515625, "epoch": 0.97, "learning_rate": 1.5546977660972405e-05, "loss": 78.0596, "step": 2938 }, { "compression_loss": 77.46356201171875, "epoch": 0.97, "learning_rate": 1.5542049934296977e-05, "loss": 77.8054, "step": 2939 }, { "compression_loss": 74.01426696777344, "epoch": 0.97, "learning_rate": 1.553712220762155e-05, "loss": 74.358, "step": 2940 }, { "compression_loss": 76.9500732421875, "epoch": 0.97, "learning_rate": 1.553219448094612e-05, "loss": 77.5953, "step": 2941 }, { "compression_loss": 77.21370697021484, "epoch": 0.97, "learning_rate": 1.55272667542707e-05, "loss": 77.566, "step": 2942 }, { "compression_loss": 78.40194702148438, "epoch": 0.97, "learning_rate": 1.5522339027595272e-05, "loss": 78.8715, "step": 2943 }, { "compression_loss": 76.72787475585938, "epoch": 0.97, "learning_rate": 1.5517411300919844e-05, "loss": 77.2405, "step": 2944 }, { "compression_loss": 73.86676025390625, "epoch": 0.97, "learning_rate": 1.5512483574244416e-05, "loss": 74.5409, "step": 2945 }, { "compression_loss": 75.3653564453125, "epoch": 0.97, "learning_rate": 1.5507555847568988e-05, "loss": 75.5923, "step": 2946 }, { "compression_loss": 77.1348876953125, "epoch": 0.97, "learning_rate": 1.5502628120893563e-05, "loss": 77.6281, "step": 2947 }, { "compression_loss": 75.79500579833984, "epoch": 0.97, "learning_rate": 1.5497700394218135e-05, "loss": 76.4776, "step": 2948 }, { "compression_loss": 77.78109741210938, "epoch": 0.97, "learning_rate": 1.5492772667542707e-05, "loss": 78.228, "step": 2949 }, { "compression_loss": 75.67218017578125, "epoch": 0.97, "learning_rate": 1.548784494086728e-05, "loss": 76.4313, "step": 2950 }, { "compression_loss": 72.30950927734375, "epoch": 0.97, "learning_rate": 1.548291721419185e-05, "loss": 72.8641, "step": 2951 }, { "compression_loss": 76.73300170898438, "epoch": 0.97, "learning_rate": 1.5477989487516426e-05, "loss": 77.1518, "step": 2952 }, { "compression_loss": 76.66668701171875, "epoch": 0.97, "learning_rate": 1.5473061760840998e-05, "loss": 77.2441, "step": 2953 }, { "compression_loss": 76.07933044433594, "epoch": 0.97, "learning_rate": 1.5468134034165574e-05, "loss": 76.7822, "step": 2954 }, { "compression_loss": 75.52034759521484, "epoch": 0.97, "learning_rate": 1.5463206307490145e-05, "loss": 75.9644, "step": 2955 }, { "compression_loss": 75.70948791503906, "epoch": 0.97, "learning_rate": 1.5458278580814717e-05, "loss": 76.0628, "step": 2956 }, { "compression_loss": 74.67134094238281, "epoch": 0.97, "learning_rate": 1.5453350854139293e-05, "loss": 74.9427, "step": 2957 }, { "compression_loss": 76.64132690429688, "epoch": 0.97, "learning_rate": 1.5448423127463865e-05, "loss": 77.0211, "step": 2958 }, { "compression_loss": 77.28073120117188, "epoch": 0.97, "learning_rate": 1.5443495400788437e-05, "loss": 77.8541, "step": 2959 }, { "compression_loss": 74.87150573730469, "epoch": 0.97, "learning_rate": 1.543856767411301e-05, "loss": 75.1299, "step": 2960 }, { "compression_loss": 74.45834350585938, "epoch": 0.97, "learning_rate": 1.543363994743758e-05, "loss": 74.87, "step": 2961 }, { "compression_loss": 75.78804779052734, "epoch": 0.97, "learning_rate": 1.5428712220762156e-05, "loss": 76.0156, "step": 2962 }, { "compression_loss": 74.5462646484375, "epoch": 0.97, "learning_rate": 1.5423784494086728e-05, "loss": 74.9898, "step": 2963 }, { "compression_loss": 78.7177505493164, "epoch": 0.97, "learning_rate": 1.54188567674113e-05, "loss": 79.4825, "step": 2964 }, { "compression_loss": 77.02359008789062, "epoch": 0.97, "learning_rate": 1.5413929040735872e-05, "loss": 77.303, "step": 2965 }, { "compression_loss": 74.22102355957031, "epoch": 0.97, "learning_rate": 1.5409001314060447e-05, "loss": 74.5972, "step": 2966 }, { "compression_loss": 76.83926391601562, "epoch": 0.97, "learning_rate": 1.5404073587385023e-05, "loss": 77.8031, "step": 2967 }, { "compression_loss": 75.22386169433594, "epoch": 0.98, "learning_rate": 1.5399145860709595e-05, "loss": 75.8872, "step": 2968 }, { "compression_loss": 75.2611312866211, "epoch": 0.98, "learning_rate": 1.5394218134034167e-05, "loss": 75.4324, "step": 2969 }, { "compression_loss": 73.8929214477539, "epoch": 0.98, "learning_rate": 1.538929040735874e-05, "loss": 74.1773, "step": 2970 }, { "compression_loss": 75.52165222167969, "epoch": 0.98, "learning_rate": 1.538436268068331e-05, "loss": 75.9992, "step": 2971 }, { "compression_loss": 75.2484130859375, "epoch": 0.98, "learning_rate": 1.5379434954007886e-05, "loss": 75.7914, "step": 2972 }, { "compression_loss": 75.53742980957031, "epoch": 0.98, "learning_rate": 1.5374507227332458e-05, "loss": 76.0641, "step": 2973 }, { "compression_loss": 75.0126953125, "epoch": 0.98, "learning_rate": 1.536957950065703e-05, "loss": 75.3959, "step": 2974 }, { "compression_loss": 74.25119018554688, "epoch": 0.98, "learning_rate": 1.5364651773981602e-05, "loss": 74.9142, "step": 2975 }, { "compression_loss": 76.3727035522461, "epoch": 0.98, "learning_rate": 1.5359724047306174e-05, "loss": 76.7518, "step": 2976 }, { "compression_loss": 75.93832397460938, "epoch": 0.98, "learning_rate": 1.535479632063075e-05, "loss": 76.2338, "step": 2977 }, { "compression_loss": 76.40843200683594, "epoch": 0.98, "learning_rate": 1.5349868593955325e-05, "loss": 76.6508, "step": 2978 }, { "compression_loss": 75.44620513916016, "epoch": 0.98, "learning_rate": 1.5344940867279896e-05, "loss": 75.8546, "step": 2979 }, { "compression_loss": 77.0383071899414, "epoch": 0.98, "learning_rate": 1.534001314060447e-05, "loss": 77.3464, "step": 2980 }, { "compression_loss": 75.74327087402344, "epoch": 0.98, "learning_rate": 1.533508541392904e-05, "loss": 76.053, "step": 2981 }, { "compression_loss": 75.97164154052734, "epoch": 0.98, "learning_rate": 1.5330157687253616e-05, "loss": 76.2375, "step": 2982 }, { "compression_loss": 74.72392272949219, "epoch": 0.98, "learning_rate": 1.5325229960578188e-05, "loss": 75.1807, "step": 2983 }, { "compression_loss": 74.3062744140625, "epoch": 0.98, "learning_rate": 1.532030223390276e-05, "loss": 74.8, "step": 2984 }, { "compression_loss": 75.4466781616211, "epoch": 0.98, "learning_rate": 1.531537450722733e-05, "loss": 75.8782, "step": 2985 }, { "compression_loss": 74.19580078125, "epoch": 0.98, "learning_rate": 1.5310446780551904e-05, "loss": 74.7224, "step": 2986 }, { "compression_loss": 77.56465148925781, "epoch": 0.98, "learning_rate": 1.530551905387648e-05, "loss": 78.0179, "step": 2987 }, { "compression_loss": 76.65431213378906, "epoch": 0.98, "learning_rate": 1.530059132720105e-05, "loss": 77.2017, "step": 2988 }, { "compression_loss": 75.05848693847656, "epoch": 0.98, "learning_rate": 1.5295663600525623e-05, "loss": 75.4392, "step": 2989 }, { "compression_loss": 74.69720458984375, "epoch": 0.98, "learning_rate": 1.5290735873850198e-05, "loss": 74.9916, "step": 2990 }, { "compression_loss": 78.33522033691406, "epoch": 0.98, "learning_rate": 1.528580814717477e-05, "loss": 78.5367, "step": 2991 }, { "compression_loss": 72.74703979492188, "epoch": 0.98, "learning_rate": 1.5280880420499346e-05, "loss": 73.0777, "step": 2992 }, { "compression_loss": 75.18325805664062, "epoch": 0.98, "learning_rate": 1.5275952693823918e-05, "loss": 75.5548, "step": 2993 }, { "compression_loss": 76.31767272949219, "epoch": 0.98, "learning_rate": 1.527102496714849e-05, "loss": 76.6059, "step": 2994 }, { "compression_loss": 76.51900482177734, "epoch": 0.98, "learning_rate": 1.526609724047306e-05, "loss": 76.9189, "step": 2995 }, { "compression_loss": 76.1085205078125, "epoch": 0.98, "learning_rate": 1.5261169513797634e-05, "loss": 76.5275, "step": 2996 }, { "compression_loss": 76.74108123779297, "epoch": 0.98, "learning_rate": 1.5256241787122209e-05, "loss": 77.3292, "step": 2997 }, { "compression_loss": 77.56096649169922, "epoch": 0.98, "learning_rate": 1.525131406044678e-05, "loss": 78.2036, "step": 2998 }, { "compression_loss": 77.26799774169922, "epoch": 0.99, "learning_rate": 1.5246386333771354e-05, "loss": 77.9729, "step": 2999 }, { "compression_loss": 75.72657775878906, "epoch": 0.99, "learning_rate": 1.5241458607095926e-05, "loss": 76.3211, "step": 3000 }, { "epoch": 0.99, "eval_exact_match": 86.29139072847683, "eval_f1": 92.56694577965702, "step": 3000 }, { "compression_loss": 78.35418701171875, "epoch": 0.99, "learning_rate": 1.5236530880420498e-05, "loss": 78.7078, "step": 3001 }, { "compression_loss": 76.69426727294922, "epoch": 0.99, "learning_rate": 1.5231603153745074e-05, "loss": 77.0409, "step": 3002 }, { "compression_loss": 74.68721008300781, "epoch": 0.99, "learning_rate": 1.5226675427069646e-05, "loss": 75.2403, "step": 3003 }, { "compression_loss": 75.81426239013672, "epoch": 0.99, "learning_rate": 1.5221747700394218e-05, "loss": 76.5242, "step": 3004 }, { "compression_loss": 77.21867370605469, "epoch": 0.99, "learning_rate": 1.5216819973718791e-05, "loss": 77.7116, "step": 3005 }, { "compression_loss": 73.79153442382812, "epoch": 0.99, "learning_rate": 1.5211892247043363e-05, "loss": 74.1126, "step": 3006 }, { "compression_loss": 76.25039672851562, "epoch": 0.99, "learning_rate": 1.5206964520367939e-05, "loss": 76.8525, "step": 3007 }, { "compression_loss": 76.96761322021484, "epoch": 0.99, "learning_rate": 1.520203679369251e-05, "loss": 77.5665, "step": 3008 }, { "compression_loss": 77.2237548828125, "epoch": 0.99, "learning_rate": 1.5197109067017083e-05, "loss": 77.4876, "step": 3009 }, { "compression_loss": 75.83143615722656, "epoch": 0.99, "learning_rate": 1.5192181340341655e-05, "loss": 76.2796, "step": 3010 }, { "compression_loss": 74.39793395996094, "epoch": 0.99, "learning_rate": 1.5187253613666228e-05, "loss": 74.6792, "step": 3011 }, { "compression_loss": 74.29598999023438, "epoch": 0.99, "learning_rate": 1.5182325886990804e-05, "loss": 74.724, "step": 3012 }, { "compression_loss": 77.94599914550781, "epoch": 0.99, "learning_rate": 1.5177398160315376e-05, "loss": 78.5746, "step": 3013 }, { "compression_loss": 77.88917541503906, "epoch": 0.99, "learning_rate": 1.5172470433639948e-05, "loss": 78.1865, "step": 3014 }, { "compression_loss": 73.94964599609375, "epoch": 0.99, "learning_rate": 1.516754270696452e-05, "loss": 74.4518, "step": 3015 }, { "compression_loss": 74.08016204833984, "epoch": 0.99, "learning_rate": 1.5162614980289092e-05, "loss": 74.4486, "step": 3016 }, { "compression_loss": 75.03977966308594, "epoch": 0.99, "learning_rate": 1.5157687253613669e-05, "loss": 75.5328, "step": 3017 }, { "compression_loss": 76.36723327636719, "epoch": 0.99, "learning_rate": 1.515275952693824e-05, "loss": 76.821, "step": 3018 }, { "compression_loss": 76.42952728271484, "epoch": 0.99, "learning_rate": 1.5147831800262813e-05, "loss": 76.8698, "step": 3019 }, { "compression_loss": 78.76699829101562, "epoch": 0.99, "learning_rate": 1.5142904073587384e-05, "loss": 79.3174, "step": 3020 }, { "compression_loss": 76.56126403808594, "epoch": 0.99, "learning_rate": 1.5137976346911956e-05, "loss": 77.1259, "step": 3021 }, { "compression_loss": 77.03938293457031, "epoch": 0.99, "learning_rate": 1.5133048620236532e-05, "loss": 77.6645, "step": 3022 }, { "compression_loss": 74.62609100341797, "epoch": 0.99, "learning_rate": 1.5128120893561105e-05, "loss": 75.0421, "step": 3023 }, { "compression_loss": 77.77528381347656, "epoch": 0.99, "learning_rate": 1.5123193166885677e-05, "loss": 78.368, "step": 3024 }, { "compression_loss": 73.45359802246094, "epoch": 0.99, "learning_rate": 1.511826544021025e-05, "loss": 74.0507, "step": 3025 }, { "compression_loss": 75.20647430419922, "epoch": 0.99, "learning_rate": 1.5113337713534821e-05, "loss": 75.4228, "step": 3026 }, { "compression_loss": 74.26277160644531, "epoch": 0.99, "learning_rate": 1.5108409986859397e-05, "loss": 74.4802, "step": 3027 }, { "compression_loss": 76.27595520019531, "epoch": 0.99, "learning_rate": 1.5103482260183969e-05, "loss": 76.8718, "step": 3028 }, { "compression_loss": 74.42536926269531, "epoch": 1.0, "learning_rate": 1.5098554533508542e-05, "loss": 75.1032, "step": 3029 }, { "compression_loss": 76.14383697509766, "epoch": 1.0, "learning_rate": 1.5093626806833114e-05, "loss": 76.3659, "step": 3030 }, { "compression_loss": 75.16004943847656, "epoch": 1.0, "learning_rate": 1.5088699080157686e-05, "loss": 75.7143, "step": 3031 }, { "compression_loss": 75.6766128540039, "epoch": 1.0, "learning_rate": 1.5083771353482262e-05, "loss": 76.0429, "step": 3032 }, { "compression_loss": 75.71842956542969, "epoch": 1.0, "learning_rate": 1.5078843626806834e-05, "loss": 76.1783, "step": 3033 }, { "compression_loss": 76.08586120605469, "epoch": 1.0, "learning_rate": 1.5073915900131406e-05, "loss": 76.3688, "step": 3034 }, { "compression_loss": 78.64971923828125, "epoch": 1.0, "learning_rate": 1.506898817345598e-05, "loss": 79.1587, "step": 3035 }, { "compression_loss": 76.43787384033203, "epoch": 1.0, "learning_rate": 1.5064060446780551e-05, "loss": 76.7388, "step": 3036 }, { "compression_loss": 73.14085388183594, "epoch": 1.0, "learning_rate": 1.5059132720105127e-05, "loss": 73.4798, "step": 3037 }, { "compression_loss": 75.37791442871094, "epoch": 1.0, "learning_rate": 1.5054204993429699e-05, "loss": 75.8585, "step": 3038 }, { "compression_loss": 78.36756134033203, "epoch": 1.0, "learning_rate": 1.504927726675427e-05, "loss": 78.7308, "step": 3039 }, { "compression_loss": 78.50086212158203, "epoch": 1.0, "learning_rate": 1.5044349540078842e-05, "loss": 78.7446, "step": 3040 }, { "compression_loss": 75.1750259399414, "epoch": 1.0, "learning_rate": 1.5039421813403416e-05, "loss": 75.8357, "step": 3041 }, { "compression_loss": 76.84888458251953, "epoch": 1.0, "learning_rate": 1.5034494086727992e-05, "loss": 77.4004, "step": 3042 }, { "compression_loss": 75.49253845214844, "epoch": 1.0, "learning_rate": 1.5029566360052563e-05, "loss": 75.6701, "step": 3043 }, { "compression_loss": 75.7218246459961, "epoch": 1.0, "learning_rate": 1.5024638633377135e-05, "loss": 76.0164, "step": 3044 }, { "compression_loss": 75.76734161376953, "epoch": 1.0, "learning_rate": 1.5019710906701707e-05, "loss": 76.0679, "step": 3045 }, { "compression_loss": 73.13593292236328, "epoch": 1.0, "learning_rate": 1.5014783180026281e-05, "loss": 73.4772, "step": 3046 }, { "compression_loss": 73.06991577148438, "epoch": 1.0, "learning_rate": 1.5009855453350856e-05, "loss": 73.3455, "step": 3047 }, { "compression_loss": 76.88356018066406, "epoch": 1.0, "learning_rate": 1.5004927726675428e-05, "loss": 77.2233, "step": 3048 }, { "compression_loss": 75.615478515625, "epoch": 1.0, "learning_rate": 1.5e-05, "loss": 75.9451, "step": 3049 }, { "compression_loss": 75.33979034423828, "epoch": 1.0, "learning_rate": 1.4995072273324572e-05, "loss": 75.5722, "step": 3050 }, { "compression_loss": 77.23886108398438, "epoch": 1.0, "learning_rate": 1.4990144546649146e-05, "loss": 77.6719, "step": 3051 }, { "compression_loss": 74.50093078613281, "epoch": 1.0, "learning_rate": 1.498521681997372e-05, "loss": 74.9137, "step": 3052 }, { "compression_loss": 73.0434799194336, "epoch": 1.0, "learning_rate": 1.4980289093298293e-05, "loss": 73.3086, "step": 3053 }, { "compression_loss": 75.64857482910156, "epoch": 1.0, "learning_rate": 1.4975361366622865e-05, "loss": 75.8568, "step": 3054 }, { "compression_loss": 74.81558227539062, "epoch": 1.0, "learning_rate": 1.4970433639947437e-05, "loss": 75.0299, "step": 3055 }, { "compression_loss": 75.26705169677734, "epoch": 1.0, "learning_rate": 1.4965505913272011e-05, "loss": 75.5632, "step": 3056 }, { "compression_loss": 76.7255859375, "epoch": 1.0, "learning_rate": 1.4960578186596583e-05, "loss": 77.3459, "step": 3057 }, { "compression_loss": 74.74442291259766, "epoch": 1.0, "learning_rate": 1.4955650459921158e-05, "loss": 75.009, "step": 3058 }, { "compression_loss": 73.831298828125, "epoch": 1.0, "learning_rate": 1.495072273324573e-05, "loss": 74.17, "step": 3059 }, { "compression_loss": 73.38060760498047, "epoch": 1.01, "learning_rate": 1.4945795006570302e-05, "loss": 73.8828, "step": 3060 }, { "compression_loss": 74.78404235839844, "epoch": 1.01, "learning_rate": 1.4940867279894876e-05, "loss": 75.0347, "step": 3061 }, { "compression_loss": 73.01480102539062, "epoch": 1.01, "learning_rate": 1.4935939553219448e-05, "loss": 73.2205, "step": 3062 }, { "compression_loss": 74.95899200439453, "epoch": 1.01, "learning_rate": 1.4931011826544022e-05, "loss": 75.3348, "step": 3063 }, { "compression_loss": 76.77436828613281, "epoch": 1.01, "learning_rate": 1.4926084099868595e-05, "loss": 77.1097, "step": 3064 }, { "compression_loss": 76.44856262207031, "epoch": 1.01, "learning_rate": 1.4921156373193167e-05, "loss": 76.702, "step": 3065 }, { "compression_loss": 72.79564666748047, "epoch": 1.01, "learning_rate": 1.491622864651774e-05, "loss": 73.1745, "step": 3066 }, { "compression_loss": 75.12013244628906, "epoch": 1.01, "learning_rate": 1.4911300919842313e-05, "loss": 75.325, "step": 3067 }, { "compression_loss": 76.67034912109375, "epoch": 1.01, "learning_rate": 1.4906373193166886e-05, "loss": 77.1674, "step": 3068 }, { "compression_loss": 76.26809692382812, "epoch": 1.01, "learning_rate": 1.4901445466491458e-05, "loss": 76.5394, "step": 3069 }, { "compression_loss": 75.24549102783203, "epoch": 1.01, "learning_rate": 1.4896517739816032e-05, "loss": 75.7073, "step": 3070 }, { "compression_loss": 75.26736450195312, "epoch": 1.01, "learning_rate": 1.4891590013140606e-05, "loss": 75.4322, "step": 3071 }, { "compression_loss": 77.63442993164062, "epoch": 1.01, "learning_rate": 1.4886662286465178e-05, "loss": 78.3579, "step": 3072 }, { "compression_loss": 76.48011779785156, "epoch": 1.01, "learning_rate": 1.4881734559789751e-05, "loss": 76.8116, "step": 3073 }, { "compression_loss": 74.6182632446289, "epoch": 1.01, "learning_rate": 1.4876806833114323e-05, "loss": 74.9563, "step": 3074 }, { "compression_loss": 75.60346984863281, "epoch": 1.01, "learning_rate": 1.4871879106438895e-05, "loss": 76.1598, "step": 3075 }, { "compression_loss": 75.74246978759766, "epoch": 1.01, "learning_rate": 1.486695137976347e-05, "loss": 76.137, "step": 3076 }, { "compression_loss": 75.59705352783203, "epoch": 1.01, "learning_rate": 1.4862023653088043e-05, "loss": 76.1017, "step": 3077 }, { "compression_loss": 76.57225036621094, "epoch": 1.01, "learning_rate": 1.4857095926412616e-05, "loss": 76.9318, "step": 3078 }, { "compression_loss": 75.60908508300781, "epoch": 1.01, "learning_rate": 1.4852168199737188e-05, "loss": 75.9082, "step": 3079 }, { "compression_loss": 75.75367736816406, "epoch": 1.01, "learning_rate": 1.484724047306176e-05, "loss": 76.118, "step": 3080 }, { "compression_loss": 74.54713439941406, "epoch": 1.01, "learning_rate": 1.4842312746386334e-05, "loss": 75.0123, "step": 3081 }, { "compression_loss": 74.27625274658203, "epoch": 1.01, "learning_rate": 1.4837385019710908e-05, "loss": 74.523, "step": 3082 }, { "compression_loss": 76.30803680419922, "epoch": 1.01, "learning_rate": 1.4832457293035481e-05, "loss": 76.7309, "step": 3083 }, { "compression_loss": 77.16230773925781, "epoch": 1.01, "learning_rate": 1.4827529566360053e-05, "loss": 77.4632, "step": 3084 }, { "compression_loss": 74.31436920166016, "epoch": 1.01, "learning_rate": 1.4822601839684625e-05, "loss": 74.8777, "step": 3085 }, { "compression_loss": 79.15486907958984, "epoch": 1.01, "learning_rate": 1.4817674113009199e-05, "loss": 79.4587, "step": 3086 }, { "compression_loss": 76.41411590576172, "epoch": 1.01, "learning_rate": 1.481274638633377e-05, "loss": 76.7798, "step": 3087 }, { "compression_loss": 75.07832336425781, "epoch": 1.01, "learning_rate": 1.4807818659658346e-05, "loss": 75.2801, "step": 3088 }, { "compression_loss": 75.2938003540039, "epoch": 1.01, "learning_rate": 1.4802890932982918e-05, "loss": 76.0101, "step": 3089 }, { "compression_loss": 77.23997497558594, "epoch": 1.02, "learning_rate": 1.479796320630749e-05, "loss": 77.5748, "step": 3090 }, { "compression_loss": 76.00066375732422, "epoch": 1.02, "learning_rate": 1.4793035479632064e-05, "loss": 76.419, "step": 3091 }, { "compression_loss": 75.98220825195312, "epoch": 1.02, "learning_rate": 1.4788107752956636e-05, "loss": 76.6544, "step": 3092 }, { "compression_loss": 74.73873901367188, "epoch": 1.02, "learning_rate": 1.478318002628121e-05, "loss": 75.2386, "step": 3093 }, { "compression_loss": 76.21705627441406, "epoch": 1.02, "learning_rate": 1.4778252299605783e-05, "loss": 76.5256, "step": 3094 }, { "compression_loss": 73.81959533691406, "epoch": 1.02, "learning_rate": 1.4773324572930355e-05, "loss": 74.172, "step": 3095 }, { "compression_loss": 74.21852111816406, "epoch": 1.02, "learning_rate": 1.4768396846254929e-05, "loss": 74.5795, "step": 3096 }, { "compression_loss": 75.29377746582031, "epoch": 1.02, "learning_rate": 1.47634691195795e-05, "loss": 75.4352, "step": 3097 }, { "compression_loss": 76.65345001220703, "epoch": 1.02, "learning_rate": 1.4758541392904074e-05, "loss": 76.9981, "step": 3098 }, { "compression_loss": 75.81827545166016, "epoch": 1.02, "learning_rate": 1.4753613666228646e-05, "loss": 75.9503, "step": 3099 }, { "compression_loss": 75.85006713867188, "epoch": 1.02, "learning_rate": 1.474868593955322e-05, "loss": 76.3322, "step": 3100 }, { "compression_loss": 76.47883605957031, "epoch": 1.02, "learning_rate": 1.4743758212877794e-05, "loss": 76.7369, "step": 3101 }, { "compression_loss": 75.27719116210938, "epoch": 1.02, "learning_rate": 1.4738830486202366e-05, "loss": 75.46, "step": 3102 }, { "compression_loss": 76.32270050048828, "epoch": 1.02, "learning_rate": 1.473390275952694e-05, "loss": 76.7461, "step": 3103 }, { "compression_loss": 75.15223693847656, "epoch": 1.02, "learning_rate": 1.4728975032851511e-05, "loss": 75.566, "step": 3104 }, { "compression_loss": 75.75655364990234, "epoch": 1.02, "learning_rate": 1.4724047306176083e-05, "loss": 76.0692, "step": 3105 }, { "compression_loss": 73.07659149169922, "epoch": 1.02, "learning_rate": 1.4719119579500659e-05, "loss": 73.3887, "step": 3106 }, { "compression_loss": 74.36385345458984, "epoch": 1.02, "learning_rate": 1.471419185282523e-05, "loss": 74.6684, "step": 3107 }, { "compression_loss": 76.24002075195312, "epoch": 1.02, "learning_rate": 1.4709264126149804e-05, "loss": 76.6531, "step": 3108 }, { "compression_loss": 73.48294830322266, "epoch": 1.02, "learning_rate": 1.4704336399474376e-05, "loss": 73.7219, "step": 3109 }, { "compression_loss": 76.19140625, "epoch": 1.02, "learning_rate": 1.4699408672798948e-05, "loss": 76.7108, "step": 3110 }, { "compression_loss": 74.35588073730469, "epoch": 1.02, "learning_rate": 1.4694480946123522e-05, "loss": 74.6817, "step": 3111 }, { "compression_loss": 77.9659652709961, "epoch": 1.02, "learning_rate": 1.4689553219448095e-05, "loss": 78.5878, "step": 3112 }, { "compression_loss": 78.97233581542969, "epoch": 1.02, "learning_rate": 1.4684625492772669e-05, "loss": 79.3102, "step": 3113 }, { "compression_loss": 74.19921875, "epoch": 1.02, "learning_rate": 1.4679697766097241e-05, "loss": 74.6026, "step": 3114 }, { "compression_loss": 74.13595581054688, "epoch": 1.02, "learning_rate": 1.4674770039421813e-05, "loss": 74.4319, "step": 3115 }, { "compression_loss": 76.58654022216797, "epoch": 1.02, "learning_rate": 1.4669842312746387e-05, "loss": 76.8713, "step": 3116 }, { "compression_loss": 73.3890380859375, "epoch": 1.02, "learning_rate": 1.4664914586070959e-05, "loss": 73.7261, "step": 3117 }, { "compression_loss": 75.54325866699219, "epoch": 1.02, "learning_rate": 1.4659986859395534e-05, "loss": 75.8646, "step": 3118 }, { "compression_loss": 74.92326354980469, "epoch": 1.02, "learning_rate": 1.4655059132720106e-05, "loss": 75.2191, "step": 3119 }, { "compression_loss": 77.0869369506836, "epoch": 1.02, "learning_rate": 1.4650131406044678e-05, "loss": 77.6626, "step": 3120 }, { "compression_loss": 78.53611755371094, "epoch": 1.03, "learning_rate": 1.4645203679369252e-05, "loss": 79.0686, "step": 3121 }, { "compression_loss": 75.83645629882812, "epoch": 1.03, "learning_rate": 1.4640275952693824e-05, "loss": 76.1472, "step": 3122 }, { "compression_loss": 76.99454498291016, "epoch": 1.03, "learning_rate": 1.4635348226018397e-05, "loss": 77.3512, "step": 3123 }, { "compression_loss": 75.06969451904297, "epoch": 1.03, "learning_rate": 1.4630420499342971e-05, "loss": 75.2624, "step": 3124 }, { "compression_loss": 76.50563049316406, "epoch": 1.03, "learning_rate": 1.4625492772667543e-05, "loss": 76.7407, "step": 3125 }, { "compression_loss": 76.74613952636719, "epoch": 1.03, "learning_rate": 1.4620565045992117e-05, "loss": 77.1448, "step": 3126 }, { "compression_loss": 74.73594665527344, "epoch": 1.03, "learning_rate": 1.4615637319316689e-05, "loss": 74.9736, "step": 3127 }, { "compression_loss": 76.66714477539062, "epoch": 1.03, "learning_rate": 1.4610709592641262e-05, "loss": 77.0268, "step": 3128 }, { "compression_loss": 79.61216735839844, "epoch": 1.03, "learning_rate": 1.4605781865965834e-05, "loss": 80.1553, "step": 3129 }, { "compression_loss": 74.23347473144531, "epoch": 1.03, "learning_rate": 1.4600854139290408e-05, "loss": 74.5449, "step": 3130 }, { "compression_loss": 78.2633056640625, "epoch": 1.03, "learning_rate": 1.4595926412614981e-05, "loss": 78.5284, "step": 3131 }, { "compression_loss": 75.4664077758789, "epoch": 1.03, "learning_rate": 1.4590998685939553e-05, "loss": 76.0043, "step": 3132 }, { "compression_loss": 75.79086303710938, "epoch": 1.03, "learning_rate": 1.4586070959264127e-05, "loss": 76.2623, "step": 3133 }, { "compression_loss": 76.0340347290039, "epoch": 1.03, "learning_rate": 1.4581143232588699e-05, "loss": 76.2783, "step": 3134 }, { "compression_loss": 74.95589447021484, "epoch": 1.03, "learning_rate": 1.4576215505913271e-05, "loss": 75.6076, "step": 3135 }, { "compression_loss": 75.08580780029297, "epoch": 1.03, "learning_rate": 1.4571287779237846e-05, "loss": 75.3721, "step": 3136 }, { "compression_loss": 75.21226501464844, "epoch": 1.03, "learning_rate": 1.4566360052562418e-05, "loss": 75.3758, "step": 3137 }, { "compression_loss": 77.17280578613281, "epoch": 1.03, "learning_rate": 1.4561432325886992e-05, "loss": 77.5147, "step": 3138 }, { "compression_loss": 77.83902740478516, "epoch": 1.03, "learning_rate": 1.4556504599211564e-05, "loss": 78.3099, "step": 3139 }, { "compression_loss": 75.2977066040039, "epoch": 1.03, "learning_rate": 1.4551576872536136e-05, "loss": 75.8361, "step": 3140 }, { "compression_loss": 74.96184539794922, "epoch": 1.03, "learning_rate": 1.454664914586071e-05, "loss": 75.328, "step": 3141 }, { "compression_loss": 76.35252380371094, "epoch": 1.03, "learning_rate": 1.4541721419185283e-05, "loss": 76.7157, "step": 3142 }, { "compression_loss": 75.1014175415039, "epoch": 1.03, "learning_rate": 1.4536793692509857e-05, "loss": 75.4067, "step": 3143 }, { "compression_loss": 76.26419067382812, "epoch": 1.03, "learning_rate": 1.4531865965834429e-05, "loss": 76.6346, "step": 3144 }, { "compression_loss": 75.14620971679688, "epoch": 1.03, "learning_rate": 1.4526938239159001e-05, "loss": 75.6818, "step": 3145 }, { "compression_loss": 77.36344146728516, "epoch": 1.03, "learning_rate": 1.4522010512483575e-05, "loss": 77.5322, "step": 3146 }, { "compression_loss": 74.7011489868164, "epoch": 1.03, "learning_rate": 1.4517082785808147e-05, "loss": 75.0569, "step": 3147 }, { "compression_loss": 77.08274841308594, "epoch": 1.03, "learning_rate": 1.4512155059132722e-05, "loss": 77.4066, "step": 3148 }, { "compression_loss": 76.27565002441406, "epoch": 1.03, "learning_rate": 1.4507227332457294e-05, "loss": 76.5628, "step": 3149 }, { "compression_loss": 74.92350769042969, "epoch": 1.03, "learning_rate": 1.4502299605781866e-05, "loss": 75.3121, "step": 3150 }, { "compression_loss": 74.1270523071289, "epoch": 1.04, "learning_rate": 1.449737187910644e-05, "loss": 74.4282, "step": 3151 }, { "compression_loss": 74.30439758300781, "epoch": 1.04, "learning_rate": 1.4492444152431011e-05, "loss": 74.7328, "step": 3152 }, { "compression_loss": 75.66256713867188, "epoch": 1.04, "learning_rate": 1.4487516425755585e-05, "loss": 76.111, "step": 3153 }, { "compression_loss": 75.39617919921875, "epoch": 1.04, "learning_rate": 1.4482588699080159e-05, "loss": 75.6603, "step": 3154 }, { "compression_loss": 77.02520751953125, "epoch": 1.04, "learning_rate": 1.447766097240473e-05, "loss": 77.3328, "step": 3155 }, { "compression_loss": 76.34950256347656, "epoch": 1.04, "learning_rate": 1.4472733245729304e-05, "loss": 76.7411, "step": 3156 }, { "compression_loss": 75.60212707519531, "epoch": 1.04, "learning_rate": 1.4467805519053876e-05, "loss": 75.9643, "step": 3157 }, { "compression_loss": 75.12751007080078, "epoch": 1.04, "learning_rate": 1.446287779237845e-05, "loss": 75.4308, "step": 3158 }, { "compression_loss": 76.41516876220703, "epoch": 1.04, "learning_rate": 1.4457950065703022e-05, "loss": 76.8969, "step": 3159 }, { "compression_loss": 76.59127044677734, "epoch": 1.04, "learning_rate": 1.4453022339027596e-05, "loss": 76.9605, "step": 3160 }, { "compression_loss": 74.5540771484375, "epoch": 1.04, "learning_rate": 1.444809461235217e-05, "loss": 74.9083, "step": 3161 }, { "compression_loss": 78.578369140625, "epoch": 1.04, "learning_rate": 1.4443166885676741e-05, "loss": 78.9893, "step": 3162 }, { "compression_loss": 76.30551147460938, "epoch": 1.04, "learning_rate": 1.4438239159001315e-05, "loss": 77.0472, "step": 3163 }, { "compression_loss": 73.69147491455078, "epoch": 1.04, "learning_rate": 1.4433311432325887e-05, "loss": 74.3765, "step": 3164 }, { "compression_loss": 76.15455627441406, "epoch": 1.04, "learning_rate": 1.4428383705650459e-05, "loss": 76.6809, "step": 3165 }, { "compression_loss": 75.92290496826172, "epoch": 1.04, "learning_rate": 1.4423455978975034e-05, "loss": 76.2653, "step": 3166 }, { "compression_loss": 75.52066802978516, "epoch": 1.04, "learning_rate": 1.4418528252299606e-05, "loss": 75.8555, "step": 3167 }, { "compression_loss": 72.83901977539062, "epoch": 1.04, "learning_rate": 1.441360052562418e-05, "loss": 73.1513, "step": 3168 }, { "compression_loss": 76.77992248535156, "epoch": 1.04, "learning_rate": 1.4408672798948752e-05, "loss": 77.4052, "step": 3169 }, { "compression_loss": 75.224365234375, "epoch": 1.04, "learning_rate": 1.4403745072273324e-05, "loss": 75.6175, "step": 3170 }, { "compression_loss": 76.70214080810547, "epoch": 1.04, "learning_rate": 1.4398817345597898e-05, "loss": 77.113, "step": 3171 }, { "compression_loss": 77.583740234375, "epoch": 1.04, "learning_rate": 1.4393889618922471e-05, "loss": 78.1577, "step": 3172 }, { "compression_loss": 74.054931640625, "epoch": 1.04, "learning_rate": 1.4388961892247045e-05, "loss": 74.4738, "step": 3173 }, { "compression_loss": 77.36629486083984, "epoch": 1.04, "learning_rate": 1.4384034165571617e-05, "loss": 77.7543, "step": 3174 }, { "compression_loss": 76.367431640625, "epoch": 1.04, "learning_rate": 1.4379106438896189e-05, "loss": 76.9609, "step": 3175 }, { "compression_loss": 75.3429946899414, "epoch": 1.04, "learning_rate": 1.4374178712220762e-05, "loss": 75.6158, "step": 3176 }, { "compression_loss": 78.38741302490234, "epoch": 1.04, "learning_rate": 1.4369250985545334e-05, "loss": 78.776, "step": 3177 }, { "compression_loss": 76.48027038574219, "epoch": 1.04, "learning_rate": 1.436432325886991e-05, "loss": 76.8406, "step": 3178 }, { "compression_loss": 76.43888854980469, "epoch": 1.04, "learning_rate": 1.4359395532194482e-05, "loss": 76.925, "step": 3179 }, { "compression_loss": 75.78968811035156, "epoch": 1.04, "learning_rate": 1.4354467805519054e-05, "loss": 76.0778, "step": 3180 }, { "compression_loss": 80.3095703125, "epoch": 1.05, "learning_rate": 1.4349540078843627e-05, "loss": 81.0586, "step": 3181 }, { "compression_loss": 76.37195587158203, "epoch": 1.05, "learning_rate": 1.43446123521682e-05, "loss": 76.8849, "step": 3182 }, { "compression_loss": 73.82388305664062, "epoch": 1.05, "learning_rate": 1.4339684625492773e-05, "loss": 74.3456, "step": 3183 }, { "compression_loss": 74.07659149169922, "epoch": 1.05, "learning_rate": 1.4334756898817347e-05, "loss": 74.4819, "step": 3184 }, { "compression_loss": 73.50682830810547, "epoch": 1.05, "learning_rate": 1.4329829172141919e-05, "loss": 73.8495, "step": 3185 }, { "compression_loss": 74.47123718261719, "epoch": 1.05, "learning_rate": 1.4324901445466492e-05, "loss": 74.9632, "step": 3186 }, { "compression_loss": 75.33113098144531, "epoch": 1.05, "learning_rate": 1.4319973718791064e-05, "loss": 75.6954, "step": 3187 }, { "compression_loss": 77.04011535644531, "epoch": 1.05, "learning_rate": 1.4315045992115638e-05, "loss": 77.3906, "step": 3188 }, { "compression_loss": 73.03001403808594, "epoch": 1.05, "learning_rate": 1.4310118265440212e-05, "loss": 73.2722, "step": 3189 }, { "compression_loss": 74.90160369873047, "epoch": 1.05, "learning_rate": 1.4305190538764784e-05, "loss": 75.2053, "step": 3190 }, { "compression_loss": 75.68270111083984, "epoch": 1.05, "learning_rate": 1.4300262812089357e-05, "loss": 76.4003, "step": 3191 }, { "compression_loss": 76.07331085205078, "epoch": 1.05, "learning_rate": 1.429533508541393e-05, "loss": 76.4316, "step": 3192 }, { "compression_loss": 76.42118835449219, "epoch": 1.05, "learning_rate": 1.4290407358738503e-05, "loss": 76.7119, "step": 3193 }, { "compression_loss": 75.45008087158203, "epoch": 1.05, "learning_rate": 1.4285479632063075e-05, "loss": 75.7751, "step": 3194 }, { "compression_loss": 75.02334594726562, "epoch": 1.05, "learning_rate": 1.4280551905387648e-05, "loss": 75.3186, "step": 3195 }, { "compression_loss": 74.61396789550781, "epoch": 1.05, "learning_rate": 1.4275624178712222e-05, "loss": 75.0801, "step": 3196 }, { "compression_loss": 75.48827362060547, "epoch": 1.05, "learning_rate": 1.4270696452036794e-05, "loss": 75.6737, "step": 3197 }, { "compression_loss": 75.37684631347656, "epoch": 1.05, "learning_rate": 1.4265768725361368e-05, "loss": 75.6528, "step": 3198 }, { "compression_loss": 76.73030853271484, "epoch": 1.05, "learning_rate": 1.426084099868594e-05, "loss": 77.129, "step": 3199 }, { "compression_loss": 76.67121124267578, "epoch": 1.05, "learning_rate": 1.4255913272010512e-05, "loss": 77.0247, "step": 3200 }, { "compression_loss": 76.22941589355469, "epoch": 1.05, "learning_rate": 1.4250985545335087e-05, "loss": 76.6354, "step": 3201 }, { "compression_loss": 73.66712951660156, "epoch": 1.05, "learning_rate": 1.4246057818659659e-05, "loss": 73.9768, "step": 3202 }, { "compression_loss": 75.27433776855469, "epoch": 1.05, "learning_rate": 1.4241130091984233e-05, "loss": 75.6958, "step": 3203 }, { "compression_loss": 77.15730285644531, "epoch": 1.05, "learning_rate": 1.4236202365308805e-05, "loss": 77.5214, "step": 3204 }, { "compression_loss": 77.16131591796875, "epoch": 1.05, "learning_rate": 1.4231274638633377e-05, "loss": 77.5255, "step": 3205 }, { "compression_loss": 74.90869140625, "epoch": 1.05, "learning_rate": 1.422634691195795e-05, "loss": 75.3877, "step": 3206 }, { "compression_loss": 75.80569458007812, "epoch": 1.05, "learning_rate": 1.4221419185282524e-05, "loss": 76.0916, "step": 3207 }, { "compression_loss": 75.77884674072266, "epoch": 1.05, "learning_rate": 1.4216491458607098e-05, "loss": 76.0305, "step": 3208 }, { "compression_loss": 75.21893310546875, "epoch": 1.05, "learning_rate": 1.421156373193167e-05, "loss": 75.4671, "step": 3209 }, { "compression_loss": 76.14268493652344, "epoch": 1.05, "learning_rate": 1.4206636005256242e-05, "loss": 76.4009, "step": 3210 }, { "compression_loss": 74.6355209350586, "epoch": 1.05, "learning_rate": 1.4201708278580815e-05, "loss": 75.024, "step": 3211 }, { "compression_loss": 75.6751708984375, "epoch": 1.06, "learning_rate": 1.4196780551905387e-05, "loss": 76.0708, "step": 3212 }, { "compression_loss": 74.87567901611328, "epoch": 1.06, "learning_rate": 1.4191852825229963e-05, "loss": 75.2995, "step": 3213 }, { "compression_loss": 77.16416931152344, "epoch": 1.06, "learning_rate": 1.4186925098554535e-05, "loss": 77.4812, "step": 3214 }, { "compression_loss": 75.4620590209961, "epoch": 1.06, "learning_rate": 1.4181997371879106e-05, "loss": 76.0398, "step": 3215 }, { "compression_loss": 74.14447021484375, "epoch": 1.06, "learning_rate": 1.417706964520368e-05, "loss": 74.7533, "step": 3216 }, { "compression_loss": 78.20365905761719, "epoch": 1.06, "learning_rate": 1.4172141918528252e-05, "loss": 79.0306, "step": 3217 }, { "compression_loss": 75.76918029785156, "epoch": 1.06, "learning_rate": 1.4167214191852826e-05, "loss": 76.1227, "step": 3218 }, { "compression_loss": 76.48028564453125, "epoch": 1.06, "learning_rate": 1.41622864651774e-05, "loss": 76.7442, "step": 3219 }, { "compression_loss": 74.83412170410156, "epoch": 1.06, "learning_rate": 1.4157358738501971e-05, "loss": 75.1067, "step": 3220 }, { "compression_loss": 73.7795181274414, "epoch": 1.06, "learning_rate": 1.4152431011826545e-05, "loss": 74.1245, "step": 3221 }, { "compression_loss": 73.89599609375, "epoch": 1.06, "learning_rate": 1.4147503285151117e-05, "loss": 74.2732, "step": 3222 }, { "compression_loss": 78.61377716064453, "epoch": 1.06, "learning_rate": 1.414257555847569e-05, "loss": 78.9471, "step": 3223 }, { "compression_loss": 77.87340545654297, "epoch": 1.06, "learning_rate": 1.4137647831800263e-05, "loss": 78.2333, "step": 3224 }, { "compression_loss": 74.7263412475586, "epoch": 1.06, "learning_rate": 1.4132720105124836e-05, "loss": 75.0999, "step": 3225 }, { "compression_loss": 74.98123931884766, "epoch": 1.06, "learning_rate": 1.412779237844941e-05, "loss": 75.5127, "step": 3226 }, { "compression_loss": 77.5269546508789, "epoch": 1.06, "learning_rate": 1.4122864651773982e-05, "loss": 77.9349, "step": 3227 }, { "compression_loss": 76.47298431396484, "epoch": 1.06, "learning_rate": 1.4117936925098556e-05, "loss": 76.8723, "step": 3228 }, { "compression_loss": 74.9513931274414, "epoch": 1.06, "learning_rate": 1.4113009198423128e-05, "loss": 75.2996, "step": 3229 }, { "compression_loss": 74.6775131225586, "epoch": 1.06, "learning_rate": 1.41080814717477e-05, "loss": 75.1277, "step": 3230 }, { "compression_loss": 75.3004150390625, "epoch": 1.06, "learning_rate": 1.4103153745072275e-05, "loss": 75.7648, "step": 3231 }, { "compression_loss": 74.82454681396484, "epoch": 1.06, "learning_rate": 1.4098226018396847e-05, "loss": 75.2864, "step": 3232 }, { "compression_loss": 77.11573791503906, "epoch": 1.06, "learning_rate": 1.409329829172142e-05, "loss": 77.4767, "step": 3233 }, { "compression_loss": 75.18433380126953, "epoch": 1.06, "learning_rate": 1.4088370565045993e-05, "loss": 75.659, "step": 3234 }, { "compression_loss": 76.86778259277344, "epoch": 1.06, "learning_rate": 1.4083442838370565e-05, "loss": 77.2437, "step": 3235 }, { "compression_loss": 73.92819213867188, "epoch": 1.06, "learning_rate": 1.4078515111695138e-05, "loss": 74.1077, "step": 3236 }, { "compression_loss": 75.19476318359375, "epoch": 1.06, "learning_rate": 1.4073587385019712e-05, "loss": 75.588, "step": 3237 }, { "compression_loss": 75.92572784423828, "epoch": 1.06, "learning_rate": 1.4068659658344286e-05, "loss": 76.2991, "step": 3238 }, { "compression_loss": 74.59718322753906, "epoch": 1.06, "learning_rate": 1.4063731931668857e-05, "loss": 74.8547, "step": 3239 }, { "compression_loss": 77.59889221191406, "epoch": 1.06, "learning_rate": 1.405880420499343e-05, "loss": 78.0777, "step": 3240 }, { "compression_loss": 75.03480529785156, "epoch": 1.06, "learning_rate": 1.4053876478318003e-05, "loss": 75.6004, "step": 3241 }, { "compression_loss": 76.64515686035156, "epoch": 1.07, "learning_rate": 1.4048948751642575e-05, "loss": 77.3217, "step": 3242 }, { "compression_loss": 76.40177917480469, "epoch": 1.07, "learning_rate": 1.4044021024967149e-05, "loss": 76.7097, "step": 3243 }, { "compression_loss": 74.29813385009766, "epoch": 1.07, "learning_rate": 1.4039093298291722e-05, "loss": 74.5799, "step": 3244 }, { "compression_loss": 73.61908721923828, "epoch": 1.07, "learning_rate": 1.4034165571616294e-05, "loss": 73.8495, "step": 3245 }, { "compression_loss": 75.47947692871094, "epoch": 1.07, "learning_rate": 1.4029237844940868e-05, "loss": 75.9674, "step": 3246 }, { "compression_loss": 76.38813018798828, "epoch": 1.07, "learning_rate": 1.402431011826544e-05, "loss": 76.8993, "step": 3247 }, { "compression_loss": 78.58035278320312, "epoch": 1.07, "learning_rate": 1.4019382391590012e-05, "loss": 79.0453, "step": 3248 }, { "compression_loss": 75.32562255859375, "epoch": 1.07, "learning_rate": 1.4014454664914587e-05, "loss": 75.8401, "step": 3249 }, { "compression_loss": 75.05097961425781, "epoch": 1.07, "learning_rate": 1.400952693823916e-05, "loss": 75.6312, "step": 3250 }, { "epoch": 1.07, "eval_exact_match": 86.27246925260171, "eval_f1": 92.59240934774718, "step": 3250 }, { "compression_loss": 77.25791931152344, "epoch": 1.07, "learning_rate": 1.4004599211563733e-05, "loss": 77.5634, "step": 3251 }, { "compression_loss": 76.00309753417969, "epoch": 1.07, "learning_rate": 1.3999671484888305e-05, "loss": 76.4383, "step": 3252 }, { "compression_loss": 76.55656433105469, "epoch": 1.07, "learning_rate": 1.3994743758212877e-05, "loss": 76.9389, "step": 3253 }, { "compression_loss": 74.62401580810547, "epoch": 1.07, "learning_rate": 1.398981603153745e-05, "loss": 74.8043, "step": 3254 }, { "compression_loss": 75.91496276855469, "epoch": 1.07, "learning_rate": 1.3984888304862024e-05, "loss": 76.4562, "step": 3255 }, { "compression_loss": 74.81614685058594, "epoch": 1.07, "learning_rate": 1.3979960578186598e-05, "loss": 75.2602, "step": 3256 }, { "compression_loss": 75.36531066894531, "epoch": 1.07, "learning_rate": 1.397503285151117e-05, "loss": 75.6851, "step": 3257 }, { "compression_loss": 75.41873168945312, "epoch": 1.07, "learning_rate": 1.3970105124835742e-05, "loss": 75.7388, "step": 3258 }, { "compression_loss": 74.1009292602539, "epoch": 1.07, "learning_rate": 1.3965177398160315e-05, "loss": 74.3923, "step": 3259 }, { "compression_loss": 73.38125610351562, "epoch": 1.07, "learning_rate": 1.3960249671484887e-05, "loss": 73.5174, "step": 3260 }, { "compression_loss": 77.42718505859375, "epoch": 1.07, "learning_rate": 1.3955321944809463e-05, "loss": 77.7774, "step": 3261 }, { "compression_loss": 76.09121704101562, "epoch": 1.07, "learning_rate": 1.3950394218134035e-05, "loss": 76.5003, "step": 3262 }, { "compression_loss": 73.86712646484375, "epoch": 1.07, "learning_rate": 1.3945466491458607e-05, "loss": 74.1679, "step": 3263 }, { "compression_loss": 76.14106750488281, "epoch": 1.07, "learning_rate": 1.394053876478318e-05, "loss": 76.5182, "step": 3264 }, { "compression_loss": 74.71839904785156, "epoch": 1.07, "learning_rate": 1.3935611038107752e-05, "loss": 75.0291, "step": 3265 }, { "compression_loss": 77.51359558105469, "epoch": 1.07, "learning_rate": 1.3930683311432326e-05, "loss": 77.9138, "step": 3266 }, { "compression_loss": 76.78097534179688, "epoch": 1.07, "learning_rate": 1.39257555847569e-05, "loss": 77.1354, "step": 3267 }, { "compression_loss": 76.15234375, "epoch": 1.07, "learning_rate": 1.3920827858081472e-05, "loss": 76.7355, "step": 3268 }, { "compression_loss": 77.58009338378906, "epoch": 1.07, "learning_rate": 1.3915900131406045e-05, "loss": 78.0134, "step": 3269 }, { "compression_loss": 77.00306701660156, "epoch": 1.07, "learning_rate": 1.3910972404730617e-05, "loss": 77.3072, "step": 3270 }, { "compression_loss": 76.66572570800781, "epoch": 1.07, "learning_rate": 1.3906044678055191e-05, "loss": 77.0154, "step": 3271 }, { "compression_loss": 76.54762268066406, "epoch": 1.07, "learning_rate": 1.3901116951379763e-05, "loss": 76.7791, "step": 3272 }, { "compression_loss": 74.60963439941406, "epoch": 1.08, "learning_rate": 1.3896189224704337e-05, "loss": 75.0077, "step": 3273 }, { "compression_loss": 76.99366760253906, "epoch": 1.08, "learning_rate": 1.389126149802891e-05, "loss": 77.2328, "step": 3274 }, { "compression_loss": 74.10089111328125, "epoch": 1.08, "learning_rate": 1.3886333771353482e-05, "loss": 74.4007, "step": 3275 }, { "compression_loss": 73.86592864990234, "epoch": 1.08, "learning_rate": 1.3881406044678056e-05, "loss": 74.22, "step": 3276 }, { "compression_loss": 76.45389556884766, "epoch": 1.08, "learning_rate": 1.3876478318002628e-05, "loss": 76.8554, "step": 3277 }, { "compression_loss": 77.311767578125, "epoch": 1.08, "learning_rate": 1.38715505913272e-05, "loss": 77.7803, "step": 3278 }, { "compression_loss": 75.77301788330078, "epoch": 1.08, "learning_rate": 1.3866622864651775e-05, "loss": 76.0939, "step": 3279 }, { "compression_loss": 74.881103515625, "epoch": 1.08, "learning_rate": 1.3861695137976347e-05, "loss": 75.2154, "step": 3280 }, { "compression_loss": 74.58872985839844, "epoch": 1.08, "learning_rate": 1.385676741130092e-05, "loss": 74.8222, "step": 3281 }, { "compression_loss": 78.11941528320312, "epoch": 1.08, "learning_rate": 1.3851839684625493e-05, "loss": 78.6023, "step": 3282 }, { "compression_loss": 76.19532775878906, "epoch": 1.08, "learning_rate": 1.3846911957950065e-05, "loss": 76.615, "step": 3283 }, { "compression_loss": 75.32916259765625, "epoch": 1.08, "learning_rate": 1.3841984231274638e-05, "loss": 75.6092, "step": 3284 }, { "compression_loss": 76.24044799804688, "epoch": 1.08, "learning_rate": 1.3837056504599212e-05, "loss": 76.4936, "step": 3285 }, { "compression_loss": 74.03977966308594, "epoch": 1.08, "learning_rate": 1.3832128777923786e-05, "loss": 74.2674, "step": 3286 }, { "compression_loss": 74.94802856445312, "epoch": 1.08, "learning_rate": 1.3827201051248358e-05, "loss": 75.3417, "step": 3287 }, { "compression_loss": 76.8419418334961, "epoch": 1.08, "learning_rate": 1.382227332457293e-05, "loss": 77.3881, "step": 3288 }, { "compression_loss": 75.04832458496094, "epoch": 1.08, "learning_rate": 1.3817345597897503e-05, "loss": 75.4511, "step": 3289 }, { "compression_loss": 74.67367553710938, "epoch": 1.08, "learning_rate": 1.3812417871222075e-05, "loss": 74.9938, "step": 3290 }, { "compression_loss": 76.27494049072266, "epoch": 1.08, "learning_rate": 1.380749014454665e-05, "loss": 76.7592, "step": 3291 }, { "compression_loss": 75.27002716064453, "epoch": 1.08, "learning_rate": 1.3802562417871223e-05, "loss": 75.5785, "step": 3292 }, { "compression_loss": 75.2228775024414, "epoch": 1.08, "learning_rate": 1.3797634691195795e-05, "loss": 75.4704, "step": 3293 }, { "compression_loss": 77.86567687988281, "epoch": 1.08, "learning_rate": 1.3792706964520368e-05, "loss": 78.4224, "step": 3294 }, { "compression_loss": 77.83024597167969, "epoch": 1.08, "learning_rate": 1.378777923784494e-05, "loss": 78.4495, "step": 3295 }, { "compression_loss": 74.96043395996094, "epoch": 1.08, "learning_rate": 1.3782851511169514e-05, "loss": 75.2857, "step": 3296 }, { "compression_loss": 77.21366119384766, "epoch": 1.08, "learning_rate": 1.3777923784494088e-05, "loss": 77.6234, "step": 3297 }, { "compression_loss": 77.20146942138672, "epoch": 1.08, "learning_rate": 1.377299605781866e-05, "loss": 77.7673, "step": 3298 }, { "compression_loss": 74.72700500488281, "epoch": 1.08, "learning_rate": 1.3768068331143233e-05, "loss": 75.0845, "step": 3299 }, { "compression_loss": 75.22114562988281, "epoch": 1.08, "learning_rate": 1.3763140604467805e-05, "loss": 75.6582, "step": 3300 }, { "compression_loss": 74.66056823730469, "epoch": 1.08, "learning_rate": 1.3758212877792379e-05, "loss": 74.9676, "step": 3301 }, { "compression_loss": 76.55508422851562, "epoch": 1.08, "learning_rate": 1.375328515111695e-05, "loss": 76.7401, "step": 3302 }, { "compression_loss": 74.45152282714844, "epoch": 1.09, "learning_rate": 1.3748357424441524e-05, "loss": 74.6176, "step": 3303 }, { "compression_loss": 75.89423370361328, "epoch": 1.09, "learning_rate": 1.3743429697766098e-05, "loss": 76.302, "step": 3304 }, { "compression_loss": 73.43158721923828, "epoch": 1.09, "learning_rate": 1.373850197109067e-05, "loss": 73.6621, "step": 3305 }, { "compression_loss": 74.58474731445312, "epoch": 1.09, "learning_rate": 1.3733574244415244e-05, "loss": 74.8519, "step": 3306 }, { "compression_loss": 73.6800537109375, "epoch": 1.09, "learning_rate": 1.3728646517739816e-05, "loss": 73.9468, "step": 3307 }, { "compression_loss": 76.61962890625, "epoch": 1.09, "learning_rate": 1.3723718791064388e-05, "loss": 76.8511, "step": 3308 }, { "compression_loss": 75.7663345336914, "epoch": 1.09, "learning_rate": 1.3718791064388963e-05, "loss": 76.3189, "step": 3309 }, { "compression_loss": 75.92530822753906, "epoch": 1.09, "learning_rate": 1.3713863337713535e-05, "loss": 76.2418, "step": 3310 }, { "compression_loss": 76.84322357177734, "epoch": 1.09, "learning_rate": 1.3708935611038109e-05, "loss": 77.116, "step": 3311 }, { "compression_loss": 74.47103118896484, "epoch": 1.09, "learning_rate": 1.370400788436268e-05, "loss": 74.7404, "step": 3312 }, { "compression_loss": 73.86911010742188, "epoch": 1.09, "learning_rate": 1.3699080157687253e-05, "loss": 74.1779, "step": 3313 }, { "compression_loss": 76.02349090576172, "epoch": 1.09, "learning_rate": 1.3694152431011826e-05, "loss": 76.1956, "step": 3314 }, { "compression_loss": 75.33415985107422, "epoch": 1.09, "learning_rate": 1.36892247043364e-05, "loss": 75.7583, "step": 3315 }, { "compression_loss": 75.6747055053711, "epoch": 1.09, "learning_rate": 1.3684296977660974e-05, "loss": 76.0437, "step": 3316 }, { "compression_loss": 75.53579711914062, "epoch": 1.09, "learning_rate": 1.3679369250985546e-05, "loss": 75.7743, "step": 3317 }, { "compression_loss": 77.00210571289062, "epoch": 1.09, "learning_rate": 1.3674441524310118e-05, "loss": 77.4086, "step": 3318 }, { "compression_loss": 76.33262634277344, "epoch": 1.09, "learning_rate": 1.3669513797634691e-05, "loss": 76.5818, "step": 3319 }, { "compression_loss": 76.94058990478516, "epoch": 1.09, "learning_rate": 1.3664586070959263e-05, "loss": 77.0932, "step": 3320 }, { "compression_loss": 78.15619659423828, "epoch": 1.09, "learning_rate": 1.3659658344283839e-05, "loss": 78.4404, "step": 3321 }, { "compression_loss": 75.15312194824219, "epoch": 1.09, "learning_rate": 1.365473061760841e-05, "loss": 75.3961, "step": 3322 }, { "compression_loss": 75.27388000488281, "epoch": 1.09, "learning_rate": 1.3649802890932982e-05, "loss": 75.4771, "step": 3323 }, { "compression_loss": 76.61244201660156, "epoch": 1.09, "learning_rate": 1.3644875164257556e-05, "loss": 77.1097, "step": 3324 }, { "compression_loss": 76.2999267578125, "epoch": 1.09, "learning_rate": 1.3639947437582128e-05, "loss": 76.6478, "step": 3325 }, { "compression_loss": 75.38169860839844, "epoch": 1.09, "learning_rate": 1.3635019710906702e-05, "loss": 75.7504, "step": 3326 }, { "compression_loss": 75.18363952636719, "epoch": 1.09, "learning_rate": 1.3630091984231275e-05, "loss": 75.458, "step": 3327 }, { "compression_loss": 74.38929748535156, "epoch": 1.09, "learning_rate": 1.3625164257555847e-05, "loss": 74.672, "step": 3328 }, { "compression_loss": 75.79301452636719, "epoch": 1.09, "learning_rate": 1.3620236530880421e-05, "loss": 76.1201, "step": 3329 }, { "compression_loss": 75.90972900390625, "epoch": 1.09, "learning_rate": 1.3615308804204993e-05, "loss": 76.6009, "step": 3330 }, { "compression_loss": 77.80803680419922, "epoch": 1.09, "learning_rate": 1.3610381077529567e-05, "loss": 78.2401, "step": 3331 }, { "compression_loss": 76.20606994628906, "epoch": 1.09, "learning_rate": 1.360545335085414e-05, "loss": 76.4764, "step": 3332 }, { "compression_loss": 75.64025115966797, "epoch": 1.09, "learning_rate": 1.3600525624178712e-05, "loss": 75.8988, "step": 3333 }, { "compression_loss": 75.40534210205078, "epoch": 1.1, "learning_rate": 1.3595597897503286e-05, "loss": 75.8206, "step": 3334 }, { "compression_loss": 75.65704345703125, "epoch": 1.1, "learning_rate": 1.3590670170827858e-05, "loss": 76.0251, "step": 3335 }, { "compression_loss": 76.17239379882812, "epoch": 1.1, "learning_rate": 1.3585742444152432e-05, "loss": 76.5882, "step": 3336 }, { "compression_loss": 74.11465454101562, "epoch": 1.1, "learning_rate": 1.3580814717477004e-05, "loss": 74.4589, "step": 3337 }, { "compression_loss": 75.50032043457031, "epoch": 1.1, "learning_rate": 1.3575886990801577e-05, "loss": 75.8467, "step": 3338 }, { "compression_loss": 75.11161041259766, "epoch": 1.1, "learning_rate": 1.3570959264126151e-05, "loss": 75.5932, "step": 3339 }, { "compression_loss": 73.09793090820312, "epoch": 1.1, "learning_rate": 1.3566031537450723e-05, "loss": 73.3795, "step": 3340 }, { "compression_loss": 74.80801391601562, "epoch": 1.1, "learning_rate": 1.3561103810775297e-05, "loss": 75.0614, "step": 3341 }, { "compression_loss": 74.16289520263672, "epoch": 1.1, "learning_rate": 1.3556176084099869e-05, "loss": 74.5142, "step": 3342 }, { "compression_loss": 76.54786682128906, "epoch": 1.1, "learning_rate": 1.355124835742444e-05, "loss": 76.7339, "step": 3343 }, { "compression_loss": 76.52278900146484, "epoch": 1.1, "learning_rate": 1.3546320630749016e-05, "loss": 76.9091, "step": 3344 }, { "compression_loss": 74.33675384521484, "epoch": 1.1, "learning_rate": 1.3541392904073588e-05, "loss": 74.5821, "step": 3345 }, { "compression_loss": 74.05293273925781, "epoch": 1.1, "learning_rate": 1.3536465177398162e-05, "loss": 74.3775, "step": 3346 }, { "compression_loss": 76.55017852783203, "epoch": 1.1, "learning_rate": 1.3531537450722733e-05, "loss": 77.1555, "step": 3347 }, { "compression_loss": 76.4200210571289, "epoch": 1.1, "learning_rate": 1.3526609724047305e-05, "loss": 76.7879, "step": 3348 }, { "compression_loss": 74.82655334472656, "epoch": 1.1, "learning_rate": 1.3521681997371879e-05, "loss": 75.3447, "step": 3349 }, { "compression_loss": 75.0546875, "epoch": 1.1, "learning_rate": 1.3516754270696453e-05, "loss": 75.372, "step": 3350 }, { "compression_loss": 76.33464050292969, "epoch": 1.1, "learning_rate": 1.3511826544021026e-05, "loss": 76.6092, "step": 3351 }, { "compression_loss": 76.63563537597656, "epoch": 1.1, "learning_rate": 1.3506898817345598e-05, "loss": 77.4021, "step": 3352 }, { "compression_loss": 75.01849365234375, "epoch": 1.1, "learning_rate": 1.350197109067017e-05, "loss": 75.4337, "step": 3353 }, { "compression_loss": 75.59208679199219, "epoch": 1.1, "learning_rate": 1.3497043363994744e-05, "loss": 75.8269, "step": 3354 }, { "compression_loss": 76.88667297363281, "epoch": 1.1, "learning_rate": 1.3492115637319316e-05, "loss": 77.3177, "step": 3355 }, { "compression_loss": 75.487548828125, "epoch": 1.1, "learning_rate": 1.3487187910643891e-05, "loss": 75.983, "step": 3356 }, { "compression_loss": 74.85263061523438, "epoch": 1.1, "learning_rate": 1.3482260183968463e-05, "loss": 75.2258, "step": 3357 }, { "compression_loss": 76.43103790283203, "epoch": 1.1, "learning_rate": 1.3477332457293035e-05, "loss": 76.7658, "step": 3358 }, { "compression_loss": 74.27384185791016, "epoch": 1.1, "learning_rate": 1.3472404730617609e-05, "loss": 74.6202, "step": 3359 }, { "compression_loss": 73.94349670410156, "epoch": 1.1, "learning_rate": 1.3467477003942181e-05, "loss": 74.3001, "step": 3360 }, { "compression_loss": 77.90507507324219, "epoch": 1.1, "learning_rate": 1.3462549277266755e-05, "loss": 78.3013, "step": 3361 }, { "compression_loss": 73.40935516357422, "epoch": 1.1, "learning_rate": 1.3457621550591328e-05, "loss": 73.8997, "step": 3362 }, { "compression_loss": 77.57333374023438, "epoch": 1.1, "learning_rate": 1.34526938239159e-05, "loss": 78.1045, "step": 3363 }, { "compression_loss": 76.21624755859375, "epoch": 1.11, "learning_rate": 1.3447766097240474e-05, "loss": 76.5156, "step": 3364 }, { "compression_loss": 75.24227142333984, "epoch": 1.11, "learning_rate": 1.3442838370565046e-05, "loss": 75.5688, "step": 3365 }, { "compression_loss": 76.69773864746094, "epoch": 1.11, "learning_rate": 1.343791064388962e-05, "loss": 77.2031, "step": 3366 }, { "compression_loss": 73.64138793945312, "epoch": 1.11, "learning_rate": 1.3432982917214191e-05, "loss": 73.9998, "step": 3367 }, { "compression_loss": 76.43032836914062, "epoch": 1.11, "learning_rate": 1.3428055190538765e-05, "loss": 76.9797, "step": 3368 }, { "compression_loss": 76.19562530517578, "epoch": 1.11, "learning_rate": 1.3423127463863339e-05, "loss": 76.4944, "step": 3369 }, { "compression_loss": 74.47592163085938, "epoch": 1.11, "learning_rate": 1.341819973718791e-05, "loss": 74.9291, "step": 3370 }, { "compression_loss": 77.18392181396484, "epoch": 1.11, "learning_rate": 1.3413272010512484e-05, "loss": 77.5896, "step": 3371 }, { "compression_loss": 75.57064819335938, "epoch": 1.11, "learning_rate": 1.3408344283837056e-05, "loss": 75.8095, "step": 3372 }, { "compression_loss": 76.15999603271484, "epoch": 1.11, "learning_rate": 1.3403416557161628e-05, "loss": 76.4897, "step": 3373 }, { "compression_loss": 71.98127746582031, "epoch": 1.11, "learning_rate": 1.3398488830486204e-05, "loss": 72.2198, "step": 3374 }, { "compression_loss": 77.452392578125, "epoch": 1.11, "learning_rate": 1.3393561103810776e-05, "loss": 77.8864, "step": 3375 }, { "compression_loss": 74.47241973876953, "epoch": 1.11, "learning_rate": 1.338863337713535e-05, "loss": 74.6085, "step": 3376 }, { "compression_loss": 74.63764953613281, "epoch": 1.11, "learning_rate": 1.3383705650459921e-05, "loss": 75.0214, "step": 3377 }, { "compression_loss": 75.29077911376953, "epoch": 1.11, "learning_rate": 1.3378777923784493e-05, "loss": 75.5578, "step": 3378 }, { "compression_loss": 77.14228820800781, "epoch": 1.11, "learning_rate": 1.3373850197109067e-05, "loss": 77.364, "step": 3379 }, { "compression_loss": 75.84553527832031, "epoch": 1.11, "learning_rate": 1.336892247043364e-05, "loss": 76.1926, "step": 3380 }, { "compression_loss": 78.39453125, "epoch": 1.11, "learning_rate": 1.3363994743758214e-05, "loss": 78.6776, "step": 3381 }, { "compression_loss": 76.23650360107422, "epoch": 1.11, "learning_rate": 1.3359067017082786e-05, "loss": 76.6858, "step": 3382 }, { "compression_loss": 74.8772964477539, "epoch": 1.11, "learning_rate": 1.3354139290407358e-05, "loss": 75.1228, "step": 3383 }, { "compression_loss": 76.11800384521484, "epoch": 1.11, "learning_rate": 1.3349211563731932e-05, "loss": 76.46, "step": 3384 }, { "compression_loss": 74.00979614257812, "epoch": 1.11, "learning_rate": 1.3344283837056504e-05, "loss": 74.3721, "step": 3385 }, { "compression_loss": 76.42771911621094, "epoch": 1.11, "learning_rate": 1.333935611038108e-05, "loss": 76.6275, "step": 3386 }, { "compression_loss": 76.99530792236328, "epoch": 1.11, "learning_rate": 1.3334428383705651e-05, "loss": 77.2483, "step": 3387 }, { "compression_loss": 76.20652770996094, "epoch": 1.11, "learning_rate": 1.3329500657030223e-05, "loss": 76.6315, "step": 3388 }, { "compression_loss": 77.44939422607422, "epoch": 1.11, "learning_rate": 1.3324572930354797e-05, "loss": 77.7824, "step": 3389 }, { "compression_loss": 75.30679321289062, "epoch": 1.11, "learning_rate": 1.3319645203679369e-05, "loss": 75.7199, "step": 3390 }, { "compression_loss": 73.60153198242188, "epoch": 1.11, "learning_rate": 1.3314717477003942e-05, "loss": 73.9292, "step": 3391 }, { "compression_loss": 72.84095001220703, "epoch": 1.11, "learning_rate": 1.3309789750328516e-05, "loss": 73.1211, "step": 3392 }, { "compression_loss": 75.95772552490234, "epoch": 1.11, "learning_rate": 1.3304862023653088e-05, "loss": 76.3001, "step": 3393 }, { "compression_loss": 76.74891662597656, "epoch": 1.11, "learning_rate": 1.3299934296977662e-05, "loss": 77.22, "step": 3394 }, { "compression_loss": 76.0572738647461, "epoch": 1.12, "learning_rate": 1.3295006570302234e-05, "loss": 76.487, "step": 3395 }, { "compression_loss": 76.17282104492188, "epoch": 1.12, "learning_rate": 1.3290078843626807e-05, "loss": 76.391, "step": 3396 }, { "compression_loss": 76.71562957763672, "epoch": 1.12, "learning_rate": 1.328515111695138e-05, "loss": 77.1088, "step": 3397 }, { "compression_loss": 75.77464294433594, "epoch": 1.12, "learning_rate": 1.3280223390275953e-05, "loss": 76.1924, "step": 3398 }, { "compression_loss": 77.96922302246094, "epoch": 1.12, "learning_rate": 1.3275295663600527e-05, "loss": 78.5464, "step": 3399 }, { "compression_loss": 74.19782257080078, "epoch": 1.12, "learning_rate": 1.3270367936925099e-05, "loss": 74.4465, "step": 3400 }, { "compression_loss": 74.67298126220703, "epoch": 1.12, "learning_rate": 1.3265440210249672e-05, "loss": 75.0765, "step": 3401 }, { "compression_loss": 76.38047790527344, "epoch": 1.12, "learning_rate": 1.3260512483574244e-05, "loss": 76.7943, "step": 3402 }, { "compression_loss": 76.54594421386719, "epoch": 1.12, "learning_rate": 1.3255584756898816e-05, "loss": 76.8188, "step": 3403 }, { "compression_loss": 75.90850830078125, "epoch": 1.12, "learning_rate": 1.3250657030223392e-05, "loss": 76.2721, "step": 3404 }, { "compression_loss": 75.86836242675781, "epoch": 1.12, "learning_rate": 1.3245729303547964e-05, "loss": 76.2088, "step": 3405 }, { "compression_loss": 75.04423522949219, "epoch": 1.12, "learning_rate": 1.3240801576872537e-05, "loss": 75.7495, "step": 3406 }, { "compression_loss": 75.35839080810547, "epoch": 1.12, "learning_rate": 1.323587385019711e-05, "loss": 75.7639, "step": 3407 }, { "compression_loss": 73.06976318359375, "epoch": 1.12, "learning_rate": 1.3230946123521681e-05, "loss": 73.464, "step": 3408 }, { "compression_loss": 74.07096099853516, "epoch": 1.12, "learning_rate": 1.3226018396846255e-05, "loss": 74.4943, "step": 3409 }, { "compression_loss": 75.39190673828125, "epoch": 1.12, "learning_rate": 1.3221090670170829e-05, "loss": 75.7351, "step": 3410 }, { "compression_loss": 74.57259368896484, "epoch": 1.12, "learning_rate": 1.3216162943495402e-05, "loss": 74.9737, "step": 3411 }, { "compression_loss": 76.72223663330078, "epoch": 1.12, "learning_rate": 1.3211235216819974e-05, "loss": 76.9426, "step": 3412 }, { "compression_loss": 77.48712921142578, "epoch": 1.12, "learning_rate": 1.3206307490144546e-05, "loss": 77.6823, "step": 3413 }, { "compression_loss": 74.669189453125, "epoch": 1.12, "learning_rate": 1.320137976346912e-05, "loss": 75.078, "step": 3414 }, { "compression_loss": 76.18392944335938, "epoch": 1.12, "learning_rate": 1.3196452036793692e-05, "loss": 76.5679, "step": 3415 }, { "compression_loss": 76.3073501586914, "epoch": 1.12, "learning_rate": 1.3191524310118267e-05, "loss": 76.6267, "step": 3416 }, { "compression_loss": 74.48053741455078, "epoch": 1.12, "learning_rate": 1.3186596583442839e-05, "loss": 74.9258, "step": 3417 }, { "compression_loss": 76.27975463867188, "epoch": 1.12, "learning_rate": 1.3181668856767411e-05, "loss": 76.7059, "step": 3418 }, { "compression_loss": 74.2370834350586, "epoch": 1.12, "learning_rate": 1.3176741130091985e-05, "loss": 74.7958, "step": 3419 }, { "compression_loss": 75.34169006347656, "epoch": 1.12, "learning_rate": 1.3171813403416557e-05, "loss": 75.6353, "step": 3420 }, { "compression_loss": 76.3885726928711, "epoch": 1.12, "learning_rate": 1.316688567674113e-05, "loss": 76.8372, "step": 3421 }, { "compression_loss": 76.26148986816406, "epoch": 1.12, "learning_rate": 1.3161957950065704e-05, "loss": 76.6006, "step": 3422 }, { "compression_loss": 72.4803466796875, "epoch": 1.12, "learning_rate": 1.3157030223390276e-05, "loss": 72.6856, "step": 3423 }, { "compression_loss": 75.03848266601562, "epoch": 1.12, "learning_rate": 1.315210249671485e-05, "loss": 75.1582, "step": 3424 }, { "compression_loss": 76.64432525634766, "epoch": 1.13, "learning_rate": 1.3147174770039422e-05, "loss": 77.1399, "step": 3425 }, { "compression_loss": 74.25895690917969, "epoch": 1.13, "learning_rate": 1.3142247043363995e-05, "loss": 74.7645, "step": 3426 }, { "compression_loss": 75.84376525878906, "epoch": 1.13, "learning_rate": 1.3137319316688567e-05, "loss": 76.1952, "step": 3427 }, { "compression_loss": 74.69464111328125, "epoch": 1.13, "learning_rate": 1.3132391590013141e-05, "loss": 75.0312, "step": 3428 }, { "compression_loss": 74.0242919921875, "epoch": 1.13, "learning_rate": 1.3127463863337715e-05, "loss": 74.5238, "step": 3429 }, { "compression_loss": 73.23843383789062, "epoch": 1.13, "learning_rate": 1.3122536136662287e-05, "loss": 73.5632, "step": 3430 }, { "compression_loss": 74.6910400390625, "epoch": 1.13, "learning_rate": 1.311760840998686e-05, "loss": 75.072, "step": 3431 }, { "compression_loss": 72.71961975097656, "epoch": 1.13, "learning_rate": 1.3112680683311432e-05, "loss": 73.0382, "step": 3432 }, { "compression_loss": 77.19062042236328, "epoch": 1.13, "learning_rate": 1.3107752956636004e-05, "loss": 77.5399, "step": 3433 }, { "compression_loss": 74.02117156982422, "epoch": 1.13, "learning_rate": 1.310282522996058e-05, "loss": 74.1674, "step": 3434 }, { "compression_loss": 76.21385192871094, "epoch": 1.13, "learning_rate": 1.3097897503285151e-05, "loss": 76.6663, "step": 3435 }, { "compression_loss": 75.57471466064453, "epoch": 1.13, "learning_rate": 1.3092969776609725e-05, "loss": 75.8912, "step": 3436 }, { "compression_loss": 75.76861572265625, "epoch": 1.13, "learning_rate": 1.3088042049934297e-05, "loss": 76.1758, "step": 3437 }, { "compression_loss": 75.04002380371094, "epoch": 1.13, "learning_rate": 1.3083114323258869e-05, "loss": 75.5061, "step": 3438 }, { "compression_loss": 75.37765502929688, "epoch": 1.13, "learning_rate": 1.3078186596583443e-05, "loss": 75.7218, "step": 3439 }, { "compression_loss": 73.4341812133789, "epoch": 1.13, "learning_rate": 1.3073258869908016e-05, "loss": 73.7829, "step": 3440 }, { "compression_loss": 76.49652862548828, "epoch": 1.13, "learning_rate": 1.306833114323259e-05, "loss": 76.6849, "step": 3441 }, { "compression_loss": 76.6728515625, "epoch": 1.13, "learning_rate": 1.3063403416557162e-05, "loss": 76.8909, "step": 3442 }, { "compression_loss": 74.94844818115234, "epoch": 1.13, "learning_rate": 1.3058475689881734e-05, "loss": 75.3758, "step": 3443 }, { "compression_loss": 73.96110534667969, "epoch": 1.13, "learning_rate": 1.3053547963206308e-05, "loss": 74.2696, "step": 3444 }, { "compression_loss": 74.79383850097656, "epoch": 1.13, "learning_rate": 1.304862023653088e-05, "loss": 75.6269, "step": 3445 }, { "compression_loss": 76.24527740478516, "epoch": 1.13, "learning_rate": 1.3043692509855455e-05, "loss": 76.594, "step": 3446 }, { "compression_loss": 76.68851470947266, "epoch": 1.13, "learning_rate": 1.3038764783180027e-05, "loss": 76.9159, "step": 3447 }, { "compression_loss": 73.39521789550781, "epoch": 1.13, "learning_rate": 1.3033837056504599e-05, "loss": 74.0234, "step": 3448 }, { "compression_loss": 74.03099060058594, "epoch": 1.13, "learning_rate": 1.3028909329829173e-05, "loss": 74.7821, "step": 3449 }, { "compression_loss": 74.89865112304688, "epoch": 1.13, "learning_rate": 1.3023981603153745e-05, "loss": 75.0883, "step": 3450 }, { "compression_loss": 76.76886749267578, "epoch": 1.13, "learning_rate": 1.3019053876478318e-05, "loss": 77.0012, "step": 3451 }, { "compression_loss": 76.47979736328125, "epoch": 1.13, "learning_rate": 1.3014126149802892e-05, "loss": 77.011, "step": 3452 }, { "compression_loss": 76.09149169921875, "epoch": 1.13, "learning_rate": 1.3009198423127464e-05, "loss": 76.3699, "step": 3453 }, { "compression_loss": 75.43803405761719, "epoch": 1.13, "learning_rate": 1.3004270696452038e-05, "loss": 75.7134, "step": 3454 }, { "compression_loss": 77.42970275878906, "epoch": 1.14, "learning_rate": 1.299934296977661e-05, "loss": 77.7394, "step": 3455 }, { "compression_loss": 75.74520111083984, "epoch": 1.14, "learning_rate": 1.2994415243101183e-05, "loss": 76.05, "step": 3456 }, { "compression_loss": 77.90695190429688, "epoch": 1.14, "learning_rate": 1.2989487516425755e-05, "loss": 78.262, "step": 3457 }, { "compression_loss": 74.7483901977539, "epoch": 1.14, "learning_rate": 1.2984559789750329e-05, "loss": 75.0182, "step": 3458 }, { "compression_loss": 76.75202941894531, "epoch": 1.14, "learning_rate": 1.2979632063074902e-05, "loss": 77.1596, "step": 3459 }, { "compression_loss": 75.33305358886719, "epoch": 1.14, "learning_rate": 1.2974704336399474e-05, "loss": 75.6599, "step": 3460 }, { "compression_loss": 76.06488037109375, "epoch": 1.14, "learning_rate": 1.2969776609724048e-05, "loss": 76.6167, "step": 3461 }, { "compression_loss": 78.57984924316406, "epoch": 1.14, "learning_rate": 1.296484888304862e-05, "loss": 79.1189, "step": 3462 }, { "compression_loss": 75.13162994384766, "epoch": 1.14, "learning_rate": 1.2959921156373194e-05, "loss": 75.3796, "step": 3463 }, { "compression_loss": 76.42718505859375, "epoch": 1.14, "learning_rate": 1.2954993429697767e-05, "loss": 76.7926, "step": 3464 }, { "compression_loss": 75.34245300292969, "epoch": 1.14, "learning_rate": 1.295006570302234e-05, "loss": 75.7428, "step": 3465 }, { "compression_loss": 75.84230041503906, "epoch": 1.14, "learning_rate": 1.2945137976346913e-05, "loss": 76.2562, "step": 3466 }, { "compression_loss": 76.57588195800781, "epoch": 1.14, "learning_rate": 1.2940210249671485e-05, "loss": 76.8419, "step": 3467 }, { "compression_loss": 71.87519836425781, "epoch": 1.14, "learning_rate": 1.2935282522996057e-05, "loss": 72.2568, "step": 3468 }, { "compression_loss": 74.60267639160156, "epoch": 1.14, "learning_rate": 1.2930354796320632e-05, "loss": 74.8046, "step": 3469 }, { "compression_loss": 74.48334503173828, "epoch": 1.14, "learning_rate": 1.2925427069645204e-05, "loss": 74.6348, "step": 3470 }, { "compression_loss": 76.4024429321289, "epoch": 1.14, "learning_rate": 1.2920499342969778e-05, "loss": 76.8138, "step": 3471 }, { "compression_loss": 75.56379699707031, "epoch": 1.14, "learning_rate": 1.291557161629435e-05, "loss": 75.8236, "step": 3472 }, { "compression_loss": 76.17965698242188, "epoch": 1.14, "learning_rate": 1.2910643889618922e-05, "loss": 76.5954, "step": 3473 }, { "compression_loss": 76.06803894042969, "epoch": 1.14, "learning_rate": 1.2905716162943496e-05, "loss": 76.3726, "step": 3474 }, { "compression_loss": 75.63187408447266, "epoch": 1.14, "learning_rate": 1.290078843626807e-05, "loss": 75.9941, "step": 3475 }, { "compression_loss": 78.00045776367188, "epoch": 1.14, "learning_rate": 1.2895860709592643e-05, "loss": 78.4309, "step": 3476 }, { "compression_loss": 75.04724884033203, "epoch": 1.14, "learning_rate": 1.2890932982917215e-05, "loss": 75.4699, "step": 3477 }, { "compression_loss": 76.28802490234375, "epoch": 1.14, "learning_rate": 1.2886005256241787e-05, "loss": 76.6845, "step": 3478 }, { "compression_loss": 77.01345825195312, "epoch": 1.14, "learning_rate": 1.288107752956636e-05, "loss": 77.423, "step": 3479 }, { "compression_loss": 77.21817016601562, "epoch": 1.14, "learning_rate": 1.2876149802890932e-05, "loss": 77.8858, "step": 3480 }, { "compression_loss": 75.05726623535156, "epoch": 1.14, "learning_rate": 1.2871222076215508e-05, "loss": 75.2764, "step": 3481 }, { "compression_loss": 75.17562866210938, "epoch": 1.14, "learning_rate": 1.286629434954008e-05, "loss": 75.3705, "step": 3482 }, { "compression_loss": 77.00979614257812, "epoch": 1.14, "learning_rate": 1.2861366622864652e-05, "loss": 77.2443, "step": 3483 }, { "compression_loss": 76.70278930664062, "epoch": 1.14, "learning_rate": 1.2856438896189225e-05, "loss": 77.2398, "step": 3484 }, { "compression_loss": 74.76189422607422, "epoch": 1.14, "learning_rate": 1.2851511169513797e-05, "loss": 75.1386, "step": 3485 }, { "compression_loss": 75.86698913574219, "epoch": 1.15, "learning_rate": 1.2846583442838371e-05, "loss": 76.2659, "step": 3486 }, { "compression_loss": 74.85392761230469, "epoch": 1.15, "learning_rate": 1.2841655716162945e-05, "loss": 75.1165, "step": 3487 }, { "compression_loss": 75.31501770019531, "epoch": 1.15, "learning_rate": 1.2836727989487517e-05, "loss": 75.8933, "step": 3488 }, { "compression_loss": 73.88604736328125, "epoch": 1.15, "learning_rate": 1.283180026281209e-05, "loss": 74.0084, "step": 3489 }, { "compression_loss": 76.87849426269531, "epoch": 1.15, "learning_rate": 1.2826872536136662e-05, "loss": 77.2367, "step": 3490 }, { "compression_loss": 74.97381591796875, "epoch": 1.15, "learning_rate": 1.2821944809461236e-05, "loss": 75.2167, "step": 3491 }, { "compression_loss": 75.63369750976562, "epoch": 1.15, "learning_rate": 1.2817017082785808e-05, "loss": 75.8717, "step": 3492 }, { "compression_loss": 75.39249420166016, "epoch": 1.15, "learning_rate": 1.2812089356110382e-05, "loss": 75.9144, "step": 3493 }, { "compression_loss": 76.70011138916016, "epoch": 1.15, "learning_rate": 1.2807161629434955e-05, "loss": 77.0425, "step": 3494 }, { "compression_loss": 74.36154174804688, "epoch": 1.15, "learning_rate": 1.2802233902759527e-05, "loss": 74.8449, "step": 3495 }, { "compression_loss": 76.22650146484375, "epoch": 1.15, "learning_rate": 1.2797306176084101e-05, "loss": 76.5093, "step": 3496 }, { "compression_loss": 75.41780090332031, "epoch": 1.15, "learning_rate": 1.2792378449408673e-05, "loss": 75.7448, "step": 3497 }, { "compression_loss": 75.43885803222656, "epoch": 1.15, "learning_rate": 1.2787450722733245e-05, "loss": 75.9755, "step": 3498 }, { "compression_loss": 74.54354858398438, "epoch": 1.15, "learning_rate": 1.278252299605782e-05, "loss": 74.7224, "step": 3499 }, { "compression_loss": 74.73841857910156, "epoch": 1.15, "learning_rate": 1.2777595269382392e-05, "loss": 75.0676, "step": 3500 }, { "epoch": 1.15, "eval_exact_match": 86.90633869441817, "eval_f1": 93.05726487138371, "step": 3500 }, { "compression_loss": 74.75942993164062, "epoch": 1.15, "learning_rate": 1.2772667542706966e-05, "loss": 75.1517, "step": 3501 }, { "compression_loss": 74.7220458984375, "epoch": 1.15, "learning_rate": 1.2767739816031538e-05, "loss": 75.4389, "step": 3502 }, { "compression_loss": 75.85050964355469, "epoch": 1.15, "learning_rate": 1.276281208935611e-05, "loss": 76.206, "step": 3503 }, { "compression_loss": 76.55743408203125, "epoch": 1.15, "learning_rate": 1.2757884362680683e-05, "loss": 76.7994, "step": 3504 }, { "compression_loss": 75.61481475830078, "epoch": 1.15, "learning_rate": 1.2752956636005257e-05, "loss": 75.9853, "step": 3505 }, { "compression_loss": 74.71389770507812, "epoch": 1.15, "learning_rate": 1.274802890932983e-05, "loss": 74.9871, "step": 3506 }, { "compression_loss": 74.05429077148438, "epoch": 1.15, "learning_rate": 1.2743101182654403e-05, "loss": 74.3653, "step": 3507 }, { "compression_loss": 75.29464721679688, "epoch": 1.15, "learning_rate": 1.2738173455978975e-05, "loss": 75.7277, "step": 3508 }, { "compression_loss": 74.98759460449219, "epoch": 1.15, "learning_rate": 1.2733245729303548e-05, "loss": 75.6027, "step": 3509 }, { "compression_loss": 74.84506225585938, "epoch": 1.15, "learning_rate": 1.272831800262812e-05, "loss": 75.185, "step": 3510 }, { "compression_loss": 76.47566986083984, "epoch": 1.15, "learning_rate": 1.2723390275952696e-05, "loss": 77.3734, "step": 3511 }, { "compression_loss": 80.39781188964844, "epoch": 1.15, "learning_rate": 1.2718462549277268e-05, "loss": 81.0383, "step": 3512 }, { "compression_loss": 74.08206176757812, "epoch": 1.15, "learning_rate": 1.271353482260184e-05, "loss": 74.4962, "step": 3513 }, { "compression_loss": 78.85717010498047, "epoch": 1.15, "learning_rate": 1.2708607095926413e-05, "loss": 79.3468, "step": 3514 }, { "compression_loss": 74.66357421875, "epoch": 1.15, "learning_rate": 1.2703679369250985e-05, "loss": 75.0914, "step": 3515 }, { "compression_loss": 77.05549621582031, "epoch": 1.16, "learning_rate": 1.2698751642575559e-05, "loss": 77.4841, "step": 3516 }, { "compression_loss": 74.14436340332031, "epoch": 1.16, "learning_rate": 1.2693823915900133e-05, "loss": 74.5447, "step": 3517 }, { "compression_loss": 75.75576782226562, "epoch": 1.16, "learning_rate": 1.2688896189224705e-05, "loss": 75.8678, "step": 3518 }, { "compression_loss": 74.68731689453125, "epoch": 1.16, "learning_rate": 1.2683968462549278e-05, "loss": 74.998, "step": 3519 }, { "compression_loss": 78.85041809082031, "epoch": 1.16, "learning_rate": 1.267904073587385e-05, "loss": 79.3019, "step": 3520 }, { "compression_loss": 77.2849349975586, "epoch": 1.16, "learning_rate": 1.2674113009198424e-05, "loss": 77.8554, "step": 3521 }, { "compression_loss": 76.88291931152344, "epoch": 1.16, "learning_rate": 1.2669185282522996e-05, "loss": 77.1996, "step": 3522 }, { "compression_loss": 75.6465072631836, "epoch": 1.16, "learning_rate": 1.266425755584757e-05, "loss": 75.9634, "step": 3523 }, { "compression_loss": 75.82815551757812, "epoch": 1.16, "learning_rate": 1.2659329829172143e-05, "loss": 76.0437, "step": 3524 }, { "compression_loss": 76.52632141113281, "epoch": 1.16, "learning_rate": 1.2654402102496715e-05, "loss": 76.9741, "step": 3525 }, { "compression_loss": 73.16691589355469, "epoch": 1.16, "learning_rate": 1.2649474375821289e-05, "loss": 73.3333, "step": 3526 }, { "compression_loss": 73.37276458740234, "epoch": 1.16, "learning_rate": 1.264454664914586e-05, "loss": 73.9165, "step": 3527 }, { "compression_loss": 76.96126556396484, "epoch": 1.16, "learning_rate": 1.2639618922470433e-05, "loss": 77.2924, "step": 3528 }, { "compression_loss": 75.40682983398438, "epoch": 1.16, "learning_rate": 1.2634691195795008e-05, "loss": 75.9484, "step": 3529 }, { "compression_loss": 74.42694091796875, "epoch": 1.16, "learning_rate": 1.262976346911958e-05, "loss": 74.9233, "step": 3530 }, { "compression_loss": 79.00483703613281, "epoch": 1.16, "learning_rate": 1.2624835742444154e-05, "loss": 79.5204, "step": 3531 }, { "compression_loss": 73.4656753540039, "epoch": 1.16, "learning_rate": 1.2619908015768726e-05, "loss": 73.8409, "step": 3532 }, { "compression_loss": 74.71206665039062, "epoch": 1.16, "learning_rate": 1.2614980289093298e-05, "loss": 75.0148, "step": 3533 }, { "compression_loss": 75.08558654785156, "epoch": 1.16, "learning_rate": 1.2610052562417871e-05, "loss": 75.4331, "step": 3534 }, { "compression_loss": 75.23149108886719, "epoch": 1.16, "learning_rate": 1.2605124835742445e-05, "loss": 75.5564, "step": 3535 }, { "compression_loss": 75.89729309082031, "epoch": 1.16, "learning_rate": 1.2600197109067019e-05, "loss": 76.1654, "step": 3536 }, { "compression_loss": 76.28643035888672, "epoch": 1.16, "learning_rate": 1.259526938239159e-05, "loss": 76.6633, "step": 3537 }, { "compression_loss": 75.91728973388672, "epoch": 1.16, "learning_rate": 1.2590341655716163e-05, "loss": 76.3559, "step": 3538 }, { "compression_loss": 74.55440521240234, "epoch": 1.16, "learning_rate": 1.2585413929040736e-05, "loss": 74.8704, "step": 3539 }, { "compression_loss": 78.0540771484375, "epoch": 1.16, "learning_rate": 1.2580486202365308e-05, "loss": 78.4394, "step": 3540 }, { "compression_loss": 75.05162811279297, "epoch": 1.16, "learning_rate": 1.2575558475689884e-05, "loss": 75.3737, "step": 3541 }, { "compression_loss": 77.20077514648438, "epoch": 1.16, "learning_rate": 1.2570630749014455e-05, "loss": 77.5878, "step": 3542 }, { "compression_loss": 76.1146240234375, "epoch": 1.16, "learning_rate": 1.2565703022339027e-05, "loss": 76.4078, "step": 3543 }, { "compression_loss": 77.51795959472656, "epoch": 1.16, "learning_rate": 1.2560775295663601e-05, "loss": 78.287, "step": 3544 }, { "compression_loss": 75.58985137939453, "epoch": 1.16, "learning_rate": 1.2555847568988173e-05, "loss": 76.0425, "step": 3545 }, { "compression_loss": 76.22422790527344, "epoch": 1.16, "learning_rate": 1.2550919842312747e-05, "loss": 76.5859, "step": 3546 }, { "compression_loss": 73.50782775878906, "epoch": 1.17, "learning_rate": 1.254599211563732e-05, "loss": 73.708, "step": 3547 }, { "compression_loss": 74.22003173828125, "epoch": 1.17, "learning_rate": 1.2541064388961892e-05, "loss": 74.561, "step": 3548 }, { "compression_loss": 79.31700897216797, "epoch": 1.17, "learning_rate": 1.2536136662286466e-05, "loss": 80.0427, "step": 3549 }, { "compression_loss": 77.18133544921875, "epoch": 1.17, "learning_rate": 1.2531208935611038e-05, "loss": 77.4384, "step": 3550 }, { "compression_loss": 73.71829223632812, "epoch": 1.17, "learning_rate": 1.2526281208935612e-05, "loss": 74.1827, "step": 3551 }, { "compression_loss": 76.49820709228516, "epoch": 1.17, "learning_rate": 1.2521353482260184e-05, "loss": 76.9352, "step": 3552 }, { "compression_loss": 75.07100677490234, "epoch": 1.17, "learning_rate": 1.2516425755584757e-05, "loss": 75.2597, "step": 3553 }, { "compression_loss": 74.54879760742188, "epoch": 1.17, "learning_rate": 1.2511498028909331e-05, "loss": 75.1153, "step": 3554 }, { "compression_loss": 74.30738830566406, "epoch": 1.17, "learning_rate": 1.2506570302233903e-05, "loss": 74.6973, "step": 3555 }, { "compression_loss": 74.30526733398438, "epoch": 1.17, "learning_rate": 1.2501642575558477e-05, "loss": 74.4558, "step": 3556 }, { "compression_loss": 77.47936248779297, "epoch": 1.17, "learning_rate": 1.2496714848883049e-05, "loss": 77.8004, "step": 3557 }, { "compression_loss": 75.69253540039062, "epoch": 1.17, "learning_rate": 1.249178712220762e-05, "loss": 76.0412, "step": 3558 }, { "compression_loss": 74.91260528564453, "epoch": 1.17, "learning_rate": 1.2486859395532196e-05, "loss": 75.2128, "step": 3559 }, { "compression_loss": 75.05329895019531, "epoch": 1.17, "learning_rate": 1.2481931668856768e-05, "loss": 75.3934, "step": 3560 }, { "compression_loss": 76.31776428222656, "epoch": 1.17, "learning_rate": 1.2477003942181342e-05, "loss": 76.5449, "step": 3561 }, { "compression_loss": 74.04158020019531, "epoch": 1.17, "learning_rate": 1.2472076215505914e-05, "loss": 74.5289, "step": 3562 }, { "compression_loss": 76.32466125488281, "epoch": 1.17, "learning_rate": 1.2467148488830485e-05, "loss": 76.7855, "step": 3563 }, { "compression_loss": 74.59584045410156, "epoch": 1.17, "learning_rate": 1.2462220762155059e-05, "loss": 75.0287, "step": 3564 }, { "compression_loss": 76.91004943847656, "epoch": 1.17, "learning_rate": 1.2457293035479633e-05, "loss": 77.8132, "step": 3565 }, { "compression_loss": 77.96133422851562, "epoch": 1.17, "learning_rate": 1.2452365308804206e-05, "loss": 78.5039, "step": 3566 }, { "compression_loss": 76.5403060913086, "epoch": 1.17, "learning_rate": 1.2447437582128778e-05, "loss": 76.9085, "step": 3567 }, { "compression_loss": 75.08697509765625, "epoch": 1.17, "learning_rate": 1.244250985545335e-05, "loss": 75.4532, "step": 3568 }, { "compression_loss": 74.32374572753906, "epoch": 1.17, "learning_rate": 1.2437582128777924e-05, "loss": 74.9653, "step": 3569 }, { "compression_loss": 76.05211639404297, "epoch": 1.17, "learning_rate": 1.2432654402102496e-05, "loss": 76.5127, "step": 3570 }, { "compression_loss": 74.32676696777344, "epoch": 1.17, "learning_rate": 1.2427726675427071e-05, "loss": 74.8187, "step": 3571 }, { "compression_loss": 76.22119140625, "epoch": 1.17, "learning_rate": 1.2422798948751643e-05, "loss": 76.4118, "step": 3572 }, { "compression_loss": 75.75469970703125, "epoch": 1.17, "learning_rate": 1.2417871222076215e-05, "loss": 76.3419, "step": 3573 }, { "compression_loss": 78.49732208251953, "epoch": 1.17, "learning_rate": 1.2412943495400789e-05, "loss": 79.0005, "step": 3574 }, { "compression_loss": 74.47740173339844, "epoch": 1.17, "learning_rate": 1.2408015768725361e-05, "loss": 74.7081, "step": 3575 }, { "compression_loss": 75.091796875, "epoch": 1.17, "learning_rate": 1.2403088042049935e-05, "loss": 75.5674, "step": 3576 }, { "compression_loss": 75.41261291503906, "epoch": 1.18, "learning_rate": 1.2398160315374508e-05, "loss": 75.8658, "step": 3577 }, { "compression_loss": 72.60482025146484, "epoch": 1.18, "learning_rate": 1.239323258869908e-05, "loss": 72.8822, "step": 3578 }, { "compression_loss": 74.22857666015625, "epoch": 1.18, "learning_rate": 1.2388304862023654e-05, "loss": 74.5596, "step": 3579 }, { "compression_loss": 75.89041137695312, "epoch": 1.18, "learning_rate": 1.2383377135348226e-05, "loss": 76.304, "step": 3580 }, { "compression_loss": 76.9853286743164, "epoch": 1.18, "learning_rate": 1.23784494086728e-05, "loss": 77.3331, "step": 3581 }, { "compression_loss": 74.35945129394531, "epoch": 1.18, "learning_rate": 1.2373521681997372e-05, "loss": 74.6236, "step": 3582 }, { "compression_loss": 74.25532531738281, "epoch": 1.18, "learning_rate": 1.2368593955321945e-05, "loss": 74.8473, "step": 3583 }, { "compression_loss": 77.2406997680664, "epoch": 1.18, "learning_rate": 1.2363666228646519e-05, "loss": 77.5538, "step": 3584 }, { "compression_loss": 75.78306579589844, "epoch": 1.18, "learning_rate": 1.235873850197109e-05, "loss": 76.0307, "step": 3585 }, { "compression_loss": 75.38787841796875, "epoch": 1.18, "learning_rate": 1.2353810775295664e-05, "loss": 75.8004, "step": 3586 }, { "compression_loss": 75.92323303222656, "epoch": 1.18, "learning_rate": 1.2348883048620236e-05, "loss": 76.3017, "step": 3587 }, { "compression_loss": 77.68983459472656, "epoch": 1.18, "learning_rate": 1.2343955321944808e-05, "loss": 78.0103, "step": 3588 }, { "compression_loss": 76.7659683227539, "epoch": 1.18, "learning_rate": 1.2339027595269384e-05, "loss": 77.3452, "step": 3589 }, { "compression_loss": 74.73663330078125, "epoch": 1.18, "learning_rate": 1.2334099868593956e-05, "loss": 75.0215, "step": 3590 }, { "compression_loss": 75.94888305664062, "epoch": 1.18, "learning_rate": 1.232917214191853e-05, "loss": 76.242, "step": 3591 }, { "compression_loss": 74.43462371826172, "epoch": 1.18, "learning_rate": 1.2324244415243101e-05, "loss": 74.6853, "step": 3592 }, { "compression_loss": 76.47586822509766, "epoch": 1.18, "learning_rate": 1.2319316688567673e-05, "loss": 76.8957, "step": 3593 }, { "compression_loss": 75.20691680908203, "epoch": 1.18, "learning_rate": 1.2314388961892247e-05, "loss": 75.6189, "step": 3594 }, { "compression_loss": 76.90045166015625, "epoch": 1.18, "learning_rate": 1.230946123521682e-05, "loss": 77.2372, "step": 3595 }, { "compression_loss": 73.18693542480469, "epoch": 1.18, "learning_rate": 1.2304533508541394e-05, "loss": 73.3601, "step": 3596 }, { "compression_loss": 75.91094207763672, "epoch": 1.18, "learning_rate": 1.2299605781865966e-05, "loss": 76.1482, "step": 3597 }, { "compression_loss": 76.39041137695312, "epoch": 1.18, "learning_rate": 1.2294678055190538e-05, "loss": 76.7002, "step": 3598 }, { "compression_loss": 77.164306640625, "epoch": 1.18, "learning_rate": 1.2289750328515112e-05, "loss": 77.6067, "step": 3599 }, { "compression_loss": 77.094482421875, "epoch": 1.18, "learning_rate": 1.2284822601839684e-05, "loss": 77.3284, "step": 3600 }, { "compression_loss": 75.77484130859375, "epoch": 1.18, "learning_rate": 1.227989487516426e-05, "loss": 76.0723, "step": 3601 }, { "compression_loss": 74.23149108886719, "epoch": 1.18, "learning_rate": 1.2274967148488831e-05, "loss": 74.3397, "step": 3602 }, { "compression_loss": 73.93644714355469, "epoch": 1.18, "learning_rate": 1.2270039421813403e-05, "loss": 74.2217, "step": 3603 }, { "compression_loss": 75.56465148925781, "epoch": 1.18, "learning_rate": 1.2265111695137977e-05, "loss": 75.9852, "step": 3604 }, { "compression_loss": 75.05172729492188, "epoch": 1.18, "learning_rate": 1.2260183968462549e-05, "loss": 75.2546, "step": 3605 }, { "compression_loss": 73.7544937133789, "epoch": 1.18, "learning_rate": 1.2255256241787122e-05, "loss": 74.0976, "step": 3606 }, { "compression_loss": 75.11688232421875, "epoch": 1.18, "learning_rate": 1.2250328515111696e-05, "loss": 75.3516, "step": 3607 }, { "compression_loss": 77.21852111816406, "epoch": 1.19, "learning_rate": 1.2245400788436268e-05, "loss": 77.5903, "step": 3608 }, { "compression_loss": 74.03791809082031, "epoch": 1.19, "learning_rate": 1.2240473061760842e-05, "loss": 74.4813, "step": 3609 }, { "compression_loss": 75.4225845336914, "epoch": 1.19, "learning_rate": 1.2235545335085414e-05, "loss": 75.7962, "step": 3610 }, { "compression_loss": 75.01346588134766, "epoch": 1.19, "learning_rate": 1.2230617608409987e-05, "loss": 75.3144, "step": 3611 }, { "compression_loss": 76.07620239257812, "epoch": 1.19, "learning_rate": 1.2225689881734561e-05, "loss": 76.3739, "step": 3612 }, { "compression_loss": 76.08778381347656, "epoch": 1.19, "learning_rate": 1.2220762155059133e-05, "loss": 76.5267, "step": 3613 }, { "compression_loss": 74.36283874511719, "epoch": 1.19, "learning_rate": 1.2215834428383707e-05, "loss": 74.5575, "step": 3614 }, { "compression_loss": 78.45268249511719, "epoch": 1.19, "learning_rate": 1.2210906701708279e-05, "loss": 79.0025, "step": 3615 }, { "compression_loss": 78.11580657958984, "epoch": 1.19, "learning_rate": 1.2205978975032852e-05, "loss": 78.566, "step": 3616 }, { "compression_loss": 75.31338500976562, "epoch": 1.19, "learning_rate": 1.2201051248357424e-05, "loss": 75.8636, "step": 3617 }, { "compression_loss": 78.32455444335938, "epoch": 1.19, "learning_rate": 1.2196123521681998e-05, "loss": 79.2022, "step": 3618 }, { "compression_loss": 76.93216705322266, "epoch": 1.19, "learning_rate": 1.2191195795006572e-05, "loss": 77.1274, "step": 3619 }, { "compression_loss": 75.64524841308594, "epoch": 1.19, "learning_rate": 1.2186268068331144e-05, "loss": 75.9276, "step": 3620 }, { "compression_loss": 76.33906555175781, "epoch": 1.19, "learning_rate": 1.2181340341655716e-05, "loss": 76.6592, "step": 3621 }, { "compression_loss": 75.07435607910156, "epoch": 1.19, "learning_rate": 1.217641261498029e-05, "loss": 75.6518, "step": 3622 }, { "compression_loss": 77.5347671508789, "epoch": 1.19, "learning_rate": 1.2171484888304861e-05, "loss": 77.9007, "step": 3623 }, { "compression_loss": 75.47059631347656, "epoch": 1.19, "learning_rate": 1.2166557161629437e-05, "loss": 75.8145, "step": 3624 }, { "compression_loss": 76.1137924194336, "epoch": 1.19, "learning_rate": 1.2161629434954009e-05, "loss": 76.4558, "step": 3625 }, { "compression_loss": 75.24311828613281, "epoch": 1.19, "learning_rate": 1.215670170827858e-05, "loss": 75.4509, "step": 3626 }, { "compression_loss": 75.45648193359375, "epoch": 1.19, "learning_rate": 1.2151773981603154e-05, "loss": 75.6577, "step": 3627 }, { "compression_loss": 74.67295837402344, "epoch": 1.19, "learning_rate": 1.2146846254927726e-05, "loss": 75.0798, "step": 3628 }, { "compression_loss": 76.55007934570312, "epoch": 1.19, "learning_rate": 1.21419185282523e-05, "loss": 76.9064, "step": 3629 }, { "compression_loss": 78.67353057861328, "epoch": 1.19, "learning_rate": 1.2136990801576873e-05, "loss": 78.9275, "step": 3630 }, { "compression_loss": 73.12100982666016, "epoch": 1.19, "learning_rate": 1.2132063074901445e-05, "loss": 73.6575, "step": 3631 }, { "compression_loss": 74.39016723632812, "epoch": 1.19, "learning_rate": 1.2127135348226019e-05, "loss": 74.6701, "step": 3632 }, { "compression_loss": 76.79254150390625, "epoch": 1.19, "learning_rate": 1.2122207621550591e-05, "loss": 77.0539, "step": 3633 }, { "compression_loss": 76.22940063476562, "epoch": 1.19, "learning_rate": 1.2117279894875165e-05, "loss": 76.8125, "step": 3634 }, { "compression_loss": 74.24625396728516, "epoch": 1.19, "learning_rate": 1.2112352168199737e-05, "loss": 74.4453, "step": 3635 }, { "compression_loss": 76.37211608886719, "epoch": 1.19, "learning_rate": 1.210742444152431e-05, "loss": 76.7664, "step": 3636 }, { "compression_loss": 76.69692993164062, "epoch": 1.19, "learning_rate": 1.2102496714848884e-05, "loss": 76.8758, "step": 3637 }, { "compression_loss": 74.73462677001953, "epoch": 1.2, "learning_rate": 1.2097568988173456e-05, "loss": 74.9684, "step": 3638 }, { "compression_loss": 76.95381927490234, "epoch": 1.2, "learning_rate": 1.209264126149803e-05, "loss": 77.479, "step": 3639 }, { "compression_loss": 75.25900268554688, "epoch": 1.2, "learning_rate": 1.2087713534822602e-05, "loss": 75.7037, "step": 3640 }, { "compression_loss": 76.40925598144531, "epoch": 1.2, "learning_rate": 1.2082785808147174e-05, "loss": 76.7304, "step": 3641 }, { "compression_loss": 78.30643463134766, "epoch": 1.2, "learning_rate": 1.2077858081471749e-05, "loss": 78.5935, "step": 3642 }, { "compression_loss": 75.90010070800781, "epoch": 1.2, "learning_rate": 1.2072930354796321e-05, "loss": 76.1027, "step": 3643 }, { "compression_loss": 76.12672424316406, "epoch": 1.2, "learning_rate": 1.2068002628120895e-05, "loss": 76.5636, "step": 3644 }, { "compression_loss": 77.28067779541016, "epoch": 1.2, "learning_rate": 1.2063074901445467e-05, "loss": 77.7646, "step": 3645 }, { "compression_loss": 77.26416778564453, "epoch": 1.2, "learning_rate": 1.2058147174770039e-05, "loss": 77.6352, "step": 3646 }, { "compression_loss": 76.11913299560547, "epoch": 1.2, "learning_rate": 1.2053219448094612e-05, "loss": 76.4595, "step": 3647 }, { "compression_loss": 77.56867218017578, "epoch": 1.2, "learning_rate": 1.2048291721419186e-05, "loss": 78.0263, "step": 3648 }, { "compression_loss": 75.82371520996094, "epoch": 1.2, "learning_rate": 1.204336399474376e-05, "loss": 76.2468, "step": 3649 }, { "compression_loss": 74.99589538574219, "epoch": 1.2, "learning_rate": 1.2038436268068331e-05, "loss": 75.4431, "step": 3650 }, { "compression_loss": 75.97245025634766, "epoch": 1.2, "learning_rate": 1.2033508541392903e-05, "loss": 76.4923, "step": 3651 }, { "compression_loss": 76.196044921875, "epoch": 1.2, "learning_rate": 1.2028580814717477e-05, "loss": 76.769, "step": 3652 }, { "compression_loss": 75.02120971679688, "epoch": 1.2, "learning_rate": 1.2023653088042049e-05, "loss": 75.5487, "step": 3653 }, { "compression_loss": 75.836669921875, "epoch": 1.2, "learning_rate": 1.2018725361366624e-05, "loss": 76.2026, "step": 3654 }, { "compression_loss": 73.15943908691406, "epoch": 1.2, "learning_rate": 1.2013797634691196e-05, "loss": 73.4867, "step": 3655 }, { "compression_loss": 77.93348693847656, "epoch": 1.2, "learning_rate": 1.2008869908015768e-05, "loss": 78.311, "step": 3656 }, { "compression_loss": 75.6724624633789, "epoch": 1.2, "learning_rate": 1.2003942181340342e-05, "loss": 75.8644, "step": 3657 }, { "compression_loss": 76.10101318359375, "epoch": 1.2, "learning_rate": 1.1999014454664914e-05, "loss": 76.5238, "step": 3658 }, { "compression_loss": 75.94198608398438, "epoch": 1.2, "learning_rate": 1.1994086727989488e-05, "loss": 76.2587, "step": 3659 }, { "compression_loss": 75.14775085449219, "epoch": 1.2, "learning_rate": 1.1989159001314061e-05, "loss": 75.5337, "step": 3660 }, { "compression_loss": 74.92640686035156, "epoch": 1.2, "learning_rate": 1.1984231274638633e-05, "loss": 75.4351, "step": 3661 }, { "compression_loss": 74.8203125, "epoch": 1.2, "learning_rate": 1.1979303547963207e-05, "loss": 75.0883, "step": 3662 }, { "compression_loss": 74.72417449951172, "epoch": 1.2, "learning_rate": 1.1974375821287779e-05, "loss": 75.0696, "step": 3663 }, { "compression_loss": 74.6166000366211, "epoch": 1.2, "learning_rate": 1.1969448094612353e-05, "loss": 75.0882, "step": 3664 }, { "compression_loss": 76.52532958984375, "epoch": 1.2, "learning_rate": 1.1964520367936925e-05, "loss": 76.9177, "step": 3665 }, { "compression_loss": 77.43585968017578, "epoch": 1.2, "learning_rate": 1.1959592641261498e-05, "loss": 77.6655, "step": 3666 }, { "compression_loss": 77.95711517333984, "epoch": 1.2, "learning_rate": 1.1954664914586072e-05, "loss": 78.4962, "step": 3667 }, { "compression_loss": 75.16763305664062, "epoch": 1.2, "learning_rate": 1.1949737187910644e-05, "loss": 75.3291, "step": 3668 }, { "compression_loss": 73.02029418945312, "epoch": 1.21, "learning_rate": 1.1944809461235218e-05, "loss": 73.2728, "step": 3669 }, { "compression_loss": 74.69436645507812, "epoch": 1.21, "learning_rate": 1.193988173455979e-05, "loss": 74.9229, "step": 3670 }, { "compression_loss": 75.95191955566406, "epoch": 1.21, "learning_rate": 1.1934954007884361e-05, "loss": 76.3014, "step": 3671 }, { "compression_loss": 76.90188598632812, "epoch": 1.21, "learning_rate": 1.1930026281208937e-05, "loss": 77.2072, "step": 3672 }, { "compression_loss": 77.54058074951172, "epoch": 1.21, "learning_rate": 1.1925098554533509e-05, "loss": 77.6832, "step": 3673 }, { "compression_loss": 75.35865020751953, "epoch": 1.21, "learning_rate": 1.1920170827858082e-05, "loss": 75.6474, "step": 3674 }, { "compression_loss": 77.08910369873047, "epoch": 1.21, "learning_rate": 1.1915243101182654e-05, "loss": 77.2939, "step": 3675 }, { "compression_loss": 77.22330474853516, "epoch": 1.21, "learning_rate": 1.1910315374507226e-05, "loss": 77.652, "step": 3676 }, { "compression_loss": 76.059814453125, "epoch": 1.21, "learning_rate": 1.19053876478318e-05, "loss": 76.4411, "step": 3677 }, { "compression_loss": 77.07597351074219, "epoch": 1.21, "learning_rate": 1.1900459921156374e-05, "loss": 77.3879, "step": 3678 }, { "compression_loss": 76.82688903808594, "epoch": 1.21, "learning_rate": 1.1895532194480947e-05, "loss": 77.1354, "step": 3679 }, { "compression_loss": 75.20849609375, "epoch": 1.21, "learning_rate": 1.189060446780552e-05, "loss": 75.8022, "step": 3680 }, { "compression_loss": 74.34529113769531, "epoch": 1.21, "learning_rate": 1.1885676741130091e-05, "loss": 74.8398, "step": 3681 }, { "compression_loss": 75.89955139160156, "epoch": 1.21, "learning_rate": 1.1880749014454665e-05, "loss": 76.1688, "step": 3682 }, { "compression_loss": 76.00444793701172, "epoch": 1.21, "learning_rate": 1.1875821287779237e-05, "loss": 76.6504, "step": 3683 }, { "compression_loss": 77.35739135742188, "epoch": 1.21, "learning_rate": 1.1870893561103812e-05, "loss": 77.8643, "step": 3684 }, { "compression_loss": 76.30980682373047, "epoch": 1.21, "learning_rate": 1.1865965834428384e-05, "loss": 76.7191, "step": 3685 }, { "compression_loss": 74.67188262939453, "epoch": 1.21, "learning_rate": 1.1861038107752956e-05, "loss": 74.9344, "step": 3686 }, { "compression_loss": 75.18110656738281, "epoch": 1.21, "learning_rate": 1.185611038107753e-05, "loss": 75.5178, "step": 3687 }, { "compression_loss": 74.14826965332031, "epoch": 1.21, "learning_rate": 1.1851182654402102e-05, "loss": 74.5055, "step": 3688 }, { "compression_loss": 77.081787109375, "epoch": 1.21, "learning_rate": 1.1846254927726676e-05, "loss": 77.6391, "step": 3689 }, { "compression_loss": 73.99781799316406, "epoch": 1.21, "learning_rate": 1.184132720105125e-05, "loss": 74.3529, "step": 3690 }, { "compression_loss": 77.11127471923828, "epoch": 1.21, "learning_rate": 1.1836399474375821e-05, "loss": 77.5011, "step": 3691 }, { "compression_loss": 73.86592102050781, "epoch": 1.21, "learning_rate": 1.1831471747700395e-05, "loss": 74.1368, "step": 3692 }, { "compression_loss": 76.65753936767578, "epoch": 1.21, "learning_rate": 1.1826544021024967e-05, "loss": 76.877, "step": 3693 }, { "compression_loss": 74.7192611694336, "epoch": 1.21, "learning_rate": 1.182161629434954e-05, "loss": 75.0392, "step": 3694 }, { "compression_loss": 76.13683319091797, "epoch": 1.21, "learning_rate": 1.1816688567674112e-05, "loss": 76.3594, "step": 3695 }, { "compression_loss": 76.98970031738281, "epoch": 1.21, "learning_rate": 1.1811760840998686e-05, "loss": 77.2829, "step": 3696 }, { "compression_loss": 76.48373413085938, "epoch": 1.21, "learning_rate": 1.180683311432326e-05, "loss": 76.8019, "step": 3697 }, { "compression_loss": 75.55101013183594, "epoch": 1.21, "learning_rate": 1.1801905387647832e-05, "loss": 76.2844, "step": 3698 }, { "compression_loss": 76.47103881835938, "epoch": 1.22, "learning_rate": 1.1796977660972405e-05, "loss": 76.9157, "step": 3699 }, { "compression_loss": 74.93089294433594, "epoch": 1.22, "learning_rate": 1.1792049934296977e-05, "loss": 75.2111, "step": 3700 }, { "compression_loss": 76.25588989257812, "epoch": 1.22, "learning_rate": 1.178712220762155e-05, "loss": 76.6123, "step": 3701 }, { "compression_loss": 75.73855590820312, "epoch": 1.22, "learning_rate": 1.1782194480946125e-05, "loss": 76.0392, "step": 3702 }, { "compression_loss": 75.41357421875, "epoch": 1.22, "learning_rate": 1.1777266754270697e-05, "loss": 75.7046, "step": 3703 }, { "compression_loss": 75.06571960449219, "epoch": 1.22, "learning_rate": 1.177233902759527e-05, "loss": 75.4262, "step": 3704 }, { "compression_loss": 77.08506774902344, "epoch": 1.22, "learning_rate": 1.1767411300919842e-05, "loss": 77.4033, "step": 3705 }, { "compression_loss": 73.22684478759766, "epoch": 1.22, "learning_rate": 1.1762483574244414e-05, "loss": 73.5066, "step": 3706 }, { "compression_loss": 76.97398376464844, "epoch": 1.22, "learning_rate": 1.1757555847568988e-05, "loss": 77.3387, "step": 3707 }, { "compression_loss": 75.91413116455078, "epoch": 1.22, "learning_rate": 1.1752628120893562e-05, "loss": 76.2809, "step": 3708 }, { "compression_loss": 73.76161193847656, "epoch": 1.22, "learning_rate": 1.1747700394218135e-05, "loss": 73.936, "step": 3709 }, { "compression_loss": 75.16361236572266, "epoch": 1.22, "learning_rate": 1.1742772667542707e-05, "loss": 75.3857, "step": 3710 }, { "compression_loss": 76.0049819946289, "epoch": 1.22, "learning_rate": 1.173784494086728e-05, "loss": 76.2496, "step": 3711 }, { "compression_loss": 74.9534683227539, "epoch": 1.22, "learning_rate": 1.1732917214191853e-05, "loss": 75.528, "step": 3712 }, { "compression_loss": 75.95277404785156, "epoch": 1.22, "learning_rate": 1.1727989487516425e-05, "loss": 76.3355, "step": 3713 }, { "compression_loss": 75.26539611816406, "epoch": 1.22, "learning_rate": 1.1723061760841e-05, "loss": 75.4254, "step": 3714 }, { "compression_loss": 74.81893157958984, "epoch": 1.22, "learning_rate": 1.1718134034165572e-05, "loss": 75.0073, "step": 3715 }, { "compression_loss": 76.59281158447266, "epoch": 1.22, "learning_rate": 1.1713206307490144e-05, "loss": 77.0917, "step": 3716 }, { "compression_loss": 76.83384704589844, "epoch": 1.22, "learning_rate": 1.1708278580814718e-05, "loss": 77.2114, "step": 3717 }, { "compression_loss": 73.89236450195312, "epoch": 1.22, "learning_rate": 1.170335085413929e-05, "loss": 74.3576, "step": 3718 }, { "compression_loss": 74.31339263916016, "epoch": 1.22, "learning_rate": 1.1698423127463863e-05, "loss": 74.4403, "step": 3719 }, { "compression_loss": 76.30117797851562, "epoch": 1.22, "learning_rate": 1.1693495400788437e-05, "loss": 76.7152, "step": 3720 }, { "compression_loss": 75.28276824951172, "epoch": 1.22, "learning_rate": 1.1688567674113009e-05, "loss": 75.5584, "step": 3721 }, { "compression_loss": 77.1649398803711, "epoch": 1.22, "learning_rate": 1.1683639947437583e-05, "loss": 77.435, "step": 3722 }, { "compression_loss": 74.3124008178711, "epoch": 1.22, "learning_rate": 1.1678712220762155e-05, "loss": 74.8054, "step": 3723 }, { "compression_loss": 76.33321380615234, "epoch": 1.22, "learning_rate": 1.1673784494086728e-05, "loss": 76.4679, "step": 3724 }, { "compression_loss": 76.08119201660156, "epoch": 1.22, "learning_rate": 1.16688567674113e-05, "loss": 76.4527, "step": 3725 }, { "compression_loss": 74.3607177734375, "epoch": 1.22, "learning_rate": 1.1663929040735874e-05, "loss": 74.9049, "step": 3726 }, { "compression_loss": 75.77269744873047, "epoch": 1.22, "learning_rate": 1.1659001314060448e-05, "loss": 76.1536, "step": 3727 }, { "compression_loss": 76.3177490234375, "epoch": 1.22, "learning_rate": 1.165407358738502e-05, "loss": 76.6899, "step": 3728 }, { "compression_loss": 75.93107604980469, "epoch": 1.23, "learning_rate": 1.1649145860709593e-05, "loss": 76.1975, "step": 3729 }, { "compression_loss": 77.58531951904297, "epoch": 1.23, "learning_rate": 1.1644218134034165e-05, "loss": 77.8772, "step": 3730 }, { "compression_loss": 75.80109405517578, "epoch": 1.23, "learning_rate": 1.1639290407358737e-05, "loss": 76.0525, "step": 3731 }, { "compression_loss": 78.45614624023438, "epoch": 1.23, "learning_rate": 1.1634362680683313e-05, "loss": 78.6839, "step": 3732 }, { "compression_loss": 75.84251403808594, "epoch": 1.23, "learning_rate": 1.1629434954007885e-05, "loss": 76.3175, "step": 3733 }, { "compression_loss": 76.3088607788086, "epoch": 1.23, "learning_rate": 1.1624507227332458e-05, "loss": 76.587, "step": 3734 }, { "compression_loss": 76.25395965576172, "epoch": 1.23, "learning_rate": 1.161957950065703e-05, "loss": 76.4713, "step": 3735 }, { "compression_loss": 75.30944061279297, "epoch": 1.23, "learning_rate": 1.1614651773981602e-05, "loss": 75.6935, "step": 3736 }, { "compression_loss": 73.90604400634766, "epoch": 1.23, "learning_rate": 1.1609724047306176e-05, "loss": 74.2352, "step": 3737 }, { "compression_loss": 74.808837890625, "epoch": 1.23, "learning_rate": 1.160479632063075e-05, "loss": 75.1998, "step": 3738 }, { "compression_loss": 73.28401184082031, "epoch": 1.23, "learning_rate": 1.1599868593955323e-05, "loss": 73.5813, "step": 3739 }, { "compression_loss": 77.44476318359375, "epoch": 1.23, "learning_rate": 1.1594940867279895e-05, "loss": 78.082, "step": 3740 }, { "compression_loss": 74.76632690429688, "epoch": 1.23, "learning_rate": 1.1590013140604467e-05, "loss": 75.201, "step": 3741 }, { "compression_loss": 75.26312255859375, "epoch": 1.23, "learning_rate": 1.158508541392904e-05, "loss": 75.5162, "step": 3742 }, { "compression_loss": 76.92359161376953, "epoch": 1.23, "learning_rate": 1.1580157687253614e-05, "loss": 77.3767, "step": 3743 }, { "compression_loss": 76.736572265625, "epoch": 1.23, "learning_rate": 1.1575229960578188e-05, "loss": 77.2045, "step": 3744 }, { "compression_loss": 75.39750671386719, "epoch": 1.23, "learning_rate": 1.157030223390276e-05, "loss": 75.6712, "step": 3745 }, { "compression_loss": 73.93112182617188, "epoch": 1.23, "learning_rate": 1.1565374507227332e-05, "loss": 74.2393, "step": 3746 }, { "compression_loss": 75.16179656982422, "epoch": 1.23, "learning_rate": 1.1560446780551906e-05, "loss": 75.5066, "step": 3747 }, { "compression_loss": 73.55319213867188, "epoch": 1.23, "learning_rate": 1.1555519053876478e-05, "loss": 73.9314, "step": 3748 }, { "compression_loss": 77.01502990722656, "epoch": 1.23, "learning_rate": 1.1550591327201053e-05, "loss": 77.4687, "step": 3749 }, { "compression_loss": 73.89677429199219, "epoch": 1.23, "learning_rate": 1.1545663600525625e-05, "loss": 74.15, "step": 3750 }, { "epoch": 1.23, "eval_exact_match": 86.49006622516556, "eval_f1": 92.72168717578025, "step": 3750 }, { "compression_loss": 76.36605834960938, "epoch": 1.23, "learning_rate": 1.1540735873850197e-05, "loss": 76.6689, "step": 3751 }, { "compression_loss": 74.91803741455078, "epoch": 1.23, "learning_rate": 1.153580814717477e-05, "loss": 75.4553, "step": 3752 }, { "compression_loss": 76.58645629882812, "epoch": 1.23, "learning_rate": 1.1530880420499343e-05, "loss": 76.8852, "step": 3753 }, { "compression_loss": 75.44781494140625, "epoch": 1.23, "learning_rate": 1.1525952693823916e-05, "loss": 75.839, "step": 3754 }, { "compression_loss": 74.51934814453125, "epoch": 1.23, "learning_rate": 1.152102496714849e-05, "loss": 74.9322, "step": 3755 }, { "compression_loss": 75.40977478027344, "epoch": 1.23, "learning_rate": 1.1516097240473062e-05, "loss": 75.6079, "step": 3756 }, { "compression_loss": 75.55890655517578, "epoch": 1.23, "learning_rate": 1.1511169513797636e-05, "loss": 75.7304, "step": 3757 }, { "compression_loss": 74.87504577636719, "epoch": 1.23, "learning_rate": 1.1506241787122207e-05, "loss": 75.1072, "step": 3758 }, { "compression_loss": 72.27474975585938, "epoch": 1.23, "learning_rate": 1.1501314060446781e-05, "loss": 72.533, "step": 3759 }, { "compression_loss": 74.8511734008789, "epoch": 1.24, "learning_rate": 1.1496386333771353e-05, "loss": 75.1466, "step": 3760 }, { "compression_loss": 75.31079864501953, "epoch": 1.24, "learning_rate": 1.1491458607095927e-05, "loss": 75.5723, "step": 3761 }, { "compression_loss": 73.75018310546875, "epoch": 1.24, "learning_rate": 1.14865308804205e-05, "loss": 74.0496, "step": 3762 }, { "compression_loss": 76.10206604003906, "epoch": 1.24, "learning_rate": 1.1481603153745072e-05, "loss": 76.3253, "step": 3763 }, { "compression_loss": 75.50191497802734, "epoch": 1.24, "learning_rate": 1.1476675427069646e-05, "loss": 75.7538, "step": 3764 }, { "compression_loss": 76.53047180175781, "epoch": 1.24, "learning_rate": 1.1471747700394218e-05, "loss": 76.728, "step": 3765 }, { "compression_loss": 78.7588882446289, "epoch": 1.24, "learning_rate": 1.146681997371879e-05, "loss": 79.2343, "step": 3766 }, { "compression_loss": 74.58650207519531, "epoch": 1.24, "learning_rate": 1.1461892247043365e-05, "loss": 74.8531, "step": 3767 }, { "compression_loss": 76.0647964477539, "epoch": 1.24, "learning_rate": 1.1456964520367937e-05, "loss": 76.4656, "step": 3768 }, { "compression_loss": 74.75848388671875, "epoch": 1.24, "learning_rate": 1.1452036793692511e-05, "loss": 75.2719, "step": 3769 }, { "compression_loss": 74.90961456298828, "epoch": 1.24, "learning_rate": 1.1447109067017083e-05, "loss": 75.2959, "step": 3770 }, { "compression_loss": 74.59296417236328, "epoch": 1.24, "learning_rate": 1.1442181340341655e-05, "loss": 74.7832, "step": 3771 }, { "compression_loss": 77.53568267822266, "epoch": 1.24, "learning_rate": 1.1437253613666229e-05, "loss": 77.8276, "step": 3772 }, { "compression_loss": 73.74694061279297, "epoch": 1.24, "learning_rate": 1.1432325886990802e-05, "loss": 74.0032, "step": 3773 }, { "compression_loss": 75.66658782958984, "epoch": 1.24, "learning_rate": 1.1427398160315376e-05, "loss": 75.9089, "step": 3774 }, { "compression_loss": 75.63674926757812, "epoch": 1.24, "learning_rate": 1.1422470433639948e-05, "loss": 76.0832, "step": 3775 }, { "compression_loss": 75.26553344726562, "epoch": 1.24, "learning_rate": 1.141754270696452e-05, "loss": 76.0646, "step": 3776 }, { "compression_loss": 76.08418273925781, "epoch": 1.24, "learning_rate": 1.1412614980289094e-05, "loss": 76.2939, "step": 3777 }, { "compression_loss": 74.47189331054688, "epoch": 1.24, "learning_rate": 1.1407687253613666e-05, "loss": 74.7608, "step": 3778 }, { "compression_loss": 74.4904556274414, "epoch": 1.24, "learning_rate": 1.1402759526938241e-05, "loss": 74.7411, "step": 3779 }, { "compression_loss": 73.99295043945312, "epoch": 1.24, "learning_rate": 1.1397831800262813e-05, "loss": 74.1816, "step": 3780 }, { "compression_loss": 74.6719741821289, "epoch": 1.24, "learning_rate": 1.1392904073587385e-05, "loss": 75.3283, "step": 3781 }, { "compression_loss": 72.38492584228516, "epoch": 1.24, "learning_rate": 1.1387976346911958e-05, "loss": 72.7585, "step": 3782 }, { "compression_loss": 75.81615447998047, "epoch": 1.24, "learning_rate": 1.138304862023653e-05, "loss": 76.0416, "step": 3783 }, { "compression_loss": 75.03636169433594, "epoch": 1.24, "learning_rate": 1.1378120893561104e-05, "loss": 75.1837, "step": 3784 }, { "compression_loss": 73.60367584228516, "epoch": 1.24, "learning_rate": 1.1373193166885678e-05, "loss": 74.1616, "step": 3785 }, { "compression_loss": 74.09309387207031, "epoch": 1.24, "learning_rate": 1.136826544021025e-05, "loss": 74.459, "step": 3786 }, { "compression_loss": 77.15200805664062, "epoch": 1.24, "learning_rate": 1.1363337713534823e-05, "loss": 77.5706, "step": 3787 }, { "compression_loss": 71.78872680664062, "epoch": 1.24, "learning_rate": 1.1358409986859395e-05, "loss": 72.0377, "step": 3788 }, { "compression_loss": 75.58557891845703, "epoch": 1.24, "learning_rate": 1.1353482260183969e-05, "loss": 75.8669, "step": 3789 }, { "compression_loss": 76.0938720703125, "epoch": 1.25, "learning_rate": 1.1348554533508541e-05, "loss": 76.3971, "step": 3790 }, { "compression_loss": 76.81855010986328, "epoch": 1.25, "learning_rate": 1.1343626806833115e-05, "loss": 77.0786, "step": 3791 }, { "compression_loss": 75.41541290283203, "epoch": 1.25, "learning_rate": 1.1338699080157688e-05, "loss": 75.604, "step": 3792 }, { "compression_loss": 74.56571197509766, "epoch": 1.25, "learning_rate": 1.133377135348226e-05, "loss": 74.8647, "step": 3793 }, { "compression_loss": 77.08122253417969, "epoch": 1.25, "learning_rate": 1.1328843626806834e-05, "loss": 77.5022, "step": 3794 }, { "compression_loss": 75.0367431640625, "epoch": 1.25, "learning_rate": 1.1323915900131406e-05, "loss": 75.4513, "step": 3795 }, { "compression_loss": 74.04886627197266, "epoch": 1.25, "learning_rate": 1.1318988173455978e-05, "loss": 74.3811, "step": 3796 }, { "compression_loss": 75.97525787353516, "epoch": 1.25, "learning_rate": 1.1314060446780553e-05, "loss": 76.3721, "step": 3797 }, { "compression_loss": 74.5714340209961, "epoch": 1.25, "learning_rate": 1.1309132720105125e-05, "loss": 74.8423, "step": 3798 }, { "compression_loss": 76.3357162475586, "epoch": 1.25, "learning_rate": 1.1304204993429699e-05, "loss": 76.7277, "step": 3799 }, { "compression_loss": 74.1329345703125, "epoch": 1.25, "learning_rate": 1.1299277266754271e-05, "loss": 74.4036, "step": 3800 }, { "compression_loss": 76.61314392089844, "epoch": 1.25, "learning_rate": 1.1294349540078843e-05, "loss": 77.0385, "step": 3801 }, { "compression_loss": 74.38359069824219, "epoch": 1.25, "learning_rate": 1.1289421813403416e-05, "loss": 74.581, "step": 3802 }, { "compression_loss": 75.86705017089844, "epoch": 1.25, "learning_rate": 1.128449408672799e-05, "loss": 76.4469, "step": 3803 }, { "compression_loss": 75.94698333740234, "epoch": 1.25, "learning_rate": 1.1279566360052564e-05, "loss": 76.0975, "step": 3804 }, { "compression_loss": 75.649169921875, "epoch": 1.25, "learning_rate": 1.1274638633377136e-05, "loss": 75.9649, "step": 3805 }, { "compression_loss": 75.3888168334961, "epoch": 1.25, "learning_rate": 1.1269710906701708e-05, "loss": 75.9128, "step": 3806 }, { "compression_loss": 74.75399780273438, "epoch": 1.25, "learning_rate": 1.1264783180026281e-05, "loss": 75.0011, "step": 3807 }, { "compression_loss": 76.050048828125, "epoch": 1.25, "learning_rate": 1.1259855453350853e-05, "loss": 76.4807, "step": 3808 }, { "compression_loss": 74.73400115966797, "epoch": 1.25, "learning_rate": 1.1254927726675429e-05, "loss": 75.0582, "step": 3809 }, { "compression_loss": 76.53268432617188, "epoch": 1.25, "learning_rate": 1.125e-05, "loss": 76.8932, "step": 3810 }, { "compression_loss": 73.5108642578125, "epoch": 1.25, "learning_rate": 1.1245072273324573e-05, "loss": 74.065, "step": 3811 }, { "compression_loss": 73.12492370605469, "epoch": 1.25, "learning_rate": 1.1240144546649146e-05, "loss": 73.349, "step": 3812 }, { "compression_loss": 75.35918426513672, "epoch": 1.25, "learning_rate": 1.1235216819973718e-05, "loss": 75.7539, "step": 3813 }, { "compression_loss": 78.32408142089844, "epoch": 1.25, "learning_rate": 1.1230289093298292e-05, "loss": 78.7306, "step": 3814 }, { "compression_loss": 75.07919311523438, "epoch": 1.25, "learning_rate": 1.1225361366622866e-05, "loss": 75.365, "step": 3815 }, { "compression_loss": 76.3564453125, "epoch": 1.25, "learning_rate": 1.1220433639947438e-05, "loss": 76.6877, "step": 3816 }, { "compression_loss": 73.91616821289062, "epoch": 1.25, "learning_rate": 1.1215505913272011e-05, "loss": 74.1718, "step": 3817 }, { "compression_loss": 75.11585998535156, "epoch": 1.25, "learning_rate": 1.1210578186596583e-05, "loss": 75.3888, "step": 3818 }, { "compression_loss": 74.33128356933594, "epoch": 1.25, "learning_rate": 1.1205650459921157e-05, "loss": 74.5502, "step": 3819 }, { "compression_loss": 75.02582550048828, "epoch": 1.25, "learning_rate": 1.1200722733245729e-05, "loss": 75.2456, "step": 3820 }, { "compression_loss": 76.06700134277344, "epoch": 1.26, "learning_rate": 1.1195795006570303e-05, "loss": 76.4272, "step": 3821 }, { "compression_loss": 77.32476806640625, "epoch": 1.26, "learning_rate": 1.1190867279894876e-05, "loss": 77.6623, "step": 3822 }, { "compression_loss": 75.4710693359375, "epoch": 1.26, "learning_rate": 1.1185939553219448e-05, "loss": 75.7881, "step": 3823 }, { "compression_loss": 78.55115509033203, "epoch": 1.26, "learning_rate": 1.1181011826544022e-05, "loss": 79.3783, "step": 3824 }, { "compression_loss": 75.03952026367188, "epoch": 1.26, "learning_rate": 1.1176084099868594e-05, "loss": 75.591, "step": 3825 }, { "compression_loss": 74.55022430419922, "epoch": 1.26, "learning_rate": 1.1171156373193166e-05, "loss": 74.9491, "step": 3826 }, { "compression_loss": 74.53814697265625, "epoch": 1.26, "learning_rate": 1.1166228646517741e-05, "loss": 74.9974, "step": 3827 }, { "compression_loss": 74.36724853515625, "epoch": 1.26, "learning_rate": 1.1161300919842313e-05, "loss": 74.8127, "step": 3828 }, { "compression_loss": 75.39788818359375, "epoch": 1.26, "learning_rate": 1.1156373193166887e-05, "loss": 75.9085, "step": 3829 }, { "compression_loss": 76.05477142333984, "epoch": 1.26, "learning_rate": 1.1151445466491459e-05, "loss": 76.3172, "step": 3830 }, { "compression_loss": 75.60958099365234, "epoch": 1.26, "learning_rate": 1.114651773981603e-05, "loss": 75.7982, "step": 3831 }, { "compression_loss": 75.14511108398438, "epoch": 1.26, "learning_rate": 1.1141590013140604e-05, "loss": 75.5581, "step": 3832 }, { "compression_loss": 75.6252212524414, "epoch": 1.26, "learning_rate": 1.1136662286465178e-05, "loss": 76.0831, "step": 3833 }, { "compression_loss": 75.91299438476562, "epoch": 1.26, "learning_rate": 1.1131734559789752e-05, "loss": 76.1196, "step": 3834 }, { "compression_loss": 75.09992980957031, "epoch": 1.26, "learning_rate": 1.1126806833114324e-05, "loss": 75.5676, "step": 3835 }, { "compression_loss": 77.01104736328125, "epoch": 1.26, "learning_rate": 1.1121879106438896e-05, "loss": 77.3996, "step": 3836 }, { "compression_loss": 76.57646179199219, "epoch": 1.26, "learning_rate": 1.111695137976347e-05, "loss": 77.0085, "step": 3837 }, { "compression_loss": 76.39973449707031, "epoch": 1.26, "learning_rate": 1.1112023653088041e-05, "loss": 76.8642, "step": 3838 }, { "compression_loss": 76.55265045166016, "epoch": 1.26, "learning_rate": 1.1107095926412617e-05, "loss": 77.2406, "step": 3839 }, { "compression_loss": 73.02508544921875, "epoch": 1.26, "learning_rate": 1.1102168199737189e-05, "loss": 73.4596, "step": 3840 }, { "compression_loss": 73.01629638671875, "epoch": 1.26, "learning_rate": 1.109724047306176e-05, "loss": 73.4108, "step": 3841 }, { "compression_loss": 74.60021209716797, "epoch": 1.26, "learning_rate": 1.1092312746386334e-05, "loss": 74.8279, "step": 3842 }, { "compression_loss": 75.92213439941406, "epoch": 1.26, "learning_rate": 1.1087385019710906e-05, "loss": 76.3727, "step": 3843 }, { "compression_loss": 75.71662902832031, "epoch": 1.26, "learning_rate": 1.108245729303548e-05, "loss": 75.9853, "step": 3844 }, { "compression_loss": 76.68399047851562, "epoch": 1.26, "learning_rate": 1.1077529566360054e-05, "loss": 77.2647, "step": 3845 }, { "compression_loss": 75.70626068115234, "epoch": 1.26, "learning_rate": 1.1072601839684625e-05, "loss": 76.4087, "step": 3846 }, { "compression_loss": 73.48812866210938, "epoch": 1.26, "learning_rate": 1.1067674113009199e-05, "loss": 73.6976, "step": 3847 }, { "compression_loss": 73.75904846191406, "epoch": 1.26, "learning_rate": 1.1062746386333771e-05, "loss": 74.4526, "step": 3848 }, { "compression_loss": 74.1412582397461, "epoch": 1.26, "learning_rate": 1.1057818659658345e-05, "loss": 74.6162, "step": 3849 }, { "compression_loss": 74.41976928710938, "epoch": 1.26, "learning_rate": 1.1052890932982917e-05, "loss": 74.9166, "step": 3850 }, { "compression_loss": 77.14227294921875, "epoch": 1.27, "learning_rate": 1.104796320630749e-05, "loss": 77.6219, "step": 3851 }, { "compression_loss": 76.80026245117188, "epoch": 1.27, "learning_rate": 1.1043035479632064e-05, "loss": 77.2399, "step": 3852 }, { "compression_loss": 75.58910369873047, "epoch": 1.27, "learning_rate": 1.1038107752956636e-05, "loss": 76.0488, "step": 3853 }, { "compression_loss": 77.51966094970703, "epoch": 1.27, "learning_rate": 1.103318002628121e-05, "loss": 78.0823, "step": 3854 }, { "compression_loss": 75.416748046875, "epoch": 1.27, "learning_rate": 1.1028252299605782e-05, "loss": 75.8716, "step": 3855 }, { "compression_loss": 73.50648498535156, "epoch": 1.27, "learning_rate": 1.1023324572930354e-05, "loss": 73.6572, "step": 3856 }, { "compression_loss": 76.86749267578125, "epoch": 1.27, "learning_rate": 1.1018396846254929e-05, "loss": 77.2202, "step": 3857 }, { "compression_loss": 73.95781707763672, "epoch": 1.27, "learning_rate": 1.1013469119579501e-05, "loss": 74.3146, "step": 3858 }, { "compression_loss": 74.58016204833984, "epoch": 1.27, "learning_rate": 1.1008541392904075e-05, "loss": 75.1204, "step": 3859 }, { "compression_loss": 77.76168823242188, "epoch": 1.27, "learning_rate": 1.1003613666228647e-05, "loss": 78.297, "step": 3860 }, { "compression_loss": 74.6428451538086, "epoch": 1.27, "learning_rate": 1.0998685939553219e-05, "loss": 75.1112, "step": 3861 }, { "compression_loss": 73.91578674316406, "epoch": 1.27, "learning_rate": 1.0993758212877792e-05, "loss": 74.4803, "step": 3862 }, { "compression_loss": 74.56402587890625, "epoch": 1.27, "learning_rate": 1.0988830486202366e-05, "loss": 75.0272, "step": 3863 }, { "compression_loss": 75.2935791015625, "epoch": 1.27, "learning_rate": 1.098390275952694e-05, "loss": 75.7175, "step": 3864 }, { "compression_loss": 76.51396942138672, "epoch": 1.27, "learning_rate": 1.0978975032851512e-05, "loss": 76.6563, "step": 3865 }, { "compression_loss": 77.60273742675781, "epoch": 1.27, "learning_rate": 1.0974047306176083e-05, "loss": 77.9137, "step": 3866 }, { "compression_loss": 75.70622253417969, "epoch": 1.27, "learning_rate": 1.0969119579500657e-05, "loss": 75.9521, "step": 3867 }, { "compression_loss": 75.2601318359375, "epoch": 1.27, "learning_rate": 1.0964191852825229e-05, "loss": 75.683, "step": 3868 }, { "compression_loss": 74.04753112792969, "epoch": 1.27, "learning_rate": 1.0959264126149804e-05, "loss": 74.283, "step": 3869 }, { "compression_loss": 76.81964874267578, "epoch": 1.27, "learning_rate": 1.0954336399474376e-05, "loss": 77.0693, "step": 3870 }, { "compression_loss": 74.81542205810547, "epoch": 1.27, "learning_rate": 1.0949408672798948e-05, "loss": 75.0584, "step": 3871 }, { "compression_loss": 76.52427673339844, "epoch": 1.27, "learning_rate": 1.0944480946123522e-05, "loss": 77.208, "step": 3872 }, { "compression_loss": 75.26922607421875, "epoch": 1.27, "learning_rate": 1.0939553219448094e-05, "loss": 75.6115, "step": 3873 }, { "compression_loss": 74.8085708618164, "epoch": 1.27, "learning_rate": 1.0934625492772668e-05, "loss": 75.6547, "step": 3874 }, { "compression_loss": 75.30635070800781, "epoch": 1.27, "learning_rate": 1.0929697766097241e-05, "loss": 75.814, "step": 3875 }, { "compression_loss": 71.7295150756836, "epoch": 1.27, "learning_rate": 1.0924770039421813e-05, "loss": 72.0439, "step": 3876 }, { "compression_loss": 77.60197448730469, "epoch": 1.27, "learning_rate": 1.0919842312746387e-05, "loss": 77.9588, "step": 3877 }, { "compression_loss": 77.2432632446289, "epoch": 1.27, "learning_rate": 1.0914914586070959e-05, "loss": 77.4773, "step": 3878 }, { "compression_loss": 77.44524383544922, "epoch": 1.27, "learning_rate": 1.0909986859395533e-05, "loss": 77.7141, "step": 3879 }, { "compression_loss": 72.3268814086914, "epoch": 1.27, "learning_rate": 1.0905059132720106e-05, "loss": 72.5714, "step": 3880 }, { "compression_loss": 74.96672821044922, "epoch": 1.27, "learning_rate": 1.0900131406044678e-05, "loss": 75.3767, "step": 3881 }, { "compression_loss": 75.53071594238281, "epoch": 1.28, "learning_rate": 1.0895203679369252e-05, "loss": 75.8983, "step": 3882 }, { "compression_loss": 75.2549057006836, "epoch": 1.28, "learning_rate": 1.0890275952693824e-05, "loss": 75.4539, "step": 3883 }, { "compression_loss": 75.63150024414062, "epoch": 1.28, "learning_rate": 1.0885348226018398e-05, "loss": 76.0813, "step": 3884 }, { "compression_loss": 73.97029113769531, "epoch": 1.28, "learning_rate": 1.088042049934297e-05, "loss": 74.2964, "step": 3885 }, { "compression_loss": 76.4461441040039, "epoch": 1.28, "learning_rate": 1.0875492772667543e-05, "loss": 76.71, "step": 3886 }, { "compression_loss": 75.54093933105469, "epoch": 1.28, "learning_rate": 1.0870565045992117e-05, "loss": 75.7008, "step": 3887 }, { "compression_loss": 76.47994995117188, "epoch": 1.28, "learning_rate": 1.0865637319316689e-05, "loss": 76.8476, "step": 3888 }, { "compression_loss": 73.72161102294922, "epoch": 1.28, "learning_rate": 1.0860709592641263e-05, "loss": 74.3302, "step": 3889 }, { "compression_loss": 75.03123474121094, "epoch": 1.28, "learning_rate": 1.0855781865965834e-05, "loss": 75.2952, "step": 3890 }, { "compression_loss": 76.34953308105469, "epoch": 1.28, "learning_rate": 1.0850854139290406e-05, "loss": 76.7458, "step": 3891 }, { "compression_loss": 75.67123413085938, "epoch": 1.28, "learning_rate": 1.0845926412614982e-05, "loss": 76.0346, "step": 3892 }, { "compression_loss": 75.74049377441406, "epoch": 1.28, "learning_rate": 1.0840998685939554e-05, "loss": 75.8484, "step": 3893 }, { "compression_loss": 74.80693054199219, "epoch": 1.28, "learning_rate": 1.0836070959264127e-05, "loss": 75.0978, "step": 3894 }, { "compression_loss": 74.69474792480469, "epoch": 1.28, "learning_rate": 1.08311432325887e-05, "loss": 75.1813, "step": 3895 }, { "compression_loss": 74.8370361328125, "epoch": 1.28, "learning_rate": 1.0826215505913271e-05, "loss": 75.5002, "step": 3896 }, { "compression_loss": 75.18977355957031, "epoch": 1.28, "learning_rate": 1.0821287779237845e-05, "loss": 75.6115, "step": 3897 }, { "compression_loss": 77.45399475097656, "epoch": 1.28, "learning_rate": 1.0816360052562419e-05, "loss": 77.9915, "step": 3898 }, { "compression_loss": 74.57125854492188, "epoch": 1.28, "learning_rate": 1.0811432325886992e-05, "loss": 74.937, "step": 3899 }, { "compression_loss": 77.79161071777344, "epoch": 1.28, "learning_rate": 1.0806504599211564e-05, "loss": 78.1716, "step": 3900 }, { "compression_loss": 75.5859375, "epoch": 1.28, "learning_rate": 1.0801576872536136e-05, "loss": 75.9567, "step": 3901 }, { "compression_loss": 75.91644287109375, "epoch": 1.28, "learning_rate": 1.079664914586071e-05, "loss": 76.3134, "step": 3902 }, { "compression_loss": 75.27316284179688, "epoch": 1.28, "learning_rate": 1.0791721419185282e-05, "loss": 75.6593, "step": 3903 }, { "compression_loss": 76.99832916259766, "epoch": 1.28, "learning_rate": 1.0786793692509857e-05, "loss": 77.2584, "step": 3904 }, { "compression_loss": 73.3730697631836, "epoch": 1.28, "learning_rate": 1.078186596583443e-05, "loss": 73.7881, "step": 3905 }, { "compression_loss": 76.93919372558594, "epoch": 1.28, "learning_rate": 1.0776938239159001e-05, "loss": 77.3908, "step": 3906 }, { "compression_loss": 74.23487091064453, "epoch": 1.28, "learning_rate": 1.0772010512483575e-05, "loss": 75.3032, "step": 3907 }, { "compression_loss": 73.97639465332031, "epoch": 1.28, "learning_rate": 1.0767082785808147e-05, "loss": 74.2424, "step": 3908 }, { "compression_loss": 73.20865631103516, "epoch": 1.28, "learning_rate": 1.076215505913272e-05, "loss": 73.4605, "step": 3909 }, { "compression_loss": 76.27169036865234, "epoch": 1.28, "learning_rate": 1.0757227332457294e-05, "loss": 76.4821, "step": 3910 }, { "compression_loss": 75.3968734741211, "epoch": 1.28, "learning_rate": 1.0752299605781866e-05, "loss": 75.6408, "step": 3911 }, { "compression_loss": 73.6760025024414, "epoch": 1.29, "learning_rate": 1.074737187910644e-05, "loss": 73.8998, "step": 3912 }, { "compression_loss": 77.91999053955078, "epoch": 1.29, "learning_rate": 1.0742444152431012e-05, "loss": 78.1979, "step": 3913 }, { "compression_loss": 77.02556610107422, "epoch": 1.29, "learning_rate": 1.0737516425755585e-05, "loss": 77.4438, "step": 3914 }, { "compression_loss": 76.88177490234375, "epoch": 1.29, "learning_rate": 1.0732588699080157e-05, "loss": 77.1452, "step": 3915 }, { "compression_loss": 75.42381286621094, "epoch": 1.29, "learning_rate": 1.0727660972404731e-05, "loss": 75.6656, "step": 3916 }, { "compression_loss": 75.42378997802734, "epoch": 1.29, "learning_rate": 1.0722733245729305e-05, "loss": 75.7727, "step": 3917 }, { "compression_loss": 77.27056121826172, "epoch": 1.29, "learning_rate": 1.0717805519053877e-05, "loss": 77.6501, "step": 3918 }, { "compression_loss": 75.76752471923828, "epoch": 1.29, "learning_rate": 1.071287779237845e-05, "loss": 75.9885, "step": 3919 }, { "compression_loss": 73.28036499023438, "epoch": 1.29, "learning_rate": 1.0707950065703022e-05, "loss": 73.4944, "step": 3920 }, { "compression_loss": 73.46316528320312, "epoch": 1.29, "learning_rate": 1.0703022339027594e-05, "loss": 73.906, "step": 3921 }, { "compression_loss": 75.40270233154297, "epoch": 1.29, "learning_rate": 1.069809461235217e-05, "loss": 75.9552, "step": 3922 }, { "compression_loss": 75.70092010498047, "epoch": 1.29, "learning_rate": 1.0693166885676742e-05, "loss": 75.9116, "step": 3923 }, { "compression_loss": 73.6670150756836, "epoch": 1.29, "learning_rate": 1.0688239159001315e-05, "loss": 73.9257, "step": 3924 }, { "compression_loss": 77.74127197265625, "epoch": 1.29, "learning_rate": 1.0683311432325887e-05, "loss": 78.1635, "step": 3925 }, { "compression_loss": 77.83592987060547, "epoch": 1.29, "learning_rate": 1.067838370565046e-05, "loss": 78.4109, "step": 3926 }, { "compression_loss": 78.14009857177734, "epoch": 1.29, "learning_rate": 1.0673455978975033e-05, "loss": 78.7572, "step": 3927 }, { "compression_loss": 74.73712158203125, "epoch": 1.29, "learning_rate": 1.0668528252299607e-05, "loss": 75.2188, "step": 3928 }, { "compression_loss": 75.61186981201172, "epoch": 1.29, "learning_rate": 1.066360052562418e-05, "loss": 75.8906, "step": 3929 }, { "compression_loss": 77.41864013671875, "epoch": 1.29, "learning_rate": 1.0658672798948752e-05, "loss": 77.6742, "step": 3930 }, { "compression_loss": 74.76347351074219, "epoch": 1.29, "learning_rate": 1.0653745072273324e-05, "loss": 75.2115, "step": 3931 }, { "compression_loss": 77.4631118774414, "epoch": 1.29, "learning_rate": 1.0648817345597898e-05, "loss": 78.0402, "step": 3932 }, { "compression_loss": 76.453125, "epoch": 1.29, "learning_rate": 1.064388961892247e-05, "loss": 76.9513, "step": 3933 }, { "compression_loss": 75.0377426147461, "epoch": 1.29, "learning_rate": 1.0638961892247045e-05, "loss": 75.4501, "step": 3934 }, { "compression_loss": 77.46382141113281, "epoch": 1.29, "learning_rate": 1.0634034165571617e-05, "loss": 77.8107, "step": 3935 }, { "compression_loss": 76.6062240600586, "epoch": 1.29, "learning_rate": 1.0629106438896189e-05, "loss": 76.8887, "step": 3936 }, { "compression_loss": 76.74871826171875, "epoch": 1.29, "learning_rate": 1.0624178712220763e-05, "loss": 77.0104, "step": 3937 }, { "compression_loss": 74.80142974853516, "epoch": 1.29, "learning_rate": 1.0619250985545335e-05, "loss": 75.1612, "step": 3938 }, { "compression_loss": 75.96575927734375, "epoch": 1.29, "learning_rate": 1.0614323258869908e-05, "loss": 76.441, "step": 3939 }, { "compression_loss": 75.87213134765625, "epoch": 1.29, "learning_rate": 1.0609395532194482e-05, "loss": 76.0705, "step": 3940 }, { "compression_loss": 74.21844482421875, "epoch": 1.29, "learning_rate": 1.0604467805519054e-05, "loss": 74.6638, "step": 3941 }, { "compression_loss": 74.23318481445312, "epoch": 1.3, "learning_rate": 1.0599540078843628e-05, "loss": 74.4115, "step": 3942 }, { "compression_loss": 75.26943969726562, "epoch": 1.3, "learning_rate": 1.05946123521682e-05, "loss": 75.6292, "step": 3943 }, { "compression_loss": 78.48323059082031, "epoch": 1.3, "learning_rate": 1.0589684625492773e-05, "loss": 78.9398, "step": 3944 }, { "compression_loss": 76.28792572021484, "epoch": 1.3, "learning_rate": 1.0584756898817345e-05, "loss": 76.535, "step": 3945 }, { "compression_loss": 74.28610229492188, "epoch": 1.3, "learning_rate": 1.0579829172141919e-05, "loss": 74.7934, "step": 3946 }, { "compression_loss": 76.40805053710938, "epoch": 1.3, "learning_rate": 1.0574901445466493e-05, "loss": 76.5687, "step": 3947 }, { "compression_loss": 73.5978775024414, "epoch": 1.3, "learning_rate": 1.0569973718791065e-05, "loss": 74.185, "step": 3948 }, { "compression_loss": 72.70531463623047, "epoch": 1.3, "learning_rate": 1.0565045992115638e-05, "loss": 72.8397, "step": 3949 }, { "compression_loss": 76.46051025390625, "epoch": 1.3, "learning_rate": 1.056011826544021e-05, "loss": 76.8266, "step": 3950 }, { "compression_loss": 73.64279174804688, "epoch": 1.3, "learning_rate": 1.0555190538764782e-05, "loss": 73.8337, "step": 3951 }, { "compression_loss": 76.12033081054688, "epoch": 1.3, "learning_rate": 1.0550262812089358e-05, "loss": 76.357, "step": 3952 }, { "compression_loss": 75.43276977539062, "epoch": 1.3, "learning_rate": 1.054533508541393e-05, "loss": 76.1351, "step": 3953 }, { "compression_loss": 76.63168334960938, "epoch": 1.3, "learning_rate": 1.0540407358738503e-05, "loss": 77.1701, "step": 3954 }, { "compression_loss": 78.41741943359375, "epoch": 1.3, "learning_rate": 1.0535479632063075e-05, "loss": 78.6785, "step": 3955 }, { "compression_loss": 74.92242431640625, "epoch": 1.3, "learning_rate": 1.0530551905387647e-05, "loss": 75.1847, "step": 3956 }, { "compression_loss": 76.05119323730469, "epoch": 1.3, "learning_rate": 1.052562417871222e-05, "loss": 76.4442, "step": 3957 }, { "compression_loss": 75.0548324584961, "epoch": 1.3, "learning_rate": 1.0520696452036794e-05, "loss": 75.3553, "step": 3958 }, { "compression_loss": 73.65679168701172, "epoch": 1.3, "learning_rate": 1.0515768725361368e-05, "loss": 73.9165, "step": 3959 }, { "compression_loss": 76.07038116455078, "epoch": 1.3, "learning_rate": 1.051084099868594e-05, "loss": 76.5969, "step": 3960 }, { "compression_loss": 75.03651428222656, "epoch": 1.3, "learning_rate": 1.0505913272010512e-05, "loss": 75.3481, "step": 3961 }, { "compression_loss": 76.30220031738281, "epoch": 1.3, "learning_rate": 1.0500985545335086e-05, "loss": 76.6751, "step": 3962 }, { "compression_loss": 74.92843627929688, "epoch": 1.3, "learning_rate": 1.0496057818659658e-05, "loss": 75.4693, "step": 3963 }, { "compression_loss": 74.739013671875, "epoch": 1.3, "learning_rate": 1.0491130091984233e-05, "loss": 75.1683, "step": 3964 }, { "compression_loss": 74.25260925292969, "epoch": 1.3, "learning_rate": 1.0486202365308805e-05, "loss": 74.5279, "step": 3965 }, { "compression_loss": 74.625732421875, "epoch": 1.3, "learning_rate": 1.0481274638633377e-05, "loss": 74.9315, "step": 3966 }, { "compression_loss": 74.10700988769531, "epoch": 1.3, "learning_rate": 1.047634691195795e-05, "loss": 74.39, "step": 3967 }, { "compression_loss": 75.97489929199219, "epoch": 1.3, "learning_rate": 1.0471419185282523e-05, "loss": 76.3342, "step": 3968 }, { "compression_loss": 73.80965423583984, "epoch": 1.3, "learning_rate": 1.0466491458607096e-05, "loss": 74.0256, "step": 3969 }, { "compression_loss": 77.33919525146484, "epoch": 1.3, "learning_rate": 1.046156373193167e-05, "loss": 77.8503, "step": 3970 }, { "compression_loss": 74.83780670166016, "epoch": 1.3, "learning_rate": 1.0456636005256242e-05, "loss": 75.1305, "step": 3971 }, { "compression_loss": 74.18309020996094, "epoch": 1.3, "learning_rate": 1.0451708278580816e-05, "loss": 74.4954, "step": 3972 }, { "compression_loss": 75.02490997314453, "epoch": 1.31, "learning_rate": 1.0446780551905388e-05, "loss": 75.33, "step": 3973 }, { "compression_loss": 75.15876770019531, "epoch": 1.31, "learning_rate": 1.0441852825229961e-05, "loss": 75.4434, "step": 3974 }, { "compression_loss": 73.92803955078125, "epoch": 1.31, "learning_rate": 1.0436925098554533e-05, "loss": 74.2539, "step": 3975 }, { "compression_loss": 77.51769256591797, "epoch": 1.31, "learning_rate": 1.0431997371879107e-05, "loss": 77.959, "step": 3976 }, { "compression_loss": 75.35803985595703, "epoch": 1.31, "learning_rate": 1.042706964520368e-05, "loss": 75.6495, "step": 3977 }, { "compression_loss": 75.91590118408203, "epoch": 1.31, "learning_rate": 1.0422141918528252e-05, "loss": 76.1899, "step": 3978 }, { "compression_loss": 76.42427062988281, "epoch": 1.31, "learning_rate": 1.0417214191852826e-05, "loss": 76.6816, "step": 3979 }, { "compression_loss": 73.15138244628906, "epoch": 1.31, "learning_rate": 1.0412286465177398e-05, "loss": 73.4853, "step": 3980 }, { "compression_loss": 76.92990112304688, "epoch": 1.31, "learning_rate": 1.040735873850197e-05, "loss": 77.5137, "step": 3981 }, { "compression_loss": 74.68722534179688, "epoch": 1.31, "learning_rate": 1.0402431011826545e-05, "loss": 74.8286, "step": 3982 }, { "compression_loss": 74.33956909179688, "epoch": 1.31, "learning_rate": 1.0397503285151117e-05, "loss": 74.873, "step": 3983 }, { "compression_loss": 75.48605346679688, "epoch": 1.31, "learning_rate": 1.0392575558475691e-05, "loss": 75.7588, "step": 3984 }, { "compression_loss": 75.39753723144531, "epoch": 1.31, "learning_rate": 1.0387647831800263e-05, "loss": 75.6699, "step": 3985 }, { "compression_loss": 72.29623413085938, "epoch": 1.31, "learning_rate": 1.0382720105124835e-05, "loss": 72.5538, "step": 3986 }, { "compression_loss": 73.9358901977539, "epoch": 1.31, "learning_rate": 1.0377792378449409e-05, "loss": 74.2305, "step": 3987 }, { "compression_loss": 76.19906616210938, "epoch": 1.31, "learning_rate": 1.0372864651773982e-05, "loss": 76.5026, "step": 3988 }, { "compression_loss": 77.89077758789062, "epoch": 1.31, "learning_rate": 1.0367936925098556e-05, "loss": 78.7229, "step": 3989 }, { "compression_loss": 76.78965759277344, "epoch": 1.31, "learning_rate": 1.0363009198423128e-05, "loss": 77.3985, "step": 3990 }, { "compression_loss": 74.20481872558594, "epoch": 1.31, "learning_rate": 1.03580814717477e-05, "loss": 74.3655, "step": 3991 }, { "compression_loss": 79.3805923461914, "epoch": 1.31, "learning_rate": 1.0353153745072274e-05, "loss": 79.6783, "step": 3992 }, { "compression_loss": 74.14088439941406, "epoch": 1.31, "learning_rate": 1.0348226018396846e-05, "loss": 74.4369, "step": 3993 }, { "compression_loss": 76.93157196044922, "epoch": 1.31, "learning_rate": 1.0343298291721421e-05, "loss": 77.5743, "step": 3994 }, { "compression_loss": 75.12550354003906, "epoch": 1.31, "learning_rate": 1.0338370565045993e-05, "loss": 75.3401, "step": 3995 }, { "compression_loss": 74.73783874511719, "epoch": 1.31, "learning_rate": 1.0333442838370565e-05, "loss": 75.0543, "step": 3996 }, { "compression_loss": 76.18977355957031, "epoch": 1.31, "learning_rate": 1.0328515111695139e-05, "loss": 76.3857, "step": 3997 }, { "compression_loss": 75.57652282714844, "epoch": 1.31, "learning_rate": 1.032358738501971e-05, "loss": 75.7881, "step": 3998 }, { "compression_loss": 75.62384033203125, "epoch": 1.31, "learning_rate": 1.0318659658344284e-05, "loss": 76.2158, "step": 3999 }, { "compression_loss": 77.02841186523438, "epoch": 1.31, "learning_rate": 1.0313731931668858e-05, "loss": 77.5862, "step": 4000 }, { "epoch": 1.31, "eval_exact_match": 86.75496688741723, "eval_f1": 92.86192972735914, "step": 4000 }, { "compression_loss": 74.90296936035156, "epoch": 1.31, "learning_rate": 1.030880420499343e-05, "loss": 75.2079, "step": 4001 }, { "compression_loss": 76.11772918701172, "epoch": 1.31, "learning_rate": 1.0303876478318003e-05, "loss": 76.3604, "step": 4002 }, { "compression_loss": 74.33897399902344, "epoch": 1.32, "learning_rate": 1.0298948751642575e-05, "loss": 74.688, "step": 4003 }, { "compression_loss": 75.43040466308594, "epoch": 1.32, "learning_rate": 1.0294021024967147e-05, "loss": 75.6914, "step": 4004 }, { "compression_loss": 72.4140625, "epoch": 1.32, "learning_rate": 1.0289093298291721e-05, "loss": 72.6496, "step": 4005 }, { "compression_loss": 76.56087493896484, "epoch": 1.32, "learning_rate": 1.0284165571616295e-05, "loss": 77.0165, "step": 4006 }, { "compression_loss": 74.91424560546875, "epoch": 1.32, "learning_rate": 1.0279237844940868e-05, "loss": 75.1868, "step": 4007 }, { "compression_loss": 74.58271789550781, "epoch": 1.32, "learning_rate": 1.027431011826544e-05, "loss": 75.1459, "step": 4008 }, { "compression_loss": 73.59942626953125, "epoch": 1.32, "learning_rate": 1.0269382391590012e-05, "loss": 73.9292, "step": 4009 }, { "compression_loss": 73.31763458251953, "epoch": 1.32, "learning_rate": 1.0264454664914586e-05, "loss": 73.6362, "step": 4010 }, { "compression_loss": 75.63191223144531, "epoch": 1.32, "learning_rate": 1.0259526938239158e-05, "loss": 76.1869, "step": 4011 }, { "compression_loss": 77.04501342773438, "epoch": 1.32, "learning_rate": 1.0254599211563733e-05, "loss": 77.4056, "step": 4012 }, { "compression_loss": 76.40419006347656, "epoch": 1.32, "learning_rate": 1.0249671484888305e-05, "loss": 76.557, "step": 4013 }, { "compression_loss": 77.00628662109375, "epoch": 1.32, "learning_rate": 1.0244743758212877e-05, "loss": 77.5958, "step": 4014 }, { "compression_loss": 74.48152160644531, "epoch": 1.32, "learning_rate": 1.0239816031537451e-05, "loss": 74.9133, "step": 4015 }, { "compression_loss": 75.99476623535156, "epoch": 1.32, "learning_rate": 1.0234888304862023e-05, "loss": 76.3377, "step": 4016 }, { "compression_loss": 77.11761474609375, "epoch": 1.32, "learning_rate": 1.0229960578186597e-05, "loss": 77.6319, "step": 4017 }, { "compression_loss": 73.60375213623047, "epoch": 1.32, "learning_rate": 1.022503285151117e-05, "loss": 74.0108, "step": 4018 }, { "compression_loss": 75.14248657226562, "epoch": 1.32, "learning_rate": 1.0220105124835742e-05, "loss": 75.5635, "step": 4019 }, { "compression_loss": 75.8639144897461, "epoch": 1.32, "learning_rate": 1.0215177398160316e-05, "loss": 76.1378, "step": 4020 }, { "compression_loss": 77.44318389892578, "epoch": 1.32, "learning_rate": 1.0210249671484888e-05, "loss": 77.7094, "step": 4021 }, { "compression_loss": 77.03965759277344, "epoch": 1.32, "learning_rate": 1.0205321944809461e-05, "loss": 77.2292, "step": 4022 }, { "compression_loss": 76.65897369384766, "epoch": 1.32, "learning_rate": 1.0200394218134035e-05, "loss": 77.1645, "step": 4023 }, { "compression_loss": 78.09188842773438, "epoch": 1.32, "learning_rate": 1.0195466491458607e-05, "loss": 78.4932, "step": 4024 }, { "compression_loss": 78.27638244628906, "epoch": 1.32, "learning_rate": 1.019053876478318e-05, "loss": 78.84, "step": 4025 }, { "compression_loss": 72.64642333984375, "epoch": 1.32, "learning_rate": 1.0185611038107753e-05, "loss": 72.969, "step": 4026 }, { "compression_loss": 77.49418640136719, "epoch": 1.32, "learning_rate": 1.0180683311432326e-05, "loss": 77.873, "step": 4027 }, { "compression_loss": 76.11045837402344, "epoch": 1.32, "learning_rate": 1.0175755584756898e-05, "loss": 76.5032, "step": 4028 }, { "compression_loss": 76.98333740234375, "epoch": 1.32, "learning_rate": 1.0170827858081472e-05, "loss": 77.3127, "step": 4029 }, { "compression_loss": 75.09538269042969, "epoch": 1.32, "learning_rate": 1.0165900131406046e-05, "loss": 75.3519, "step": 4030 }, { "compression_loss": 76.8856201171875, "epoch": 1.32, "learning_rate": 1.0160972404730618e-05, "loss": 77.5165, "step": 4031 }, { "compression_loss": 76.54971313476562, "epoch": 1.32, "learning_rate": 1.0156044678055191e-05, "loss": 76.8403, "step": 4032 }, { "compression_loss": 73.4167251586914, "epoch": 1.32, "learning_rate": 1.0151116951379763e-05, "loss": 73.7543, "step": 4033 }, { "compression_loss": 76.97528076171875, "epoch": 1.33, "learning_rate": 1.0146189224704335e-05, "loss": 77.4448, "step": 4034 }, { "compression_loss": 76.52345275878906, "epoch": 1.33, "learning_rate": 1.014126149802891e-05, "loss": 77.0395, "step": 4035 }, { "compression_loss": 74.93978881835938, "epoch": 1.33, "learning_rate": 1.0136333771353483e-05, "loss": 75.5468, "step": 4036 }, { "compression_loss": 75.99075317382812, "epoch": 1.33, "learning_rate": 1.0131406044678056e-05, "loss": 76.3972, "step": 4037 }, { "compression_loss": 76.75952911376953, "epoch": 1.33, "learning_rate": 1.0126478318002628e-05, "loss": 77.1641, "step": 4038 }, { "compression_loss": 77.24048614501953, "epoch": 1.33, "learning_rate": 1.01215505913272e-05, "loss": 77.4886, "step": 4039 }, { "compression_loss": 77.24729919433594, "epoch": 1.33, "learning_rate": 1.0116622864651774e-05, "loss": 77.5998, "step": 4040 }, { "compression_loss": 75.88380432128906, "epoch": 1.33, "learning_rate": 1.0111695137976347e-05, "loss": 76.1955, "step": 4041 }, { "compression_loss": 75.94196319580078, "epoch": 1.33, "learning_rate": 1.0106767411300921e-05, "loss": 76.2022, "step": 4042 }, { "compression_loss": 75.10679626464844, "epoch": 1.33, "learning_rate": 1.0101839684625493e-05, "loss": 75.4449, "step": 4043 }, { "compression_loss": 77.26696014404297, "epoch": 1.33, "learning_rate": 1.0096911957950065e-05, "loss": 77.6006, "step": 4044 }, { "compression_loss": 74.95943450927734, "epoch": 1.33, "learning_rate": 1.0091984231274639e-05, "loss": 75.3254, "step": 4045 }, { "compression_loss": 75.10321044921875, "epoch": 1.33, "learning_rate": 1.008705650459921e-05, "loss": 75.3306, "step": 4046 }, { "compression_loss": 74.21595001220703, "epoch": 1.33, "learning_rate": 1.0082128777923786e-05, "loss": 74.479, "step": 4047 }, { "compression_loss": 75.49917602539062, "epoch": 1.33, "learning_rate": 1.0077201051248358e-05, "loss": 75.8447, "step": 4048 }, { "compression_loss": 76.51348114013672, "epoch": 1.33, "learning_rate": 1.007227332457293e-05, "loss": 76.8725, "step": 4049 }, { "compression_loss": 74.74662780761719, "epoch": 1.33, "learning_rate": 1.0067345597897504e-05, "loss": 75.0017, "step": 4050 }, { "compression_loss": 74.87127685546875, "epoch": 1.33, "learning_rate": 1.0062417871222076e-05, "loss": 75.0225, "step": 4051 }, { "compression_loss": 78.29646301269531, "epoch": 1.33, "learning_rate": 1.005749014454665e-05, "loss": 78.6044, "step": 4052 }, { "compression_loss": 73.11892700195312, "epoch": 1.33, "learning_rate": 1.0052562417871223e-05, "loss": 73.4512, "step": 4053 }, { "compression_loss": 74.73112487792969, "epoch": 1.33, "learning_rate": 1.0047634691195795e-05, "loss": 75.2595, "step": 4054 }, { "compression_loss": 74.202392578125, "epoch": 1.33, "learning_rate": 1.0042706964520369e-05, "loss": 74.4174, "step": 4055 }, { "compression_loss": 75.56712341308594, "epoch": 1.33, "learning_rate": 1.003777923784494e-05, "loss": 75.8173, "step": 4056 }, { "compression_loss": 74.31686401367188, "epoch": 1.33, "learning_rate": 1.0032851511169514e-05, "loss": 74.6334, "step": 4057 }, { "compression_loss": 74.16253662109375, "epoch": 1.33, "learning_rate": 1.0027923784494086e-05, "loss": 74.4899, "step": 4058 }, { "compression_loss": 74.92500305175781, "epoch": 1.33, "learning_rate": 1.002299605781866e-05, "loss": 75.0529, "step": 4059 }, { "compression_loss": 75.66758728027344, "epoch": 1.33, "learning_rate": 1.0018068331143234e-05, "loss": 75.9108, "step": 4060 }, { "compression_loss": 75.07440185546875, "epoch": 1.33, "learning_rate": 1.0013140604467806e-05, "loss": 75.6609, "step": 4061 }, { "compression_loss": 75.76335906982422, "epoch": 1.33, "learning_rate": 1.000821287779238e-05, "loss": 76.0576, "step": 4062 }, { "compression_loss": 75.80149841308594, "epoch": 1.33, "learning_rate": 1.0003285151116951e-05, "loss": 76.131, "step": 4063 }, { "compression_loss": 77.10747528076172, "epoch": 1.34, "learning_rate": 9.998357424441523e-06, "loss": 77.4534, "step": 4064 }, { "compression_loss": 74.24429321289062, "epoch": 1.34, "learning_rate": 9.993429697766098e-06, "loss": 74.4516, "step": 4065 }, { "compression_loss": 75.52983093261719, "epoch": 1.34, "learning_rate": 9.98850197109067e-06, "loss": 75.7902, "step": 4066 }, { "compression_loss": 76.67770385742188, "epoch": 1.34, "learning_rate": 9.983574244415244e-06, "loss": 77.0161, "step": 4067 }, { "compression_loss": 74.72583770751953, "epoch": 1.34, "learning_rate": 9.978646517739816e-06, "loss": 75.4586, "step": 4068 }, { "compression_loss": 74.74270629882812, "epoch": 1.34, "learning_rate": 9.973718791064388e-06, "loss": 75.4861, "step": 4069 }, { "compression_loss": 75.87886047363281, "epoch": 1.34, "learning_rate": 9.968791064388962e-06, "loss": 76.315, "step": 4070 }, { "compression_loss": 75.43717956542969, "epoch": 1.34, "learning_rate": 9.963863337713535e-06, "loss": 75.6431, "step": 4071 }, { "compression_loss": 73.17573547363281, "epoch": 1.34, "learning_rate": 9.958935611038109e-06, "loss": 73.4612, "step": 4072 }, { "compression_loss": 77.53068542480469, "epoch": 1.34, "learning_rate": 9.954007884362681e-06, "loss": 78.195, "step": 4073 }, { "compression_loss": 75.44953918457031, "epoch": 1.34, "learning_rate": 9.949080157687253e-06, "loss": 75.7138, "step": 4074 }, { "compression_loss": 78.35479736328125, "epoch": 1.34, "learning_rate": 9.944152431011827e-06, "loss": 78.7096, "step": 4075 }, { "compression_loss": 75.35833740234375, "epoch": 1.34, "learning_rate": 9.939224704336399e-06, "loss": 75.9569, "step": 4076 }, { "compression_loss": 76.05143737792969, "epoch": 1.34, "learning_rate": 9.934296977660974e-06, "loss": 76.3709, "step": 4077 }, { "compression_loss": 76.65362548828125, "epoch": 1.34, "learning_rate": 9.929369250985546e-06, "loss": 77.0171, "step": 4078 }, { "compression_loss": 77.12458038330078, "epoch": 1.34, "learning_rate": 9.924441524310118e-06, "loss": 77.3276, "step": 4079 }, { "compression_loss": 75.77854919433594, "epoch": 1.34, "learning_rate": 9.919513797634692e-06, "loss": 76.0922, "step": 4080 }, { "compression_loss": 72.71815490722656, "epoch": 1.34, "learning_rate": 9.914586070959264e-06, "loss": 73.4577, "step": 4081 }, { "compression_loss": 78.50361633300781, "epoch": 1.34, "learning_rate": 9.909658344283837e-06, "loss": 78.7769, "step": 4082 }, { "compression_loss": 74.1322021484375, "epoch": 1.34, "learning_rate": 9.904730617608411e-06, "loss": 74.4729, "step": 4083 }, { "compression_loss": 74.20369720458984, "epoch": 1.34, "learning_rate": 9.899802890932983e-06, "loss": 74.6814, "step": 4084 }, { "compression_loss": 75.47637939453125, "epoch": 1.34, "learning_rate": 9.894875164257556e-06, "loss": 75.8182, "step": 4085 }, { "compression_loss": 74.46490478515625, "epoch": 1.34, "learning_rate": 9.889947437582128e-06, "loss": 74.7718, "step": 4086 }, { "compression_loss": 74.58992767333984, "epoch": 1.34, "learning_rate": 9.885019710906702e-06, "loss": 75.1902, "step": 4087 }, { "compression_loss": 73.75533294677734, "epoch": 1.34, "learning_rate": 9.880091984231274e-06, "loss": 74.0632, "step": 4088 }, { "compression_loss": 75.3575439453125, "epoch": 1.34, "learning_rate": 9.875164257555848e-06, "loss": 75.7884, "step": 4089 }, { "compression_loss": 77.15031433105469, "epoch": 1.34, "learning_rate": 9.870236530880421e-06, "loss": 77.9897, "step": 4090 }, { "compression_loss": 74.24691772460938, "epoch": 1.34, "learning_rate": 9.865308804204993e-06, "loss": 74.866, "step": 4091 }, { "compression_loss": 77.6796646118164, "epoch": 1.34, "learning_rate": 9.860381077529567e-06, "loss": 77.8777, "step": 4092 }, { "compression_loss": 74.3775634765625, "epoch": 1.34, "learning_rate": 9.855453350854139e-06, "loss": 75.1014, "step": 4093 }, { "compression_loss": 75.48031616210938, "epoch": 1.34, "learning_rate": 9.850525624178711e-06, "loss": 75.7978, "step": 4094 }, { "compression_loss": 77.23861694335938, "epoch": 1.35, "learning_rate": 9.845597897503286e-06, "loss": 77.6835, "step": 4095 }, { "compression_loss": 73.51698303222656, "epoch": 1.35, "learning_rate": 9.840670170827858e-06, "loss": 73.8274, "step": 4096 }, { "compression_loss": 72.6375961303711, "epoch": 1.35, "learning_rate": 9.835742444152432e-06, "loss": 73.0295, "step": 4097 }, { "compression_loss": 75.48787689208984, "epoch": 1.35, "learning_rate": 9.830814717477004e-06, "loss": 75.8691, "step": 4098 }, { "compression_loss": 73.74092102050781, "epoch": 1.35, "learning_rate": 9.825886990801576e-06, "loss": 74.2328, "step": 4099 }, { "compression_loss": 75.47211456298828, "epoch": 1.35, "learning_rate": 9.82095926412615e-06, "loss": 76.0077, "step": 4100 }, { "compression_loss": 74.74520111083984, "epoch": 1.35, "learning_rate": 9.816031537450723e-06, "loss": 75.0237, "step": 4101 }, { "compression_loss": 77.96798706054688, "epoch": 1.35, "learning_rate": 9.811103810775297e-06, "loss": 78.5981, "step": 4102 }, { "compression_loss": 75.74700164794922, "epoch": 1.35, "learning_rate": 9.806176084099869e-06, "loss": 76.1886, "step": 4103 }, { "compression_loss": 73.06710815429688, "epoch": 1.35, "learning_rate": 9.80124835742444e-06, "loss": 73.3327, "step": 4104 }, { "compression_loss": 76.67642211914062, "epoch": 1.35, "learning_rate": 9.796320630749015e-06, "loss": 77.0747, "step": 4105 }, { "compression_loss": 76.64073181152344, "epoch": 1.35, "learning_rate": 9.791392904073586e-06, "loss": 77.055, "step": 4106 }, { "compression_loss": 73.01336669921875, "epoch": 1.35, "learning_rate": 9.786465177398162e-06, "loss": 73.3688, "step": 4107 }, { "compression_loss": 75.88108825683594, "epoch": 1.35, "learning_rate": 9.781537450722734e-06, "loss": 76.3806, "step": 4108 }, { "compression_loss": 77.53547668457031, "epoch": 1.35, "learning_rate": 9.776609724047306e-06, "loss": 78.2297, "step": 4109 }, { "compression_loss": 74.3563003540039, "epoch": 1.35, "learning_rate": 9.77168199737188e-06, "loss": 74.7138, "step": 4110 }, { "compression_loss": 77.1269302368164, "epoch": 1.35, "learning_rate": 9.766754270696451e-06, "loss": 77.441, "step": 4111 }, { "compression_loss": 75.76234436035156, "epoch": 1.35, "learning_rate": 9.761826544021025e-06, "loss": 76.1536, "step": 4112 }, { "compression_loss": 75.9363784790039, "epoch": 1.35, "learning_rate": 9.756898817345599e-06, "loss": 76.2261, "step": 4113 }, { "compression_loss": 75.19779968261719, "epoch": 1.35, "learning_rate": 9.75197109067017e-06, "loss": 75.8573, "step": 4114 }, { "compression_loss": 77.36675262451172, "epoch": 1.35, "learning_rate": 9.747043363994744e-06, "loss": 77.9095, "step": 4115 }, { "compression_loss": 76.19081115722656, "epoch": 1.35, "learning_rate": 9.742115637319316e-06, "loss": 76.4104, "step": 4116 }, { "compression_loss": 73.66929626464844, "epoch": 1.35, "learning_rate": 9.73718791064389e-06, "loss": 73.9702, "step": 4117 }, { "compression_loss": 74.08853149414062, "epoch": 1.35, "learning_rate": 9.732260183968462e-06, "loss": 74.5568, "step": 4118 }, { "compression_loss": 76.12289428710938, "epoch": 1.35, "learning_rate": 9.727332457293036e-06, "loss": 76.6972, "step": 4119 }, { "compression_loss": 76.4320068359375, "epoch": 1.35, "learning_rate": 9.72240473061761e-06, "loss": 76.9392, "step": 4120 }, { "compression_loss": 73.44463348388672, "epoch": 1.35, "learning_rate": 9.717477003942181e-06, "loss": 73.7025, "step": 4121 }, { "compression_loss": 76.55274963378906, "epoch": 1.35, "learning_rate": 9.712549277266755e-06, "loss": 76.9487, "step": 4122 }, { "compression_loss": 75.37535095214844, "epoch": 1.35, "learning_rate": 9.707621550591327e-06, "loss": 75.684, "step": 4123 }, { "compression_loss": 73.93405151367188, "epoch": 1.35, "learning_rate": 9.702693823915899e-06, "loss": 74.2263, "step": 4124 }, { "compression_loss": 72.20289611816406, "epoch": 1.36, "learning_rate": 9.697766097240474e-06, "loss": 72.4649, "step": 4125 }, { "compression_loss": 75.07551574707031, "epoch": 1.36, "learning_rate": 9.692838370565046e-06, "loss": 75.4854, "step": 4126 }, { "compression_loss": 77.12078857421875, "epoch": 1.36, "learning_rate": 9.68791064388962e-06, "loss": 77.474, "step": 4127 }, { "compression_loss": 77.71177673339844, "epoch": 1.36, "learning_rate": 9.682982917214192e-06, "loss": 78.3163, "step": 4128 }, { "compression_loss": 76.78913879394531, "epoch": 1.36, "learning_rate": 9.678055190538764e-06, "loss": 77.4111, "step": 4129 }, { "compression_loss": 74.27027893066406, "epoch": 1.36, "learning_rate": 9.673127463863337e-06, "loss": 74.6091, "step": 4130 }, { "compression_loss": 75.37120819091797, "epoch": 1.36, "learning_rate": 9.668199737187911e-06, "loss": 75.8904, "step": 4131 }, { "compression_loss": 74.40867614746094, "epoch": 1.36, "learning_rate": 9.663272010512485e-06, "loss": 74.6891, "step": 4132 }, { "compression_loss": 76.66586303710938, "epoch": 1.36, "learning_rate": 9.658344283837057e-06, "loss": 77.3563, "step": 4133 }, { "compression_loss": 74.74101257324219, "epoch": 1.36, "learning_rate": 9.653416557161629e-06, "loss": 75.5562, "step": 4134 }, { "compression_loss": 75.59283447265625, "epoch": 1.36, "learning_rate": 9.648488830486202e-06, "loss": 75.8455, "step": 4135 }, { "compression_loss": 75.6530990600586, "epoch": 1.36, "learning_rate": 9.643561103810774e-06, "loss": 76.0319, "step": 4136 }, { "compression_loss": 74.23977661132812, "epoch": 1.36, "learning_rate": 9.63863337713535e-06, "loss": 74.5001, "step": 4137 }, { "compression_loss": 78.63446044921875, "epoch": 1.36, "learning_rate": 9.633705650459922e-06, "loss": 79.4008, "step": 4138 }, { "compression_loss": 73.68536376953125, "epoch": 1.36, "learning_rate": 9.628777923784494e-06, "loss": 74.0338, "step": 4139 }, { "compression_loss": 75.22941589355469, "epoch": 1.36, "learning_rate": 9.623850197109067e-06, "loss": 75.5944, "step": 4140 }, { "compression_loss": 76.3095932006836, "epoch": 1.36, "learning_rate": 9.61892247043364e-06, "loss": 76.6852, "step": 4141 }, { "compression_loss": 72.41957092285156, "epoch": 1.36, "learning_rate": 9.613994743758213e-06, "loss": 72.7441, "step": 4142 }, { "compression_loss": 75.35433197021484, "epoch": 1.36, "learning_rate": 9.609067017082787e-06, "loss": 75.6819, "step": 4143 }, { "compression_loss": 76.85521697998047, "epoch": 1.36, "learning_rate": 9.604139290407359e-06, "loss": 77.2165, "step": 4144 }, { "compression_loss": 75.76655578613281, "epoch": 1.36, "learning_rate": 9.599211563731932e-06, "loss": 76.095, "step": 4145 }, { "compression_loss": 75.80448150634766, "epoch": 1.36, "learning_rate": 9.594283837056504e-06, "loss": 76.065, "step": 4146 }, { "compression_loss": 78.23475646972656, "epoch": 1.36, "learning_rate": 9.589356110381078e-06, "loss": 78.6415, "step": 4147 }, { "compression_loss": 73.78550720214844, "epoch": 1.36, "learning_rate": 9.58442838370565e-06, "loss": 74.0818, "step": 4148 }, { "compression_loss": 74.99952697753906, "epoch": 1.36, "learning_rate": 9.579500657030223e-06, "loss": 75.3128, "step": 4149 }, { "compression_loss": 74.82718658447266, "epoch": 1.36, "learning_rate": 9.574572930354797e-06, "loss": 75.1969, "step": 4150 }, { "compression_loss": 78.0660171508789, "epoch": 1.36, "learning_rate": 9.569645203679369e-06, "loss": 78.6061, "step": 4151 }, { "compression_loss": 75.73556518554688, "epoch": 1.36, "learning_rate": 9.564717477003943e-06, "loss": 76.0775, "step": 4152 }, { "compression_loss": 75.29975891113281, "epoch": 1.36, "learning_rate": 9.559789750328515e-06, "loss": 75.579, "step": 4153 }, { "compression_loss": 76.05193328857422, "epoch": 1.36, "learning_rate": 9.554862023653088e-06, "loss": 76.2857, "step": 4154 }, { "compression_loss": 75.87814331054688, "epoch": 1.36, "learning_rate": 9.549934296977662e-06, "loss": 76.14, "step": 4155 }, { "compression_loss": 74.66033935546875, "epoch": 1.37, "learning_rate": 9.545006570302234e-06, "loss": 74.9068, "step": 4156 }, { "compression_loss": 77.63726806640625, "epoch": 1.37, "learning_rate": 9.540078843626808e-06, "loss": 78.179, "step": 4157 }, { "compression_loss": 74.9180908203125, "epoch": 1.37, "learning_rate": 9.53515111695138e-06, "loss": 75.3669, "step": 4158 }, { "compression_loss": 75.0244140625, "epoch": 1.37, "learning_rate": 9.530223390275952e-06, "loss": 75.3687, "step": 4159 }, { "compression_loss": 77.95262145996094, "epoch": 1.37, "learning_rate": 9.525295663600527e-06, "loss": 78.3374, "step": 4160 }, { "compression_loss": 74.37466430664062, "epoch": 1.37, "learning_rate": 9.520367936925099e-06, "loss": 74.5124, "step": 4161 }, { "compression_loss": 76.17080688476562, "epoch": 1.37, "learning_rate": 9.515440210249673e-06, "loss": 76.5924, "step": 4162 }, { "compression_loss": 76.2353744506836, "epoch": 1.37, "learning_rate": 9.510512483574245e-06, "loss": 76.4502, "step": 4163 }, { "compression_loss": 75.5520248413086, "epoch": 1.37, "learning_rate": 9.505584756898817e-06, "loss": 75.7368, "step": 4164 }, { "compression_loss": 76.64222717285156, "epoch": 1.37, "learning_rate": 9.50065703022339e-06, "loss": 77.0396, "step": 4165 }, { "compression_loss": 76.80408477783203, "epoch": 1.37, "learning_rate": 9.495729303547964e-06, "loss": 77.2868, "step": 4166 }, { "compression_loss": 75.61824035644531, "epoch": 1.37, "learning_rate": 9.490801576872538e-06, "loss": 75.844, "step": 4167 }, { "compression_loss": 76.17678833007812, "epoch": 1.37, "learning_rate": 9.48587385019711e-06, "loss": 76.4261, "step": 4168 }, { "compression_loss": 74.86189270019531, "epoch": 1.37, "learning_rate": 9.480946123521682e-06, "loss": 75.1812, "step": 4169 }, { "compression_loss": 79.37541961669922, "epoch": 1.37, "learning_rate": 9.476018396846255e-06, "loss": 79.6814, "step": 4170 }, { "compression_loss": 76.86327362060547, "epoch": 1.37, "learning_rate": 9.471090670170827e-06, "loss": 77.2434, "step": 4171 }, { "compression_loss": 77.1970443725586, "epoch": 1.37, "learning_rate": 9.466162943495403e-06, "loss": 77.5735, "step": 4172 }, { "compression_loss": 76.08241271972656, "epoch": 1.37, "learning_rate": 9.461235216819974e-06, "loss": 76.4575, "step": 4173 }, { "compression_loss": 74.85580444335938, "epoch": 1.37, "learning_rate": 9.456307490144546e-06, "loss": 75.1508, "step": 4174 }, { "compression_loss": 75.36466979980469, "epoch": 1.37, "learning_rate": 9.45137976346912e-06, "loss": 76.0599, "step": 4175 }, { "compression_loss": 76.09803009033203, "epoch": 1.37, "learning_rate": 9.446452036793692e-06, "loss": 76.3041, "step": 4176 }, { "compression_loss": 75.11951446533203, "epoch": 1.37, "learning_rate": 9.441524310118266e-06, "loss": 75.4097, "step": 4177 }, { "compression_loss": 76.7960433959961, "epoch": 1.37, "learning_rate": 9.43659658344284e-06, "loss": 77.5362, "step": 4178 }, { "compression_loss": 75.47525024414062, "epoch": 1.37, "learning_rate": 9.431668856767411e-06, "loss": 76.0991, "step": 4179 }, { "compression_loss": 77.13127899169922, "epoch": 1.37, "learning_rate": 9.426741130091985e-06, "loss": 77.4133, "step": 4180 }, { "compression_loss": 77.02019500732422, "epoch": 1.37, "learning_rate": 9.421813403416557e-06, "loss": 77.4842, "step": 4181 }, { "compression_loss": 76.837646484375, "epoch": 1.37, "learning_rate": 9.41688567674113e-06, "loss": 77.1519, "step": 4182 }, { "compression_loss": 77.21278381347656, "epoch": 1.37, "learning_rate": 9.411957950065703e-06, "loss": 77.5115, "step": 4183 }, { "compression_loss": 77.9531478881836, "epoch": 1.37, "learning_rate": 9.407030223390276e-06, "loss": 78.2444, "step": 4184 }, { "compression_loss": 75.91619873046875, "epoch": 1.37, "learning_rate": 9.40210249671485e-06, "loss": 76.294, "step": 4185 }, { "compression_loss": 74.03849792480469, "epoch": 1.38, "learning_rate": 9.397174770039422e-06, "loss": 74.2294, "step": 4186 }, { "compression_loss": 74.03688049316406, "epoch": 1.38, "learning_rate": 9.392247043363996e-06, "loss": 74.2204, "step": 4187 }, { "compression_loss": 77.94082641601562, "epoch": 1.38, "learning_rate": 9.387319316688568e-06, "loss": 78.2534, "step": 4188 }, { "compression_loss": 74.47746276855469, "epoch": 1.38, "learning_rate": 9.38239159001314e-06, "loss": 74.9253, "step": 4189 }, { "compression_loss": 74.30654907226562, "epoch": 1.38, "learning_rate": 9.377463863337715e-06, "loss": 74.6748, "step": 4190 }, { "compression_loss": 76.46552276611328, "epoch": 1.38, "learning_rate": 9.372536136662287e-06, "loss": 76.8269, "step": 4191 }, { "compression_loss": 74.13790893554688, "epoch": 1.38, "learning_rate": 9.36760840998686e-06, "loss": 74.4126, "step": 4192 }, { "compression_loss": 72.40630340576172, "epoch": 1.38, "learning_rate": 9.362680683311432e-06, "loss": 72.6957, "step": 4193 }, { "compression_loss": 75.55635833740234, "epoch": 1.38, "learning_rate": 9.357752956636004e-06, "loss": 75.9684, "step": 4194 }, { "compression_loss": 79.6227035522461, "epoch": 1.38, "learning_rate": 9.352825229960578e-06, "loss": 80.5556, "step": 4195 }, { "compression_loss": 76.72725677490234, "epoch": 1.38, "learning_rate": 9.347897503285152e-06, "loss": 77.2683, "step": 4196 }, { "compression_loss": 74.25408172607422, "epoch": 1.38, "learning_rate": 9.342969776609725e-06, "loss": 74.5617, "step": 4197 }, { "compression_loss": 74.99888610839844, "epoch": 1.38, "learning_rate": 9.338042049934297e-06, "loss": 75.2452, "step": 4198 }, { "compression_loss": 75.9849853515625, "epoch": 1.38, "learning_rate": 9.33311432325887e-06, "loss": 76.2812, "step": 4199 }, { "compression_loss": 74.09384155273438, "epoch": 1.38, "learning_rate": 9.328186596583443e-06, "loss": 74.4321, "step": 4200 }, { "compression_loss": 78.38671875, "epoch": 1.38, "learning_rate": 9.323258869908015e-06, "loss": 78.8431, "step": 4201 }, { "compression_loss": 79.03773498535156, "epoch": 1.38, "learning_rate": 9.31833114323259e-06, "loss": 79.5211, "step": 4202 }, { "compression_loss": 78.40274047851562, "epoch": 1.38, "learning_rate": 9.313403416557162e-06, "loss": 78.6808, "step": 4203 }, { "compression_loss": 75.66395568847656, "epoch": 1.38, "learning_rate": 9.308475689881734e-06, "loss": 75.9033, "step": 4204 }, { "compression_loss": 74.21729278564453, "epoch": 1.38, "learning_rate": 9.303547963206308e-06, "loss": 74.4685, "step": 4205 }, { "compression_loss": 74.94802856445312, "epoch": 1.38, "learning_rate": 9.29862023653088e-06, "loss": 75.3772, "step": 4206 }, { "compression_loss": 78.96295166015625, "epoch": 1.38, "learning_rate": 9.293692509855454e-06, "loss": 79.1816, "step": 4207 }, { "compression_loss": 74.0183334350586, "epoch": 1.38, "learning_rate": 9.288764783180027e-06, "loss": 74.243, "step": 4208 }, { "compression_loss": 73.952392578125, "epoch": 1.38, "learning_rate": 9.2838370565046e-06, "loss": 74.2726, "step": 4209 }, { "compression_loss": 75.96597290039062, "epoch": 1.38, "learning_rate": 9.278909329829173e-06, "loss": 76.4914, "step": 4210 }, { "compression_loss": 73.19717407226562, "epoch": 1.38, "learning_rate": 9.273981603153745e-06, "loss": 73.3138, "step": 4211 }, { "compression_loss": 76.43836975097656, "epoch": 1.38, "learning_rate": 9.269053876478319e-06, "loss": 76.6999, "step": 4212 }, { "compression_loss": 76.8775405883789, "epoch": 1.38, "learning_rate": 9.26412614980289e-06, "loss": 77.7748, "step": 4213 }, { "compression_loss": 74.9063720703125, "epoch": 1.38, "learning_rate": 9.259198423127464e-06, "loss": 75.3222, "step": 4214 }, { "compression_loss": 76.5011978149414, "epoch": 1.38, "learning_rate": 9.254270696452038e-06, "loss": 76.7379, "step": 4215 }, { "compression_loss": 75.05536651611328, "epoch": 1.39, "learning_rate": 9.24934296977661e-06, "loss": 75.3222, "step": 4216 }, { "compression_loss": 74.95208740234375, "epoch": 1.39, "learning_rate": 9.244415243101183e-06, "loss": 75.3684, "step": 4217 }, { "compression_loss": 74.23797607421875, "epoch": 1.39, "learning_rate": 9.239487516425755e-06, "loss": 74.5381, "step": 4218 }, { "compression_loss": 74.37831115722656, "epoch": 1.39, "learning_rate": 9.234559789750327e-06, "loss": 74.6282, "step": 4219 }, { "compression_loss": 76.47624969482422, "epoch": 1.39, "learning_rate": 9.229632063074903e-06, "loss": 76.7413, "step": 4220 }, { "compression_loss": 75.38956451416016, "epoch": 1.39, "learning_rate": 9.224704336399475e-06, "loss": 75.6545, "step": 4221 }, { "compression_loss": 79.36585235595703, "epoch": 1.39, "learning_rate": 9.219776609724048e-06, "loss": 79.9111, "step": 4222 }, { "compression_loss": 75.27477264404297, "epoch": 1.39, "learning_rate": 9.21484888304862e-06, "loss": 75.5352, "step": 4223 }, { "compression_loss": 76.75773620605469, "epoch": 1.39, "learning_rate": 9.209921156373192e-06, "loss": 76.9632, "step": 4224 }, { "compression_loss": 77.70992279052734, "epoch": 1.39, "learning_rate": 9.204993429697766e-06, "loss": 78.2999, "step": 4225 }, { "compression_loss": 75.88034057617188, "epoch": 1.39, "learning_rate": 9.20006570302234e-06, "loss": 76.2907, "step": 4226 }, { "compression_loss": 75.91911315917969, "epoch": 1.39, "learning_rate": 9.195137976346913e-06, "loss": 76.1988, "step": 4227 }, { "compression_loss": 75.92372131347656, "epoch": 1.39, "learning_rate": 9.190210249671485e-06, "loss": 76.2907, "step": 4228 }, { "compression_loss": 75.07179260253906, "epoch": 1.39, "learning_rate": 9.185282522996057e-06, "loss": 75.3458, "step": 4229 }, { "compression_loss": 75.74602508544922, "epoch": 1.39, "learning_rate": 9.180354796320631e-06, "loss": 76.4061, "step": 4230 }, { "compression_loss": 75.17984008789062, "epoch": 1.39, "learning_rate": 9.175427069645203e-06, "loss": 75.4929, "step": 4231 }, { "compression_loss": 75.94752502441406, "epoch": 1.39, "learning_rate": 9.170499342969778e-06, "loss": 76.4929, "step": 4232 }, { "compression_loss": 75.310302734375, "epoch": 1.39, "learning_rate": 9.16557161629435e-06, "loss": 75.6862, "step": 4233 }, { "compression_loss": 75.89241790771484, "epoch": 1.39, "learning_rate": 9.160643889618922e-06, "loss": 76.2562, "step": 4234 }, { "compression_loss": 75.31031799316406, "epoch": 1.39, "learning_rate": 9.155716162943496e-06, "loss": 75.4863, "step": 4235 }, { "compression_loss": 75.13160705566406, "epoch": 1.39, "learning_rate": 9.150788436268068e-06, "loss": 75.4837, "step": 4236 }, { "compression_loss": 75.97694396972656, "epoch": 1.39, "learning_rate": 9.145860709592641e-06, "loss": 76.1126, "step": 4237 }, { "compression_loss": 74.9768295288086, "epoch": 1.39, "learning_rate": 9.140932982917215e-06, "loss": 75.3965, "step": 4238 }, { "compression_loss": 76.30684661865234, "epoch": 1.39, "learning_rate": 9.136005256241787e-06, "loss": 76.9951, "step": 4239 }, { "compression_loss": 73.73506164550781, "epoch": 1.39, "learning_rate": 9.13107752956636e-06, "loss": 74.1095, "step": 4240 }, { "compression_loss": 76.19361877441406, "epoch": 1.39, "learning_rate": 9.126149802890933e-06, "loss": 76.5968, "step": 4241 }, { "compression_loss": 76.41901397705078, "epoch": 1.39, "learning_rate": 9.121222076215506e-06, "loss": 77.0605, "step": 4242 }, { "compression_loss": 75.90937042236328, "epoch": 1.39, "learning_rate": 9.116294349540078e-06, "loss": 76.2392, "step": 4243 }, { "compression_loss": 75.4488754272461, "epoch": 1.39, "learning_rate": 9.111366622864652e-06, "loss": 75.865, "step": 4244 }, { "compression_loss": 73.13139343261719, "epoch": 1.39, "learning_rate": 9.106438896189226e-06, "loss": 73.3513, "step": 4245 }, { "compression_loss": 75.10374450683594, "epoch": 1.39, "learning_rate": 9.101511169513798e-06, "loss": 75.5239, "step": 4246 }, { "compression_loss": 78.39555358886719, "epoch": 1.4, "learning_rate": 9.096583442838371e-06, "loss": 78.9852, "step": 4247 }, { "compression_loss": 75.40971374511719, "epoch": 1.4, "learning_rate": 9.091655716162943e-06, "loss": 76.1589, "step": 4248 }, { "compression_loss": 75.88746643066406, "epoch": 1.4, "learning_rate": 9.086727989487515e-06, "loss": 76.1639, "step": 4249 }, { "compression_loss": 75.39401245117188, "epoch": 1.4, "learning_rate": 9.08180026281209e-06, "loss": 75.655, "step": 4250 }, { "epoch": 1.4, "eval_exact_match": 86.7360454115421, "eval_f1": 92.9469777016182, "step": 4250 }, { "compression_loss": 74.45848846435547, "epoch": 1.4, "learning_rate": 9.076872536136663e-06, "loss": 74.8792, "step": 4251 }, { "compression_loss": 76.88591766357422, "epoch": 1.4, "learning_rate": 9.071944809461236e-06, "loss": 77.1325, "step": 4252 }, { "compression_loss": 77.10183715820312, "epoch": 1.4, "learning_rate": 9.067017082785808e-06, "loss": 77.5103, "step": 4253 }, { "compression_loss": 73.59307861328125, "epoch": 1.4, "learning_rate": 9.06208935611038e-06, "loss": 73.7464, "step": 4254 }, { "compression_loss": 75.86552429199219, "epoch": 1.4, "learning_rate": 9.057161629434954e-06, "loss": 76.1961, "step": 4255 }, { "compression_loss": 75.00886535644531, "epoch": 1.4, "learning_rate": 9.052233902759528e-06, "loss": 75.4438, "step": 4256 }, { "compression_loss": 75.36558532714844, "epoch": 1.4, "learning_rate": 9.047306176084101e-06, "loss": 75.5816, "step": 4257 }, { "compression_loss": 75.4257583618164, "epoch": 1.4, "learning_rate": 9.042378449408673e-06, "loss": 76.1282, "step": 4258 }, { "compression_loss": 72.98800659179688, "epoch": 1.4, "learning_rate": 9.037450722733245e-06, "loss": 73.2665, "step": 4259 }, { "compression_loss": 73.58082580566406, "epoch": 1.4, "learning_rate": 9.032522996057819e-06, "loss": 74.2351, "step": 4260 }, { "compression_loss": 74.5683364868164, "epoch": 1.4, "learning_rate": 9.02759526938239e-06, "loss": 74.9214, "step": 4261 }, { "compression_loss": 74.43685913085938, "epoch": 1.4, "learning_rate": 9.022667542706966e-06, "loss": 74.6456, "step": 4262 }, { "compression_loss": 76.31382751464844, "epoch": 1.4, "learning_rate": 9.017739816031538e-06, "loss": 76.5394, "step": 4263 }, { "compression_loss": 75.64016723632812, "epoch": 1.4, "learning_rate": 9.01281208935611e-06, "loss": 76.2323, "step": 4264 }, { "compression_loss": 75.4041976928711, "epoch": 1.4, "learning_rate": 9.007884362680684e-06, "loss": 75.6891, "step": 4265 }, { "compression_loss": 75.73760986328125, "epoch": 1.4, "learning_rate": 9.002956636005256e-06, "loss": 75.9554, "step": 4266 }, { "compression_loss": 75.10369873046875, "epoch": 1.4, "learning_rate": 8.99802890932983e-06, "loss": 75.5261, "step": 4267 }, { "compression_loss": 75.39616394042969, "epoch": 1.4, "learning_rate": 8.993101182654403e-06, "loss": 75.6399, "step": 4268 }, { "compression_loss": 76.32881927490234, "epoch": 1.4, "learning_rate": 8.988173455978975e-06, "loss": 76.6618, "step": 4269 }, { "compression_loss": 76.25389099121094, "epoch": 1.4, "learning_rate": 8.983245729303549e-06, "loss": 76.4954, "step": 4270 }, { "compression_loss": 76.43524932861328, "epoch": 1.4, "learning_rate": 8.97831800262812e-06, "loss": 76.7562, "step": 4271 }, { "compression_loss": 77.99520874023438, "epoch": 1.4, "learning_rate": 8.973390275952694e-06, "loss": 78.3598, "step": 4272 }, { "compression_loss": 76.37858581542969, "epoch": 1.4, "learning_rate": 8.968462549277266e-06, "loss": 77.1281, "step": 4273 }, { "compression_loss": 75.5953369140625, "epoch": 1.4, "learning_rate": 8.96353482260184e-06, "loss": 76.0163, "step": 4274 }, { "compression_loss": 73.77407836914062, "epoch": 1.4, "learning_rate": 8.958607095926414e-06, "loss": 74.0867, "step": 4275 }, { "compression_loss": 75.60368347167969, "epoch": 1.4, "learning_rate": 8.953679369250986e-06, "loss": 75.9345, "step": 4276 }, { "compression_loss": 75.62161254882812, "epoch": 1.41, "learning_rate": 8.94875164257556e-06, "loss": 76.0875, "step": 4277 }, { "compression_loss": 74.68400573730469, "epoch": 1.41, "learning_rate": 8.943823915900131e-06, "loss": 75.1754, "step": 4278 }, { "compression_loss": 77.0492935180664, "epoch": 1.41, "learning_rate": 8.938896189224703e-06, "loss": 77.3402, "step": 4279 }, { "compression_loss": 76.03877258300781, "epoch": 1.41, "learning_rate": 8.933968462549279e-06, "loss": 76.3638, "step": 4280 }, { "compression_loss": 74.5597915649414, "epoch": 1.41, "learning_rate": 8.92904073587385e-06, "loss": 74.7042, "step": 4281 }, { "compression_loss": 75.44747161865234, "epoch": 1.41, "learning_rate": 8.924113009198424e-06, "loss": 75.8844, "step": 4282 }, { "compression_loss": 77.12153625488281, "epoch": 1.41, "learning_rate": 8.919185282522996e-06, "loss": 77.541, "step": 4283 }, { "compression_loss": 77.87985229492188, "epoch": 1.41, "learning_rate": 8.914257555847568e-06, "loss": 78.2413, "step": 4284 }, { "compression_loss": 75.48198699951172, "epoch": 1.41, "learning_rate": 8.909329829172142e-06, "loss": 75.7513, "step": 4285 }, { "compression_loss": 76.90068054199219, "epoch": 1.41, "learning_rate": 8.904402102496715e-06, "loss": 77.4682, "step": 4286 }, { "compression_loss": 73.89570617675781, "epoch": 1.41, "learning_rate": 8.899474375821289e-06, "loss": 74.0969, "step": 4287 }, { "compression_loss": 73.25718688964844, "epoch": 1.41, "learning_rate": 8.894546649145861e-06, "loss": 73.7552, "step": 4288 }, { "compression_loss": 73.35360717773438, "epoch": 1.41, "learning_rate": 8.889618922470433e-06, "loss": 73.6907, "step": 4289 }, { "compression_loss": 74.55205535888672, "epoch": 1.41, "learning_rate": 8.884691195795007e-06, "loss": 74.9885, "step": 4290 }, { "compression_loss": 74.44450378417969, "epoch": 1.41, "learning_rate": 8.879763469119579e-06, "loss": 74.8693, "step": 4291 }, { "compression_loss": 76.57142639160156, "epoch": 1.41, "learning_rate": 8.874835742444154e-06, "loss": 77.2978, "step": 4292 }, { "compression_loss": 74.4224853515625, "epoch": 1.41, "learning_rate": 8.869908015768726e-06, "loss": 74.704, "step": 4293 }, { "compression_loss": 76.68972778320312, "epoch": 1.41, "learning_rate": 8.864980289093298e-06, "loss": 77.0385, "step": 4294 }, { "compression_loss": 73.9992446899414, "epoch": 1.41, "learning_rate": 8.860052562417872e-06, "loss": 74.3372, "step": 4295 }, { "compression_loss": 73.08070373535156, "epoch": 1.41, "learning_rate": 8.855124835742444e-06, "loss": 73.3189, "step": 4296 }, { "compression_loss": 75.50161743164062, "epoch": 1.41, "learning_rate": 8.850197109067017e-06, "loss": 75.7676, "step": 4297 }, { "compression_loss": 74.54583740234375, "epoch": 1.41, "learning_rate": 8.845269382391591e-06, "loss": 74.9142, "step": 4298 }, { "compression_loss": 73.65679931640625, "epoch": 1.41, "learning_rate": 8.840341655716163e-06, "loss": 73.9444, "step": 4299 }, { "compression_loss": 75.38574981689453, "epoch": 1.41, "learning_rate": 8.835413929040737e-06, "loss": 75.6127, "step": 4300 }, { "compression_loss": 75.72200012207031, "epoch": 1.41, "learning_rate": 8.830486202365308e-06, "loss": 76.0239, "step": 4301 }, { "compression_loss": 78.04254150390625, "epoch": 1.41, "learning_rate": 8.825558475689882e-06, "loss": 78.5833, "step": 4302 }, { "compression_loss": 76.27459716796875, "epoch": 1.41, "learning_rate": 8.820630749014456e-06, "loss": 76.8527, "step": 4303 }, { "compression_loss": 76.17608642578125, "epoch": 1.41, "learning_rate": 8.815703022339028e-06, "loss": 76.486, "step": 4304 }, { "compression_loss": 73.34971618652344, "epoch": 1.41, "learning_rate": 8.810775295663601e-06, "loss": 73.5379, "step": 4305 }, { "compression_loss": 73.6554946899414, "epoch": 1.41, "learning_rate": 8.805847568988173e-06, "loss": 73.9221, "step": 4306 }, { "compression_loss": 74.38163757324219, "epoch": 1.41, "learning_rate": 8.800919842312747e-06, "loss": 74.6672, "step": 4307 }, { "compression_loss": 76.20033264160156, "epoch": 1.42, "learning_rate": 8.795992115637319e-06, "loss": 76.5557, "step": 4308 }, { "compression_loss": 75.671142578125, "epoch": 1.42, "learning_rate": 8.791064388961893e-06, "loss": 75.8909, "step": 4309 }, { "compression_loss": 75.97480010986328, "epoch": 1.42, "learning_rate": 8.786136662286466e-06, "loss": 76.3718, "step": 4310 }, { "compression_loss": 77.25090026855469, "epoch": 1.42, "learning_rate": 8.781208935611038e-06, "loss": 77.5816, "step": 4311 }, { "compression_loss": 77.07223510742188, "epoch": 1.42, "learning_rate": 8.776281208935612e-06, "loss": 77.2496, "step": 4312 }, { "compression_loss": 74.89286041259766, "epoch": 1.42, "learning_rate": 8.771353482260184e-06, "loss": 75.3022, "step": 4313 }, { "compression_loss": 74.52772521972656, "epoch": 1.42, "learning_rate": 8.766425755584756e-06, "loss": 74.8244, "step": 4314 }, { "compression_loss": 76.7837142944336, "epoch": 1.42, "learning_rate": 8.761498028909331e-06, "loss": 77.0846, "step": 4315 }, { "compression_loss": 77.1468734741211, "epoch": 1.42, "learning_rate": 8.756570302233903e-06, "loss": 77.4657, "step": 4316 }, { "compression_loss": 76.58403778076172, "epoch": 1.42, "learning_rate": 8.751642575558477e-06, "loss": 77.2591, "step": 4317 }, { "compression_loss": 75.73607635498047, "epoch": 1.42, "learning_rate": 8.746714848883049e-06, "loss": 76.0448, "step": 4318 }, { "compression_loss": 76.03120422363281, "epoch": 1.42, "learning_rate": 8.741787122207621e-06, "loss": 76.2246, "step": 4319 }, { "compression_loss": 73.95491790771484, "epoch": 1.42, "learning_rate": 8.736859395532195e-06, "loss": 74.1764, "step": 4320 }, { "compression_loss": 75.0513687133789, "epoch": 1.42, "learning_rate": 8.731931668856768e-06, "loss": 75.4752, "step": 4321 }, { "compression_loss": 73.97331237792969, "epoch": 1.42, "learning_rate": 8.727003942181342e-06, "loss": 74.1792, "step": 4322 }, { "compression_loss": 77.72145080566406, "epoch": 1.42, "learning_rate": 8.722076215505914e-06, "loss": 78.0149, "step": 4323 }, { "compression_loss": 73.74617004394531, "epoch": 1.42, "learning_rate": 8.717148488830486e-06, "loss": 74.1671, "step": 4324 }, { "compression_loss": 74.6556396484375, "epoch": 1.42, "learning_rate": 8.71222076215506e-06, "loss": 75.1022, "step": 4325 }, { "compression_loss": 74.76193237304688, "epoch": 1.42, "learning_rate": 8.707293035479631e-06, "loss": 74.8838, "step": 4326 }, { "compression_loss": 76.56448364257812, "epoch": 1.42, "learning_rate": 8.702365308804207e-06, "loss": 76.9417, "step": 4327 }, { "compression_loss": 75.048828125, "epoch": 1.42, "learning_rate": 8.697437582128779e-06, "loss": 75.2387, "step": 4328 }, { "compression_loss": 73.95133972167969, "epoch": 1.42, "learning_rate": 8.69250985545335e-06, "loss": 74.3333, "step": 4329 }, { "compression_loss": 75.10482025146484, "epoch": 1.42, "learning_rate": 8.687582128777924e-06, "loss": 75.5631, "step": 4330 }, { "compression_loss": 76.13111877441406, "epoch": 1.42, "learning_rate": 8.682654402102496e-06, "loss": 76.5504, "step": 4331 }, { "compression_loss": 75.90545654296875, "epoch": 1.42, "learning_rate": 8.67772667542707e-06, "loss": 76.2789, "step": 4332 }, { "compression_loss": 75.10521697998047, "epoch": 1.42, "learning_rate": 8.672798948751644e-06, "loss": 75.3316, "step": 4333 }, { "compression_loss": 75.3390121459961, "epoch": 1.42, "learning_rate": 8.667871222076216e-06, "loss": 75.713, "step": 4334 }, { "compression_loss": 74.45570373535156, "epoch": 1.42, "learning_rate": 8.66294349540079e-06, "loss": 74.9622, "step": 4335 }, { "compression_loss": 73.1803207397461, "epoch": 1.42, "learning_rate": 8.658015768725361e-06, "loss": 73.4208, "step": 4336 }, { "compression_loss": 77.58873748779297, "epoch": 1.42, "learning_rate": 8.653088042049935e-06, "loss": 77.9067, "step": 4337 }, { "compression_loss": 76.26597595214844, "epoch": 1.43, "learning_rate": 8.648160315374507e-06, "loss": 76.5632, "step": 4338 }, { "compression_loss": 76.052734375, "epoch": 1.43, "learning_rate": 8.64323258869908e-06, "loss": 76.4589, "step": 4339 }, { "compression_loss": 77.54048156738281, "epoch": 1.43, "learning_rate": 8.638304862023654e-06, "loss": 78.1888, "step": 4340 }, { "compression_loss": 74.20198059082031, "epoch": 1.43, "learning_rate": 8.633377135348226e-06, "loss": 74.5817, "step": 4341 }, { "compression_loss": 76.18965148925781, "epoch": 1.43, "learning_rate": 8.6284494086728e-06, "loss": 76.7122, "step": 4342 }, { "compression_loss": 76.53068542480469, "epoch": 1.43, "learning_rate": 8.623521681997372e-06, "loss": 77.0513, "step": 4343 }, { "compression_loss": 75.64057159423828, "epoch": 1.43, "learning_rate": 8.618593955321944e-06, "loss": 75.9679, "step": 4344 }, { "compression_loss": 77.6495132446289, "epoch": 1.43, "learning_rate": 8.61366622864652e-06, "loss": 78.0021, "step": 4345 }, { "compression_loss": 75.49249267578125, "epoch": 1.43, "learning_rate": 8.608738501971091e-06, "loss": 75.9283, "step": 4346 }, { "compression_loss": 75.72366333007812, "epoch": 1.43, "learning_rate": 8.603810775295665e-06, "loss": 76.0362, "step": 4347 }, { "compression_loss": 75.56365966796875, "epoch": 1.43, "learning_rate": 8.598883048620237e-06, "loss": 75.7685, "step": 4348 }, { "compression_loss": 75.12234497070312, "epoch": 1.43, "learning_rate": 8.593955321944809e-06, "loss": 75.473, "step": 4349 }, { "compression_loss": 75.57212829589844, "epoch": 1.43, "learning_rate": 8.589027595269382e-06, "loss": 75.8571, "step": 4350 }, { "compression_loss": 77.68721008300781, "epoch": 1.43, "learning_rate": 8.584099868593956e-06, "loss": 78.184, "step": 4351 }, { "compression_loss": 75.40115356445312, "epoch": 1.43, "learning_rate": 8.57917214191853e-06, "loss": 75.6177, "step": 4352 }, { "compression_loss": 74.08015441894531, "epoch": 1.43, "learning_rate": 8.574244415243102e-06, "loss": 74.6344, "step": 4353 }, { "compression_loss": 73.99986267089844, "epoch": 1.43, "learning_rate": 8.569316688567674e-06, "loss": 74.1216, "step": 4354 }, { "compression_loss": 73.89257049560547, "epoch": 1.43, "learning_rate": 8.564388961892247e-06, "loss": 74.4692, "step": 4355 }, { "compression_loss": 75.22103881835938, "epoch": 1.43, "learning_rate": 8.55946123521682e-06, "loss": 75.8071, "step": 4356 }, { "compression_loss": 77.529052734375, "epoch": 1.43, "learning_rate": 8.554533508541395e-06, "loss": 77.9191, "step": 4357 }, { "compression_loss": 74.68663024902344, "epoch": 1.43, "learning_rate": 8.549605781865967e-06, "loss": 75.0264, "step": 4358 }, { "compression_loss": 76.02479553222656, "epoch": 1.43, "learning_rate": 8.544678055190539e-06, "loss": 76.5355, "step": 4359 }, { "compression_loss": 76.78834533691406, "epoch": 1.43, "learning_rate": 8.539750328515112e-06, "loss": 77.0453, "step": 4360 }, { "compression_loss": 74.52340698242188, "epoch": 1.43, "learning_rate": 8.534822601839684e-06, "loss": 74.8282, "step": 4361 }, { "compression_loss": 76.41590881347656, "epoch": 1.43, "learning_rate": 8.529894875164258e-06, "loss": 76.701, "step": 4362 }, { "compression_loss": 73.58705139160156, "epoch": 1.43, "learning_rate": 8.524967148488832e-06, "loss": 73.9723, "step": 4363 }, { "compression_loss": 75.52064514160156, "epoch": 1.43, "learning_rate": 8.520039421813404e-06, "loss": 75.7382, "step": 4364 }, { "compression_loss": 74.92279052734375, "epoch": 1.43, "learning_rate": 8.515111695137977e-06, "loss": 75.4367, "step": 4365 }, { "compression_loss": 75.437744140625, "epoch": 1.43, "learning_rate": 8.51018396846255e-06, "loss": 75.9495, "step": 4366 }, { "compression_loss": 76.3502426147461, "epoch": 1.43, "learning_rate": 8.505256241787123e-06, "loss": 76.6132, "step": 4367 }, { "compression_loss": 75.17749786376953, "epoch": 1.43, "learning_rate": 8.500328515111695e-06, "loss": 75.4182, "step": 4368 }, { "compression_loss": 75.86787414550781, "epoch": 1.44, "learning_rate": 8.495400788436268e-06, "loss": 76.3436, "step": 4369 }, { "compression_loss": 77.33757781982422, "epoch": 1.44, "learning_rate": 8.490473061760842e-06, "loss": 77.652, "step": 4370 }, { "compression_loss": 72.72158813476562, "epoch": 1.44, "learning_rate": 8.485545335085414e-06, "loss": 73.0084, "step": 4371 }, { "compression_loss": 73.39479064941406, "epoch": 1.44, "learning_rate": 8.480617608409988e-06, "loss": 73.6326, "step": 4372 }, { "compression_loss": 73.3172378540039, "epoch": 1.44, "learning_rate": 8.47568988173456e-06, "loss": 73.531, "step": 4373 }, { "compression_loss": 73.77225494384766, "epoch": 1.44, "learning_rate": 8.470762155059132e-06, "loss": 74.0729, "step": 4374 }, { "compression_loss": 75.38619232177734, "epoch": 1.44, "learning_rate": 8.465834428383707e-06, "loss": 75.7736, "step": 4375 }, { "compression_loss": 75.69367980957031, "epoch": 1.44, "learning_rate": 8.460906701708279e-06, "loss": 76.0126, "step": 4376 }, { "compression_loss": 77.18064880371094, "epoch": 1.44, "learning_rate": 8.455978975032853e-06, "loss": 77.4958, "step": 4377 }, { "compression_loss": 75.24784851074219, "epoch": 1.44, "learning_rate": 8.451051248357425e-06, "loss": 75.4495, "step": 4378 }, { "compression_loss": 75.58809661865234, "epoch": 1.44, "learning_rate": 8.446123521681997e-06, "loss": 76.0232, "step": 4379 }, { "compression_loss": 76.08808898925781, "epoch": 1.44, "learning_rate": 8.44119579500657e-06, "loss": 76.465, "step": 4380 }, { "compression_loss": 77.19857788085938, "epoch": 1.44, "learning_rate": 8.436268068331144e-06, "loss": 77.4556, "step": 4381 }, { "compression_loss": 76.10247039794922, "epoch": 1.44, "learning_rate": 8.431340341655716e-06, "loss": 76.4121, "step": 4382 }, { "compression_loss": 76.56877136230469, "epoch": 1.44, "learning_rate": 8.42641261498029e-06, "loss": 76.961, "step": 4383 }, { "compression_loss": 74.65615844726562, "epoch": 1.44, "learning_rate": 8.421484888304862e-06, "loss": 74.8987, "step": 4384 }, { "compression_loss": 74.41127014160156, "epoch": 1.44, "learning_rate": 8.416557161629435e-06, "loss": 74.7227, "step": 4385 }, { "compression_loss": 73.9400634765625, "epoch": 1.44, "learning_rate": 8.411629434954007e-06, "loss": 74.3867, "step": 4386 }, { "compression_loss": 75.63121032714844, "epoch": 1.44, "learning_rate": 8.40670170827858e-06, "loss": 75.9096, "step": 4387 }, { "compression_loss": 76.77003479003906, "epoch": 1.44, "learning_rate": 8.401773981603155e-06, "loss": 77.0499, "step": 4388 }, { "compression_loss": 75.18572998046875, "epoch": 1.44, "learning_rate": 8.396846254927726e-06, "loss": 75.5665, "step": 4389 }, { "compression_loss": 73.36708068847656, "epoch": 1.44, "learning_rate": 8.3919185282523e-06, "loss": 73.6263, "step": 4390 }, { "compression_loss": 77.16046905517578, "epoch": 1.44, "learning_rate": 8.386990801576872e-06, "loss": 77.5422, "step": 4391 }, { "compression_loss": 76.57321166992188, "epoch": 1.44, "learning_rate": 8.382063074901444e-06, "loss": 77.1063, "step": 4392 }, { "compression_loss": 75.92262268066406, "epoch": 1.44, "learning_rate": 8.37713534822602e-06, "loss": 76.4332, "step": 4393 }, { "compression_loss": 75.00239562988281, "epoch": 1.44, "learning_rate": 8.372207621550591e-06, "loss": 75.5844, "step": 4394 }, { "compression_loss": 78.9233627319336, "epoch": 1.44, "learning_rate": 8.367279894875165e-06, "loss": 79.6634, "step": 4395 }, { "compression_loss": 75.08946228027344, "epoch": 1.44, "learning_rate": 8.362352168199737e-06, "loss": 75.6393, "step": 4396 }, { "compression_loss": 76.3470458984375, "epoch": 1.44, "learning_rate": 8.357424441524309e-06, "loss": 76.7804, "step": 4397 }, { "compression_loss": 75.06243133544922, "epoch": 1.44, "learning_rate": 8.352496714848883e-06, "loss": 75.4156, "step": 4398 }, { "compression_loss": 76.41667938232422, "epoch": 1.45, "learning_rate": 8.347568988173456e-06, "loss": 76.8658, "step": 4399 }, { "compression_loss": 77.2727279663086, "epoch": 1.45, "learning_rate": 8.34264126149803e-06, "loss": 77.9472, "step": 4400 }, { "compression_loss": 74.34107971191406, "epoch": 1.45, "learning_rate": 8.337713534822602e-06, "loss": 74.5815, "step": 4401 }, { "compression_loss": 72.82707977294922, "epoch": 1.45, "learning_rate": 8.332785808147174e-06, "loss": 73.1199, "step": 4402 }, { "compression_loss": 75.81188201904297, "epoch": 1.45, "learning_rate": 8.327858081471748e-06, "loss": 76.4673, "step": 4403 }, { "compression_loss": 74.85255432128906, "epoch": 1.45, "learning_rate": 8.32293035479632e-06, "loss": 75.242, "step": 4404 }, { "compression_loss": 76.741455078125, "epoch": 1.45, "learning_rate": 8.318002628120895e-06, "loss": 77.0241, "step": 4405 }, { "compression_loss": 76.94251251220703, "epoch": 1.45, "learning_rate": 8.313074901445467e-06, "loss": 77.3324, "step": 4406 }, { "compression_loss": 74.29093170166016, "epoch": 1.45, "learning_rate": 8.308147174770039e-06, "loss": 74.6048, "step": 4407 }, { "compression_loss": 77.04141235351562, "epoch": 1.45, "learning_rate": 8.303219448094613e-06, "loss": 77.5081, "step": 4408 }, { "compression_loss": 75.843505859375, "epoch": 1.45, "learning_rate": 8.298291721419184e-06, "loss": 76.0063, "step": 4409 }, { "compression_loss": 75.51016235351562, "epoch": 1.45, "learning_rate": 8.293363994743758e-06, "loss": 75.8317, "step": 4410 }, { "compression_loss": 74.51052856445312, "epoch": 1.45, "learning_rate": 8.288436268068332e-06, "loss": 74.6744, "step": 4411 }, { "compression_loss": 76.42539978027344, "epoch": 1.45, "learning_rate": 8.283508541392904e-06, "loss": 76.7139, "step": 4412 }, { "compression_loss": 74.04736328125, "epoch": 1.45, "learning_rate": 8.278580814717477e-06, "loss": 74.4937, "step": 4413 }, { "compression_loss": 74.0925064086914, "epoch": 1.45, "learning_rate": 8.27365308804205e-06, "loss": 74.5769, "step": 4414 }, { "compression_loss": 74.9970703125, "epoch": 1.45, "learning_rate": 8.268725361366623e-06, "loss": 75.2606, "step": 4415 }, { "compression_loss": 73.70099639892578, "epoch": 1.45, "learning_rate": 8.263797634691195e-06, "loss": 73.9814, "step": 4416 }, { "compression_loss": 73.29978942871094, "epoch": 1.45, "learning_rate": 8.258869908015769e-06, "loss": 73.9251, "step": 4417 }, { "compression_loss": 74.09477233886719, "epoch": 1.45, "learning_rate": 8.253942181340342e-06, "loss": 74.5647, "step": 4418 }, { "compression_loss": 73.57806396484375, "epoch": 1.45, "learning_rate": 8.249014454664914e-06, "loss": 73.8622, "step": 4419 }, { "compression_loss": 77.38966369628906, "epoch": 1.45, "learning_rate": 8.244086727989488e-06, "loss": 77.7663, "step": 4420 }, { "compression_loss": 73.48344421386719, "epoch": 1.45, "learning_rate": 8.23915900131406e-06, "loss": 73.8301, "step": 4421 }, { "compression_loss": 75.84657287597656, "epoch": 1.45, "learning_rate": 8.234231274638632e-06, "loss": 76.3597, "step": 4422 }, { "compression_loss": 75.74491882324219, "epoch": 1.45, "learning_rate": 8.229303547963207e-06, "loss": 76.1279, "step": 4423 }, { "compression_loss": 76.84963989257812, "epoch": 1.45, "learning_rate": 8.22437582128778e-06, "loss": 77.1888, "step": 4424 }, { "compression_loss": 74.59716796875, "epoch": 1.45, "learning_rate": 8.219448094612353e-06, "loss": 74.8981, "step": 4425 }, { "compression_loss": 75.95292663574219, "epoch": 1.45, "learning_rate": 8.214520367936925e-06, "loss": 76.5814, "step": 4426 }, { "compression_loss": 74.68983459472656, "epoch": 1.45, "learning_rate": 8.209592641261497e-06, "loss": 75.1852, "step": 4427 }, { "compression_loss": 77.22120666503906, "epoch": 1.45, "learning_rate": 8.20466491458607e-06, "loss": 77.5817, "step": 4428 }, { "compression_loss": 77.6015853881836, "epoch": 1.45, "learning_rate": 8.199737187910644e-06, "loss": 77.9076, "step": 4429 }, { "compression_loss": 75.297119140625, "epoch": 1.46, "learning_rate": 8.194809461235218e-06, "loss": 75.8466, "step": 4430 }, { "compression_loss": 73.96074676513672, "epoch": 1.46, "learning_rate": 8.18988173455979e-06, "loss": 74.2699, "step": 4431 }, { "compression_loss": 76.18521118164062, "epoch": 1.46, "learning_rate": 8.184954007884362e-06, "loss": 76.6073, "step": 4432 }, { "compression_loss": 76.5681381225586, "epoch": 1.46, "learning_rate": 8.180026281208935e-06, "loss": 76.7909, "step": 4433 }, { "compression_loss": 77.25804138183594, "epoch": 1.46, "learning_rate": 8.175098554533509e-06, "loss": 78.1598, "step": 4434 }, { "compression_loss": 74.4791259765625, "epoch": 1.46, "learning_rate": 8.170170827858083e-06, "loss": 75.1622, "step": 4435 }, { "compression_loss": 74.84210205078125, "epoch": 1.46, "learning_rate": 8.165243101182655e-06, "loss": 75.2797, "step": 4436 }, { "compression_loss": 77.15666961669922, "epoch": 1.46, "learning_rate": 8.160315374507227e-06, "loss": 77.4659, "step": 4437 }, { "compression_loss": 74.4284439086914, "epoch": 1.46, "learning_rate": 8.1553876478318e-06, "loss": 74.743, "step": 4438 }, { "compression_loss": 75.50455474853516, "epoch": 1.46, "learning_rate": 8.150459921156372e-06, "loss": 75.7395, "step": 4439 }, { "compression_loss": 77.22410583496094, "epoch": 1.46, "learning_rate": 8.145532194480948e-06, "loss": 77.5216, "step": 4440 }, { "compression_loss": 76.23684692382812, "epoch": 1.46, "learning_rate": 8.14060446780552e-06, "loss": 76.6883, "step": 4441 }, { "compression_loss": 76.84538269042969, "epoch": 1.46, "learning_rate": 8.135676741130092e-06, "loss": 77.1088, "step": 4442 }, { "compression_loss": 77.00768280029297, "epoch": 1.46, "learning_rate": 8.130749014454665e-06, "loss": 77.5946, "step": 4443 }, { "compression_loss": 75.69993591308594, "epoch": 1.46, "learning_rate": 8.125821287779237e-06, "loss": 76.2059, "step": 4444 }, { "compression_loss": 76.01459503173828, "epoch": 1.46, "learning_rate": 8.120893561103811e-06, "loss": 76.6012, "step": 4445 }, { "compression_loss": 74.68033599853516, "epoch": 1.46, "learning_rate": 8.115965834428385e-06, "loss": 75.0561, "step": 4446 }, { "compression_loss": 75.20719146728516, "epoch": 1.46, "learning_rate": 8.111038107752957e-06, "loss": 75.488, "step": 4447 }, { "compression_loss": 76.54963684082031, "epoch": 1.46, "learning_rate": 8.10611038107753e-06, "loss": 77.0784, "step": 4448 }, { "compression_loss": 75.40148162841797, "epoch": 1.46, "learning_rate": 8.101182654402102e-06, "loss": 75.7087, "step": 4449 }, { "compression_loss": 75.55645751953125, "epoch": 1.46, "learning_rate": 8.096254927726676e-06, "loss": 76.1494, "step": 4450 }, { "compression_loss": 76.1068115234375, "epoch": 1.46, "learning_rate": 8.091327201051248e-06, "loss": 76.4618, "step": 4451 }, { "compression_loss": 77.1488265991211, "epoch": 1.46, "learning_rate": 8.086399474375822e-06, "loss": 77.5, "step": 4452 }, { "compression_loss": 75.92932891845703, "epoch": 1.46, "learning_rate": 8.081471747700395e-06, "loss": 76.4047, "step": 4453 }, { "compression_loss": 76.22671508789062, "epoch": 1.46, "learning_rate": 8.076544021024967e-06, "loss": 76.534, "step": 4454 }, { "compression_loss": 75.47093200683594, "epoch": 1.46, "learning_rate": 8.07161629434954e-06, "loss": 75.6802, "step": 4455 }, { "compression_loss": 78.12214660644531, "epoch": 1.46, "learning_rate": 8.066688567674113e-06, "loss": 78.5013, "step": 4456 }, { "compression_loss": 72.42308044433594, "epoch": 1.46, "learning_rate": 8.061760840998685e-06, "loss": 72.8557, "step": 4457 }, { "compression_loss": 74.65949249267578, "epoch": 1.46, "learning_rate": 8.05683311432326e-06, "loss": 75.0848, "step": 4458 }, { "compression_loss": 76.64946746826172, "epoch": 1.46, "learning_rate": 8.051905387647832e-06, "loss": 77.0619, "step": 4459 }, { "compression_loss": 76.84713745117188, "epoch": 1.47, "learning_rate": 8.046977660972406e-06, "loss": 77.0926, "step": 4460 }, { "compression_loss": 76.5900650024414, "epoch": 1.47, "learning_rate": 8.042049934296978e-06, "loss": 77.1676, "step": 4461 }, { "compression_loss": 76.60780334472656, "epoch": 1.47, "learning_rate": 8.03712220762155e-06, "loss": 76.8224, "step": 4462 }, { "compression_loss": 76.9488525390625, "epoch": 1.47, "learning_rate": 8.032194480946123e-06, "loss": 77.5152, "step": 4463 }, { "compression_loss": 75.33164978027344, "epoch": 1.47, "learning_rate": 8.027266754270697e-06, "loss": 75.5893, "step": 4464 }, { "compression_loss": 74.47616577148438, "epoch": 1.47, "learning_rate": 8.02233902759527e-06, "loss": 74.7198, "step": 4465 }, { "compression_loss": 74.77142333984375, "epoch": 1.47, "learning_rate": 8.017411300919843e-06, "loss": 75.1635, "step": 4466 }, { "compression_loss": 77.00871276855469, "epoch": 1.47, "learning_rate": 8.012483574244415e-06, "loss": 77.545, "step": 4467 }, { "compression_loss": 76.86743927001953, "epoch": 1.47, "learning_rate": 8.007555847568988e-06, "loss": 77.1864, "step": 4468 }, { "compression_loss": 77.19503784179688, "epoch": 1.47, "learning_rate": 8.00262812089356e-06, "loss": 77.493, "step": 4469 }, { "compression_loss": 75.32083892822266, "epoch": 1.47, "learning_rate": 7.997700394218136e-06, "loss": 75.6201, "step": 4470 }, { "compression_loss": 77.3345947265625, "epoch": 1.47, "learning_rate": 7.992772667542708e-06, "loss": 77.6199, "step": 4471 }, { "compression_loss": 76.57260131835938, "epoch": 1.47, "learning_rate": 7.98784494086728e-06, "loss": 76.7664, "step": 4472 }, { "compression_loss": 75.35022735595703, "epoch": 1.47, "learning_rate": 7.982917214191853e-06, "loss": 75.5225, "step": 4473 }, { "compression_loss": 76.6412582397461, "epoch": 1.47, "learning_rate": 7.977989487516425e-06, "loss": 76.8346, "step": 4474 }, { "compression_loss": 75.17718505859375, "epoch": 1.47, "learning_rate": 7.973061760840999e-06, "loss": 75.419, "step": 4475 }, { "compression_loss": 77.54826354980469, "epoch": 1.47, "learning_rate": 7.968134034165572e-06, "loss": 77.7377, "step": 4476 }, { "compression_loss": 73.986572265625, "epoch": 1.47, "learning_rate": 7.963206307490144e-06, "loss": 74.5084, "step": 4477 }, { "compression_loss": 75.13743591308594, "epoch": 1.47, "learning_rate": 7.958278580814718e-06, "loss": 75.402, "step": 4478 }, { "compression_loss": 78.9431381225586, "epoch": 1.47, "learning_rate": 7.95335085413929e-06, "loss": 79.3049, "step": 4479 }, { "compression_loss": 76.5770263671875, "epoch": 1.47, "learning_rate": 7.948423127463864e-06, "loss": 76.9419, "step": 4480 }, { "compression_loss": 75.54627990722656, "epoch": 1.47, "learning_rate": 7.943495400788436e-06, "loss": 75.7767, "step": 4481 }, { "compression_loss": 72.80355834960938, "epoch": 1.47, "learning_rate": 7.93856767411301e-06, "loss": 73.0459, "step": 4482 }, { "compression_loss": 75.88595581054688, "epoch": 1.47, "learning_rate": 7.933639947437583e-06, "loss": 76.2151, "step": 4483 }, { "compression_loss": 76.0378189086914, "epoch": 1.47, "learning_rate": 7.928712220762155e-06, "loss": 76.7405, "step": 4484 }, { "compression_loss": 76.60010528564453, "epoch": 1.47, "learning_rate": 7.923784494086729e-06, "loss": 77.1702, "step": 4485 }, { "compression_loss": 75.00410461425781, "epoch": 1.47, "learning_rate": 7.9188567674113e-06, "loss": 75.1629, "step": 4486 }, { "compression_loss": 76.04521179199219, "epoch": 1.47, "learning_rate": 7.913929040735873e-06, "loss": 76.4065, "step": 4487 }, { "compression_loss": 77.19505310058594, "epoch": 1.47, "learning_rate": 7.909001314060448e-06, "loss": 77.558, "step": 4488 }, { "compression_loss": 74.4534912109375, "epoch": 1.47, "learning_rate": 7.90407358738502e-06, "loss": 74.8877, "step": 4489 }, { "compression_loss": 75.6074447631836, "epoch": 1.48, "learning_rate": 7.899145860709594e-06, "loss": 76.179, "step": 4490 }, { "compression_loss": 75.15528106689453, "epoch": 1.48, "learning_rate": 7.894218134034166e-06, "loss": 75.7006, "step": 4491 }, { "compression_loss": 75.79702758789062, "epoch": 1.48, "learning_rate": 7.889290407358738e-06, "loss": 76.2077, "step": 4492 }, { "compression_loss": 74.78510284423828, "epoch": 1.48, "learning_rate": 7.884362680683311e-06, "loss": 75.2265, "step": 4493 }, { "compression_loss": 75.1790771484375, "epoch": 1.48, "learning_rate": 7.879434954007885e-06, "loss": 75.3667, "step": 4494 }, { "compression_loss": 74.78810119628906, "epoch": 1.48, "learning_rate": 7.874507227332459e-06, "loss": 75.0839, "step": 4495 }, { "compression_loss": 76.24575805664062, "epoch": 1.48, "learning_rate": 7.86957950065703e-06, "loss": 76.4279, "step": 4496 }, { "compression_loss": 76.06120300292969, "epoch": 1.48, "learning_rate": 7.864651773981602e-06, "loss": 76.2555, "step": 4497 }, { "compression_loss": 74.71146392822266, "epoch": 1.48, "learning_rate": 7.859724047306176e-06, "loss": 74.9993, "step": 4498 }, { "compression_loss": 74.63123321533203, "epoch": 1.48, "learning_rate": 7.854796320630748e-06, "loss": 75.1416, "step": 4499 }, { "compression_loss": 74.63452911376953, "epoch": 1.48, "learning_rate": 7.849868593955323e-06, "loss": 75.0266, "step": 4500 }, { "epoch": 1.48, "eval_exact_match": 87.24692526017029, "eval_f1": 93.25905342519843, "step": 4500 } ], "max_steps": 6088, "num_train_epochs": 2, "total_flos": 1.33734116081664e+17, "trial_name": null, "trial_params": null }