{ "best_metric": 0.6102247834205627, "best_model_checkpoint": "./luke-marian-training-1/checkpoint-123440", "epoch": 10.0, "global_step": 123440, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.04, "learning_rate": 8.052495139338951e-07, "loss": 0.7856, "step": 500 }, { "epoch": 0.08, "learning_rate": 1.6153596889176928e-06, "loss": 0.748, "step": 1000 }, { "epoch": 0.12, "learning_rate": 2.4254698639014908e-06, "loss": 0.7809, "step": 1500 }, { "epoch": 0.16, "learning_rate": 3.2355800388852886e-06, "loss": 0.7923, "step": 2000 }, { "epoch": 0.2, "learning_rate": 4.044069993519119e-06, "loss": 0.7799, "step": 2500 }, { "epoch": 0.24, "learning_rate": 4.854180168502917e-06, "loss": 0.8085, "step": 3000 }, { "epoch": 0.28, "learning_rate": 5.664290343486715e-06, "loss": 0.7759, "step": 3500 }, { "epoch": 0.32, "learning_rate": 6.474400518470513e-06, "loss": 0.7833, "step": 4000 }, { "epoch": 0.36, "learning_rate": 7.2828904731043424e-06, "loss": 0.7804, "step": 4500 }, { "epoch": 0.41, "learning_rate": 8.09300064808814e-06, "loss": 0.8169, "step": 5000 }, { "epoch": 0.45, "learning_rate": 8.903110823071939e-06, "loss": 0.8058, "step": 5500 }, { "epoch": 0.49, "learning_rate": 9.713220998055737e-06, "loss": 0.8265, "step": 6000 }, { "epoch": 0.53, "learning_rate": 9.972541528805813e-06, "loss": 0.7904, "step": 6500 }, { "epoch": 0.57, "learning_rate": 9.929904151175087e-06, "loss": 0.7829, "step": 7000 }, { "epoch": 0.61, "learning_rate": 9.88726677354436e-06, "loss": 0.8126, "step": 7500 }, { "epoch": 0.65, "learning_rate": 9.844629395913635e-06, "loss": 0.8133, "step": 8000 }, { "epoch": 0.69, "learning_rate": 9.802077293038171e-06, "loss": 0.7928, "step": 8500 }, { "epoch": 0.73, "learning_rate": 9.759525190162705e-06, "loss": 0.8564, "step": 9000 }, { "epoch": 0.77, "learning_rate": 9.716887812531979e-06, "loss": 0.7624, "step": 9500 }, { "epoch": 0.81, "learning_rate": 9.674250434901252e-06, "loss": 0.7662, "step": 10000 }, { "epoch": 0.85, "learning_rate": 9.631613057270527e-06, "loss": 0.8363, "step": 10500 }, { "epoch": 0.89, "learning_rate": 9.5889756796398e-06, "loss": 0.7793, "step": 11000 }, { "epoch": 0.93, "learning_rate": 9.546338302009073e-06, "loss": 0.7561, "step": 11500 }, { "epoch": 0.97, "learning_rate": 9.50378619913361e-06, "loss": 0.774, "step": 12000 }, { "epoch": 1.0, "eval_BLEU": 32.0811169920738, "eval_BLEU-Bigram-Precision": 37.1092045950269, "eval_BLEU-Trigram-Precision": 26.487573476332887, "eval_BLEU-Unigram-Precision": 56.60325589903054, "eval_ROUGE-2": 25.923225512351976, "eval_ROUGE-L": 40.443599801980355, "eval_Sacre-Bigram-Precision": 34.86367988711385, "eval_Sacre-Trigram-Precision": 26.20122541115769, "eval_Sacre-Unigram-Precision": 54.42125237191651, "eval_SacreBLEU": 30.767354925943692, "eval_loss": 0.9131098985671997, "eval_runtime": 174.9979, "eval_samples_per_second": 7.069, "eval_steps_per_second": 7.069, "step": 12344 }, { "epoch": 1.01, "learning_rate": 9.461148821502884e-06, "loss": 0.7846, "step": 12500 }, { "epoch": 1.05, "learning_rate": 9.418511443872157e-06, "loss": 0.7368, "step": 13000 }, { "epoch": 1.09, "learning_rate": 9.37587406624143e-06, "loss": 0.7279, "step": 13500 }, { "epoch": 1.13, "learning_rate": 9.333236688610704e-06, "loss": 0.7253, "step": 14000 }, { "epoch": 1.17, "learning_rate": 9.290599310979979e-06, "loss": 0.7259, "step": 14500 }, { "epoch": 1.22, "learning_rate": 9.247961933349252e-06, "loss": 0.7432, "step": 15000 }, { "epoch": 1.26, "learning_rate": 9.205324555718525e-06, "loss": 0.7437, "step": 15500 }, { "epoch": 1.3, "learning_rate": 9.1626871780878e-06, "loss": 0.6641, "step": 16000 }, { "epoch": 1.34, "learning_rate": 9.120135075212335e-06, "loss": 0.7239, "step": 16500 }, { "epoch": 1.38, "learning_rate": 9.077497697581608e-06, "loss": 0.7007, "step": 17000 }, { "epoch": 1.42, "learning_rate": 9.034860319950883e-06, "loss": 0.7107, "step": 17500 }, { "epoch": 1.46, "learning_rate": 8.992222942320156e-06, "loss": 0.7033, "step": 18000 }, { "epoch": 1.5, "learning_rate": 8.949670839444692e-06, "loss": 0.6932, "step": 18500 }, { "epoch": 1.54, "learning_rate": 8.907118736569226e-06, "loss": 0.7351, "step": 19000 }, { "epoch": 1.58, "learning_rate": 8.8644813589385e-06, "loss": 0.7511, "step": 19500 }, { "epoch": 1.62, "learning_rate": 8.821843981307775e-06, "loss": 0.6994, "step": 20000 }, { "epoch": 1.66, "learning_rate": 8.779206603677048e-06, "loss": 0.6792, "step": 20500 }, { "epoch": 1.7, "learning_rate": 8.736569226046321e-06, "loss": 0.6701, "step": 21000 }, { "epoch": 1.74, "learning_rate": 8.693931848415596e-06, "loss": 0.6885, "step": 21500 }, { "epoch": 1.78, "learning_rate": 8.65129447078487e-06, "loss": 0.6796, "step": 22000 }, { "epoch": 1.82, "learning_rate": 8.608657093154144e-06, "loss": 0.6812, "step": 22500 }, { "epoch": 1.86, "learning_rate": 8.566104990278679e-06, "loss": 0.704, "step": 23000 }, { "epoch": 1.9, "learning_rate": 8.523467612647952e-06, "loss": 0.6443, "step": 23500 }, { "epoch": 1.94, "learning_rate": 8.480830235017227e-06, "loss": 0.6513, "step": 24000 }, { "epoch": 1.98, "learning_rate": 8.4381928573865e-06, "loss": 0.6532, "step": 24500 }, { "epoch": 2.0, "eval_BLEU": 34.14564345469897, "eval_BLEU-Bigram-Precision": 42.337954289996254, "eval_BLEU-Trigram-Precision": 32.169876203576344, "eval_BLEU-Unigram-Precision": 61.27510040160643, "eval_ROUGE-2": 28.971835230430752, "eval_ROUGE-L": 43.230417531342944, "eval_Sacre-Bigram-Precision": 40.19043401240035, "eval_Sacre-Trigram-Precision": 31.907054139181078, "eval_Sacre-Unigram-Precision": 59.04875395057251, "eval_SacreBLEU": 33.00279853845508, "eval_loss": 0.8338403701782227, "eval_runtime": 167.0515, "eval_samples_per_second": 7.405, "eval_steps_per_second": 7.405, "step": 24688 }, { "epoch": 2.03, "learning_rate": 8.395555479755773e-06, "loss": 0.6209, "step": 25000 }, { "epoch": 2.07, "learning_rate": 8.35300337688031e-06, "loss": 0.6246, "step": 25500 }, { "epoch": 2.11, "learning_rate": 8.310365999249583e-06, "loss": 0.6061, "step": 26000 }, { "epoch": 2.15, "learning_rate": 8.267728621618858e-06, "loss": 0.6347, "step": 26500 }, { "epoch": 2.19, "learning_rate": 8.22509124398813e-06, "loss": 0.609, "step": 27000 }, { "epoch": 2.23, "learning_rate": 8.182539141112665e-06, "loss": 0.6059, "step": 27500 }, { "epoch": 2.27, "learning_rate": 8.13990176348194e-06, "loss": 0.6301, "step": 28000 }, { "epoch": 2.31, "learning_rate": 8.097264385851213e-06, "loss": 0.5859, "step": 28500 }, { "epoch": 2.35, "learning_rate": 8.054627008220487e-06, "loss": 0.6133, "step": 29000 }, { "epoch": 2.39, "learning_rate": 8.012074905345023e-06, "loss": 0.5813, "step": 29500 }, { "epoch": 2.43, "learning_rate": 7.969437527714296e-06, "loss": 0.6049, "step": 30000 }, { "epoch": 2.47, "learning_rate": 7.92680015008357e-06, "loss": 0.5982, "step": 30500 }, { "epoch": 2.51, "learning_rate": 7.884162772452844e-06, "loss": 0.5807, "step": 31000 }, { "epoch": 2.55, "learning_rate": 7.841525394822117e-06, "loss": 0.5809, "step": 31500 }, { "epoch": 2.59, "learning_rate": 7.798973291946653e-06, "loss": 0.5984, "step": 32000 }, { "epoch": 2.63, "learning_rate": 7.756335914315927e-06, "loss": 0.6308, "step": 32500 }, { "epoch": 2.67, "learning_rate": 7.7136985366852e-06, "loss": 0.5964, "step": 33000 }, { "epoch": 2.71, "learning_rate": 7.671061159054473e-06, "loss": 0.5783, "step": 33500 }, { "epoch": 2.75, "learning_rate": 7.62850905617901e-06, "loss": 0.5914, "step": 34000 }, { "epoch": 2.79, "learning_rate": 7.585871678548283e-06, "loss": 0.5716, "step": 34500 }, { "epoch": 2.84, "learning_rate": 7.543319575672818e-06, "loss": 0.613, "step": 35000 }, { "epoch": 2.88, "learning_rate": 7.500682198042092e-06, "loss": 0.561, "step": 35500 }, { "epoch": 2.92, "learning_rate": 7.458044820411367e-06, "loss": 0.5731, "step": 36000 }, { "epoch": 2.96, "learning_rate": 7.41540744278064e-06, "loss": 0.565, "step": 36500 }, { "epoch": 3.0, "learning_rate": 7.372855339905175e-06, "loss": 0.6118, "step": 37000 }, { "epoch": 3.0, "eval_BLEU": 39.5916929052258, "eval_BLEU-Bigram-Precision": 44.30874450415242, "eval_BLEU-Trigram-Precision": 34.66957832891385, "eval_BLEU-Unigram-Precision": 61.68977749113189, "eval_ROUGE-2": 33.461963127391705, "eval_ROUGE-L": 46.65653806735692, "eval_Sacre-Bigram-Precision": 42.369437106279214, "eval_Sacre-Trigram-Precision": 34.54220198406245, "eval_Sacre-Unigram-Precision": 59.68165957650208, "eval_SacreBLEU": 38.27100461755832, "eval_loss": 0.7759082317352295, "eval_runtime": 173.3948, "eval_samples_per_second": 7.134, "eval_steps_per_second": 7.134, "step": 37032 }, { "epoch": 3.04, "learning_rate": 7.33030323702971e-06, "loss": 0.4871, "step": 37500 }, { "epoch": 3.08, "learning_rate": 7.287665859398984e-06, "loss": 0.4991, "step": 38000 }, { "epoch": 3.12, "learning_rate": 7.245028481768258e-06, "loss": 0.5436, "step": 38500 }, { "epoch": 3.16, "learning_rate": 7.202391104137532e-06, "loss": 0.5165, "step": 39000 }, { "epoch": 3.2, "learning_rate": 7.159753726506805e-06, "loss": 0.516, "step": 39500 }, { "epoch": 3.24, "learning_rate": 7.11711634887608e-06, "loss": 0.5173, "step": 40000 }, { "epoch": 3.28, "learning_rate": 7.074478971245353e-06, "loss": 0.5239, "step": 40500 }, { "epoch": 3.32, "learning_rate": 7.031841593614626e-06, "loss": 0.5112, "step": 41000 }, { "epoch": 3.36, "learning_rate": 6.9892042159839e-06, "loss": 0.516, "step": 41500 }, { "epoch": 3.4, "learning_rate": 6.9465668383531745e-06, "loss": 0.5384, "step": 42000 }, { "epoch": 3.44, "learning_rate": 6.903929460722449e-06, "loss": 0.5194, "step": 42500 }, { "epoch": 3.48, "learning_rate": 6.861292083091722e-06, "loss": 0.5006, "step": 43000 }, { "epoch": 3.52, "learning_rate": 6.818739980216257e-06, "loss": 0.5362, "step": 43500 }, { "epoch": 3.56, "learning_rate": 6.776102602585531e-06, "loss": 0.5505, "step": 44000 }, { "epoch": 3.6, "learning_rate": 6.733465224954805e-06, "loss": 0.5247, "step": 44500 }, { "epoch": 3.65, "learning_rate": 6.6908278473240785e-06, "loss": 0.5104, "step": 45000 }, { "epoch": 3.69, "learning_rate": 6.648275744448614e-06, "loss": 0.516, "step": 45500 }, { "epoch": 3.73, "learning_rate": 6.605638366817888e-06, "loss": 0.5097, "step": 46000 }, { "epoch": 3.77, "learning_rate": 6.563000989187162e-06, "loss": 0.5147, "step": 46500 }, { "epoch": 3.81, "learning_rate": 6.520363611556435e-06, "loss": 0.4988, "step": 47000 }, { "epoch": 3.85, "learning_rate": 6.47781150868097e-06, "loss": 0.5143, "step": 47500 }, { "epoch": 3.89, "learning_rate": 6.435174131050244e-06, "loss": 0.5146, "step": 48000 }, { "epoch": 3.93, "learning_rate": 6.3925367534195185e-06, "loss": 0.4949, "step": 48500 }, { "epoch": 3.97, "learning_rate": 6.349899375788792e-06, "loss": 0.5327, "step": 49000 }, { "epoch": 4.0, "eval_BLEU": 41.583977153720554, "eval_BLEU-Bigram-Precision": 47.79210232463709, "eval_BLEU-Trigram-Precision": 38.44570809639859, "eval_BLEU-Unigram-Precision": 64.70223028797936, "eval_ROUGE-2": 35.92233849645701, "eval_ROUGE-L": 48.89087193511837, "eval_Sacre-Bigram-Precision": 46.017043349388665, "eval_Sacre-Trigram-Precision": 38.29859537834164, "eval_Sacre-Unigram-Precision": 62.97069051167412, "eval_SacreBLEU": 40.202754965999326, "eval_loss": 0.734529972076416, "eval_runtime": 168.3041, "eval_samples_per_second": 7.35, "eval_steps_per_second": 7.35, "step": 49376 }, { "epoch": 4.01, "learning_rate": 6.307347272913327e-06, "loss": 0.515, "step": 49500 }, { "epoch": 4.05, "learning_rate": 6.264709895282601e-06, "loss": 0.4603, "step": 50000 }, { "epoch": 4.09, "learning_rate": 6.222072517651875e-06, "loss": 0.4565, "step": 50500 }, { "epoch": 4.13, "learning_rate": 6.179435140021148e-06, "loss": 0.4557, "step": 51000 }, { "epoch": 4.17, "learning_rate": 6.136797762390423e-06, "loss": 0.4586, "step": 51500 }, { "epoch": 4.21, "learning_rate": 6.094245659514958e-06, "loss": 0.4295, "step": 52000 }, { "epoch": 4.25, "learning_rate": 6.051608281884232e-06, "loss": 0.4774, "step": 52500 }, { "epoch": 4.29, "learning_rate": 6.009056179008766e-06, "loss": 0.4555, "step": 53000 }, { "epoch": 4.33, "learning_rate": 5.966418801378041e-06, "loss": 0.4448, "step": 53500 }, { "epoch": 4.37, "learning_rate": 5.923781423747314e-06, "loss": 0.4481, "step": 54000 }, { "epoch": 4.42, "learning_rate": 5.881144046116588e-06, "loss": 0.4288, "step": 54500 }, { "epoch": 4.46, "learning_rate": 5.838506668485862e-06, "loss": 0.4459, "step": 55000 }, { "epoch": 4.5, "learning_rate": 5.7958692908551366e-06, "loss": 0.4616, "step": 55500 }, { "epoch": 4.54, "learning_rate": 5.75323191322441e-06, "loss": 0.4775, "step": 56000 }, { "epoch": 4.58, "learning_rate": 5.710594535593683e-06, "loss": 0.4414, "step": 56500 }, { "epoch": 4.62, "learning_rate": 5.668042432718218e-06, "loss": 0.483, "step": 57000 }, { "epoch": 4.66, "learning_rate": 5.625405055087493e-06, "loss": 0.4671, "step": 57500 }, { "epoch": 4.7, "learning_rate": 5.5827676774567664e-06, "loss": 0.4696, "step": 58000 }, { "epoch": 4.74, "learning_rate": 5.54013029982604e-06, "loss": 0.4628, "step": 58500 }, { "epoch": 4.78, "learning_rate": 5.497578196950575e-06, "loss": 0.4591, "step": 59000 }, { "epoch": 4.82, "learning_rate": 5.45494081931985e-06, "loss": 0.4362, "step": 59500 }, { "epoch": 4.86, "learning_rate": 5.412303441689123e-06, "loss": 0.4767, "step": 60000 }, { "epoch": 4.9, "learning_rate": 5.369666064058396e-06, "loss": 0.4642, "step": 60500 }, { "epoch": 4.94, "learning_rate": 5.3271139611829315e-06, "loss": 0.4348, "step": 61000 }, { "epoch": 4.98, "learning_rate": 5.2844765835522064e-06, "loss": 0.4365, "step": 61500 }, { "epoch": 5.0, "eval_BLEU": 43.53907049710156, "eval_BLEU-Bigram-Precision": 49.90573248407643, "eval_BLEU-Trigram-Precision": 40.825538394605175, "eval_BLEU-Unigram-Precision": 66.37906241012367, "eval_ROUGE-2": 38.22782051808487, "eval_ROUGE-L": 51.36307879612767, "eval_Sacre-Bigram-Precision": 48.25423818887177, "eval_Sacre-Trigram-Precision": 40.784982935153586, "eval_Sacre-Unigram-Precision": 64.79006681958711, "eval_SacreBLEU": 42.18614400441328, "eval_loss": 0.7037733197212219, "eval_runtime": 168.6409, "eval_samples_per_second": 7.335, "eval_steps_per_second": 7.335, "step": 61720 }, { "epoch": 5.02, "learning_rate": 5.241924480676741e-06, "loss": 0.4345, "step": 62000 }, { "epoch": 5.06, "learning_rate": 5.199372377801276e-06, "loss": 0.4124, "step": 62500 }, { "epoch": 5.1, "learning_rate": 5.156735000170549e-06, "loss": 0.3992, "step": 63000 }, { "epoch": 5.14, "learning_rate": 5.114097622539824e-06, "loss": 0.3928, "step": 63500 }, { "epoch": 5.18, "learning_rate": 5.0714602449090974e-06, "loss": 0.4428, "step": 64000 }, { "epoch": 5.23, "learning_rate": 5.0288228672783715e-06, "loss": 0.4003, "step": 64500 }, { "epoch": 5.27, "learning_rate": 4.986185489647646e-06, "loss": 0.4077, "step": 65000 }, { "epoch": 5.31, "learning_rate": 4.943548112016919e-06, "loss": 0.4076, "step": 65500 }, { "epoch": 5.35, "learning_rate": 4.900910734386193e-06, "loss": 0.4138, "step": 66000 }, { "epoch": 5.39, "learning_rate": 4.858273356755466e-06, "loss": 0.3901, "step": 66500 }, { "epoch": 5.43, "learning_rate": 4.81563597912474e-06, "loss": 0.4412, "step": 67000 }, { "epoch": 5.47, "learning_rate": 4.7729986014940135e-06, "loss": 0.4128, "step": 67500 }, { "epoch": 5.51, "learning_rate": 4.730361223863288e-06, "loss": 0.4382, "step": 68000 }, { "epoch": 5.55, "learning_rate": 4.687809120987823e-06, "loss": 0.3855, "step": 68500 }, { "epoch": 5.59, "learning_rate": 4.645171743357097e-06, "loss": 0.4116, "step": 69000 }, { "epoch": 5.63, "learning_rate": 4.60253436572637e-06, "loss": 0.3977, "step": 69500 }, { "epoch": 5.67, "learning_rate": 4.559896988095644e-06, "loss": 0.4232, "step": 70000 }, { "epoch": 5.71, "learning_rate": 4.517259610464918e-06, "loss": 0.4194, "step": 70500 }, { "epoch": 5.75, "learning_rate": 4.4746222328341924e-06, "loss": 0.389, "step": 71000 }, { "epoch": 5.79, "learning_rate": 4.4319848552034665e-06, "loss": 0.4051, "step": 71500 }, { "epoch": 5.83, "learning_rate": 4.38934747757274e-06, "loss": 0.4337, "step": 72000 }, { "epoch": 5.87, "learning_rate": 4.346795374697275e-06, "loss": 0.4143, "step": 72500 }, { "epoch": 5.91, "learning_rate": 4.304157997066549e-06, "loss": 0.4146, "step": 73000 }, { "epoch": 5.95, "learning_rate": 4.261520619435823e-06, "loss": 0.3878, "step": 73500 }, { "epoch": 5.99, "learning_rate": 4.218883241805096e-06, "loss": 0.3955, "step": 74000 }, { "epoch": 6.0, "eval_BLEU": 46.18598569567297, "eval_BLEU-Bigram-Precision": 51.835734293717486, "eval_BLEU-Trigram-Precision": 42.84724073580379, "eval_BLEU-Unigram-Precision": 67.89297658862876, "eval_ROUGE-2": 40.77288879803008, "eval_ROUGE-L": 53.47443573476642, "eval_Sacre-Bigram-Precision": 50.16359387172163, "eval_Sacre-Trigram-Precision": 42.76279276279276, "eval_Sacre-Unigram-Precision": 66.28928362287722, "eval_SacreBLEU": 44.9656942267261, "eval_loss": 0.6645232439041138, "eval_runtime": 169.8377, "eval_samples_per_second": 7.283, "eval_steps_per_second": 7.283, "step": 74064 }, { "epoch": 6.04, "learning_rate": 4.176331138929632e-06, "loss": 0.3766, "step": 74500 }, { "epoch": 6.08, "learning_rate": 4.133693761298906e-06, "loss": 0.3556, "step": 75000 }, { "epoch": 6.12, "learning_rate": 4.09105638366818e-06, "loss": 0.3511, "step": 75500 }, { "epoch": 6.16, "learning_rate": 4.048504280792714e-06, "loss": 0.4021, "step": 76000 }, { "epoch": 6.2, "learning_rate": 4.005952177917249e-06, "loss": 0.3931, "step": 76500 }, { "epoch": 6.24, "learning_rate": 3.9633148002865234e-06, "loss": 0.3704, "step": 77000 }, { "epoch": 6.28, "learning_rate": 3.9206774226557975e-06, "loss": 0.3515, "step": 77500 }, { "epoch": 6.32, "learning_rate": 3.878040045025071e-06, "loss": 0.3907, "step": 78000 }, { "epoch": 6.36, "learning_rate": 3.835402667394345e-06, "loss": 0.3643, "step": 78500 }, { "epoch": 6.4, "learning_rate": 3.7927652897636185e-06, "loss": 0.3708, "step": 79000 }, { "epoch": 6.44, "learning_rate": 3.7501279121328926e-06, "loss": 0.3697, "step": 79500 }, { "epoch": 6.48, "learning_rate": 3.7074905345021663e-06, "loss": 0.3486, "step": 80000 }, { "epoch": 6.52, "learning_rate": 3.6648531568714404e-06, "loss": 0.3552, "step": 80500 }, { "epoch": 6.56, "learning_rate": 3.6222157792407136e-06, "loss": 0.3812, "step": 81000 }, { "epoch": 6.6, "learning_rate": 3.5795784016099877e-06, "loss": 0.3622, "step": 81500 }, { "epoch": 6.64, "learning_rate": 3.5369410239792614e-06, "loss": 0.391, "step": 82000 }, { "epoch": 6.68, "learning_rate": 3.494388921103797e-06, "loss": 0.4055, "step": 82500 }, { "epoch": 6.72, "learning_rate": 3.4517515434730702e-06, "loss": 0.3967, "step": 83000 }, { "epoch": 6.76, "learning_rate": 3.4091141658423443e-06, "loss": 0.3949, "step": 83500 }, { "epoch": 6.8, "learning_rate": 3.3664767882116184e-06, "loss": 0.3898, "step": 84000 }, { "epoch": 6.85, "learning_rate": 3.3239246853361536e-06, "loss": 0.3997, "step": 84500 }, { "epoch": 6.89, "learning_rate": 3.281287307705427e-06, "loss": 0.4075, "step": 85000 }, { "epoch": 6.93, "learning_rate": 3.238649930074701e-06, "loss": 0.3704, "step": 85500 }, { "epoch": 6.97, "learning_rate": 3.196012552443975e-06, "loss": 0.3809, "step": 86000 }, { "epoch": 7.0, "eval_BLEU": 48.53041153339425, "eval_BLEU-Bigram-Precision": 52.751822317890515, "eval_BLEU-Trigram-Precision": 44.54280358258696, "eval_BLEU-Unigram-Precision": 67.82913552910786, "eval_ROUGE-2": 42.93136082972705, "eval_ROUGE-L": 54.935588774300534, "eval_Sacre-Bigram-Precision": 51.321117635148816, "eval_Sacre-Trigram-Precision": 44.494843134078515, "eval_Sacre-Unigram-Precision": 66.2363645024532, "eval_SacreBLEU": 47.488510994186036, "eval_loss": 0.6420064568519592, "eval_runtime": 175.4955, "eval_samples_per_second": 7.049, "eval_steps_per_second": 7.049, "step": 86408 }, { "epoch": 7.01, "learning_rate": 3.1534604495685103e-06, "loss": 0.3682, "step": 86500 }, { "epoch": 7.05, "learning_rate": 3.110908346693045e-06, "loss": 0.3759, "step": 87000 }, { "epoch": 7.09, "learning_rate": 3.068270969062319e-06, "loss": 0.3607, "step": 87500 }, { "epoch": 7.13, "learning_rate": 3.025633591431593e-06, "loss": 0.3478, "step": 88000 }, { "epoch": 7.17, "learning_rate": 2.982996213800867e-06, "loss": 0.3474, "step": 88500 }, { "epoch": 7.21, "learning_rate": 2.94035883617014e-06, "loss": 0.3824, "step": 89000 }, { "epoch": 7.25, "learning_rate": 2.8977214585394142e-06, "loss": 0.3469, "step": 89500 }, { "epoch": 7.29, "learning_rate": 2.8550840809086883e-06, "loss": 0.3599, "step": 90000 }, { "epoch": 7.33, "learning_rate": 2.8124467032779616e-06, "loss": 0.3332, "step": 90500 }, { "epoch": 7.37, "learning_rate": 2.7698093256472356e-06, "loss": 0.3508, "step": 91000 }, { "epoch": 7.41, "learning_rate": 2.727257222771771e-06, "loss": 0.3382, "step": 91500 }, { "epoch": 7.45, "learning_rate": 2.684619845141045e-06, "loss": 0.3583, "step": 92000 }, { "epoch": 7.49, "learning_rate": 2.641982467510318e-06, "loss": 0.3392, "step": 92500 }, { "epoch": 7.53, "learning_rate": 2.5993450898795923e-06, "loss": 0.3548, "step": 93000 }, { "epoch": 7.57, "learning_rate": 2.556707712248866e-06, "loss": 0.3372, "step": 93500 }, { "epoch": 7.62, "learning_rate": 2.5141556093734016e-06, "loss": 0.3386, "step": 94000 }, { "epoch": 7.66, "learning_rate": 2.471518231742675e-06, "loss": 0.362, "step": 94500 }, { "epoch": 7.7, "learning_rate": 2.428880854111949e-06, "loss": 0.3454, "step": 95000 }, { "epoch": 7.74, "learning_rate": 2.3862434764812226e-06, "loss": 0.3757, "step": 95500 }, { "epoch": 7.78, "learning_rate": 2.3436913736057578e-06, "loss": 0.3785, "step": 96000 }, { "epoch": 7.82, "learning_rate": 2.3010539959750314e-06, "loss": 0.3656, "step": 96500 }, { "epoch": 7.86, "learning_rate": 2.2584166183443055e-06, "loss": 0.36, "step": 97000 }, { "epoch": 7.9, "learning_rate": 2.2157792407135796e-06, "loss": 0.3522, "step": 97500 }, { "epoch": 7.94, "learning_rate": 2.1732271378381144e-06, "loss": 0.3511, "step": 98000 }, { "epoch": 7.98, "learning_rate": 2.130589760207388e-06, "loss": 0.3349, "step": 98500 }, { "epoch": 8.0, "eval_BLEU": 49.57199736076417, "eval_BLEU-Bigram-Precision": 55.58697298661954, "eval_BLEU-Trigram-Precision": 47.143087942269375, "eval_BLEU-Unigram-Precision": 70.68244463454045, "eval_ROUGE-2": 44.4865922111988, "eval_ROUGE-L": 56.71133192128113, "eval_Sacre-Bigram-Precision": 53.98119122257053, "eval_Sacre-Trigram-Precision": 47.064737753449144, "eval_Sacre-Unigram-Precision": 69.11714187564411, "eval_SacreBLEU": 48.47735653922784, "eval_loss": 0.6274014711380005, "eval_runtime": 169.1165, "eval_samples_per_second": 7.314, "eval_steps_per_second": 7.314, "step": 98752 }, { "epoch": 8.02, "learning_rate": 2.087952382576662e-06, "loss": 0.3635, "step": 99000 }, { "epoch": 8.06, "learning_rate": 2.0453150049459363e-06, "loss": 0.3149, "step": 99500 }, { "epoch": 8.1, "learning_rate": 2.00267762731521e-06, "loss": 0.3348, "step": 100000 }, { "epoch": 8.14, "learning_rate": 1.9600402496844836e-06, "loss": 0.3358, "step": 100500 }, { "epoch": 8.18, "learning_rate": 1.9174028720537573e-06, "loss": 0.3405, "step": 101000 }, { "epoch": 8.22, "learning_rate": 1.8747654944230311e-06, "loss": 0.328, "step": 101500 }, { "epoch": 8.26, "learning_rate": 1.8322133915475666e-06, "loss": 0.3487, "step": 102000 }, { "epoch": 8.3, "learning_rate": 1.7895760139168402e-06, "loss": 0.3012, "step": 102500 }, { "epoch": 8.34, "learning_rate": 1.746938636286114e-06, "loss": 0.3456, "step": 103000 }, { "epoch": 8.38, "learning_rate": 1.7043012586553878e-06, "loss": 0.3472, "step": 103500 }, { "epoch": 8.43, "learning_rate": 1.6617491557799232e-06, "loss": 0.3476, "step": 104000 }, { "epoch": 8.47, "learning_rate": 1.6191117781491968e-06, "loss": 0.3262, "step": 104500 }, { "epoch": 8.51, "learning_rate": 1.5765596752737323e-06, "loss": 0.3607, "step": 105000 }, { "epoch": 8.55, "learning_rate": 1.533922297643006e-06, "loss": 0.341, "step": 105500 }, { "epoch": 8.59, "learning_rate": 1.4912849200122798e-06, "loss": 0.3164, "step": 106000 }, { "epoch": 8.63, "learning_rate": 1.4486475423815535e-06, "loss": 0.354, "step": 106500 }, { "epoch": 8.67, "learning_rate": 1.4060101647508274e-06, "loss": 0.3282, "step": 107000 }, { "epoch": 8.71, "learning_rate": 1.363372787120101e-06, "loss": 0.3419, "step": 107500 }, { "epoch": 8.75, "learning_rate": 1.320735409489375e-06, "loss": 0.3524, "step": 108000 }, { "epoch": 8.79, "learning_rate": 1.2780980318586486e-06, "loss": 0.3372, "step": 108500 }, { "epoch": 8.83, "learning_rate": 1.2354606542279224e-06, "loss": 0.358, "step": 109000 }, { "epoch": 8.87, "learning_rate": 1.1929085513524577e-06, "loss": 0.3426, "step": 109500 }, { "epoch": 8.91, "learning_rate": 1.150356448476993e-06, "loss": 0.3449, "step": 110000 }, { "epoch": 8.95, "learning_rate": 1.1077190708462667e-06, "loss": 0.3547, "step": 110500 }, { "epoch": 8.99, "learning_rate": 1.0650816932155406e-06, "loss": 0.347, "step": 111000 }, { "epoch": 9.0, "eval_BLEU": 50.42765483523289, "eval_BLEU-Bigram-Precision": 55.31231468669697, "eval_BLEU-Trigram-Precision": 47.11826938259908, "eval_BLEU-Unigram-Precision": 69.82722488706749, "eval_ROUGE-2": 44.880113409125535, "eval_ROUGE-L": 56.82064108206701, "eval_Sacre-Bigram-Precision": 53.78361795684631, "eval_Sacre-Trigram-Precision": 47.04437530702472, "eval_Sacre-Unigram-Precision": 68.36258716037509, "eval_SacreBLEU": 49.36901330124516, "eval_loss": 0.6147586703300476, "eval_runtime": 169.9129, "eval_samples_per_second": 7.28, "eval_steps_per_second": 7.28, "step": 111096 }, { "epoch": 9.03, "learning_rate": 1.0224443155848143e-06, "loss": 0.3398, "step": 111500 }, { "epoch": 9.07, "learning_rate": 9.798069379540882e-07, "loss": 0.3255, "step": 112000 }, { "epoch": 9.11, "learning_rate": 9.37169560323362e-07, "loss": 0.3178, "step": 112500 }, { "epoch": 9.15, "learning_rate": 8.946174574478971e-07, "loss": 0.3281, "step": 113000 }, { "epoch": 9.19, "learning_rate": 8.520653545724324e-07, "loss": 0.3261, "step": 113500 }, { "epoch": 9.24, "learning_rate": 8.094279769417062e-07, "loss": 0.3275, "step": 114000 }, { "epoch": 9.28, "learning_rate": 7.6679059931098e-07, "loss": 0.3359, "step": 114500 }, { "epoch": 9.32, "learning_rate": 7.241532216802538e-07, "loss": 0.3165, "step": 115000 }, { "epoch": 9.36, "learning_rate": 6.815158440495275e-07, "loss": 0.3199, "step": 115500 }, { "epoch": 9.4, "learning_rate": 6.388784664188015e-07, "loss": 0.332, "step": 116000 }, { "epoch": 9.44, "learning_rate": 5.962410887880752e-07, "loss": 0.3227, "step": 116500 }, { "epoch": 9.48, "learning_rate": 5.536037111573491e-07, "loss": 0.3203, "step": 117000 }, { "epoch": 9.52, "learning_rate": 5.109663335266228e-07, "loss": 0.347, "step": 117500 }, { "epoch": 9.56, "learning_rate": 4.6832895589589667e-07, "loss": 0.3485, "step": 118000 }, { "epoch": 9.6, "learning_rate": 4.2569157826517044e-07, "loss": 0.3244, "step": 118500 }, { "epoch": 9.64, "learning_rate": 3.830542006344442e-07, "loss": 0.3351, "step": 119000 }, { "epoch": 9.68, "learning_rate": 3.4041682300371804e-07, "loss": 0.3341, "step": 119500 }, { "epoch": 9.72, "learning_rate": 2.9786472012825325e-07, "loss": 0.328, "step": 120000 }, { "epoch": 9.76, "learning_rate": 2.5522734249752707e-07, "loss": 0.3447, "step": 120500 }, { "epoch": 9.8, "learning_rate": 2.1258996486680084e-07, "loss": 0.3326, "step": 121000 }, { "epoch": 9.84, "learning_rate": 1.6995258723607467e-07, "loss": 0.3259, "step": 121500 }, { "epoch": 9.88, "learning_rate": 1.274004843606099e-07, "loss": 0.3355, "step": 122000 }, { "epoch": 9.92, "learning_rate": 8.47631067298837e-08, "loss": 0.3653, "step": 122500 }, { "epoch": 9.96, "learning_rate": 4.212572909915749e-08, "loss": 0.3359, "step": 123000 }, { "epoch": 10.0, "eval_BLEU": 50.8713809636162, "eval_BLEU-Bigram-Precision": 55.25529935869192, "eval_BLEU-Trigram-Precision": 47.15997915581032, "eval_BLEU-Unigram-Precision": 69.56702363367799, "eval_ROUGE-2": 45.59251780348448, "eval_ROUGE-L": 57.14390377702154, "eval_Sacre-Bigram-Precision": 53.736780853109344, "eval_Sacre-Trigram-Precision": 47.052790672568285, "eval_Sacre-Unigram-Precision": 68.06666666666666, "eval_SacreBLEU": 49.82287256259556, "eval_loss": 0.6102247834205627, "eval_runtime": 172.6687, "eval_samples_per_second": 7.164, "eval_steps_per_second": 7.164, "step": 123440 }, { "epoch": 10.0, "step": 123440, "total_flos": 3771355808563200.0, "train_loss": 0.4878676912032082, "train_runtime": 11854.8374, "train_samples_per_second": 20.824, "train_steps_per_second": 10.413 } ], "max_steps": 123440, "num_train_epochs": 10, "total_flos": 3771355808563200.0, "trial_name": null, "trial_params": null }