MarianCG-CoNaLa / trainer_state.json
AhmedSSoliman's picture
Upload trainer_state.json
0e61c45
{
"best_metric": 0.06889855116605759,
"best_model_checkpoint": "./models/1/checkpoint-9737",
"epoch": 10.0,
"global_step": 13910,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.36,
"learning_rate": 3.591954022988506e-05,
"loss": 0.0876,
"step": 500
},
{
"epoch": 0.72,
"learning_rate": 4.884970485848343e-05,
"loss": 0.0879,
"step": 1000
},
{
"epoch": 1.0,
"eval_BLEU": 8.064725756490624,
"eval_BLEU-Bigram-Precision": 10.432874977383753,
"eval_BLEU-Trigram-Precision": 5.407291933269737,
"eval_BLEU-Unigram-Precision": 21.631135458605595,
"eval_ROUGE-2": 12.39926519031995,
"eval_ROUGE-L": 24.731962013373813,
"eval_Sacre-Bigram-Precision": 8.552870639534884,
"eval_Sacre-Trigram-Precision": 4.851033999298983,
"eval_Sacre-Unigram-Precision": 19.295765314011796,
"eval_SacreBLEU": 7.1292786955103695,
"eval_loss": 0.0844602882862091,
"eval_runtime": 523.4808,
"eval_samples_per_second": 2.363,
"eval_steps_per_second": 0.296,
"step": 1391
},
{
"epoch": 1.08,
"learning_rate": 4.6961555925533526e-05,
"loss": 0.0829,
"step": 1500
},
{
"epoch": 1.44,
"learning_rate": 4.506962312698653e-05,
"loss": 0.0747,
"step": 2000
},
{
"epoch": 1.8,
"learning_rate": 4.317769032843954e-05,
"loss": 0.0738,
"step": 2500
},
{
"epoch": 2.0,
"eval_BLEU": 7.726645595861129,
"eval_BLEU-Bigram-Precision": 9.779983147645352,
"eval_BLEU-Trigram-Precision": 5.376859377396105,
"eval_BLEU-Unigram-Precision": 19.00918707221551,
"eval_ROUGE-2": 14.567403936228773,
"eval_ROUGE-L": 26.962145889007623,
"eval_Sacre-Bigram-Precision": 8.340425531914894,
"eval_Sacre-Trigram-Precision": 5.06410008133545,
"eval_Sacre-Unigram-Precision": 17.33626552335896,
"eval_SacreBLEU": 7.08253441311445,
"eval_loss": 0.07733357697725296,
"eval_runtime": 594.316,
"eval_samples_per_second": 2.081,
"eval_steps_per_second": 0.261,
"step": 2782
},
{
"epoch": 2.16,
"learning_rate": 4.1285757529892535e-05,
"loss": 0.0669,
"step": 3000
},
{
"epoch": 2.52,
"learning_rate": 3.939382473134555e-05,
"loss": 0.0631,
"step": 3500
},
{
"epoch": 2.88,
"learning_rate": 3.7505675798395645e-05,
"loss": 0.0616,
"step": 4000
},
{
"epoch": 3.0,
"eval_BLEU": 13.75944614514078,
"eval_BLEU-Bigram-Precision": 17.05563909774436,
"eval_BLEU-Trigram-Precision": 9.895979758223222,
"eval_BLEU-Unigram-Precision": 31.226259170122074,
"eval_ROUGE-2": 16.762154637582928,
"eval_ROUGE-L": 29.733231785857146,
"eval_Sacre-Bigram-Precision": 14.784847370356749,
"eval_Sacre-Trigram-Precision": 9.445382434455736,
"eval_Sacre-Unigram-Precision": 28.888232688616565,
"eval_SacreBLEU": 12.80566461172721,
"eval_loss": 0.07288952171802521,
"eval_runtime": 656.3315,
"eval_samples_per_second": 1.885,
"eval_steps_per_second": 0.236,
"step": 4173
},
{
"epoch": 3.24,
"learning_rate": 3.561374299984865e-05,
"loss": 0.0556,
"step": 4500
},
{
"epoch": 3.59,
"learning_rate": 3.372181020130165e-05,
"loss": 0.0522,
"step": 5000
},
{
"epoch": 3.95,
"learning_rate": 3.1829877402754655e-05,
"loss": 0.0541,
"step": 5500
},
{
"epoch": 4.0,
"eval_BLEU": 12.584990506585802,
"eval_BLEU-Bigram-Precision": 15.284490474363894,
"eval_BLEU-Trigram-Precision": 9.189817259955635,
"eval_BLEU-Unigram-Precision": 27.802290416935204,
"eval_ROUGE-2": 18.2427503756888,
"eval_ROUGE-L": 31.248445460832464,
"eval_Sacre-Bigram-Precision": 13.375350140056023,
"eval_Sacre-Trigram-Precision": 8.759613065730592,
"eval_Sacre-Unigram-Precision": 25.743221169552434,
"eval_SacreBLEU": 11.731554660360798,
"eval_loss": 0.07118678092956543,
"eval_runtime": 463.4231,
"eval_samples_per_second": 2.669,
"eval_steps_per_second": 0.334,
"step": 5564
},
{
"epoch": 4.31,
"learning_rate": 2.9937944604207657e-05,
"loss": 0.0464,
"step": 6000
},
{
"epoch": 4.67,
"learning_rate": 2.8046011805660665e-05,
"loss": 0.0459,
"step": 6500
},
{
"epoch": 5.0,
"eval_BLEU": 15.1776190124192,
"eval_BLEU-Bigram-Precision": 18.32776890695311,
"eval_BLEU-Trigram-Precision": 11.247289809684414,
"eval_BLEU-Unigram-Precision": 32.42811501597444,
"eval_ROUGE-2": 17.936785005872903,
"eval_ROUGE-L": 31.307371141735135,
"eval_Sacre-Bigram-Precision": 16.26334519572954,
"eval_Sacre-Trigram-Precision": 10.753317119724164,
"eval_Sacre-Unigram-Precision": 30.265755070515205,
"eval_SacreBLEU": 14.242434580808037,
"eval_loss": 0.06986960768699646,
"eval_runtime": 398.2357,
"eval_samples_per_second": 3.106,
"eval_steps_per_second": 0.389,
"step": 6955
},
{
"epoch": 5.03,
"learning_rate": 2.6154079007113667e-05,
"loss": 0.0455,
"step": 7000
},
{
"epoch": 5.39,
"learning_rate": 2.4262146208566673e-05,
"loss": 0.0393,
"step": 7500
},
{
"epoch": 5.75,
"learning_rate": 2.2370213410019678e-05,
"loss": 0.0405,
"step": 8000
},
{
"epoch": 6.0,
"eval_BLEU": 17.03804435280673,
"eval_BLEU-Bigram-Precision": 20.16791440608614,
"eval_BLEU-Trigram-Precision": 12.76924045849603,
"eval_BLEU-Unigram-Precision": 34.79696172947707,
"eval_ROUGE-2": 20.00404092356461,
"eval_ROUGE-L": 33.44410880996549,
"eval_Sacre-Bigram-Precision": 18.140892297184017,
"eval_Sacre-Trigram-Precision": 12.388609066253391,
"eval_Sacre-Unigram-Precision": 32.66758356754171,
"eval_SacreBLEU": 16.183245622226913,
"eval_loss": 0.06919773668050766,
"eval_runtime": 389.7744,
"eval_samples_per_second": 3.174,
"eval_steps_per_second": 0.398,
"step": 8346
},
{
"epoch": 6.11,
"learning_rate": 2.047828061147268e-05,
"loss": 0.0395,
"step": 8500
},
{
"epoch": 6.47,
"learning_rate": 1.8586347812925686e-05,
"loss": 0.0351,
"step": 9000
},
{
"epoch": 6.83,
"learning_rate": 1.669441501437869e-05,
"loss": 0.0356,
"step": 9500
},
{
"epoch": 7.0,
"eval_BLEU": 18.45635959837559,
"eval_BLEU-Bigram-Precision": 21.599746072052056,
"eval_BLEU-Trigram-Precision": 13.98176291793313,
"eval_BLEU-Unigram-Precision": 36.946429662207564,
"eval_ROUGE-2": 20.663766964343406,
"eval_ROUGE-L": 33.85196384000605,
"eval_Sacre-Bigram-Precision": 19.43390720747009,
"eval_Sacre-Trigram-Precision": 13.581706410862664,
"eval_Sacre-Unigram-Precision": 34.678927680798004,
"eval_SacreBLEU": 17.534804534318486,
"eval_loss": 0.06889855116605759,
"eval_runtime": 289.9351,
"eval_samples_per_second": 4.266,
"eval_steps_per_second": 0.535,
"step": 9737
},
{
"epoch": 7.19,
"learning_rate": 1.4802482215831695e-05,
"loss": 0.0324,
"step": 10000
},
{
"epoch": 7.55,
"learning_rate": 1.2910549417284698e-05,
"loss": 0.032,
"step": 10500
},
{
"epoch": 7.91,
"learning_rate": 1.1018616618737702e-05,
"loss": 0.0316,
"step": 11000
},
{
"epoch": 8.0,
"eval_BLEU": 23.036610051092996,
"eval_BLEU-Bigram-Precision": 26.881516587677723,
"eval_BLEU-Trigram-Precision": 17.706261436868488,
"eval_BLEU-Unigram-Precision": 44.332174416920175,
"eval_ROUGE-2": 21.29088469404566,
"eval_ROUGE-L": 35.036834725440904,
"eval_Sacre-Bigram-Precision": 24.360966513868387,
"eval_Sacre-Trigram-Precision": 17.249540910018364,
"eval_Sacre-Unigram-Precision": 41.78435079135624,
"eval_SacreBLEU": 21.988824762940034,
"eval_loss": 0.0689924880862236,
"eval_runtime": 259.2001,
"eval_samples_per_second": 4.772,
"eval_steps_per_second": 0.598,
"step": 11128
},
{
"epoch": 8.27,
"learning_rate": 9.126683820190707e-06,
"loss": 0.0295,
"step": 11500
},
{
"epoch": 8.63,
"learning_rate": 7.238534887240806e-06,
"loss": 0.0283,
"step": 12000
},
{
"epoch": 8.99,
"learning_rate": 5.3466020886938095e-06,
"loss": 0.0287,
"step": 12500
},
{
"epoch": 9.0,
"eval_BLEU": 26.995788994612766,
"eval_BLEU-Bigram-Precision": 30.950083340643914,
"eval_BLEU-Trigram-Precision": 20.98232920550995,
"eval_BLEU-Unigram-Precision": 50.522155190347405,
"eval_ROUGE-2": 22.105956241069418,
"eval_ROUGE-L": 35.85099346344104,
"eval_Sacre-Bigram-Precision": 28.175123119324088,
"eval_Sacre-Trigram-Precision": 20.573315467075037,
"eval_Sacre-Unigram-Precision": 47.89045785194694,
"eval_SacreBLEU": 25.924953054404842,
"eval_loss": 0.06922509521245956,
"eval_runtime": 198.6398,
"eval_samples_per_second": 6.227,
"eval_steps_per_second": 0.78,
"step": 12519
},
{
"epoch": 9.35,
"learning_rate": 3.454669290146814e-06,
"loss": 0.0271,
"step": 13000
},
{
"epoch": 9.71,
"learning_rate": 1.5627364915998186e-06,
"loss": 0.0265,
"step": 13500
},
{
"epoch": 10.0,
"eval_BLEU": 26.712396231914138,
"eval_BLEU-Bigram-Precision": 30.61851211072664,
"eval_BLEU-Trigram-Precision": 20.719279806242287,
"eval_BLEU-Unigram-Precision": 49.895307303855155,
"eval_ROUGE-2": 22.458649124626863,
"eval_ROUGE-L": 36.256812107261126,
"eval_Sacre-Bigram-Precision": 27.962698554345884,
"eval_Sacre-Trigram-Precision": 20.387249114521843,
"eval_Sacre-Unigram-Precision": 47.333925324538036,
"eval_SacreBLEU": 25.70732578879903,
"eval_loss": 0.06933248043060303,
"eval_runtime": 190.4819,
"eval_samples_per_second": 6.494,
"eval_steps_per_second": 0.814,
"step": 13910
}
],
"max_steps": 13910,
"num_train_epochs": 10,
"total_flos": 3.016954478592e+16,
"trial_name": null,
"trial_params": null
}