mt5-small-dequad-ae / eval /metric.first.answer.paragraph_sentence.answer.lmqg_qg_dequad.default.json
asahi417's picture
model update
9787d2a
raw
history blame
682 Bytes
{"validation": {"Bleu_1": 0.24304991686964997, "Bleu_2": 0.16743695376367998, "Bleu_3": 0.11338395369130196, "Bleu_4": 0.07649351478718919, "METEOR": 0.26688192488935053, "ROUGE_L": 0.23337152266287337, "BERTScore": 0.7811309026584651, "MoverScore": 0.6005610418847153, "AnswerF1Score": 45.73879822286475, "AnswerExactMatch": 11.88747731397459}, "test": {"Bleu_1": 0.17773225186051603, "Bleu_2": 0.11729974329463738, "Bleu_3": 0.07735644104838064, "Bleu_4": 0.05113872186645225, "METEOR": 0.21088348840079052, "ROUGE_L": 0.17542653733220975, "BERTScore": 0.74031749895069, "MoverScore": 0.5682240358751466, "AnswerF1Score": 36.06844480141066, "AnswerExactMatch": 8.802177858439201}}