mt5-small-dequad-qg-ae-trimmed-50000 / eval /metric.first.answer.paragraph_sentence.answer.lmqg_qg_dequad.default.json
asahi417's picture
add eval
dca775a
raw history blame
No virus
683 Bytes
{"validation": {"Bleu_1": 0.2512808570097713, "Bleu_2": 0.17200666979618856, "Bleu_3": 0.11601037389714632, "Bleu_4": 0.07833093186130648, "METEOR": 0.27241316698312057, "ROUGE_L": 0.2442186146594222, "BERTScore": 0.7912570112089063, "MoverScore": 0.6111877039173896, "AnswerF1Score": 47.97914329982987, "AnswerExactMatch": 13.203266787658801}, "test": {"Bleu_1": 0.19504937685970955, "Bleu_2": 0.13013436077051674, "Bleu_3": 0.08631932690075304, "Bleu_4": 0.056732079759815104, "METEOR": 0.2266378098231599, "ROUGE_L": 0.18189892033201369, "BERTScore": 0.7504787875336009, "MoverScore": 0.5779213304061341, "AnswerF1Score": 38.08332842395432, "AnswerExactMatch": 9.936479128856623}}