mt5-base-koquad-qg / eval /metric.first.answer.paragraph.questions_answers.lmqg_qg_koquad.default.json
asahi417's picture
add model
ae99cb7
raw
history blame
1.16 kB
{"test": {"QAAlignedF1Score (BERTScore)": 0.8879773283296034, "QAAlignedRecall (BERTScore)": 0.8876320042113061, "QAAlignedPrecision (BERTScore)": 0.8883508750018496, "QAAlignedF1Score (MoverScore)": 0.8593153010367547, "QAAlignedRecall (MoverScore)": 0.8586677990508506, "QAAlignedPrecision (MoverScore)": 0.8601344155347183, "Bleu_1": 0.4394748521330362, "Bleu_2": 0.32465203754289673, "Bleu_3": 0.20337457563910152, "Bleu_4": 0.12077139165694208, "METEOR": 0.38749180601824296, "ROUGE_L": 0.4869929002020682, "BERTScore": 0.884951471504141, "MoverScore": 0.85179271311584}, "validation": {"QAAlignedF1Score (BERTScore)": 0.8883957551986126, "QAAlignedRecall (BERTScore)": 0.8848137852149922, "QAAlignedPrecision (BERTScore)": 0.8921166759873519, "QAAlignedF1Score (MoverScore)": 0.902176629872783, "QAAlignedRecall (MoverScore)": 0.8953159851651873, "QAAlignedPrecision (MoverScore)": 0.9096823630920005, "Bleu_1": 0.4824664083497727, "Bleu_2": 0.39437396741633923, "Bleu_3": 0.32204888184110353, "Bleu_4": 0.25860908999010057, "METEOR": 0.38855366878584546, "ROUGE_L": 0.501580491566324, "BERTScore": 0.8590216088419159, "MoverScore": 0.7976418468525899}}