mbart-large-cc25-dequad-qg / eval_pipeline /metric.first.answer.paragraph.questions_answers.lmqg_qg_dequad.default.lmqg_mbart-large-cc25-dequad-ae.json
asahi417's picture
add model
f9ae9f3
raw
history blame
800 Bytes
{"test": {"QAAlignedF1Score (BERTScore)": 0.0, "QAAlignedRecall (BERTScore)": 0.0, "QAAlignedPrecision (BERTScore)": 0.0, "QAAlignedF1Score (MoverScore)": 0.0, "QAAlignedRecall (MoverScore)": 0.0, "QAAlignedPrecision (MoverScore)": 0.0, "Bleu_1": 0.0, "Bleu_2": 0.0, "Bleu_3": 0.0, "Bleu_4": 0.0, "METEOR": 0.0, "ROUGE_L": 0.006427099173143384, "BERTScore": 0.0, "MoverScore": 0.4288403010807885}, "validation": {"QAAlignedF1Score (BERTScore)": 0.0, "QAAlignedRecall (BERTScore)": 0.0, "QAAlignedPrecision (BERTScore)": 0.0, "QAAlignedF1Score (MoverScore)": 0.0, "QAAlignedRecall (MoverScore)": 0.0, "QAAlignedPrecision (MoverScore)": 0.0, "Bleu_1": 0.0, "Bleu_2": 0.0, "Bleu_3": 0.0, "Bleu_4": 0.0, "METEOR": 0.0, "ROUGE_L": 0.011010381923259495, "BERTScore": 0.0, "MoverScore": 0.4333211633438749}}