mt5-base-frquad-qag / eval /metric.first.answer.paragraph.questions_answers.lmqg_qag_frquad.default.json
asahi417's picture
model update
c6f0d6e
raw
history blame
1.17 kB
{"validation": {"Bleu_1": 0.22517424732341337, "Bleu_2": 0.10286865005351412, "Bleu_3": 0.025911984283044087, "Bleu_4": 0.008628308013417189, "METEOR": 0.16391078547569096, "ROUGE_L": 0.22500306125834743, "BERTScore": 0.6937287280646464, "MoverScore": 0.5009121050083953, "QAAlignedF1Score (BERTScore)": 0.7932147704389645, "QAAlignedRecall (BERTScore)": 0.7820358078685348, "QAAlignedPrecision (BERTScore)": 0.8048385219622611, "QAAlignedF1Score (MoverScore)": 0.5252035901899432, "QAAlignedRecall (MoverScore)": 0.5182132561321993, "QAAlignedPrecision (MoverScore)": 0.5324733137198794}, "test": {"Bleu_1": 0.0721618976075026, "Bleu_2": 0.02909491652705291, "Bleu_3": 0.013053227853060874, "Bleu_4": 0.004174278072327164, "METEOR": 0.12473174901360141, "ROUGE_L": 0.12571268329583962, "BERTScore": 0.6249730437314149, "MoverScore": 0.4902676659649159, "QAAlignedF1Score (BERTScore)": 0.7828301996020194, "QAAlignedRecall (BERTScore)": 0.7820959003608982, "QAAlignedPrecision (BERTScore)": 0.7836373862835541, "QAAlignedF1Score (MoverScore)": 0.5165951049621047, "QAAlignedRecall (MoverScore)": 0.5159255457120212, "QAAlignedPrecision (MoverScore)": 0.5172958670342138}}