asahi417 commited on
Commit
b28087d
1 Parent(s): 39d2415
eval_pipeline/metric.first.answer.paragraph.questions_answers.lmqg_qg_dequad.default.lmqg_mt5-small-dequad-ae.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"test": {"QAAlignedF1Score (BERTScore)": 0.8118804193536356, "QAAlignedRecall (BERTScore)": 0.824596523132044, "QAAlignedPrecision (BERTScore)": 0.8000467229518743, "QAAlignedF1Score (MoverScore)": 0.5429548995572906, "QAAlignedRecall (MoverScore)": 0.5459173346459335, "QAAlignedPrecision (MoverScore)": 0.5403561420586401, "Bleu_1": 0.14647601032873211, "Bleu_2": 0.071486353544086, "Bleu_3": 0.022272293555467507, "Bleu_4": 0.00885217825293781, "METEOR": 0.20322221925643377, "ROUGE_L": 0.16851015313328144, "BERTScore": 0.7315713914134835, "MoverScore": 0.5263126138033825}, "validation": {"QAAlignedF1Score (BERTScore)": 0.8096359747582013, "QAAlignedRecall (BERTScore)": 0.8379781697123785, "QAAlignedPrecision (BERTScore)": 0.7836877909563646, "QAAlignedF1Score (MoverScore)": 0.5389558164814849, "QAAlignedRecall (MoverScore)": 0.5521757089230347, "QAAlignedPrecision (MoverScore)": 0.5269144537976915, "Bleu_1": 0.07281397671337796, "Bleu_2": 0.029468376728496903, "Bleu_3": 0.0091111496579868, "Bleu_4": 0.00352671887506014, "METEOR": 0.16116780031098193, "ROUGE_L": 0.10880543984627178, "BERTScore": 0.6731072370275896, "MoverScore": 0.5097959351635637}}
eval_pipeline/samples.test.hyp.paragraph.questions_answers.lmqg_qg_dequad.default.lmqg_mt5-small-dequad-ae.txt ADDED
The diff for this file is too large to render. See raw diff
 
eval_pipeline/samples.validation.hyp.paragraph.questions_answers.lmqg_qg_dequad.default.lmqg_mt5-small-dequad-ae.txt ADDED
The diff for this file is too large to render. See raw diff