asahi417 commited on
Commit
c8534cc
1 Parent(s): 52e73f1
eval/metric.first.answer.paragraph_question.answer.lmqg_qg_frquad.default.json CHANGED
@@ -1 +1 @@
1
- {"validation": {"Bleu_1": 0.2637493389740703, "Bleu_2": 0.21066068639051885, "Bleu_3": 0.17206804116857294, "Bleu_4": 0.14276567713384303, "METEOR": 0.17844974741518205, "ROUGE_L": 0.2627081546062905, "BERTScore": 0.7839096197052163, "MoverScore": 0.5959817925573929, "AnswerF1Score": 36.5236808721106, "AnswerExactMatch": 13.676286072772898}, "test": {"Bleu_1": 0.230710939652781, "Bleu_2": 0.18158538794732307, "Bleu_3": 0.14770036160029532, "Bleu_4": 0.12147648827514447, "METEOR": 0.18559904458777157, "ROUGE_L": 0.25279893726224967, "BERTScore": 0.786680847266163, "MoverScore": 0.6173578350334938, "AnswerF1Score": 36.756352271513876, "AnswerExactMatch": 19.040150564617313}}
 
1
+ {"validation": {"Bleu_1": 0.40855381475963315, "Bleu_2": 0.36332457798239165, "Bleu_3": 0.3285593716275254, "Bleu_4": 0.29723063946667544, "METEOR": 0.2953107841875028, "ROUGE_L": 0.39258261253473803, "BERTScore": 0.8661458669709441, "MoverScore": 0.7099813465360924, "AnswerF1Score": 61.63741999952982, "AnswerExactMatch": 35.53952321204517}, "test": {"Bleu_1": 0.3856713195033249, "Bleu_2": 0.33900509275323054, "Bleu_3": 0.30536763566832525, "Bleu_4": 0.27646202677726517, "METEOR": 0.3184462230882313, "ROUGE_L": 0.384840006984747, "BERTScore": 0.8666755957353669, "MoverScore": 0.7286766962208026, "AnswerF1Score": 61.093253683030376, "AnswerExactMatch": 39.89962358845671}}
eval/samples.test.hyp.paragraph_question.answer.lmqg_qg_frquad.default.txt CHANGED
The diff for this file is too large to render. See raw diff
 
eval/samples.validation.hyp.paragraph_question.answer.lmqg_qg_frquad.default.txt CHANGED
The diff for this file is too large to render. See raw diff