add eval
Browse files
eval/metric.first.answer.paragraph_question.answer.lmqg_qg_ruquad.default.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"validation": {"Bleu_1": 0.4945611485444799, "Bleu_2": 0.4346738248850003, "Bleu_3": 0.3810974312090755, "Bleu_4": 0.32977021275327834, "METEOR": 0.4161452376499569, "ROUGE_L": 0.5651467746610948, "BERTScore": 0.9019812614057062, "MoverScore": 0.7864831029648792, "AnswerF1Score": 75.01416765127489, "AnswerExactMatch": 52.16441620333598}, "test": {"Bleu_1": 0.48214205891553874, "Bleu_2": 0.42108103391914214, "Bleu_3": 0.36684642587019944, "Bleu_4": 0.3152335901828712, "METEOR": 0.4129547849186712, "ROUGE_L": 0.5520388285909801, "BERTScore": 0.897205683620938, "MoverScore": 0.7830196570751001, "AnswerF1Score": 73.82039508473352, "AnswerExactMatch": 51.389992057188245}}
|
eval/samples.test.hyp.paragraph_question.answer.lmqg_qg_ruquad.default.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval/samples.validation.hyp.paragraph_question.answer.lmqg_qg_ruquad.default.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|