add eval
Browse files
eval/metric.first.answer.paragraph_question.answer.lmqg_qg_frquad.default.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"validation": {"Bleu_1": 0.20630618421646085, "Bleu_2": 0.1716433062676975, "Bleu_3": 0.14680747202796066, "Bleu_4": 0.12707841211176438, "METEOR": 0.20920864291094007, "ROUGE_L": 0.253747180116882, "BERTScore": 0.
|
|
|
1 |
+
{"validation": {"Bleu_1": 0.20630618421646085, "Bleu_2": 0.1716433062676975, "Bleu_3": 0.14680747202796066, "Bleu_4": 0.12707841211176438, "METEOR": 0.20920864291094007, "ROUGE_L": 0.253747180116882, "BERTScore": 0.7719727669990946, "MoverScore": 0.6084405239603016, "AnswerF1Score": 38.15330964667721, "AnswerExactMatch": 16.844416562107906}, "test": {"Bleu_1": 0.19028751396930862, "Bleu_2": 0.15474020614488332, "Bleu_3": 0.13008980990866909, "Bleu_4": 0.10972487229837086, "METEOR": 0.2105031358590392, "ROUGE_L": 0.2566066040631531, "BERTScore": 0.7772796637982766, "MoverScore": 0.6283564679101659, "AnswerF1Score": 39.53952694670434, "AnswerExactMatch": 22.396486825595986}}
|