asahi417 commited on
Commit
70c0120
1 Parent(s): d770b29
eval/metric.first.answer.paragraph_question.answer.lmqg_qg_squad.default.json CHANGED
@@ -1 +1 @@
1
- {"validation": {"Bleu_1": 0.07891057297554473, "Bleu_2": 0.03464868675374984, "Bleu_3": 0.01634889499019619, "Bleu_4": 0.008113909423249847, "METEOR": 0.057786868650397914, "ROUGE_L": 0.13587226480436446, "BERTScore": 0.8517762487636158, "MoverScore": 0.5822567537286933, "AnswerF1Score": 12.723727625739254, "AnswerExactMatch": 5.146641438032167}, "test": {"Bleu_1": 0.09227894109510289, "Bleu_2": 0.04400800264499248, "Bleu_3": 0.022221474281658965, "Bleu_4": 0.011408101358060582, "METEOR": 0.06864302117467744, "ROUGE_L": 0.1646497404271194, "BERTScore": 0.8565609899245299, "MoverScore": 0.5999087363667507, "AnswerF1Score": 15.681779041571348, "AnswerExactMatch": 7.451376610255115}}
 
1
+ {"validation": {"Bleu_1": 0.6656879677604876, "Bleu_2": 0.6172511817641129, "Bleu_3": 0.5734198683580137, "Bleu_4": 0.5352610964286729, "METEOR": 0.409054002223627, "ROUGE_L": 0.7222111409086849, "BERTScore": 0.9259103927560426, "MoverScore": 0.8161147888838659, "AnswerF1Score": 73.82930982741765, "AnswerExactMatch": 56.74550614947966}, "test": {"Bleu_1": 0.6976384158527121, "Bleu_2": 0.650603492248607, "Bleu_3": 0.6057204607806849, "Bleu_4": 0.5666236819074097, "METEOR": 0.43171629039719916, "ROUGE_L": 0.7477038656182274, "BERTScore": 0.9271523304094799, "MoverScore": 0.8405408862275872, "AnswerF1Score": 77.0082614591304, "AnswerExactMatch": 62.71785804496085}}
eval/samples.test.hyp.paragraph_question.answer.lmqg_qg_squad.default.txt CHANGED
The diff for this file is too large to render. See raw diff
 
eval/samples.validation.hyp.paragraph_question.answer.lmqg_qg_squad.default.txt CHANGED
The diff for this file is too large to render. See raw diff