asahi417 commited on
Commit
c1e8c3d
1 Parent(s): 081cd8c
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "facebook/mbart-large-cc25",
3
  "_num_labels": 3,
4
  "activation_dropout": 0.0,
5
  "activation_function": "gelu",
@@ -54,7 +54,7 @@
54
  }
55
  },
56
  "torch_dtype": "float32",
57
- "transformers_version": "4.18.0.dev0",
58
  "use_cache": true,
59
  "vocab_size": 250028
60
  }
 
1
  {
2
+ "_name_or_path": "mbart-large-cc25-esquad",
3
  "_num_labels": 3,
4
  "activation_dropout": 0.0,
5
  "activation_function": "gelu",
 
54
  }
55
  },
56
  "torch_dtype": "float32",
57
+ "transformers_version": "4.21.2",
58
  "use_cache": true,
59
  "vocab_size": 250028
60
  }
eval/metric.first.answer.paragraph.questions_answers.lmqg_qg_esquad.default.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"test": {"QAAlignedF1Score (BERTScore)": 0.8898626829475537, "QAAlignedF1Score (MoverScore)": 0.6347281318957313}, "validation": {"QAAlignedF1Score (BERTScore)": 0.8896467358321192, "QAAlignedF1Score (MoverScore)": 0.6299539346242563}}
eval/samples.test.hyp.paragraph.questions_answers.lmqg_qg_esquad.default.txt ADDED
The diff for this file is too large to render. See raw diff
 
eval/samples.validation.hyp.paragraph.questions_answers.lmqg_qg_esquad.default.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ea0df80cd7f2dbf1cd1932ffa669b2907f5961a534296bceb011f99352f8b64
3
- size 2444604857
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb349e84c0f87f0a46469779bf246c90dbde5dbb188b5d040be1bed6aeff6779
3
+ size 2444580125