asahi417 commited on
Commit
1744e25
1 Parent(s): d22ab8c
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "lmqg_output/mbart_large_cc25_itquad/model_dytzou/epoch_5",
3
  "_num_labels": 3,
4
  "activation_dropout": 0.0,
5
  "activation_function": "gelu",
@@ -54,7 +54,7 @@
54
  }
55
  },
56
  "torch_dtype": "float32",
57
- "transformers_version": "4.18.0.dev0",
58
  "use_cache": true,
59
  "vocab_size": 250028
60
  }
 
1
  {
2
+ "_name_or_path": "mbart-large-cc25-itquad",
3
  "_num_labels": 3,
4
  "activation_dropout": 0.0,
5
  "activation_function": "gelu",
 
54
  }
55
  },
56
  "torch_dtype": "float32",
57
+ "transformers_version": "4.21.2",
58
  "use_cache": true,
59
  "vocab_size": 250028
60
  }
eval/metric.first.answer.paragraph.questions_answers.lmqg_qg_itquad.default.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"test": {"QAAlignedF1Score (BERTScore)": 0.875555217179592, "QAAlignedF1Score (MoverScore)": 0.6170629909692065}, "validation": {"QAAlignedF1Score (BERTScore)": 0.8788252032406213, "QAAlignedF1Score (MoverScore)": 0.6235306562971487}}
eval/samples.test.hyp.paragraph.questions_answers.lmqg_qg_itquad.default.txt ADDED
The diff for this file is too large to render. See raw diff
 
eval/samples.validation.hyp.paragraph.questions_answers.lmqg_qg_itquad.default.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c43c7f44f6a7c9280ed3d9b23ecda9d5551221ccd776c581e56920ef9d1d72c
3
- size 2444604857
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f81469ca728b6106bda0ea8d5e3c51e706141e55425771bbcf99e4635b862cb1
3
+ size 2444580125