model update
Browse files- config.json +1 -1
- eval/metric.first.answer.paragraph_answer.question.asahi417_qg_esquad.default.json +1 -0
- eval/metric.first.sentence.paragraph_answer.question.asahi417_qg_esquad.default.json +1 -0
- eval/samples.test.hyp.paragraph_answer.question.asahi417_qg_esquad.default.txt +0 -0
- eval/samples.validation.hyp.paragraph_answer.question.asahi417_qg_esquad.default.txt +0 -0
- pytorch_model.bin +2 -2
- tokenizer_config.json +1 -1
- trainer_config.json +1 -0
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"_num_labels": 3,
|
4 |
"activation_dropout": 0.0,
|
5 |
"activation_function": "gelu",
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "facebook/mbart-large-cc25",
|
3 |
"_num_labels": 3,
|
4 |
"activation_dropout": 0.0,
|
5 |
"activation_function": "gelu",
|
eval/metric.first.answer.paragraph_answer.question.asahi417_qg_esquad.default.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"validation": {"Bleu_1": 0.24590404196235538, "Bleu_2": 0.16548301272325436, "Bleu_3": 0.11907285456027517, "Bleu_4": 0.08821029247980323}, "test": {"Bleu_1": 0.24927547676392584, "Bleu_2": 0.16861023137100342, "Bleu_3": 0.12227582559531344, "Bleu_4": 0.09135186262656596}}
|
eval/metric.first.sentence.paragraph_answer.question.asahi417_qg_esquad.default.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"validation": {"Bleu_1": 0.25707402091449444, "Bleu_2": 0.1743504114263434, "Bleu_3": 0.12625976443161002, "Bleu_4": 0.09394013238192139, "METEOR": 0.2280763366224896, "ROUGE_L": 0.24469255710139245, "BERTScore": 0.8317060463417784, "MoverScore": 0.5834830953332713}, "test": {"Bleu_1": 0.25016765128532453, "Bleu_2": 0.16927851225254284, "Bleu_3": 0.12282901763682275, "Bleu_4": 0.09181162637548275, "METEOR": 0.2294718580484058, "ROUGE_L": 0.24257746651996367, "BERTScore": 0.8358409249373304, "MoverScore": 0.5890867174367108}}
|
eval/samples.test.hyp.paragraph_answer.question.asahi417_qg_esquad.default.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval/samples.validation.hyp.paragraph_answer.question.asahi417_qg_esquad.default.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7ea0df80cd7f2dbf1cd1932ffa669b2907f5961a534296bceb011f99352f8b64
|
3 |
+
size 2444604857
|
tokenizer_config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "src_lang": null, "tgt_lang": null, "additional_special_tokens": null, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "
|
|
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "src_lang": null, "tgt_lang": null, "additional_special_tokens": null, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "facebook/mbart-large-cc25", "tokenizer_class": "MBartTokenizer"}
|
trainer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dataset_path": "asahi417/qg_esquad", "dataset_name": "default", "input_types": ["paragraph_answer"], "output_types": ["question"], "prefix_types": null, "model": "facebook/mbart-large-cc25", "max_length": 512, "max_length_output": 32, "epoch": 4, "batch": 4, "lr": 0.0001, "fp16": false, "random_seed": 1, "gradient_accumulation_steps": 16, "label_smoothing": 0.15}
|