add model
Browse files- eval/metric.first.answer.paragraph.questions_answers.lmqg_qag_tweetqa.default.json +1 -0
- eval/metric.first.sentence.paragraph.questions_answers.lmqg_qag_tweetqa.default.json +1 -0
- eval/samples.test.hyp.paragraph.questions_answers.lmqg_qag_tweetqa.default.txt +0 -0
- eval/samples.validation.hyp.paragraph.questions_answers.lmqg_qag_tweetqa.default.txt +0 -0
- tokenizer_config.json +1 -1
- trainer_config.json +1 -0
eval/metric.first.answer.paragraph.questions_answers.lmqg_qag_tweetqa.default.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"validation": {"Bleu_1": 0.0005127722896463237, "Bleu_2": 9.431228240232113e-05, "Bleu_3": 2.7076339389272802e-05, "Bleu_4": 1.2863502556787301e-05}, "test": {"Bleu_1": 0.00018828430858505577, "Bleu_2": 2.4223408658113176e-05, "Bleu_3": 4.712010310605882e-11, "Bleu_4": 6.621851429116013e-14}}
|
eval/metric.first.sentence.paragraph.questions_answers.lmqg_qag_tweetqa.default.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"validation": {"Bleu_1": 7.2276591445965386e-09, "Bleu_2": 1.6679000201921992e-09, "Bleu_3": 3.804491341366487e-10, "Bleu_4": 3.254677449875007e-14, "METEOR": 0.0013770117051312212, "ROUGE_L": 0.002269460780863736, "BERTScore": 0.013776266364335488, "MoverScore": 0.45626310068280873}, "test": {"Bleu_1": 0.0003232232106450975, "Bleu_2": 5.3147257264241544e-05, "Bleu_3": 1.0831463269382036e-05, "Bleu_4": 8.760988157508385e-10, "METEOR": 0.0028425982865562323, "ROUGE_L": 0.004925321600227637, "BERTScore": 0.03894749127879461, "MoverScore": 0.45595221244168144}}
|
eval/samples.test.hyp.paragraph.questions_answers.lmqg_qag_tweetqa.default.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval/samples.validation.hyp.paragraph.questions_answers.lmqg_qag_tweetqa.default.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
CHANGED
@@ -104,7 +104,7 @@
|
|
104 |
"eos_token": "</s>",
|
105 |
"extra_ids": 100,
|
106 |
"model_max_length": 512,
|
107 |
-
"name_or_path": "lmqg_output/t5_base_tweetqa/
|
108 |
"pad_token": "<pad>",
|
109 |
"special_tokens_map_file": null,
|
110 |
"tokenizer_class": "T5Tokenizer",
|
|
|
104 |
"eos_token": "</s>",
|
105 |
"extra_ids": 100,
|
106 |
"model_max_length": 512,
|
107 |
+
"name_or_path": "lmqg_output/t5_base_tweetqa/model_btmsre/epoch_10",
|
108 |
"pad_token": "<pad>",
|
109 |
"special_tokens_map_file": null,
|
110 |
"tokenizer_class": "T5Tokenizer",
|
trainer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dataset_path": "lmqg/qag_tweetqa", "dataset_name": "default", "input_types": ["paragraph"], "output_types": ["questions_answers"], "prefix_types": ["qag"], "model": "t5-base", "max_length": 256, "max_length_output": 128, "epoch": 12, "batch": 64, "lr": 0.0001, "fp16": false, "random_seed": 1, "gradient_accumulation_steps": 2, "label_smoothing": 0.15}
|