khuangaf commited on
Commit
b6562c0
1 Parent(s): 901e04b

upload everything

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ optimizer.pt filter=lfs diff=lfs merge=lfs -text
37
+ pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: en
3
+ widget:
4
+ - text: What is Night of the Living Dead? \n Night of the Living Dead is a 1968 American independent horror film , directed by George A. Romero , starring Duane Jones and Judith O'Dea . George A. Romero George A. Romero Duane Jones Duane Jones Judith O'Dea Judith O'Dea independent Independent film horror film horror film.
5
+ ---
6
+
7
+ # Domain-adapted QA Model From ZeroFEC
8
+
9
+ ZeroFEC is a faithful and interpetable factual error correction framework introduced in the paper [Zero-shot Faithful Factual Error Correction](https://aclanthology.org/2023.acl-long.311/). It involves a QA component, which is a UnifiedQA model continue fine-tuned on two additional biomedical QA datasets. The associated code is released in [this](https://github.com/khuangaf/ZeroFEC) repository.
10
+
11
+ ### How to use
12
+ Using Huggingface pipeline abstraction:
13
+ ```python
14
+ from transformers import pipeline
15
+
16
+ nlp = pipeline("text2text-generation", model='khhuang/zerofec-daqa-t5-base', tokenizer='khhuang/zerofec-daqa-t5-base')
17
+
18
+ QUESTION = "What is Night of the Living Dead?"
19
+ CONTEXT = "Night of the Living Dead is a 1968 American independent horror film , directed by George A. Romero , starring Duane Jones and Judith O'Dea . George A. Romero George A. Romero Duane Jones Duane Jones Judith O'Dea Judith O'Dea independent Independent film horror film horror film."
20
+
21
+ def format_inputs(context: str, question: str):
22
+ return f"{question} \n {context}"
23
+
24
+ text = format_inputs(CONTEXT, QUESTION)
25
+
26
+ nlp(text)
27
+ # should output [{'generated_text': 'a 1968 american independent horror film'}]
28
+ ```
29
+
30
+ Using the pre-trained model directly:
31
+ ```python
32
+ from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
33
+
34
+ tokenizer = AutoTokenizer.from_pretrained('khhuang/zerofec-daqa-t5-base')
35
+ model = AutoModelForSeq2SeqLM.from_pretrained('khhuang/zerofec-daqa-t5-base')
36
+
37
+ QUESTION = "What is Night of the Living Dead?"
38
+ CONTEXT = "Night of the Living Dead is a 1968 American independent horror film , directed by George A. Romero , starring Duane Jones and Judith O'Dea . George A. Romero George A. Romero Duane Jones Duane Jones Judith O'Dea Judith O'Dea independent Independent film horror film horror film."
39
+
40
+ def format_inputs(context: str, question: str):
41
+ return f"{question} \n {context}"
42
+
43
+ text = format_inputs(CONTEXT, QUESTION)
44
+
45
+
46
+ input_ids = tokenizer(text, return_tensors="pt").input_ids
47
+ generated_ids = model.generate(input_ids, max_length=32, num_beams=4)
48
+ output = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)
49
+ print(output)
50
+ # should output "a 1968 american independent horror film"
51
+ ```
52
+
53
+ ### Citation
54
+ ```
55
+ @inproceedings{huang-etal-2023-zero,
56
+ title = "Zero-shot Faithful Factual Error Correction",
57
+ author = "Huang, Kung-Hsiang and
58
+ Chan, Hou Pong and
59
+ Ji, Heng",
60
+ booktitle = "Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
61
+ month = jul,
62
+ year = "2023",
63
+ address = "Toronto, Canada",
64
+ publisher = "Association for Computational Linguistics",
65
+ url = "https://aclanthology.org/2023.acl-long.311",
66
+ doi = "10.18653/v1/2023.acl-long.311",
67
+ pages = "5660--5676",
68
+ }
69
+ ```
config.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "allenai/unifiedqa-v2-t5-base-1251000",
3
+ "architectures": [
4
+ "T5ForConditionalGeneration"
5
+ ],
6
+ "d_ff": 3072,
7
+ "d_kv": 64,
8
+ "d_model": 768,
9
+ "decoder_start_token_id": 0,
10
+ "dropout_rate": 0.1,
11
+ "eos_token_id": 1,
12
+ "feed_forward_proj": "relu",
13
+ "gradient_checkpointing": false,
14
+ "initializer_factor": 1.0,
15
+ "is_encoder_decoder": true,
16
+ "layer_norm_epsilon": 1e-06,
17
+ "model_type": "t5",
18
+ "n_positions": 512,
19
+ "num_decoder_layers": 12,
20
+ "num_heads": 12,
21
+ "num_layers": 12,
22
+ "output_past": true,
23
+ "pad_token_id": 0,
24
+ "relative_attention_max_distance": 128,
25
+ "relative_attention_num_buckets": 32,
26
+ "task_specific_params": {
27
+ "summarization": {
28
+ "early_stopping": true,
29
+ "length_penalty": 2.0,
30
+ "max_length": 200,
31
+ "min_length": 30,
32
+ "no_repeat_ngram_size": 3,
33
+ "num_beams": 4,
34
+ "prefix": "summarize: "
35
+ },
36
+ "translation_en_to_de": {
37
+ "early_stopping": true,
38
+ "max_length": 300,
39
+ "num_beams": 4,
40
+ "prefix": "translate English to German: "
41
+ },
42
+ "translation_en_to_fr": {
43
+ "early_stopping": true,
44
+ "max_length": 300,
45
+ "num_beams": 4,
46
+ "prefix": "translate English to French: "
47
+ },
48
+ "translation_en_to_ro": {
49
+ "early_stopping": true,
50
+ "max_length": 300,
51
+ "num_beams": 4,
52
+ "prefix": "translate English to Romanian: "
53
+ }
54
+ },
55
+ "torch_dtype": "float32",
56
+ "transformers_version": "4.19.2",
57
+ "use_cache": true,
58
+ "vocab_size": 32128
59
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da2cd3167734405f65ae2f756238da68b925b3f3d5aab21ea6b1c2781ef9070c
3
+ size 1783379133
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fd9a05c6323a6ddb5ad922529d506b6e86591d73d0aafe711abe59e3178040f
3
+ size 891700799
rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:673b71f54c0c70f4e0bd565013e8a7f3456fc29723c29c988dac718275580bbc
3
+ size 14503
scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca2359fd6bb85576e36c95dd2d4356c709ca53c6847e7aab7798843fd12175a2
3
+ size 559
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a836542adc98c03c89bf40db3637ba926035990d136f6780cd5043d550fe9a5d
3
+ size 623
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "additional_special_tokens": ["<extra_id_0>", "<extra_id_1>", "<extra_id_2>", "<extra_id_3>", "<extra_id_4>", "<extra_id_5>", "<extra_id_6>", "<extra_id_7>", "<extra_id_8>", "<extra_id_9>", "<extra_id_10>", "<extra_id_11>", "<extra_id_12>", "<extra_id_13>", "<extra_id_14>", "<extra_id_15>", "<extra_id_16>", "<extra_id_17>", "<extra_id_18>", "<extra_id_19>", "<extra_id_20>", "<extra_id_21>", "<extra_id_22>", "<extra_id_23>", "<extra_id_24>", "<extra_id_25>", "<extra_id_26>", "<extra_id_27>", "<extra_id_28>", "<extra_id_29>", "<extra_id_30>", "<extra_id_31>", "<extra_id_32>", "<extra_id_33>", "<extra_id_34>", "<extra_id_35>", "<extra_id_36>", "<extra_id_37>", "<extra_id_38>", "<extra_id_39>", "<extra_id_40>", "<extra_id_41>", "<extra_id_42>", "<extra_id_43>", "<extra_id_44>", "<extra_id_45>", "<extra_id_46>", "<extra_id_47>", "<extra_id_48>", "<extra_id_49>", "<extra_id_50>", "<extra_id_51>", "<extra_id_52>", "<extra_id_53>", "<extra_id_54>", "<extra_id_55>", "<extra_id_56>", "<extra_id_57>", "<extra_id_58>", "<extra_id_59>", "<extra_id_60>", "<extra_id_61>", "<extra_id_62>", "<extra_id_63>", "<extra_id_64>", "<extra_id_65>", "<extra_id_66>", "<extra_id_67>", "<extra_id_68>", "<extra_id_69>", "<extra_id_70>", "<extra_id_71>", "<extra_id_72>", "<extra_id_73>", "<extra_id_74>", "<extra_id_75>", "<extra_id_76>", "<extra_id_77>", "<extra_id_78>", "<extra_id_79>", "<extra_id_80>", "<extra_id_81>", "<extra_id_82>", "<extra_id_83>", "<extra_id_84>", "<extra_id_85>", "<extra_id_86>", "<extra_id_87>", "<extra_id_88>", "<extra_id_89>", "<extra_id_90>", "<extra_id_91>", "<extra_id_92>", "<extra_id_93>", "<extra_id_94>", "<extra_id_95>", "<extra_id_96>", "<extra_id_97>", "<extra_id_98>", "<extra_id_99>"]}
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86
3
+ size 791656
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "extra_ids": 100, "additional_special_tokens": ["<extra_id_0>", "<extra_id_1>", "<extra_id_2>", "<extra_id_3>", "<extra_id_4>", "<extra_id_5>", "<extra_id_6>", "<extra_id_7>", "<extra_id_8>", "<extra_id_9>", "<extra_id_10>", "<extra_id_11>", "<extra_id_12>", "<extra_id_13>", "<extra_id_14>", "<extra_id_15>", "<extra_id_16>", "<extra_id_17>", "<extra_id_18>", "<extra_id_19>", "<extra_id_20>", "<extra_id_21>", "<extra_id_22>", "<extra_id_23>", "<extra_id_24>", "<extra_id_25>", "<extra_id_26>", "<extra_id_27>", "<extra_id_28>", "<extra_id_29>", "<extra_id_30>", "<extra_id_31>", "<extra_id_32>", "<extra_id_33>", "<extra_id_34>", "<extra_id_35>", "<extra_id_36>", "<extra_id_37>", "<extra_id_38>", "<extra_id_39>", "<extra_id_40>", "<extra_id_41>", "<extra_id_42>", "<extra_id_43>", "<extra_id_44>", "<extra_id_45>", "<extra_id_46>", "<extra_id_47>", "<extra_id_48>", "<extra_id_49>", "<extra_id_50>", "<extra_id_51>", "<extra_id_52>", "<extra_id_53>", "<extra_id_54>", "<extra_id_55>", "<extra_id_56>", "<extra_id_57>", "<extra_id_58>", "<extra_id_59>", "<extra_id_60>", "<extra_id_61>", "<extra_id_62>", "<extra_id_63>", "<extra_id_64>", "<extra_id_65>", "<extra_id_66>", "<extra_id_67>", "<extra_id_68>", "<extra_id_69>", "<extra_id_70>", "<extra_id_71>", "<extra_id_72>", "<extra_id_73>", "<extra_id_74>", "<extra_id_75>", "<extra_id_76>", "<extra_id_77>", "<extra_id_78>", "<extra_id_79>", "<extra_id_80>", "<extra_id_81>", "<extra_id_82>", "<extra_id_83>", "<extra_id_84>", "<extra_id_85>", "<extra_id_86>", "<extra_id_87>", "<extra_id_88>", "<extra_id_89>", "<extra_id_90>", "<extra_id_91>", "<extra_id_92>", "<extra_id_93>", "<extra_id_94>", "<extra_id_95>", "<extra_id_96>", "<extra_id_97>", "<extra_id_98>", "<extra_id_99>"], "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "t5-base", "tokenizer_class": "T5Tokenizer"}
trainer_state.json ADDED
@@ -0,0 +1,253 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 64.0,
3
+ "best_model_checkpoint": "pubmedqa_unifiedqa/checkpoint-1100",
4
+ "epoch": 39.283185840707965,
5
+ "global_step": 1100,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.04,
12
+ "learning_rate": 0.0,
13
+ "loss": 5.2557,
14
+ "step": 1
15
+ },
16
+ {
17
+ "epoch": 3.57,
18
+ "learning_rate": 1.44e-06,
19
+ "loss": 5.5451,
20
+ "step": 100
21
+ },
22
+ {
23
+ "epoch": 3.57,
24
+ "eval_bleu": 3.515852791538636,
25
+ "eval_gen_len": 2.44,
26
+ "eval_loss": 5.634477615356445,
27
+ "eval_meteor": 0.25,
28
+ "eval_rouge1": 50.0,
29
+ "eval_rouge2": 0.0,
30
+ "eval_rougeL": 50.0,
31
+ "eval_rougeLsum": 50.0,
32
+ "eval_runtime": 3.8236,
33
+ "eval_samples_per_second": 13.077,
34
+ "eval_steps_per_second": 1.831,
35
+ "step": 100
36
+ },
37
+ {
38
+ "epoch": 7.14,
39
+ "learning_rate": 2.925e-06,
40
+ "loss": 4.9008,
41
+ "step": 200
42
+ },
43
+ {
44
+ "epoch": 7.14,
45
+ "eval_bleu": 3.4801542313344793,
46
+ "eval_gen_len": 2.44,
47
+ "eval_loss": 4.566105842590332,
48
+ "eval_meteor": 0.24,
49
+ "eval_rouge1": 48.0,
50
+ "eval_rouge2": 0.0,
51
+ "eval_rougeL": 48.0,
52
+ "eval_rougeLsum": 48.0,
53
+ "eval_runtime": 1.6121,
54
+ "eval_samples_per_second": 31.016,
55
+ "eval_steps_per_second": 4.342,
56
+ "step": 200
57
+ },
58
+ {
59
+ "epoch": 10.71,
60
+ "learning_rate": 2.9425e-06,
61
+ "loss": 3.9884,
62
+ "step": 300
63
+ },
64
+ {
65
+ "epoch": 10.71,
66
+ "eval_bleu": 5.48568366540206,
67
+ "eval_gen_len": 2.24,
68
+ "eval_loss": 3.5699760913848877,
69
+ "eval_meteor": 0.28,
70
+ "eval_rouge1": 56.0,
71
+ "eval_rouge2": 0.0,
72
+ "eval_rougeL": 56.0,
73
+ "eval_rougeLsum": 56.0,
74
+ "eval_runtime": 1.5002,
75
+ "eval_samples_per_second": 33.328,
76
+ "eval_steps_per_second": 4.666,
77
+ "step": 300
78
+ },
79
+ {
80
+ "epoch": 14.28,
81
+ "learning_rate": 2.88125e-06,
82
+ "loss": 3.2254,
83
+ "step": 400
84
+ },
85
+ {
86
+ "epoch": 14.28,
87
+ "eval_bleu": 0.0,
88
+ "eval_gen_len": 2.0,
89
+ "eval_loss": 2.9374730587005615,
90
+ "eval_meteor": 0.3,
91
+ "eval_rouge1": 60.0,
92
+ "eval_rouge2": 0.0,
93
+ "eval_rougeL": 60.0,
94
+ "eval_rougeLsum": 60.0,
95
+ "eval_runtime": 1.3736,
96
+ "eval_samples_per_second": 36.402,
97
+ "eval_steps_per_second": 5.096,
98
+ "step": 400
99
+ },
100
+ {
101
+ "epoch": 17.85,
102
+ "learning_rate": 2.81875e-06,
103
+ "loss": 2.6924,
104
+ "step": 500
105
+ },
106
+ {
107
+ "epoch": 17.85,
108
+ "eval_bleu": 0.0,
109
+ "eval_gen_len": 2.0,
110
+ "eval_loss": 2.5620052814483643,
111
+ "eval_meteor": 0.29,
112
+ "eval_rouge1": 58.0,
113
+ "eval_rouge2": 0.0,
114
+ "eval_rougeL": 58.0,
115
+ "eval_rougeLsum": 58.0,
116
+ "eval_runtime": 1.3734,
117
+ "eval_samples_per_second": 36.405,
118
+ "eval_steps_per_second": 5.097,
119
+ "step": 500
120
+ },
121
+ {
122
+ "epoch": 21.42,
123
+ "learning_rate": 2.75625e-06,
124
+ "loss": 2.4122,
125
+ "step": 600
126
+ },
127
+ {
128
+ "epoch": 21.42,
129
+ "eval_bleu": 0.0,
130
+ "eval_gen_len": 2.0,
131
+ "eval_loss": 2.3537566661834717,
132
+ "eval_meteor": 0.28,
133
+ "eval_rouge1": 56.0,
134
+ "eval_rouge2": 0.0,
135
+ "eval_rougeL": 56.0,
136
+ "eval_rougeLsum": 56.0,
137
+ "eval_runtime": 1.3768,
138
+ "eval_samples_per_second": 36.317,
139
+ "eval_steps_per_second": 5.084,
140
+ "step": 600
141
+ },
142
+ {
143
+ "epoch": 24.99,
144
+ "learning_rate": 2.695e-06,
145
+ "loss": 2.2719,
146
+ "step": 700
147
+ },
148
+ {
149
+ "epoch": 24.99,
150
+ "eval_bleu": 0.0,
151
+ "eval_gen_len": 2.0,
152
+ "eval_loss": 2.212905168533325,
153
+ "eval_meteor": 0.29,
154
+ "eval_rouge1": 58.0,
155
+ "eval_rouge2": 0.0,
156
+ "eval_rougeL": 58.0,
157
+ "eval_rougeLsum": 58.0,
158
+ "eval_runtime": 1.3724,
159
+ "eval_samples_per_second": 36.432,
160
+ "eval_steps_per_second": 5.101,
161
+ "step": 700
162
+ },
163
+ {
164
+ "epoch": 28.57,
165
+ "learning_rate": 2.6324999999999998e-06,
166
+ "loss": 2.1539,
167
+ "step": 800
168
+ },
169
+ {
170
+ "epoch": 28.57,
171
+ "eval_bleu": 0.0,
172
+ "eval_gen_len": 2.0,
173
+ "eval_loss": 2.123441219329834,
174
+ "eval_meteor": 0.29,
175
+ "eval_rouge1": 58.0,
176
+ "eval_rouge2": 0.0,
177
+ "eval_rougeL": 58.0,
178
+ "eval_rougeLsum": 58.0,
179
+ "eval_runtime": 1.3738,
180
+ "eval_samples_per_second": 36.397,
181
+ "eval_steps_per_second": 5.096,
182
+ "step": 800
183
+ },
184
+ {
185
+ "epoch": 32.14,
186
+ "learning_rate": 2.57125e-06,
187
+ "loss": 2.1149,
188
+ "step": 900
189
+ },
190
+ {
191
+ "epoch": 32.14,
192
+ "eval_bleu": 0.0,
193
+ "eval_gen_len": 2.0,
194
+ "eval_loss": 2.0748918056488037,
195
+ "eval_meteor": 0.29,
196
+ "eval_rouge1": 58.0,
197
+ "eval_rouge2": 0.0,
198
+ "eval_rougeL": 58.0,
199
+ "eval_rougeLsum": 58.0,
200
+ "eval_runtime": 1.3748,
201
+ "eval_samples_per_second": 36.37,
202
+ "eval_steps_per_second": 5.092,
203
+ "step": 900
204
+ },
205
+ {
206
+ "epoch": 35.71,
207
+ "learning_rate": 2.509375e-06,
208
+ "loss": 2.0404,
209
+ "step": 1000
210
+ },
211
+ {
212
+ "epoch": 35.71,
213
+ "eval_bleu": 0.0,
214
+ "eval_gen_len": 2.0,
215
+ "eval_loss": 2.0326242446899414,
216
+ "eval_meteor": 0.31,
217
+ "eval_rouge1": 62.0,
218
+ "eval_rouge2": 0.0,
219
+ "eval_rougeL": 62.0,
220
+ "eval_rougeLsum": 62.0,
221
+ "eval_runtime": 1.3725,
222
+ "eval_samples_per_second": 36.429,
223
+ "eval_steps_per_second": 5.1,
224
+ "step": 1000
225
+ },
226
+ {
227
+ "epoch": 39.28,
228
+ "learning_rate": 2.446875e-06,
229
+ "loss": 1.9918,
230
+ "step": 1100
231
+ },
232
+ {
233
+ "epoch": 39.28,
234
+ "eval_bleu": 0.0,
235
+ "eval_gen_len": 2.0,
236
+ "eval_loss": 2.012605667114258,
237
+ "eval_meteor": 0.32,
238
+ "eval_rouge1": 64.0,
239
+ "eval_rouge2": 0.0,
240
+ "eval_rougeL": 64.0,
241
+ "eval_rougeLsum": 64.0,
242
+ "eval_runtime": 1.3788,
243
+ "eval_samples_per_second": 36.265,
244
+ "eval_steps_per_second": 5.077,
245
+ "step": 1100
246
+ }
247
+ ],
248
+ "max_steps": 5000,
249
+ "num_train_epochs": 179,
250
+ "total_flos": 2.153031549530931e+16,
251
+ "trial_name": null,
252
+ "trial_params": null
253
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f9334b33e22eb3319970cb9153ba10497580ada38c636d4807cece434bae65b
3
+ size 3311