mr-oogway commited on
Commit
ce7560e
1 Parent(s): 60126e8

Upload 8 files

Browse files
config.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/flan-t5-base",
3
+ "architectures": [
4
+ "T5ForConditionalGeneration"
5
+ ],
6
+ "d_ff": 2048,
7
+ "d_kv": 64,
8
+ "d_model": 768,
9
+ "decoder_start_token_id": 0,
10
+ "dense_act_fn": "gelu_new",
11
+ "dropout_rate": 0.1,
12
+ "eos_token_id": 1,
13
+ "feed_forward_proj": "gated-gelu",
14
+ "initializer_factor": 1.0,
15
+ "is_encoder_decoder": true,
16
+ "is_gated_act": true,
17
+ "layer_norm_epsilon": 1e-06,
18
+ "model_type": "t5",
19
+ "n_positions": 512,
20
+ "num_decoder_layers": 12,
21
+ "num_heads": 12,
22
+ "num_layers": 12,
23
+ "output_past": true,
24
+ "pad_token_id": 0,
25
+ "relative_attention_max_distance": 128,
26
+ "relative_attention_num_buckets": 32,
27
+ "task_specific_params": {
28
+ "summarization": {
29
+ "early_stopping": true,
30
+ "length_penalty": 2.0,
31
+ "max_length": 200,
32
+ "min_length": 30,
33
+ "no_repeat_ngram_size": 3,
34
+ "num_beams": 4,
35
+ "prefix": "summarize: "
36
+ },
37
+ "translation_en_to_de": {
38
+ "early_stopping": true,
39
+ "max_length": 300,
40
+ "num_beams": 4,
41
+ "prefix": "translate English to German: "
42
+ },
43
+ "translation_en_to_fr": {
44
+ "early_stopping": true,
45
+ "max_length": 300,
46
+ "num_beams": 4,
47
+ "prefix": "translate English to French: "
48
+ },
49
+ "translation_en_to_ro": {
50
+ "early_stopping": true,
51
+ "max_length": 300,
52
+ "num_beams": 4,
53
+ "prefix": "translate English to Romanian: "
54
+ }
55
+ },
56
+ "tie_word_embeddings": false,
57
+ "torch_dtype": "float32",
58
+ "transformers_version": "4.27.4",
59
+ "use_cache": true,
60
+ "vocab_size": 32128
61
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "decoder_start_token_id": 0,
3
+ "eos_token_id": 1,
4
+ "pad_token_id": 0,
5
+ "transformers_version": "4.27.4"
6
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc447d3128d384f4763b0ba438f963a9a0a1c108eedb6a422d56a6c47ed18b74
3
+ size 1980790149
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c05eb1790e28aabd51bcc067e17000f70539972fe6b1c148d03fa831de6e9956
3
+ size 990408885
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d364961fd41257fd9e5e4f8e623af755a4ec8ea4459987ad0f0d186e5bec9f80
3
+ size 14575
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:991f3abde745b4b10cc13ea33a9d402812630a3981fb5f9b03b850aee28cedfc
3
+ size 627
trainer_state.json ADDED
@@ -0,0 +1,201 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 1.9500832557678223,
3
+ "best_model_checkpoint": "flan-t5-qa/checkpoint-10000",
4
+ "epoch": 5.0,
5
+ "global_step": 10000,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.25,
12
+ "learning_rate": 4.75e-05,
13
+ "loss": 3.0646,
14
+ "step": 500
15
+ },
16
+ {
17
+ "epoch": 0.5,
18
+ "learning_rate": 4.5e-05,
19
+ "loss": 2.5167,
20
+ "step": 1000
21
+ },
22
+ {
23
+ "epoch": 0.75,
24
+ "learning_rate": 4.25e-05,
25
+ "loss": 2.3151,
26
+ "step": 1500
27
+ },
28
+ {
29
+ "epoch": 1.0,
30
+ "learning_rate": 4e-05,
31
+ "loss": 2.2112,
32
+ "step": 2000
33
+ },
34
+ {
35
+ "epoch": 1.0,
36
+ "eval_gen_len": 17.726,
37
+ "eval_loss": 2.2397749423980713,
38
+ "eval_rouge1": 31.2944,
39
+ "eval_rouge2": 16.8097,
40
+ "eval_rougeL": 29.7158,
41
+ "eval_rougeLsum": 29.856,
42
+ "eval_runtime": 111.783,
43
+ "eval_samples_per_second": 8.946,
44
+ "eval_steps_per_second": 1.789,
45
+ "step": 2000
46
+ },
47
+ {
48
+ "epoch": 1.25,
49
+ "learning_rate": 3.7500000000000003e-05,
50
+ "loss": 2.0983,
51
+ "step": 2500
52
+ },
53
+ {
54
+ "epoch": 1.5,
55
+ "learning_rate": 3.5e-05,
56
+ "loss": 2.0534,
57
+ "step": 3000
58
+ },
59
+ {
60
+ "epoch": 1.75,
61
+ "learning_rate": 3.2500000000000004e-05,
62
+ "loss": 2.0071,
63
+ "step": 3500
64
+ },
65
+ {
66
+ "epoch": 2.0,
67
+ "learning_rate": 3e-05,
68
+ "loss": 1.9424,
69
+ "step": 4000
70
+ },
71
+ {
72
+ "epoch": 2.0,
73
+ "eval_gen_len": 17.264,
74
+ "eval_loss": 2.061356544494629,
75
+ "eval_rouge1": 32.2914,
76
+ "eval_rouge2": 17.2352,
77
+ "eval_rougeL": 30.6559,
78
+ "eval_rougeLsum": 30.7999,
79
+ "eval_runtime": 109.8715,
80
+ "eval_samples_per_second": 9.102,
81
+ "eval_steps_per_second": 1.82,
82
+ "step": 4000
83
+ },
84
+ {
85
+ "epoch": 2.25,
86
+ "learning_rate": 2.7500000000000004e-05,
87
+ "loss": 1.9267,
88
+ "step": 4500
89
+ },
90
+ {
91
+ "epoch": 2.5,
92
+ "learning_rate": 2.5e-05,
93
+ "loss": 1.8592,
94
+ "step": 5000
95
+ },
96
+ {
97
+ "epoch": 2.75,
98
+ "learning_rate": 2.25e-05,
99
+ "loss": 1.8036,
100
+ "step": 5500
101
+ },
102
+ {
103
+ "epoch": 3.0,
104
+ "learning_rate": 2e-05,
105
+ "loss": 1.8168,
106
+ "step": 6000
107
+ },
108
+ {
109
+ "epoch": 3.0,
110
+ "eval_gen_len": 17.385,
111
+ "eval_loss": 1.995050072669983,
112
+ "eval_rouge1": 32.1343,
113
+ "eval_rouge2": 17.6839,
114
+ "eval_rougeL": 30.9136,
115
+ "eval_rougeLsum": 31.0279,
116
+ "eval_runtime": 110.3548,
117
+ "eval_samples_per_second": 9.062,
118
+ "eval_steps_per_second": 1.812,
119
+ "step": 6000
120
+ },
121
+ {
122
+ "epoch": 3.25,
123
+ "learning_rate": 1.75e-05,
124
+ "loss": 1.7743,
125
+ "step": 6500
126
+ },
127
+ {
128
+ "epoch": 3.5,
129
+ "learning_rate": 1.5e-05,
130
+ "loss": 1.788,
131
+ "step": 7000
132
+ },
133
+ {
134
+ "epoch": 3.75,
135
+ "learning_rate": 1.25e-05,
136
+ "loss": 1.7067,
137
+ "step": 7500
138
+ },
139
+ {
140
+ "epoch": 4.0,
141
+ "learning_rate": 1e-05,
142
+ "loss": 1.7924,
143
+ "step": 8000
144
+ },
145
+ {
146
+ "epoch": 4.0,
147
+ "eval_gen_len": 17.302,
148
+ "eval_loss": 1.9565637111663818,
149
+ "eval_rouge1": 32.6305,
150
+ "eval_rouge2": 18.4737,
151
+ "eval_rougeL": 31.4423,
152
+ "eval_rougeLsum": 31.5984,
153
+ "eval_runtime": 111.5662,
154
+ "eval_samples_per_second": 8.963,
155
+ "eval_steps_per_second": 1.793,
156
+ "step": 8000
157
+ },
158
+ {
159
+ "epoch": 4.25,
160
+ "learning_rate": 7.5e-06,
161
+ "loss": 1.7113,
162
+ "step": 8500
163
+ },
164
+ {
165
+ "epoch": 4.5,
166
+ "learning_rate": 5e-06,
167
+ "loss": 1.7379,
168
+ "step": 9000
169
+ },
170
+ {
171
+ "epoch": 4.75,
172
+ "learning_rate": 2.5e-06,
173
+ "loss": 1.7181,
174
+ "step": 9500
175
+ },
176
+ {
177
+ "epoch": 5.0,
178
+ "learning_rate": 0.0,
179
+ "loss": 1.6977,
180
+ "step": 10000
181
+ },
182
+ {
183
+ "epoch": 5.0,
184
+ "eval_gen_len": 17.333,
185
+ "eval_loss": 1.9500832557678223,
186
+ "eval_rouge1": 32.6382,
187
+ "eval_rouge2": 18.8113,
188
+ "eval_rougeL": 31.4434,
189
+ "eval_rougeLsum": 31.55,
190
+ "eval_runtime": 111.7348,
191
+ "eval_samples_per_second": 8.95,
192
+ "eval_steps_per_second": 1.79,
193
+ "step": 10000
194
+ }
195
+ ],
196
+ "max_steps": 10000,
197
+ "num_train_epochs": 5,
198
+ "total_flos": 1.65839671296e+16,
199
+ "trial_name": null,
200
+ "trial_params": null
201
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfa1bc2238d8d154910e52f1c748f7d4ca324a3fb8a199c64258e783d7e06ac2
3
+ size 3643