Mousumi44 commited on
Commit
92603c5
1 Parent(s): d638798
config.json ADDED
@@ -0,0 +1,124 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/pegasus-large",
3
+ "activation_dropout": 0.1,
4
+ "activation_function": "relu",
5
+ "add_bias_logits": false,
6
+ "add_final_layer_norm": true,
7
+ "architectures": [
8
+ "PegasusForConditionalGeneration"
9
+ ],
10
+ "attention_dropout": 0.1,
11
+ "bos_token_id": 0,
12
+ "classif_dropout": 0.0,
13
+ "classifier_dropout": 0.0,
14
+ "d_model": 1024,
15
+ "decoder_attention_heads": 16,
16
+ "decoder_ffn_dim": 4096,
17
+ "decoder_layerdrop": 0.0,
18
+ "decoder_layers": 16,
19
+ "decoder_start_token_id": 0,
20
+ "dropout": 0.1,
21
+ "encoder_attention_heads": 16,
22
+ "encoder_ffn_dim": 4096,
23
+ "encoder_layerdrop": 0.0,
24
+ "encoder_layers": 16,
25
+ "eos_token_id": 1,
26
+ "extra_pos_embeddings": 1,
27
+ "force_bos_token_to_be_generated": false,
28
+ "forced_eos_token_id": 1,
29
+ "gradient_checkpointing": false,
30
+ "id2label": {
31
+ "0": "LABEL_0",
32
+ "1": "LABEL_1",
33
+ "2": "LABEL_2"
34
+ },
35
+ "init_std": 0.02,
36
+ "is_encoder_decoder": true,
37
+ "label2id": {
38
+ "LABEL_0": 0,
39
+ "LABEL_1": 1,
40
+ "LABEL_2": 2
41
+ },
42
+ "length_penalty": 0.8,
43
+ "max_length": 256,
44
+ "max_position_embeddings": 1024,
45
+ "model_type": "pegasus",
46
+ "normalize_before": true,
47
+ "normalize_embedding": false,
48
+ "num_beams": 8,
49
+ "num_hidden_layers": 16,
50
+ "pad_token_id": 0,
51
+ "scale_embedding": true,
52
+ "static_position_embeddings": true,
53
+ "task_specific_params": {
54
+ "summarization_aeslc": {
55
+ "length_penalty": 0.6,
56
+ "max_length": 32,
57
+ "max_position_embeddings": 512
58
+ },
59
+ "summarization_arxiv": {
60
+ "length_penalty": 0.8,
61
+ "max_length": 256,
62
+ "max_position_embeddings": 1024
63
+ },
64
+ "summarization_big_patent": {
65
+ "length_penalty": 0.7,
66
+ "max_length": 256,
67
+ "max_position_embeddings": 1024
68
+ },
69
+ "summarization_billsum": {
70
+ "length_penalty": 0.6,
71
+ "max_length": 256,
72
+ "max_position_embeddings": 1024
73
+ },
74
+ "summarization_cnn_dailymail": {
75
+ "length_penalty": 0.8,
76
+ "max_length": 128,
77
+ "max_position_embeddings": 1024
78
+ },
79
+ "summarization_gigaword": {
80
+ "length_penalty": 0.6,
81
+ "max_length": 32,
82
+ "max_position_embeddings": 128
83
+ },
84
+ "summarization_large": {
85
+ "length_penalty": 0.8,
86
+ "max_length": 256,
87
+ "max_position_embeddings": 1024
88
+ },
89
+ "summarization_multi_news": {
90
+ "length_penalty": 0.8,
91
+ "max_length": 256,
92
+ "max_position_embeddings": 1024
93
+ },
94
+ "summarization_newsroom": {
95
+ "length_penalty": 0.8,
96
+ "max_length": 128,
97
+ "max_position_embeddings": 512
98
+ },
99
+ "summarization_pubmed": {
100
+ "length_penalty": 0.8,
101
+ "max_length": 256,
102
+ "max_position_embeddings": 1024
103
+ },
104
+ "summarization_reddit_tifu": {
105
+ "length_penalty": 0.6,
106
+ "max_length": 128,
107
+ "max_position_embeddings": 512
108
+ },
109
+ "summarization_wikihow": {
110
+ "length_penalty": 0.6,
111
+ "max_length": 256,
112
+ "max_position_embeddings": 512
113
+ },
114
+ "summarization_xsum": {
115
+ "length_penalty": 0.8,
116
+ "max_length": 64,
117
+ "max_position_embeddings": 512
118
+ }
119
+ },
120
+ "torch_dtype": "float32",
121
+ "transformers_version": "4.11.2",
122
+ "use_cache": true,
123
+ "vocab_size": 96103
124
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0ff98b3b3cc1ddac5427268378f850e41c71b5c9435fde7c68ea8508ca6dc61
3
+ size 4549998179
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d64c6152d1f6f509f21e80c705d8cb0813deb19ed1476cdf41eb779683889703
3
+ size 2283825905
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfa26a201d09a229664e508b4cb7f7bb5ffe127478ff9d74210c71d956d1e082
3
+ size 14503
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27cfed17e5d3a4b92744fc5e719235b700be5cf6ef9177a64a4958f203c9d962
3
+ size 623
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": "<mask_2>", "additional_special_tokens": ["<mask_1>", "<unk_2>", "<unk_3>", "<unk_4>", "<unk_5>", "<unk_6>", "<unk_7>", "<unk_8>", "<unk_9>", "<unk_10>", "<unk_11>", "<unk_12>", "<unk_13>", "<unk_14>", "<unk_15>", "<unk_16>", "<unk_17>", "<unk_18>", "<unk_19>", "<unk_20>", "<unk_21>", "<unk_22>", "<unk_23>", "<unk_24>", "<unk_25>", "<unk_26>", "<unk_27>", "<unk_28>", "<unk_29>", "<unk_30>", "<unk_31>", "<unk_32>", "<unk_33>", "<unk_34>", "<unk_35>", "<unk_36>", "<unk_37>", "<unk_38>", "<unk_39>", "<unk_40>", "<unk_41>", "<unk_42>", "<unk_43>", "<unk_44>", "<unk_45>", "<unk_46>", "<unk_47>", "<unk_48>", "<unk_49>", "<unk_50>", "<unk_51>", "<unk_52>", "<unk_53>", "<unk_54>", "<unk_55>", "<unk_56>", "<unk_57>", "<unk_58>", "<unk_59>", "<unk_60>", "<unk_61>", "<unk_62>", "<unk_63>", "<unk_64>", "<unk_65>", "<unk_66>", "<unk_67>", "<unk_68>", "<unk_69>", "<unk_70>", "<unk_71>", "<unk_72>", "<unk_73>", "<unk_74>", "<unk_75>", "<unk_76>", "<unk_77>", "<unk_78>", "<unk_79>", "<unk_80>", "<unk_81>", "<unk_82>", "<unk_83>", "<unk_84>", "<unk_85>", "<unk_86>", "<unk_87>", "<unk_88>", "<unk_89>", "<unk_90>", "<unk_91>", "<unk_92>", "<unk_93>", "<unk_94>", "<unk_95>", "<unk_96>", "<unk_97>", "<unk_98>", "<unk_99>", "<unk_100>", "<unk_101>", "<unk_102>"]}
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0015189ef36359283fec8b93cf6d9ce51bca37eb1101defc68a53b394913b96c
3
+ size 1912529
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "mask_token": "<mask_2>", "pad_token": "<pad>", "mask_token_sent": "<mask_1>", "offset": 103, "additional_special_tokens": ["<mask_1>", "<unk_2>", "<unk_3>", "<unk_4>", "<unk_5>", "<unk_6>", "<unk_7>", "<unk_8>", "<unk_9>", "<unk_10>", "<unk_11>", "<unk_12>", "<unk_13>", "<unk_14>", "<unk_15>", "<unk_16>", "<unk_17>", "<unk_18>", "<unk_19>", "<unk_20>", "<unk_21>", "<unk_22>", "<unk_23>", "<unk_24>", "<unk_25>", "<unk_26>", "<unk_27>", "<unk_28>", "<unk_29>", "<unk_30>", "<unk_31>", "<unk_32>", "<unk_33>", "<unk_34>", "<unk_35>", "<unk_36>", "<unk_37>", "<unk_38>", "<unk_39>", "<unk_40>", "<unk_41>", "<unk_42>", "<unk_43>", "<unk_44>", "<unk_45>", "<unk_46>", "<unk_47>", "<unk_48>", "<unk_49>", "<unk_50>", "<unk_51>", "<unk_52>", "<unk_53>", "<unk_54>", "<unk_55>", "<unk_56>", "<unk_57>", "<unk_58>", "<unk_59>", "<unk_60>", "<unk_61>", "<unk_62>", "<unk_63>", "<unk_64>", "<unk_65>", "<unk_66>", "<unk_67>", "<unk_68>", "<unk_69>", "<unk_70>", "<unk_71>", "<unk_72>", "<unk_73>", "<unk_74>", "<unk_75>", "<unk_76>", "<unk_77>", "<unk_78>", "<unk_79>", "<unk_80>", "<unk_81>", "<unk_82>", "<unk_83>", "<unk_84>", "<unk_85>", "<unk_86>", "<unk_87>", "<unk_88>", "<unk_89>", "<unk_90>", "<unk_91>", "<unk_92>", "<unk_93>", "<unk_94>", "<unk_95>", "<unk_96>", "<unk_97>", "<unk_98>", "<unk_99>", "<unk_100>", "<unk_101>", "<unk_102>"], "sp_model_kwargs": {}, "model_max_length": 1024, "special_tokens_map_file": null, "full_tokenizer_file": null, "tokenizer_file": null, "name_or_path": "google/pegasus-large", "tokenizer_class": "PegasusTokenizer"}
trainer_state.json ADDED
@@ -0,0 +1,238 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.9216,
5
+ "global_step": 75,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.05,
12
+ "learning_rate": 2e-05,
13
+ "loss": 7.7788,
14
+ "step": 2
15
+ },
16
+ {
17
+ "epoch": 0.1,
18
+ "learning_rate": 4e-05,
19
+ "loss": 7.6098,
20
+ "step": 4
21
+ },
22
+ {
23
+ "epoch": 0.15,
24
+ "learning_rate": 4.9315068493150684e-05,
25
+ "loss": 7.3839,
26
+ "step": 6
27
+ },
28
+ {
29
+ "epoch": 0.2,
30
+ "learning_rate": 4.794520547945205e-05,
31
+ "loss": 7.1239,
32
+ "step": 8
33
+ },
34
+ {
35
+ "epoch": 0.26,
36
+ "learning_rate": 4.657534246575342e-05,
37
+ "loss": 6.844,
38
+ "step": 10
39
+ },
40
+ {
41
+ "epoch": 0.31,
42
+ "learning_rate": 4.520547945205479e-05,
43
+ "loss": 6.8217,
44
+ "step": 12
45
+ },
46
+ {
47
+ "epoch": 0.36,
48
+ "learning_rate": 4.383561643835617e-05,
49
+ "loss": 6.5852,
50
+ "step": 14
51
+ },
52
+ {
53
+ "epoch": 0.41,
54
+ "learning_rate": 4.2465753424657536e-05,
55
+ "loss": 6.4827,
56
+ "step": 16
57
+ },
58
+ {
59
+ "epoch": 0.46,
60
+ "learning_rate": 4.1095890410958905e-05,
61
+ "loss": 6.4289,
62
+ "step": 18
63
+ },
64
+ {
65
+ "epoch": 0.51,
66
+ "learning_rate": 3.9726027397260274e-05,
67
+ "loss": 6.275,
68
+ "step": 20
69
+ },
70
+ {
71
+ "epoch": 0.56,
72
+ "learning_rate": 3.8356164383561644e-05,
73
+ "loss": 6.2334,
74
+ "step": 22
75
+ },
76
+ {
77
+ "epoch": 0.61,
78
+ "learning_rate": 3.698630136986301e-05,
79
+ "loss": 6.254,
80
+ "step": 24
81
+ },
82
+ {
83
+ "epoch": 0.67,
84
+ "learning_rate": 3.561643835616438e-05,
85
+ "loss": 6.1053,
86
+ "step": 26
87
+ },
88
+ {
89
+ "epoch": 0.72,
90
+ "learning_rate": 3.424657534246575e-05,
91
+ "loss": 6.1087,
92
+ "step": 28
93
+ },
94
+ {
95
+ "epoch": 0.77,
96
+ "learning_rate": 3.287671232876712e-05,
97
+ "loss": 6.0152,
98
+ "step": 30
99
+ },
100
+ {
101
+ "epoch": 0.82,
102
+ "learning_rate": 3.1506849315068496e-05,
103
+ "loss": 5.9039,
104
+ "step": 32
105
+ },
106
+ {
107
+ "epoch": 0.87,
108
+ "learning_rate": 3.0136986301369862e-05,
109
+ "loss": 5.9487,
110
+ "step": 34
111
+ },
112
+ {
113
+ "epoch": 0.92,
114
+ "learning_rate": 2.8767123287671234e-05,
115
+ "loss": 5.9647,
116
+ "step": 36
117
+ },
118
+ {
119
+ "epoch": 0.97,
120
+ "learning_rate": 2.7397260273972603e-05,
121
+ "loss": 5.9559,
122
+ "step": 38
123
+ },
124
+ {
125
+ "epoch": 1.03,
126
+ "learning_rate": 2.6027397260273973e-05,
127
+ "loss": 6.0371,
128
+ "step": 40
129
+ },
130
+ {
131
+ "epoch": 1.08,
132
+ "learning_rate": 2.4657534246575342e-05,
133
+ "loss": 5.8627,
134
+ "step": 42
135
+ },
136
+ {
137
+ "epoch": 1.13,
138
+ "learning_rate": 2.328767123287671e-05,
139
+ "loss": 5.8017,
140
+ "step": 44
141
+ },
142
+ {
143
+ "epoch": 1.18,
144
+ "learning_rate": 2.1917808219178083e-05,
145
+ "loss": 5.8414,
146
+ "step": 46
147
+ },
148
+ {
149
+ "epoch": 1.23,
150
+ "learning_rate": 2.0547945205479453e-05,
151
+ "loss": 5.8787,
152
+ "step": 48
153
+ },
154
+ {
155
+ "epoch": 1.28,
156
+ "learning_rate": 1.9178082191780822e-05,
157
+ "loss": 5.8423,
158
+ "step": 50
159
+ },
160
+ {
161
+ "epoch": 1.33,
162
+ "learning_rate": 1.780821917808219e-05,
163
+ "loss": 5.7113,
164
+ "step": 52
165
+ },
166
+ {
167
+ "epoch": 1.38,
168
+ "learning_rate": 1.643835616438356e-05,
169
+ "loss": 5.8619,
170
+ "step": 54
171
+ },
172
+ {
173
+ "epoch": 1.44,
174
+ "learning_rate": 1.5068493150684931e-05,
175
+ "loss": 5.7901,
176
+ "step": 56
177
+ },
178
+ {
179
+ "epoch": 1.49,
180
+ "learning_rate": 1.3698630136986302e-05,
181
+ "loss": 5.6903,
182
+ "step": 58
183
+ },
184
+ {
185
+ "epoch": 1.54,
186
+ "learning_rate": 1.2328767123287671e-05,
187
+ "loss": 5.6354,
188
+ "step": 60
189
+ },
190
+ {
191
+ "epoch": 1.59,
192
+ "learning_rate": 1.0958904109589042e-05,
193
+ "loss": 5.7176,
194
+ "step": 62
195
+ },
196
+ {
197
+ "epoch": 1.64,
198
+ "learning_rate": 9.589041095890411e-06,
199
+ "loss": 5.7527,
200
+ "step": 64
201
+ },
202
+ {
203
+ "epoch": 1.69,
204
+ "learning_rate": 8.21917808219178e-06,
205
+ "loss": 5.8004,
206
+ "step": 66
207
+ },
208
+ {
209
+ "epoch": 1.74,
210
+ "learning_rate": 6.849315068493151e-06,
211
+ "loss": 5.7,
212
+ "step": 68
213
+ },
214
+ {
215
+ "epoch": 1.79,
216
+ "learning_rate": 5.479452054794521e-06,
217
+ "loss": 5.7405,
218
+ "step": 70
219
+ },
220
+ {
221
+ "epoch": 1.84,
222
+ "learning_rate": 4.10958904109589e-06,
223
+ "loss": 5.782,
224
+ "step": 72
225
+ },
226
+ {
227
+ "epoch": 1.9,
228
+ "learning_rate": 2.7397260273972604e-06,
229
+ "loss": 5.6801,
230
+ "step": 74
231
+ }
232
+ ],
233
+ "max_steps": 78,
234
+ "num_train_epochs": 2,
235
+ "total_flos": 5.552394818342093e+16,
236
+ "trial_name": null,
237
+ "trial_params": null
238
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffa9d71104228418bed764f35eec881bde117a61133971f097ce93cd2a92e4bb
3
+ size 2799