oMateos2020 commited on
Commit
5a5b6ce
1 Parent(s): 366b48f

Model save

Browse files
last-checkpoint/config.json DELETED
@@ -1,56 +0,0 @@
1
- {
2
- "_name_or_path": "oMateos2020/pegasus-newsroom-cnn_full-adafactor-bs6",
3
- "activation_dropout": 0.1,
4
- "activation_function": "relu",
5
- "add_bias_logits": false,
6
- "add_final_layer_norm": true,
7
- "architectures": [
8
- "PegasusForConditionalGeneration"
9
- ],
10
- "attention_dropout": 0.1,
11
- "bos_token_id": 0,
12
- "classif_dropout": 0.0,
13
- "classifier_dropout": 0.0,
14
- "d_model": 1024,
15
- "decoder_attention_heads": 16,
16
- "decoder_ffn_dim": 4096,
17
- "decoder_layerdrop": 0.0,
18
- "decoder_layers": 16,
19
- "decoder_start_token_id": 0,
20
- "dropout": 0.0,
21
- "encoder_attention_heads": 16,
22
- "encoder_ffn_dim": 4096,
23
- "encoder_layerdrop": 0.0,
24
- "encoder_layers": 16,
25
- "eos_token_id": 1,
26
- "extra_pos_embeddings": 1,
27
- "forced_eos_token_id": 1,
28
- "id2label": {
29
- "0": "LABEL_0",
30
- "1": "LABEL_1",
31
- "2": "LABEL_2"
32
- },
33
- "init_std": 0.02,
34
- "is_encoder_decoder": true,
35
- "label2id": {
36
- "LABEL_0": 0,
37
- "LABEL_1": 1,
38
- "LABEL_2": 2
39
- },
40
- "length_penalty": 0.8,
41
- "max_length": 128,
42
- "max_position_embeddings": 512,
43
- "min_length": 32,
44
- "model_type": "pegasus",
45
- "normalize_before": true,
46
- "normalize_embedding": false,
47
- "num_beams": 2,
48
- "num_hidden_layers": 16,
49
- "pad_token_id": 0,
50
- "scale_embedding": true,
51
- "static_position_embeddings": true,
52
- "torch_dtype": "float32",
53
- "transformers_version": "4.20.1",
54
- "use_cache": true,
55
- "vocab_size": 96103
56
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:92fcb081a65a24f5203f68033129916e8f3de1f9ddb98648805bbc5104873aba
3
- size 5840398
 
 
 
 
last-checkpoint/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b3dce194dc738268dd2c1f5262d8d2f2ba2fe4941f8e5db5a86356c6e0edd8b
3
- size 2279605745
 
 
 
 
last-checkpoint/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:87aaf30dcfcb440c49f986834c15f5d1924efd3c01470dc1364bf02404df01d7
3
- size 14503
 
 
 
 
last-checkpoint/scaler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8ca221068257890306847f8d3ee335b25a16ad3c8d4a625f911e0389b9d1b6f
3
- size 559
 
 
 
 
last-checkpoint/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e1f0b1cc1d5e456531ad0948086aa7319d517038ef91e91e924651ec8a9048f
3
- size 623
 
 
 
 
last-checkpoint/special_tokens_map.json DELETED
@@ -1,110 +0,0 @@
1
- {
2
- "additional_special_tokens": [
3
- "<mask_1>",
4
- "<unk_2>",
5
- "<unk_3>",
6
- "<unk_4>",
7
- "<unk_5>",
8
- "<unk_6>",
9
- "<unk_7>",
10
- "<unk_8>",
11
- "<unk_9>",
12
- "<unk_10>",
13
- "<unk_11>",
14
- "<unk_12>",
15
- "<unk_13>",
16
- "<unk_14>",
17
- "<unk_15>",
18
- "<unk_16>",
19
- "<unk_17>",
20
- "<unk_18>",
21
- "<unk_19>",
22
- "<unk_20>",
23
- "<unk_21>",
24
- "<unk_22>",
25
- "<unk_23>",
26
- "<unk_24>",
27
- "<unk_25>",
28
- "<unk_26>",
29
- "<unk_27>",
30
- "<unk_28>",
31
- "<unk_29>",
32
- "<unk_30>",
33
- "<unk_31>",
34
- "<unk_32>",
35
- "<unk_33>",
36
- "<unk_34>",
37
- "<unk_35>",
38
- "<unk_36>",
39
- "<unk_37>",
40
- "<unk_38>",
41
- "<unk_39>",
42
- "<unk_40>",
43
- "<unk_41>",
44
- "<unk_42>",
45
- "<unk_43>",
46
- "<unk_44>",
47
- "<unk_45>",
48
- "<unk_46>",
49
- "<unk_47>",
50
- "<unk_48>",
51
- "<unk_49>",
52
- "<unk_50>",
53
- "<unk_51>",
54
- "<unk_52>",
55
- "<unk_53>",
56
- "<unk_54>",
57
- "<unk_55>",
58
- "<unk_56>",
59
- "<unk_57>",
60
- "<unk_58>",
61
- "<unk_59>",
62
- "<unk_60>",
63
- "<unk_61>",
64
- "<unk_62>",
65
- "<unk_63>",
66
- "<unk_64>",
67
- "<unk_65>",
68
- "<unk_66>",
69
- "<unk_67>",
70
- "<unk_68>",
71
- "<unk_69>",
72
- "<unk_70>",
73
- "<unk_71>",
74
- "<unk_72>",
75
- "<unk_73>",
76
- "<unk_74>",
77
- "<unk_75>",
78
- "<unk_76>",
79
- "<unk_77>",
80
- "<unk_78>",
81
- "<unk_79>",
82
- "<unk_80>",
83
- "<unk_81>",
84
- "<unk_82>",
85
- "<unk_83>",
86
- "<unk_84>",
87
- "<unk_85>",
88
- "<unk_86>",
89
- "<unk_87>",
90
- "<unk_88>",
91
- "<unk_89>",
92
- "<unk_90>",
93
- "<unk_91>",
94
- "<unk_92>",
95
- "<unk_93>",
96
- "<unk_94>",
97
- "<unk_95>",
98
- "<unk_96>",
99
- "<unk_97>",
100
- "<unk_98>",
101
- "<unk_99>",
102
- "<unk_100>",
103
- "<unk_101>",
104
- "<unk_102>"
105
- ],
106
- "eos_token": "</s>",
107
- "mask_token": "<mask_2>",
108
- "pad_token": "<pad>",
109
- "unk_token": "<unk>"
110
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/spiece.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:0015189ef36359283fec8b93cf6d9ce51bca37eb1101defc68a53b394913b96c
3
- size 1912529
 
 
 
 
last-checkpoint/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
last-checkpoint/tokenizer_config.json DELETED
@@ -1,118 +0,0 @@
1
- {
2
- "additional_special_tokens": [
3
- "<mask_1>",
4
- "<unk_2>",
5
- "<unk_3>",
6
- "<unk_4>",
7
- "<unk_5>",
8
- "<unk_6>",
9
- "<unk_7>",
10
- "<unk_8>",
11
- "<unk_9>",
12
- "<unk_10>",
13
- "<unk_11>",
14
- "<unk_12>",
15
- "<unk_13>",
16
- "<unk_14>",
17
- "<unk_15>",
18
- "<unk_16>",
19
- "<unk_17>",
20
- "<unk_18>",
21
- "<unk_19>",
22
- "<unk_20>",
23
- "<unk_21>",
24
- "<unk_22>",
25
- "<unk_23>",
26
- "<unk_24>",
27
- "<unk_25>",
28
- "<unk_26>",
29
- "<unk_27>",
30
- "<unk_28>",
31
- "<unk_29>",
32
- "<unk_30>",
33
- "<unk_31>",
34
- "<unk_32>",
35
- "<unk_33>",
36
- "<unk_34>",
37
- "<unk_35>",
38
- "<unk_36>",
39
- "<unk_37>",
40
- "<unk_38>",
41
- "<unk_39>",
42
- "<unk_40>",
43
- "<unk_41>",
44
- "<unk_42>",
45
- "<unk_43>",
46
- "<unk_44>",
47
- "<unk_45>",
48
- "<unk_46>",
49
- "<unk_47>",
50
- "<unk_48>",
51
- "<unk_49>",
52
- "<unk_50>",
53
- "<unk_51>",
54
- "<unk_52>",
55
- "<unk_53>",
56
- "<unk_54>",
57
- "<unk_55>",
58
- "<unk_56>",
59
- "<unk_57>",
60
- "<unk_58>",
61
- "<unk_59>",
62
- "<unk_60>",
63
- "<unk_61>",
64
- "<unk_62>",
65
- "<unk_63>",
66
- "<unk_64>",
67
- "<unk_65>",
68
- "<unk_66>",
69
- "<unk_67>",
70
- "<unk_68>",
71
- "<unk_69>",
72
- "<unk_70>",
73
- "<unk_71>",
74
- "<unk_72>",
75
- "<unk_73>",
76
- "<unk_74>",
77
- "<unk_75>",
78
- "<unk_76>",
79
- "<unk_77>",
80
- "<unk_78>",
81
- "<unk_79>",
82
- "<unk_80>",
83
- "<unk_81>",
84
- "<unk_82>",
85
- "<unk_83>",
86
- "<unk_84>",
87
- "<unk_85>",
88
- "<unk_86>",
89
- "<unk_87>",
90
- "<unk_88>",
91
- "<unk_89>",
92
- "<unk_90>",
93
- "<unk_91>",
94
- "<unk_92>",
95
- "<unk_93>",
96
- "<unk_94>",
97
- "<unk_95>",
98
- "<unk_96>",
99
- "<unk_97>",
100
- "<unk_98>",
101
- "<unk_99>",
102
- "<unk_100>",
103
- "<unk_101>",
104
- "<unk_102>"
105
- ],
106
- "eos_token": "</s>",
107
- "full_tokenizer_file": null,
108
- "mask_token": "<mask_2>",
109
- "mask_token_sent": "<mask_1>",
110
- "model_max_length": 512,
111
- "name_or_path": "oMateos2020/pegasus-newsroom-cnn_full-adafactor-bs6",
112
- "offset": 103,
113
- "pad_token": "<pad>",
114
- "sp_model_kwargs": {},
115
- "special_tokens_map_file": null,
116
- "tokenizer_class": "PegasusTokenizer",
117
- "unk_token": "<unk>"
118
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/trainer_state.json DELETED
@@ -1,59 +0,0 @@
1
- {
2
- "best_metric": 43.9038,
3
- "best_model_checkpoint": "pegasus-newsroom-cnn_full-adafactor-bs6/checkpoint-560",
4
- "epoch": 0.49931038325972776,
5
- "global_step": 560,
6
- "is_hyper_param_search": false,
7
- "is_local_process_zero": true,
8
- "is_world_process_zero": true,
9
- "log_history": [
10
- {
11
- "epoch": 0.09,
12
- "learning_rate": 1.28e-05,
13
- "loss": 3.0058,
14
- "step": 100
15
- },
16
- {
17
- "epoch": 0.18,
18
- "learning_rate": 2.56e-05,
19
- "loss": 2.931,
20
- "step": 200
21
- },
22
- {
23
- "epoch": 0.27,
24
- "learning_rate": 3.84e-05,
25
- "loss": 2.9422,
26
- "step": 300
27
- },
28
- {
29
- "epoch": 0.36,
30
- "learning_rate": 5.12e-05,
31
- "loss": 2.9606,
32
- "step": 400
33
- },
34
- {
35
- "epoch": 0.45,
36
- "learning_rate": 6.4e-05,
37
- "loss": 2.9656,
38
- "step": 500
39
- },
40
- {
41
- "epoch": 0.5,
42
- "eval_gen_len": 69.515,
43
- "eval_loss": 2.8840489387512207,
44
- "eval_rouge1": 43.9038,
45
- "eval_rouge2": 21.2396,
46
- "eval_rougeL": 31.099,
47
- "eval_rougeLsum": 40.8409,
48
- "eval_runtime": 8242.0465,
49
- "eval_samples_per_second": 1.622,
50
- "eval_steps_per_second": 0.405,
51
- "step": 560
52
- }
53
- ],
54
- "max_steps": 1121,
55
- "num_train_epochs": 1,
56
- "total_flos": 2.070798218085335e+17,
57
- "trial_name": null,
58
- "trial_params": null
59
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c74e3795b3c07d65c60dc9c1ac7384595751fa7d2fd4cd81dc4acde86aa7fbb1
3
- size 3567
 
 
 
 
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b3dce194dc738268dd2c1f5262d8d2f2ba2fe4941f8e5db5a86356c6e0edd8b
3
  size 2279605745
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4042f04d7c8ddfdfa52e5cddd47f1a6e3fa58520dbe86f5cc5ca0071c42e8eb3
3
  size 2279605745
trainer_state.json CHANGED
@@ -1,120 +1,117 @@
1
  {
2
- "best_metric": 39.4265,
3
- "best_model_checkpoint": "pegasus-newsroom-cnn_full-adafactor-bs6/checkpoint-598",
4
- "epoch": 0.49986416734582995,
5
- "global_step": 1495,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
- "epoch": 0.1,
12
- "learning_rate": 7.839464882943145e-05,
13
- "loss": 3.2894,
14
- "step": 299
15
  },
16
  {
17
- "epoch": 0.1,
18
- "eval_gen_len": 64.6904,
19
- "eval_loss": 2.9463675022125244,
20
- "eval_rouge1": 39.4079,
21
- "eval_rouge2": 18.3064,
22
- "eval_rougeL": 28.093,
23
- "eval_rougeLsum": 36.5182,
24
- "eval_runtime": 1204.537,
25
- "eval_samples_per_second": 2.075,
26
- "eval_steps_per_second": 0.346,
27
- "step": 299
28
  },
29
  {
30
- "epoch": 0.2,
31
- "learning_rate": 0.00015812709030100337,
32
- "loss": 3.0427,
33
- "step": 598
34
  },
35
  {
36
- "epoch": 0.2,
37
- "eval_gen_len": 60.5696,
38
- "eval_loss": 2.93066143989563,
39
- "eval_rouge1": 39.4265,
40
- "eval_rouge2": 18.2924,
41
- "eval_rougeL": 28.247,
42
- "eval_rougeLsum": 36.6382,
43
- "eval_runtime": 1153.059,
44
- "eval_samples_per_second": 2.168,
45
- "eval_steps_per_second": 0.362,
46
- "step": 598
47
  },
48
  {
49
- "epoch": 0.3,
50
- "learning_rate": 0.00014086956521739132,
51
- "loss": 3.1017,
52
- "step": 897
53
  },
54
  {
55
- "epoch": 0.3,
56
- "eval_gen_len": 58.5172,
57
- "eval_loss": 2.9890542030334473,
58
- "eval_rouge1": 39.0977,
59
- "eval_rouge2": 17.9198,
60
- "eval_rougeL": 27.9078,
61
- "eval_rougeLsum": 36.2363,
62
- "eval_runtime": 1087.3782,
63
- "eval_samples_per_second": 2.299,
64
- "eval_steps_per_second": 0.383,
65
- "step": 897
66
  },
67
  {
68
- "epoch": 0.4,
69
- "learning_rate": 0.00012100334448160536,
70
- "loss": 3.2891,
71
- "step": 1196
72
  },
73
  {
74
- "epoch": 0.4,
75
- "eval_gen_len": 45.0232,
76
- "eval_loss": 3.57556414604187,
77
- "eval_rouge1": 29.5555,
78
- "eval_rouge2": 11.7552,
79
- "eval_rougeL": 22.4675,
80
- "eval_rougeLsum": 27.2432,
81
- "eval_runtime": 981.4008,
82
- "eval_samples_per_second": 2.547,
83
- "eval_steps_per_second": 0.425,
84
- "step": 1196
85
  },
86
  {
87
- "epoch": 0.5,
88
- "learning_rate": 0.00011110367892976589,
89
- "loss": 637.0317,
90
- "step": 1495
91
  },
92
  {
93
- "epoch": 0.5,
94
- "eval_gen_len": 1.0,
95
- "eval_loss": NaN,
96
- "eval_rouge1": 0.0,
97
- "eval_rouge2": 0.0,
98
- "eval_rougeL": 0.0,
99
- "eval_rougeLsum": 0.0,
100
- "eval_runtime": 1564.693,
101
- "eval_samples_per_second": 1.598,
102
- "eval_steps_per_second": 0.267,
103
- "step": 1495
104
  },
105
  {
106
- "epoch": 0.5,
107
- "step": 1495,
108
- "total_flos": 2.0734783091992166e+17,
109
- "train_loss": 129.9509090768055,
110
- "train_runtime": 31793.2991,
111
- "train_samples_per_second": 9.031,
112
- "train_steps_per_second": 0.094
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
113
  }
114
  ],
115
- "max_steps": 2990,
116
  "num_train_epochs": 1,
117
- "total_flos": 2.0734783091992166e+17,
118
  "trial_name": null,
119
  "trial_params": null
120
  }
 
1
  {
2
+ "best_metric": 43.9897,
3
+ "best_model_checkpoint": "pegasus-newsroom-cnn_full-adafactor-bs6/checkpoint-1120",
4
+ "epoch": 0.9995123922038479,
5
+ "global_step": 1121,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
+ "epoch": 0.09,
12
+ "learning_rate": 1.28e-05,
13
+ "loss": 3.0058,
14
+ "step": 100
15
  },
16
  {
17
+ "epoch": 0.18,
18
+ "learning_rate": 2.56e-05,
19
+ "loss": 2.931,
20
+ "step": 200
 
 
 
 
 
 
 
21
  },
22
  {
23
+ "epoch": 0.27,
24
+ "learning_rate": 3.84e-05,
25
+ "loss": 2.9422,
26
+ "step": 300
27
  },
28
  {
29
+ "epoch": 0.36,
30
+ "learning_rate": 5.12e-05,
31
+ "loss": 2.9606,
32
+ "step": 400
 
 
 
 
 
 
 
33
  },
34
  {
35
+ "epoch": 0.45,
36
+ "learning_rate": 6.4e-05,
37
+ "loss": 2.9656,
38
+ "step": 500
39
  },
40
  {
41
+ "epoch": 0.5,
42
+ "eval_gen_len": 69.515,
43
+ "eval_loss": 2.8840489387512207,
44
+ "eval_rouge1": 43.9038,
45
+ "eval_rouge2": 21.2396,
46
+ "eval_rougeL": 31.099,
47
+ "eval_rougeLsum": 40.8409,
48
+ "eval_runtime": 8242.0465,
49
+ "eval_samples_per_second": 1.622,
50
+ "eval_steps_per_second": 0.405,
51
+ "step": 560
52
  },
53
  {
54
+ "epoch": 0.53,
55
+ "learning_rate": 5.369404186795491e-05,
56
+ "loss": 2.9878,
57
+ "step": 600
58
  },
59
  {
60
+ "epoch": 0.62,
61
+ "learning_rate": 4.338808373590982e-05,
62
+ "loss": 3.0051,
63
+ "step": 700
 
 
 
 
 
 
 
64
  },
65
  {
66
+ "epoch": 0.71,
67
+ "learning_rate": 3.308212560386473e-05,
68
+ "loss": 3.0096,
69
+ "step": 800
70
  },
71
  {
72
+ "epoch": 0.8,
73
+ "learning_rate": 2.2776167471819646e-05,
74
+ "loss": 3.0016,
75
+ "step": 900
 
 
 
 
 
 
 
76
  },
77
  {
78
+ "epoch": 0.89,
79
+ "learning_rate": 1.2470209339774557e-05,
80
+ "loss": 3.0071,
81
+ "step": 1000
82
+ },
83
+ {
84
+ "epoch": 0.98,
85
+ "learning_rate": 2.1642512077294685e-06,
86
+ "loss": 2.9976,
87
+ "step": 1100
88
+ },
89
+ {
90
+ "epoch": 1.0,
91
+ "eval_gen_len": 72.362,
92
+ "eval_loss": 2.8759312629699707,
93
+ "eval_rouge1": 43.9897,
94
+ "eval_rouge2": 21.287,
95
+ "eval_rougeL": 31.1064,
96
+ "eval_rougeLsum": 40.9035,
97
+ "eval_runtime": 8420.7089,
98
+ "eval_samples_per_second": 1.588,
99
+ "eval_steps_per_second": 0.397,
100
+ "step": 1120
101
+ },
102
+ {
103
+ "epoch": 1.0,
104
+ "step": 1121,
105
+ "total_flos": 4.145227793147167e+17,
106
+ "train_loss": 2.983753987022216,
107
+ "train_runtime": 51642.0714,
108
+ "train_samples_per_second": 5.56,
109
+ "train_steps_per_second": 0.022
110
  }
111
  ],
112
+ "max_steps": 1121,
113
  "num_train_epochs": 1,
114
+ "total_flos": 4.145227793147167e+17,
115
  "trial_name": null,
116
  "trial_params": null
117
  }