bhpardo commited on
Commit
8cd4846
1 Parent(s): 7d8512f

Training in progress, step 500

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "google/pegasus-large",
3
  "activation_dropout": 0.1,
4
  "activation_function": "relu",
5
  "add_bias_logits": false,
@@ -17,13 +17,14 @@
17
  "decoder_layerdrop": 0.0,
18
  "decoder_layers": 16,
19
  "decoder_start_token_id": 0,
 
20
  "dropout": 0.1,
21
  "encoder_attention_heads": 16,
22
  "encoder_ffn_dim": 4096,
23
  "encoder_layerdrop": 0.0,
24
  "encoder_layers": 16,
25
  "eos_token_id": 1,
26
- "extra_pos_embeddings": 1,
27
  "force_bos_token_to_be_generated": false,
28
  "forced_eos_token_id": 1,
29
  "gradient_checkpointing": false,
@@ -39,9 +40,9 @@
39
  "LABEL_1": 1,
40
  "LABEL_2": 2
41
  },
42
- "length_penalty": 0.8,
43
- "max_length": 256,
44
- "max_position_embeddings": 1024,
45
  "model_type": "pegasus",
46
  "normalize_before": true,
47
  "normalize_embedding": false,
@@ -50,73 +51,6 @@
50
  "pad_token_id": 0,
51
  "scale_embedding": true,
52
  "static_position_embeddings": true,
53
- "task_specific_params": {
54
- "summarization_aeslc": {
55
- "length_penalty": 0.6,
56
- "max_length": 32,
57
- "max_position_embeddings": 512
58
- },
59
- "summarization_arxiv": {
60
- "length_penalty": 0.8,
61
- "max_length": 256,
62
- "max_position_embeddings": 1024
63
- },
64
- "summarization_big_patent": {
65
- "length_penalty": 0.7,
66
- "max_length": 256,
67
- "max_position_embeddings": 1024
68
- },
69
- "summarization_billsum": {
70
- "length_penalty": 0.6,
71
- "max_length": 256,
72
- "max_position_embeddings": 1024
73
- },
74
- "summarization_cnn_dailymail": {
75
- "length_penalty": 0.8,
76
- "max_length": 128,
77
- "max_position_embeddings": 1024
78
- },
79
- "summarization_gigaword": {
80
- "length_penalty": 0.6,
81
- "max_length": 32,
82
- "max_position_embeddings": 128
83
- },
84
- "summarization_large": {
85
- "length_penalty": 0.8,
86
- "max_length": 256,
87
- "max_position_embeddings": 1024
88
- },
89
- "summarization_multi_news": {
90
- "length_penalty": 0.8,
91
- "max_length": 256,
92
- "max_position_embeddings": 1024
93
- },
94
- "summarization_newsroom": {
95
- "length_penalty": 0.8,
96
- "max_length": 128,
97
- "max_position_embeddings": 512
98
- },
99
- "summarization_pubmed": {
100
- "length_penalty": 0.8,
101
- "max_length": 256,
102
- "max_position_embeddings": 1024
103
- },
104
- "summarization_reddit_tifu": {
105
- "length_penalty": 0.6,
106
- "max_length": 128,
107
- "max_position_embeddings": 512
108
- },
109
- "summarization_wikihow": {
110
- "length_penalty": 0.6,
111
- "max_length": 256,
112
- "max_position_embeddings": 512
113
- },
114
- "summarization_xsum": {
115
- "length_penalty": 0.8,
116
- "max_length": 64,
117
- "max_position_embeddings": 512
118
- }
119
- },
120
  "torch_dtype": "float32",
121
  "transformers_version": "4.26.1",
122
  "use_cache": true,
 
1
  {
2
+ "_name_or_path": "google/pegasus-xsum",
3
  "activation_dropout": 0.1,
4
  "activation_function": "relu",
5
  "add_bias_logits": false,
 
17
  "decoder_layerdrop": 0.0,
18
  "decoder_layers": 16,
19
  "decoder_start_token_id": 0,
20
+ "do_blenderbot_90_layernorm": false,
21
  "dropout": 0.1,
22
  "encoder_attention_heads": 16,
23
  "encoder_ffn_dim": 4096,
24
  "encoder_layerdrop": 0.0,
25
  "encoder_layers": 16,
26
  "eos_token_id": 1,
27
+ "extra_pos_embeddings": 0,
28
  "force_bos_token_to_be_generated": false,
29
  "forced_eos_token_id": 1,
30
  "gradient_checkpointing": false,
 
40
  "LABEL_1": 1,
41
  "LABEL_2": 2
42
  },
43
+ "length_penalty": 0.6,
44
+ "max_length": 64,
45
+ "max_position_embeddings": 512,
46
  "model_type": "pegasus",
47
  "normalize_before": true,
48
  "normalize_embedding": false,
 
51
  "pad_token_id": 0,
52
  "scale_embedding": true,
53
  "static_position_embeddings": true,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
54
  "torch_dtype": "float32",
55
  "transformers_version": "4.26.1",
56
  "use_cache": true,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fce6f49c422cbc4e29152de1a85d27ff412a6f6b20cd573e3f0a4838ec3863ad
3
- size 2283804653
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9014d68692d30326f79504fbc70d22e8cc95ed0e6248fa1e532089e7239a3468
3
+ size 2279610349
runs/Mar05_09-38-01_c77a45fa24a4/1678009653.0877988/events.out.tfevents.1678009653.c77a45fa24a4.330.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb72d3cf31485d6fc40363f653b8a25cc9813804f43fad1888b558f26b3f97cb
3
+ size 5878
runs/Mar05_09-38-01_c77a45fa24a4/events.out.tfevents.1678009653.c77a45fa24a4.330.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f62debf76b9bf677740052c6647337d319205cf9be551ce64209952add00401a
3
+ size 4838
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -107,11 +107,10 @@
107
  "full_tokenizer_file": null,
108
  "mask_token": "<mask_2>",
109
  "mask_token_sent": "<mask_1>",
110
- "model_max_length": 1024,
111
- "name_or_path": "google/pegasus-large",
112
  "offset": 103,
113
  "pad_token": "<pad>",
114
- "sp_model_kwargs": {},
115
  "special_tokens_map_file": null,
116
  "tokenizer_class": "PegasusTokenizer",
117
  "unk_token": "<unk>"
 
107
  "full_tokenizer_file": null,
108
  "mask_token": "<mask_2>",
109
  "mask_token_sent": "<mask_1>",
110
+ "model_max_length": 512,
111
+ "name_or_path": "google/pegasus-xsum",
112
  "offset": 103,
113
  "pad_token": "<pad>",
 
114
  "special_tokens_map_file": null,
115
  "tokenizer_class": "PegasusTokenizer",
116
  "unk_token": "<unk>"
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ac718486f50cdda5d8fe4106204b7cbf0b1548f1883ca5882dcdf76fe54fefe
3
  size 3579
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42d6387ec40f6df4532b1c96fdda4398c940248e508b00a290942f9c676e45b2
3
  size 3579