debbiesoon commited on
Commit
7474a41
·
1 Parent(s): 22c283e

End of training

Browse files
Files changed (5) hide show
  1. config.json +19 -16
  2. merges.txt +1 -1
  3. pytorch_model.bin +2 -2
  4. tokenizer_config.json +2 -3
  5. vocab.json +0 -0
config.json CHANGED
@@ -1,14 +1,13 @@
1
  {
2
- "_name_or_path": "sshleifer/distilbart-xsum-12-3",
3
  "_num_labels": 3,
4
  "activation_dropout": 0.0,
5
  "activation_function": "gelu",
6
- "add_bias_logits": false,
7
  "add_final_layer_norm": false,
8
  "architectures": [
9
  "BartForConditionalGeneration"
10
  ],
11
- "attention_dropout": 0.1,
12
  "bos_token_id": 0,
13
  "classif_dropout": 0.0,
14
  "classifier_dropout": 0.0,
@@ -16,7 +15,7 @@
16
  "decoder_attention_heads": 16,
17
  "decoder_ffn_dim": 4096,
18
  "decoder_layerdrop": 0.0,
19
- "decoder_layers": 3,
20
  "decoder_start_token_id": 2,
21
  "dropout": 0.1,
22
  "early_stopping": true,
@@ -25,11 +24,8 @@
25
  "encoder_layerdrop": 0.0,
26
  "encoder_layers": 12,
27
  "eos_token_id": 2,
28
- "eos_token_ids": [
29
- 2
30
- ],
31
- "extra_pos_embeddings": 2,
32
- "force_bos_token_to_be_generated": false,
33
  "forced_eos_token_id": 2,
34
  "gradient_checkpointing": false,
35
  "id2label": {
@@ -44,22 +40,29 @@
44
  "LABEL_1": 1,
45
  "LABEL_2": 2
46
  },
47
- "max_length": 62,
 
48
  "max_position_embeddings": 1024,
49
- "min_length": 11,
50
  "model_type": "bart",
51
  "no_repeat_ngram_size": 3,
52
  "normalize_before": false,
53
- "normalize_embedding": true,
54
- "num_beams": 6,
55
  "num_hidden_layers": 12,
56
  "output_past": true,
57
  "pad_token_id": 1,
58
  "prefix": " ",
59
- "save_step": 58,
60
  "scale_embedding": false,
61
- "static_position_embeddings": false,
62
- "task_specific_params": {},
 
 
 
 
 
 
 
 
63
  "torch_dtype": "float32",
64
  "transformers_version": "4.23.1",
65
  "use_cache": true,
 
1
  {
2
+ "_name_or_path": "facebook/bart-large-cnn",
3
  "_num_labels": 3,
4
  "activation_dropout": 0.0,
5
  "activation_function": "gelu",
 
6
  "add_final_layer_norm": false,
7
  "architectures": [
8
  "BartForConditionalGeneration"
9
  ],
10
+ "attention_dropout": 0.0,
11
  "bos_token_id": 0,
12
  "classif_dropout": 0.0,
13
  "classifier_dropout": 0.0,
 
15
  "decoder_attention_heads": 16,
16
  "decoder_ffn_dim": 4096,
17
  "decoder_layerdrop": 0.0,
18
+ "decoder_layers": 12,
19
  "decoder_start_token_id": 2,
20
  "dropout": 0.1,
21
  "early_stopping": true,
 
24
  "encoder_layerdrop": 0.0,
25
  "encoder_layers": 12,
26
  "eos_token_id": 2,
27
+ "force_bos_token_to_be_generated": true,
28
+ "forced_bos_token_id": 0,
 
 
 
29
  "forced_eos_token_id": 2,
30
  "gradient_checkpointing": false,
31
  "id2label": {
 
40
  "LABEL_1": 1,
41
  "LABEL_2": 2
42
  },
43
+ "length_penalty": 2.0,
44
+ "max_length": 142,
45
  "max_position_embeddings": 1024,
46
+ "min_length": 56,
47
  "model_type": "bart",
48
  "no_repeat_ngram_size": 3,
49
  "normalize_before": false,
50
+ "num_beams": 4,
 
51
  "num_hidden_layers": 12,
52
  "output_past": true,
53
  "pad_token_id": 1,
54
  "prefix": " ",
 
55
  "scale_embedding": false,
56
+ "task_specific_params": {
57
+ "summarization": {
58
+ "early_stopping": true,
59
+ "length_penalty": 2.0,
60
+ "max_length": 142,
61
+ "min_length": 56,
62
+ "no_repeat_ngram_size": 3,
63
+ "num_beams": 4
64
+ }
65
+ },
66
  "torch_dtype": "float32",
67
  "transformers_version": "4.23.1",
68
  "use_cache": true,
merges.txt CHANGED
@@ -1,4 +1,4 @@
1
- #version: 0.2 - Trained by `huggingface/tokenizers`
2
  Ġ t
3
  Ġ a
4
  h e
 
1
+ #version: 0.2
2
  Ġ t
3
  Ġ a
4
  h e
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:684468bb56035050a220c19f71f253ad3957f4cc34661e0c92bfde5825369d42
3
- size 1020774805
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab503efc55df0c3c9b653a6c3fc42b2630369ad48ad43f22c28dc6552d7ed70a
3
+ size 1625533697
tokenizer_config.json CHANGED
@@ -34,7 +34,7 @@
34
  "single_word": false
35
  },
36
  "model_max_length": 1024,
37
- "name_or_path": "sshleifer/distilbart-xsum-12-3",
38
  "pad_token": {
39
  "__type": "AddedToken",
40
  "content": "<pad>",
@@ -51,9 +51,8 @@
51
  "rstrip": false,
52
  "single_word": false
53
  },
54
- "special_tokens_map_file": "/root/.cache/huggingface/hub/models--sshleifer--distilbart-xsum-12-3/snapshots/1d2bfbc16dcdd28720f9f1d37be764e5cc5c78c8/special_tokens_map.json",
55
  "tokenizer_class": "BartTokenizer",
56
- "trim_offsets": true,
57
  "unk_token": {
58
  "__type": "AddedToken",
59
  "content": "<unk>",
 
34
  "single_word": false
35
  },
36
  "model_max_length": 1024,
37
+ "name_or_path": "facebook/bart-large-cnn",
38
  "pad_token": {
39
  "__type": "AddedToken",
40
  "content": "<pad>",
 
51
  "rstrip": false,
52
  "single_word": false
53
  },
54
+ "special_tokens_map_file": null,
55
  "tokenizer_class": "BartTokenizer",
 
56
  "unk_token": {
57
  "__type": "AddedToken",
58
  "content": "<unk>",
vocab.json CHANGED
The diff for this file is too large to render. See raw diff