patrickvonplaten commited on
Commit
9aba671
1 Parent(s): bd58652

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +1 -47
config.json CHANGED
@@ -2,43 +2,12 @@
2
  "_num_labels": 3,
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
- "add_bias_logits": false,
6
- "add_final_layer_norm": false,
7
  "architectures": [
8
  "BartForConditionalGeneration"
9
  ],
10
- "attention_dilation": [
11
- 1,
12
- 1,
13
- 1,
14
- 1,
15
- 1,
16
- 1,
17
- 1,
18
- 1,
19
- 1,
20
- 1,
21
- 1,
22
- 1
23
- ],
24
  "attention_dropout": 0.0,
25
- "attention_mode": "sliding_chunks",
26
  "attention_probs_dropout_prob": 0.0,
27
- "attention_window": [
28
- 512,
29
- 512,
30
- 512,
31
- 512,
32
- 512,
33
- 512,
34
- 512,
35
- 512,
36
- 512,
37
- 512,
38
- 512,
39
- 512
40
- ],
41
- "autoregressive": false,
42
  "bos_token_id": 0,
43
  "classif_dropout": 0.0,
44
  "d_model": 1024,
@@ -53,7 +22,6 @@
53
  "encoder_layerdrop": 0.0,
54
  "encoder_layers": 12,
55
  "eos_token_id": 2,
56
- "extra_pos_embeddings": 2,
57
  "gradient_checkpointing": false,
58
  "id2label": {
59
  "0": "LABEL_0",
@@ -70,23 +38,9 @@
70
  "max_decoder_position_embeddings": 1024,
71
  "max_encoder_position_embeddings": 16384,
72
  "model_type": "bart",
73
- "normalize_before": false,
74
- "normalize_embedding": true,
75
  "num_hidden_layers": 12,
76
  "output_past": false,
77
  "pad_token_id": 1,
78
  "prefix": " ",
79
- "scale_embedding": false,
80
- "static_position_embeddings": false,
81
- "task_specific_params": {
82
- "summarization": {
83
- "early_stopping": true,
84
- "length_penalty": 2.0,
85
- "max_length": 142,
86
- "min_length": 56,
87
- "no_repeat_ngram_size": 3,
88
- "num_beams": 4
89
- }
90
- },
91
  "vocab_size": 50265
92
  }
 
2
  "_num_labels": 3,
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
 
 
5
  "architectures": [
6
  "BartForConditionalGeneration"
7
  ],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "attention_dropout": 0.0,
 
9
  "attention_probs_dropout_prob": 0.0,
10
+ "attention_window": 1024,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  "bos_token_id": 0,
12
  "classif_dropout": 0.0,
13
  "d_model": 1024,
 
22
  "encoder_layerdrop": 0.0,
23
  "encoder_layers": 12,
24
  "eos_token_id": 2,
 
25
  "gradient_checkpointing": false,
26
  "id2label": {
27
  "0": "LABEL_0",
 
38
  "max_decoder_position_embeddings": 1024,
39
  "max_encoder_position_embeddings": 16384,
40
  "model_type": "bart",
 
 
41
  "num_hidden_layers": 12,
42
  "output_past": false,
43
  "pad_token_id": 1,
44
  "prefix": " ",
 
 
 
 
 
 
 
 
 
 
 
 
45
  "vocab_size": 50265
46
  }