patrickvonplaten commited on
Commit
228e307
1 Parent(s): 9b413d4

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +1 -25
config.json CHANGED
@@ -1,31 +1,12 @@
1
  {
2
  "activation_dropout": 0.0,
3
  "activation_function": "gelu",
4
- "add_bias_logits": false,
5
- "add_final_layer_norm": false,
6
  "architectures": [
7
  "BartForConditionalGeneration"
8
  ],
9
- "attention_dilation": [
10
- 1,
11
- 1,
12
- 1,
13
- 1,
14
- 1,
15
- 1
16
- ],
17
  "attention_dropout": 0.0,
18
- "attention_mode": "sliding_chunks",
19
  "attention_probs_dropout_prob": 0.0,
20
- "attention_window": [
21
- 512,
22
- 512,
23
- 512,
24
- 512,
25
- 512,
26
- 512
27
- ],
28
- "autoregressive": false,
29
  "bos_token_id": 0,
30
  "classif_dropout": 0.0,
31
  "d_model": 768,
@@ -39,7 +20,6 @@
39
  "encoder_layerdrop": 0.0,
40
  "encoder_layers": 6,
41
  "eos_token_id": 2,
42
- "extra_pos_embeddings": 2,
43
  "gradient_checkpointing": false,
44
  "id2label": {
45
  "0": "LABEL_0",
@@ -56,11 +36,7 @@
56
  "max_decoder_position_embeddings": 1024,
57
  "max_encoder_position_embeddings": 16384,
58
  "model_type": "bart",
59
- "normalize_before": false,
60
- "normalize_embedding": true,
61
  "num_hidden_layers": 6,
62
  "pad_token_id": 1,
63
- "scale_embedding": false,
64
- "static_position_embeddings": false,
65
  "vocab_size": 50265
66
  }
 
1
  {
2
  "activation_dropout": 0.0,
3
  "activation_function": "gelu",
 
 
4
  "architectures": [
5
  "BartForConditionalGeneration"
6
  ],
 
 
 
 
 
 
 
 
7
  "attention_dropout": 0.0,
 
8
  "attention_probs_dropout_prob": 0.0,
9
+ "attention_window": 1024,
 
 
 
 
 
 
 
 
10
  "bos_token_id": 0,
11
  "classif_dropout": 0.0,
12
  "d_model": 768,
 
20
  "encoder_layerdrop": 0.0,
21
  "encoder_layers": 6,
22
  "eos_token_id": 2,
 
23
  "gradient_checkpointing": false,
24
  "id2label": {
25
  "0": "LABEL_0",
 
36
  "max_decoder_position_embeddings": 1024,
37
  "max_encoder_position_embeddings": 16384,
38
  "model_type": "bart",
 
 
39
  "num_hidden_layers": 6,
40
  "pad_token_id": 1,
 
 
41
  "vocab_size": 50265
42
  }