Update config.json
Browse files- config.json +2 -2
config.json
CHANGED
@@ -22,7 +22,7 @@
|
|
22 |
"encoder_attention_heads": 32,
|
23 |
"encoder_ffn_dim": 5120,
|
24 |
"encoder_layerdrop": 0.0,
|
25 |
-
"encoder_layers":
|
26 |
"encoder_no_repeat_ngram_size": 3,
|
27 |
"eos_token_id": 2,
|
28 |
"extra_layer_norm": false,
|
@@ -52,7 +52,7 @@
|
|
52 |
"normalize_before": true,
|
53 |
"normalize_embedding": false,
|
54 |
"num_beams": 10,
|
55 |
-
"num_hidden_layers":
|
56 |
"pad_token_id": 0,
|
57 |
"scale_embedding": true,
|
58 |
"static_position_embeddings": false,
|
|
|
22 |
"encoder_attention_heads": 32,
|
23 |
"encoder_ffn_dim": 5120,
|
24 |
"encoder_layerdrop": 0.0,
|
25 |
+
"encoder_layers": 12,
|
26 |
"encoder_no_repeat_ngram_size": 3,
|
27 |
"eos_token_id": 2,
|
28 |
"extra_layer_norm": false,
|
|
|
52 |
"normalize_before": true,
|
53 |
"normalize_embedding": false,
|
54 |
"num_beams": 10,
|
55 |
+
"num_hidden_layers": 12,
|
56 |
"pad_token_id": 0,
|
57 |
"scale_embedding": true,
|
58 |
"static_position_embeddings": false,
|