Michael-Vptn commited on
Commit
724623f
1 Parent(s): 6413eb1

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +4 -4
config.json CHANGED
@@ -18,16 +18,16 @@
18
  "layer_norm_epsilon": 1e-06,
19
  "model_type": "t5",
20
  "n_positions": 512,
21
- "num_decoder_layers": 6,
22
- "num_heads": 8,
23
- "num_layers": 6,
24
  "output_past": true,
25
  "pad_token_id": 0,
26
  "relative_attention_max_distance": 128,
27
  "relative_attention_num_buckets": 32,
28
  "task_specific_params": {
29
  "summarization": {
30
- "early_stopping": true,
31
  "length_penalty": 2.0,
32
  "max_length": 200,
33
  "min_length": 30,
 
18
  "layer_norm_epsilon": 1e-06,
19
  "model_type": "t5",
20
  "n_positions": 512,
21
+ "num_decoder_layers": 12,
22
+ "num_heads": 12,
23
+ "num_layers": 12,
24
  "output_past": true,
25
  "pad_token_id": 0,
26
  "relative_attention_max_distance": 128,
27
  "relative_attention_num_buckets": 32,
28
  "task_specific_params": {
29
  "summarization": {
30
+ "early_stopping": false,
31
  "length_penalty": 2.0,
32
  "max_length": 200,
33
  "min_length": 30,