TheAIchemist13 commited on
Commit
d708f56
1 Parent(s): 7f3b454

Upload 2 files

Browse files
Files changed (2) hide show
  1. config.json +3 -3
  2. preprocessor_config.json +0 -0
config.json CHANGED
@@ -13,15 +13,15 @@
13
  ],
14
  "bos_token_id": 50257,
15
  "classifier_proj_size": 256,
16
- "d_model": 768,
17
  "decoder_attention_heads": 12,
18
- "decoder_ffn_dim": 3072,
19
  "decoder_layerdrop": 0.0,
20
  "decoder_layers": 12,
21
  "decoder_start_token_id": 50258,
22
  "dropout": 0.0,
23
  "encoder_attention_heads": 12,
24
- "encoder_ffn_dim": 3072,
25
  "encoder_layerdrop": 0.0,
26
  "encoder_layers": 12,
27
  "eos_token_id": 50257,
 
13
  ],
14
  "bos_token_id": 50257,
15
  "classifier_proj_size": 256,
16
+ "d_model": 384,
17
  "decoder_attention_heads": 12,
18
+ "decoder_ffn_dim": 1536,
19
  "decoder_layerdrop": 0.0,
20
  "decoder_layers": 12,
21
  "decoder_start_token_id": 50258,
22
  "dropout": 0.0,
23
  "encoder_attention_heads": 12,
24
+ "encoder_ffn_dim": 1536,
25
  "encoder_layerdrop": 0.0,
26
  "encoder_layers": 12,
27
  "eos_token_id": 50257,
preprocessor_config.json CHANGED
The diff for this file is too large to render. See raw diff