Files changed (1) hide show
  1. config.json +1 -1
config.json CHANGED
@@ -14,7 +14,7 @@
14
  "intermediate_size": null,
15
  "layer_norm_epsilon": 1e-05,
16
  "model_type": "rwkv5",
17
- "num_attention_heads": 64,
18
  "num_hidden_layers": 24,
19
  "rescale_every": 6,
20
  "tie_word_embeddings": false,
 
14
  "intermediate_size": null,
15
  "layer_norm_epsilon": 1e-05,
16
  "model_type": "rwkv5",
17
+ "num_attention_heads": 32,
18
  "num_hidden_layers": 24,
19
  "rescale_every": 6,
20
  "tie_word_embeddings": false,