mixtral-8x7b-32kseqlen / params.json
someone13574's picture
Upload 3 files
aa0c59b
raw
history blame
193 Bytes
{"dim": 4096, "n_layers": 32, "head_dim": 128, "hidden_dim": 14336, "n_heads": 32, "n_kv_heads": 8, "norm_eps": 1e-05, "vocab_size": 32000, "moe": {"num_experts_per_tok": 2, "num_experts": 8}}