{ "dim": 4096, "n_layers": 32, "n_heads": 32, "vocab_size": 50432, "norm_eps": 1e-5, "seq_len": 2048, "weight_tying": false, "apply_qk_norm": false, "norm_type": "gain_only_layer_norm", "positional_embedding_type": "rotary", "ffn_type": "swiglu" }