{ | |
"hidden_dim": 2048, | |
"n_layers": 24, | |
"n_heads": 16, | |
"seq_len": 2048, | |
"vocab_size": 50432, | |
"post_embed_norm": false, | |
"weight_tying": false, | |
"qk_norm": false, | |
"ffn_type": "swiglu", | |
"model_norm": "default_layer_norm", | |
"positional_embedding_type": "head_rotary" | |
} |