{"dropout": 0.0, "identity_init": true, "include_input": true, "rank": null, "shared_mlp_hidden_sizes": [], "sublayers": false, "num_layers": 12, "vocab_size": 50304, "bias": true, "d_model": 768, "mlp_hidden_sizes": []}