{ "init_args": [ { "vocab_size": 12000, "hidden_size": 768, "intermediate_size": 1536, "num_hidden_layers": 24, "max_position_embeddings": 512, "type_vocab_size": 2, "initializer_range": 0.02, "attention_key_size": 128, "norm_eps": 1e-12, "pad_token_id": 0, "hidden_dropout_prob": 0.1, "attention_probs_dropout_prob": 0.1, "hidden_act": "swish", "use_bias": false, "normalization": "softmax_plus", "attention_scale": true, "init_class": "GAUAlphaModel" } ], "init_class": "GAUAlphaForMaskedLM" }