{ "num_layers":40, "vocab_size":50176, "hidden_size":3072, "num_attention_heads":48, "embedding_dropout_prob":0.1, "attention_dropout_prob":0.1, "output_dropout_prob":0.1, "max_sequence_length":1024, "max_memory_length":512, "checkpoint_activations":false, "checkpoint_num_layers":1, "parallel_output":true, "relative_encoding":true }