{ "auto_map": { "AutoConfig": "configuration_btlm.BTLMConfig", "AutoModel": "modeling_btlm.BTLMModel", "AutoModelForSequenceClassification": "modeling_btlm.BTLMForSequenceClassification", "AutoModelForTokenClassification": "modeling_btlm.BTLMForTokenClassification", "AutoModelForQuestionAnswering": "modeling_btlm.BTLMForQuestionAnswering", "AutoModelForCausalLM": "modeling_btlm.BTLMLMHeadModel" }, "architectures": [ "BTLMLMHeadModel" ], "attn_pdrop": 0.0, "bos_token_id": 50256, "eos_token_id": 50256, "scale_attn_weights": true, "resid_pdrop": 0.0, "embeddings_scale": 14.6, "n_inner": 6826, "n_embd": 2560, "layer_norm_epsilon": 1e-05, "n_positions": 8192, "activation_function": "swiglu", "n_head": 32, "n_layer": 32, "model_type": "btlm", "width_scale": 0.1, "output_logits_scale": 0.22200000000000003, "position_embedding_type": "alibi", "scale_qk_dot_by_d": true, "tie_word_embeddings": true, "vocab_size": 50257, "embd_pdrop": 0.0 }