distil_bert_base_en / config.json
Divyasreepat's picture
Upload folder using huggingface_hub
1df14ac verified
raw
history blame
515 Bytes
{
"module": "keras_nlp.src.models.distil_bert.distil_bert_backbone",
"class_name": "DistilBertBackbone",
"config": {
"name": "distil_bert_backbone",
"trainable": true,
"vocabulary_size": 28996,
"num_layers": 6,
"num_heads": 12,
"hidden_dim": 768,
"intermediate_dim": 3072,
"dropout": 0.1,
"max_sequence_length": 512
},
"registered_name": "keras_nlp>DistilBertBackbone",
"assets": [],
"weights": "model.weights.h5"
}