pm_task_gpt2 / head_config.json
selimyagci's picture
Upload model
a57a1b3 verified
{
"config": {
"activation_function": "gelu",
"bias": true,
"dropout_prob": null,
"embedding_size": 768,
"head_type": "causal_lm",
"label2id": null,
"layer_norm": true,
"layers": 2,
"shift_labels": true,
"vocab_size": 50257
},
"hidden_size": 768,
"model_class": "GPT2AdapterModel",
"model_name": "distilgpt2",
"model_type": "gpt2",
"name": "pm_task",
"version": "adapters.1.0.0"
}