nerugm-lora-r2a0d0.05 / nerugm-lora /adapter_config.json
apwic's picture
Model save
e80b583 verified
raw
history blame contribute delete
No virus
562 Bytes
{
"config": {
"alpha": 0,
"architecture": "lora",
"attn_matrices": [
"q",
"v"
],
"composition_mode": "add",
"dropout": 0.05,
"init_weights": "lora",
"intermediate_lora": false,
"leave_out": [],
"output_lora": false,
"r": 2,
"selfattn_lora": true,
"use_gating": false
},
"config_id": "8106482e5aeaad89",
"hidden_size": 768,
"model_class": "BertForTokenClassification",
"model_name": "indolem/indobert-base-uncased",
"model_type": "bert",
"name": "nerugm-lora",
"version": "0.2.0"
}