{ "config": { "alpha": 1, "architecture": "lora", "attn_matrices": [ "q", "v" ], "composition_mode": "add", "dropout": 0.15, "init_weights": "lora", "intermediate_lora": false, "leave_out": [], "output_lora": false, "r": 8, "selfattn_lora": true, "use_gating": false }, "config_id": "d7a928f95cb5f33d", "hidden_size": 768, "model_class": "BertForTokenClassification", "model_name": "indolem/indobert-base-uncased", "model_type": "bert", "name": "nerugm-lora", "version": "0.2.0" }