xlora / adapter_config.json
msfm's picture
Upload model
04cb952 verified
raw
history blame contribute delete
607 Bytes
{
"adapters": {
"0": "msfm/llm-jp-3-1.8b_adapter-1",
"1": "msfm/llm-jp-3-1.8b_adapter-2"
},
"auto_mapping": null,
"base_model_name_or_path": "llm-jp/llm-jp-3-1.8b",
"enable_softmax": true,
"enable_softmax_topk": false,
"global_scaling_weight": 1.0,
"hidden_size": 2048,
"inference_mode": true,
"layerwise_scalings": false,
"peft_type": "XLORA",
"revision": null,
"scaling_pass_value": 0.0,
"softmax_temperature": 1.0,
"task_type": "CAUSAL_LM",
"top_k_lora": null,
"use_trainable_adapters": false,
"xlora_depth": 4,
"xlora_dropout_p": 0.2,
"xlora_size": 2048
}