|
{ |
|
"adapters": { |
|
"0": "msfm/llm-jp-3-1.8b_adapter-1", |
|
"1": "msfm/llm-jp-3-1.8b_adapter-2" |
|
}, |
|
"auto_mapping": null, |
|
"base_model_name_or_path": "llm-jp/llm-jp-3-1.8b", |
|
"enable_softmax": true, |
|
"enable_softmax_topk": false, |
|
"global_scaling_weight": 1.0, |
|
"hidden_size": 2048, |
|
"inference_mode": true, |
|
"layerwise_scalings": false, |
|
"peft_type": "XLORA", |
|
"revision": null, |
|
"scaling_pass_value": 0.0, |
|
"softmax_temperature": 1.0, |
|
"task_type": "CAUSAL_LM", |
|
"top_k_lora": null, |
|
"use_trainable_adapters": false, |
|
"xlora_depth": 4, |
|
"xlora_dropout_p": 0.2, |
|
"xlora_size": 2048 |
|
} |