File size: 194 Bytes
0f5f3fe
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
{
  "finetuning_type": "lora",
  "lora_alpha": 32.0,
  "lora_dropout": 0.1,
  "lora_rank": 8,
  "lora_target": [
    "W_pack"
  ],
  "name_module_trainable": "mlp",
  "num_layer_trainable": 3
}