JiaweiGuan commited on
Commit
b5340ea
1 Parent(s): 2eaa9a4

Upload model

Browse files
Files changed (2) hide show
  1. adapter_config.json +4 -1
  2. adapter_model.bin +2 -2
adapter_config.json CHANGED
@@ -9,7 +9,10 @@
9
  "layers_to_transform": null,
10
  "lora_alpha": 32,
11
  "lora_dropout": 0.05,
12
- "modules_to_save": null,
 
 
 
13
  "peft_type": "LORA",
14
  "r": 16,
15
  "revision": null,
 
9
  "layers_to_transform": null,
10
  "lora_alpha": 32,
11
  "lora_dropout": 0.05,
12
+ "modules_to_save": [
13
+ "layer_norm_q",
14
+ "layer_norm_k"
15
+ ],
16
  "peft_type": "LORA",
17
  "r": 16,
18
  "revision": null,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4904613f1f82ba4af2cb2d8d1305252953a115550ecf22b0b6b8ee010932e37a
3
- size 75604554
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfac5b87ba6d47a481b60d587861c483c267a1004e9f1d685cf263297b509f7c
3
+ size 75694282