CindyXWu commited on
Commit
6e65211
1 Parent(s): 75e508f

Upload MistralForCausalLM

Browse files
Files changed (2) hide show
  1. adapter_config.json +3 -3
  2. adapter_model.safetensors +1 -1
adapter_config.json CHANGED
@@ -22,11 +22,11 @@
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
25
- "down_proj",
26
  "k_proj",
27
- "q_proj",
28
  "v_proj",
29
- "up_proj"
 
30
  ],
31
  "task_type": null,
32
  "use_rslora": false
 
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
25
+ "up_proj",
26
  "k_proj",
 
27
  "v_proj",
28
+ "q_proj",
29
+ "down_proj"
30
  ],
31
  "task_type": null,
32
  "use_rslora": false
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34fb627e9cdc88ee82e3eae6bf4fce4f0ff522aaf85f3841d2f417ddff395b96
3
  size 226535760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1149073ba5d02b33ccdc03708a08a5759ddad820e7b33e6c4ec1233c46f807a0
3
  size 226535760