chocopiee commited on
Commit
5df96e8
1 Parent(s): d07a288

Training in progress, step 20

Browse files
adapter_config.json CHANGED
@@ -20,14 +20,14 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "o_proj",
24
- "gate_proj",
25
- "down_proj",
26
  "v_proj",
 
27
  "k_proj",
 
 
28
  "lm_head",
29
- "q_proj",
30
- "up_proj"
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "q_proj",
 
 
24
  "v_proj",
25
+ "up_proj",
26
  "k_proj",
27
+ "o_proj",
28
+ "gate_proj",
29
  "lm_head",
30
+ "down_proj"
 
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:445050b7132e1bbb70cdc56ad9d3c3a7107c1d0f1242941705116044ba5c45ef
3
  size 1310658800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:851c21fb33964d2f158257028f79a00df225591d713793a150e4f20e5ea7ba87
3
  size 1310658800
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:03baff4ce24c308e28b05158206a37c4ffb6b56254b80c4249fc5e91131ba9a1
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45a5c68e7b6be2d1203ca35666ae7962b20cf1d4dd726c0eb3a7698ebc6e4927
3
  size 5304