aengusl commited on
Commit
160cb5f
1 Parent(s): f5fc82a

Model save

Browse files
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:80142649f2850c3a32e5ffbc85181d2b0e93830d4648b4b069b50a3d8d790b3d
3
  size 258001832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddad0df15def252e51cb6c283b6d2046e5a732f5b45af6c2c32ba8ba7ae4f6d5
3
  size 258001832
runs/Jun05_15-18-34_702dae2383a0/events.out.tfevents.1717600740.702dae2383a0.127543.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c5d4149d93c10898b31448b5f93aba14fa9b1fda41e653e9ed766987e9325e42
3
- size 9521
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f4942ab104f1bf7159c0d92b0a629bbe6c201faa28aff0f8d7ee9d006b4d86a
3
+ size 9835
step_1200/adapter_config.json CHANGED
@@ -16,12 +16,12 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
 
19
  "k_proj",
20
- "down_proj",
21
- "v_proj",
22
- "q_proj",
23
  "up_proj",
24
- "o_proj"
 
 
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
+ "o_proj",
20
  "k_proj",
 
 
 
21
  "up_proj",
22
+ "q_proj",
23
+ "down_proj",
24
+ "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
step_1200/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b7e145e1d6e5063d0b58e6bbabe0ee76af3c4399e27f7cfc44b9ed0d433c4086
3
  size 258001832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddad0df15def252e51cb6c283b6d2046e5a732f5b45af6c2c32ba8ba7ae4f6d5
3
  size 258001832
step_1200/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fddcb958cb044b08eb2784936cc2d61c7603663890e064c856baab404ea4a540
3
- size 6200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0d652444bf2350cdb3e94cef2e9faf380e4e1e4c94136bbe6f149254b88f14d
3
+ size 6136