aengusl commited on
Commit
7cdd5cf
1 Parent(s): 2c50170

Model save

Browse files
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e76d8637f7e5fe251f73f607b2a93f29a1b9cd853f4541097ba0fb1cefcda5f
3
  size 258001832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:544f0e53651383f3fee0e167303b5b7cf0d5d0d130382011fae029bd8fd132c4
3
  size 258001832
runs/Jun13_19-09-54_dccc3444dd3c/events.out.tfevents.1718305870.dccc3444dd3c.137614.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:73f4a9a37d0ba32188bcc2cd3b11f0b1b60fbbdbe0c9d9969f0d179a7d0a8d6d
3
- size 142860
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29718f97580bccf1d5bcedf77ca28ad3039ddf089fd6188b8ee7bd91a9717668
3
+ size 146000
step_900/adapter_config.json CHANGED
@@ -16,12 +16,12 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "up_proj",
20
  "v_proj",
21
- "q_proj",
22
  "k_proj",
23
- "o_proj",
24
- "down_proj"
 
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
+ "down_proj",
20
  "v_proj",
 
21
  "k_proj",
22
+ "q_proj",
23
+ "up_proj",
24
+ "o_proj"
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
step_900/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a40ed31444a4ab8eee5d227c618bd7b6110098e54a3179e9785aa73e35bedb63
3
  size 258001832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71d541b27f3786ec3825515836b4bf626a85e2e670d254f444dc93d194dc0013
3
  size 258001832
step_900/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:141a3bbe872c69bee25c29bea550a0a65e47e19272018c2d758e0bf4fd444e54
3
  size 6200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30d0ae3dfd32b6142d983e67ef62dec8157f29777e7f12c400f04900af76ecfa
3
  size 6200