EmmaGthn commited on
Commit
e2cdea4
1 Parent(s): 24d9804

Training in progress, step 200

Browse files
adapter_config.json CHANGED
@@ -19,8 +19,8 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "v_proj",
23
- "q_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "q_proj",
23
+ "v_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:611692720b5c1815ff58124633a277a942bef42dc89d8c9dc8cb6b9882db0a5c
3
  size 67126104
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9800428008760ed36f07a67f7acdfb8eb3ee944a76a1d94dea71365c5f327ffe
3
  size 67126104
runs/Dec19_16-12-50_calcul-bigcpu-lahc-4/events.out.tfevents.1702998774.calcul-bigcpu-lahc-4 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a77609516b2dbcd6483893cf3503dd89e502449fa1a913f19b6412d5d3ff21b0
3
- size 33443
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27ed7054167786a4a0d2863ad18006a18603c22db23960c242ba7a5cedec6473
3
+ size 34496
runs/Dec19_20-08-49_calcul-bigcpu-lahc-4/events.out.tfevents.1703012932.calcul-bigcpu-lahc-4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93c2f01edc31aa85d96557eb6f84944ce9904f39f6c60af85cae213c139488e0
3
+ size 5194
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6bfdcd8c1ffb674c57fa585f3e5efc1cc8ea64345979388f23bd941786389291
3
  size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ba8387a1e8d3b1cfcf00aaff66a89275dd23de6a25a896e875af4b89fa189ca
3
  size 4600