LeoTungAnh commited on
Commit
abf41d0
1 Parent(s): eddac74

Training in progress, step 1500

Browse files
adapter_config.json CHANGED
@@ -20,12 +20,12 @@
20
  "revision": null,
21
  "target_modules": [
22
  "v_proj",
23
- "q_proj",
24
- "o_proj",
25
- "up_proj",
26
  "k_proj",
27
  "down_proj",
28
- "gate_proj"
 
 
 
29
  ],
30
  "task_type": "CAUSAL_LM"
31
  }
 
20
  "revision": null,
21
  "target_modules": [
22
  "v_proj",
 
 
 
23
  "k_proj",
24
  "down_proj",
25
+ "o_proj",
26
+ "q_proj",
27
+ "gate_proj",
28
+ "up_proj"
29
  ],
30
  "task_type": "CAUSAL_LM"
31
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f6d869f9cc5c5e364702fbd8a21244cdccee866e96e86d2f4c4fd5d379383eda
3
  size 671150064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13936c442ae58c87149398a4853c04510431b8e6e9019e7d98ecc0bfd2ecd55d
3
  size 671150064
runs/May01_15-27-17_gpu4-119-4/events.out.tfevents.1714541656.gpu4-119-4.1739961.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a64ffa6120b90d7e5959af80efddb983850f86c561336d183289c0a114ce72b
3
- size 104181
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76f98b8497da04e0786c22d34f38f288ca13285600955df352a4d4dc953903da
3
+ size 106083
runs/May03_17-30-05_gpu4-119-4/events.out.tfevents.1714721665.gpu4-119-4.289261.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62dd28b8fcc6d664f9c9d70967ca88453b52d123ff1a100763c31c0fa6b2dabc
3
+ size 11663
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e476b02274bc5801cdc0eb0130a720d895b9c524cdf641b840bdc4bf3c438566
3
  size 5816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e178ee89b42ff9725e5f2440d59c1bdda581a20e27803e425c924bdcb868f10
3
  size 5816