gprashant17 commited on
Commit
2660d22
1 Parent(s): 24e4352

Training in progress, step 9800

Browse files
adapter_config.json CHANGED
@@ -22,8 +22,8 @@
22
  "target_modules": [
23
  "c_fc",
24
  "c_attn",
25
- "c_proj",
26
- "q_attn"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
22
  "target_modules": [
23
  "c_fc",
24
  "c_attn",
25
+ "q_attn",
26
+ "c_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bca01f44a6b607b3c64211e2c42b60e4d6105d85547c288695433aaab1bf9c1
3
  size 29506408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afd26e43cc2ec477c2851b492e3983827ca99a17997bfc28dc3fe4b160157198
3
  size 29506408
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d0be47ea9fee0c17c7916b5f8478db46c21151b04ac980e5e02e5aab73c39318
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42e1a87e359f873d0589b2ac96dd82718c728dbbc0e68ffa940186ac2f01e924
3
  size 5176