fazeelzafar commited on
Commit
d4e3d16
1 Parent(s): 65755b8

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -16,14 +16,14 @@
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
- "r": 16,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
  "v_proj",
24
- "k_proj",
25
  "o_proj",
26
- "q_proj"
 
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
+ "r": 32,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
  "v_proj",
 
24
  "o_proj",
25
+ "q_proj",
26
+ "k_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:efd31350fd4baf902083078115a05080390002e9ed044a9a924431baac1f62b4
3
- size 67143296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26ca28454ba94e4216c8a472b7c59f527791dbd45953a056ae0357a1239e6c8a
3
+ size 134252336
runs/Mar31_11-45-49_9505c06e5711/events.out.tfevents.1711885552.9505c06e5711.4975.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:330c1dac0e5a715f8899f5ec077b72043b271dde075fabecc7ec7937c4e3eee9
3
+ size 5291
runs/Mar31_11-47-20_9505c06e5711/events.out.tfevents.1711885643.9505c06e5711.4975.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebd6c39020eff80b8c9554f0b70168368779124945b4d19d0bfe1205279cf7d8
3
+ size 9917
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c90d7f9af3be7486c84451bf52e67e0a5282457c28687f718948491b22e49506
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd122d695c388684582957ad10e004ddb12639d3351b41e64ea1c673c6228d56
3
  size 4920