davidr99 commited on
Commit
e82dfed
·
verified ·
1 Parent(s): fd430a5

End of training

Browse files
README.md CHANGED
@@ -34,15 +34,15 @@ More information needed
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 2e-05
37
- - train_batch_size: 3
38
  - eval_batch_size: 8
39
  - seed: 42
40
  - gradient_accumulation_steps: 3
41
- - total_train_batch_size: 9
42
  - optimizer: Use paged_adamw_8bit with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
  - lr_scheduler_warmup_steps: 2
45
- - num_epochs: 3
46
 
47
  ### Training results
48
 
 
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 2e-05
37
+ - train_batch_size: 4
38
  - eval_batch_size: 8
39
  - seed: 42
40
  - gradient_accumulation_steps: 3
41
+ - total_train_batch_size: 12
42
  - optimizer: Use paged_adamw_8bit with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
  - lr_scheduler_warmup_steps: 2
45
+ - num_epochs: 150
46
 
47
  ### Training results
48
 
adapter_config.json CHANGED
@@ -23,14 +23,14 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "q_proj",
27
- "gate_proj",
28
  "down_proj",
29
- "k_proj",
30
  "lora_megnitude_vector",
 
 
 
31
  "v_proj",
32
  "up_proj",
33
- "o_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "use_dora": true,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
 
26
  "down_proj",
 
27
  "lora_megnitude_vector",
28
+ "gate_proj",
29
+ "k_proj",
30
+ "o_proj",
31
  "v_proj",
32
  "up_proj",
33
+ "q_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "use_dora": true,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:599f06c2fa8058e1ed38914d7ced7bbc57c88fbc351ba7f3ae1c4a7d70d2e401
3
  size 48502952
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c6a2efd4cac3b5ceaebbd4ba3907d31a71a352ac80380cb91e09dc5bb2c4bde
3
  size 48502952
runs/Dec24_02-37-55_ml-ubuntu/events.out.tfevents.1735007876.ml-ubuntu.4180.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5901c59ed3e3188875e9ac93a68c2845ccebe5b36f2ed4a562f8fea36005585
3
+ size 5842
runs/Dec24_02-39-09_ml-ubuntu/events.out.tfevents.1735007950.ml-ubuntu.7797.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:649cd7a41c20ceeaa46b228d671aadb09dc92fe4cb9d44d59153a89ab6d70aeb
3
+ size 46879
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d559b667c6e209289535f82e4382127ad95244184b80cefd833153468951a014
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e63eec21761e8ccebac187a7da84b9c87e2496b5551bd2cc2044c81e7ddcd61c
3
  size 5368