MohamedAhmedAE commited on
Commit
879581f
1 Parent(s): 0a3fca9

Training in progress, step 20000

Browse files
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9307e2d5144712eb0edc45d55fc646b836c00d2492ad6daf59bbd2e923ded630
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb5d4a8e80692092cbb6ea69a8cd6902d7306436bd2ef4db3d268beb6d254345
3
  size 167832240
last-checkpoint/adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
  "q_proj",
 
 
25
  "gate_proj",
26
  "down_proj",
27
  "up_proj",
28
- "k_proj",
29
- "o_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "q_proj",
24
+ "o_proj",
25
+ "k_proj",
26
  "gate_proj",
27
  "down_proj",
28
  "up_proj",
29
+ "v_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9307e2d5144712eb0edc45d55fc646b836c00d2492ad6daf59bbd2e923ded630
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30377e3553033d932e69daa7064c259e8a7335e5cc0b78f6a1f4f648bae4b97e
3
  size 167832240
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3adc34a0679276e58fb29420dd020117da79f9d5028057f766a197ee4ed5bb6
3
  size 85736914
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:727033a6eb5dc35337334fe86877b4a8995db5404205b07aeda22261d1e1f288
3
  size 85736914
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c38f461f74746273b931bc8b8109f841c37b776aac16fde6bc2dd08dda9b359f
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b7de94d4e3bf0d0cab47bcde24f41cd7eb97ed0885fd609023fab02c6a16228
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b355cdb0578b7da5f2a64e28d8aa72c18a706e6b2dc8a01d319ef156a2841bf3
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf318569a9b4769e81e4790eccd85435a6e6935dc6ee58b1ecc55958c1b71bd0
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.029449057221154238,
5
  "eval_steps": 2000,
6
- "global_step": 19800,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -672,34 +672,6 @@
672
  "learning_rate": 1.9998426349986698e-05,
673
  "loss": 1.5367,
674
  "step": 19000
675
- },
676
- {
677
- "epoch": 0.03,
678
- "grad_norm": 3.7834692001342773,
679
- "learning_rate": 1.9998393186348416e-05,
680
- "loss": 1.5619,
681
- "step": 19200
682
- },
683
- {
684
- "epoch": 0.03,
685
- "grad_norm": 2.51110577583313,
686
- "learning_rate": 1.9998359507659452e-05,
687
- "loss": 1.5688,
688
- "step": 19400
689
- },
690
- {
691
- "epoch": 0.03,
692
- "grad_norm": 2.6746368408203125,
693
- "learning_rate": 1.999832547969925e-05,
694
- "loss": 1.5216,
695
- "step": 19600
696
- },
697
- {
698
- "epoch": 0.03,
699
- "grad_norm": 2.6558778285980225,
700
- "learning_rate": 1.9998291102469e-05,
701
- "loss": 1.5335,
702
- "step": 19800
703
  }
704
  ],
705
  "logging_steps": 200,
@@ -707,7 +679,7 @@
707
  "num_input_tokens_seen": 0,
708
  "num_train_epochs": 5,
709
  "save_steps": 200,
710
- "total_flos": 3.6451122002325504e+17,
711
  "train_batch_size": 1,
712
  "trial_name": null,
713
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.028259196323329825,
5
  "eval_steps": 2000,
6
+ "global_step": 19000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
672
  "learning_rate": 1.9998426349986698e-05,
673
  "loss": 1.5367,
674
  "step": 19000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
675
  }
676
  ],
677
  "logging_steps": 200,
 
679
  "num_input_tokens_seen": 0,
680
  "num_train_epochs": 5,
681
  "save_steps": 200,
682
+ "total_flos": 3.4393054952298086e+17,
683
  "train_batch_size": 1,
684
  "trial_name": null,
685
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7da6c64555cac2ee9fc43605b134636b4610ed6cd244e07ab7cb4ce3c058548c
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc9e75826f834526adf57daa4ee7a58f88bf2ec9679f7599af2037d01589eb4f
3
  size 4920