MohamedAhmedAE commited on
Commit
0a31dd4
1 Parent(s): f54e427

Training in progress, step 26600

Browse files
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:be84fb38c3787d4d9a0a842fe86f337de52b2ffe7a5d69a4f2be10ca449aaede
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a66649ed360eedf7f38576a034cb81a9f0d0d85b4e760a08a7af7c644b300130
3
  size 167832240
last-checkpoint/adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
- "o_proj",
25
- "gate_proj",
26
  "k_proj",
 
27
  "up_proj",
28
  "q_proj",
29
- "down_proj"
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
23
  "k_proj",
24
+ "down_proj",
25
  "up_proj",
26
  "q_proj",
27
+ "gate_proj",
28
+ "o_proj",
29
+ "v_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:be84fb38c3787d4d9a0a842fe86f337de52b2ffe7a5d69a4f2be10ca449aaede
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09e7569231f9d48775234e444dc3002609a2aaf23f9fbb03afc31dd08b174acd
3
  size 167832240
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c96cb8f0f1a681c15e8ed2760660f6d33cb92861d81a47b036784d07497c5099
3
  size 85736914
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1795d0f5907365f50ee8bf592e588c51cf0c1336607a8958053b6df1866c85ef
3
  size 85736914
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:20c68222f3372262392758b273879ed4f661c3fb7cc6b05e1cf45273c18d2035
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1e0be0fbdc67baa64bde42b3f62fb0cbfa6b616dea815616465446e745a61cd
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:40e9ee5e09ade4ca277911e9ac0e424525fe43efa67c8554ac479a544373b55d
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afd4eb6473181bbf6da6b1613c063360450d1f8ec75384d51547bd12664e8b3a
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.31412327702564524,
5
  "eval_steps": 2000,
6
- "global_step": 26400,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -910,35 +910,14 @@
910
  "learning_rate": 1.998839744298062e-05,
911
  "loss": 1.5509,
912
  "step": 25800
913
- },
914
- {
915
- "epoch": 0.31,
916
- "grad_norm": 0.8327608704566956,
917
- "learning_rate": 1.9812018045085563e-05,
918
- "loss": 1.5483,
919
- "step": 26000
920
- },
921
- {
922
- "epoch": 0.31,
923
- "grad_norm": 0.7521975636482239,
924
- "learning_rate": 1.980912147882786e-05,
925
- "loss": 1.5296,
926
- "step": 26200
927
- },
928
- {
929
- "epoch": 0.31,
930
- "grad_norm": 0.7781311869621277,
931
- "learning_rate": 1.9806202981642514e-05,
932
- "loss": 1.5346,
933
- "step": 26400
934
  }
935
  ],
936
  "logging_steps": 200,
937
- "max_steps": 420215,
938
  "num_input_tokens_seen": 0,
939
  "num_train_epochs": 5,
940
  "save_steps": 200,
941
- "total_flos": 7.248865905092444e+17,
942
  "train_batch_size": 1,
943
  "trial_name": null,
944
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.07674602790967469,
5
  "eval_steps": 2000,
6
+ "global_step": 25800,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
910
  "learning_rate": 1.998839744298062e-05,
911
  "loss": 1.5509,
912
  "step": 25800
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
913
  }
914
  ],
915
  "logging_steps": 200,
916
+ "max_steps": 1680865,
917
  "num_input_tokens_seen": 0,
918
  "num_train_epochs": 5,
919
  "save_steps": 200,
920
+ "total_flos": 5.996940170325443e+17,
921
  "train_batch_size": 1,
922
  "trial_name": null,
923
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c09e4d0caa94d3d4acabbf865ee69cc816a077a03533c04bf2c5e62bf08ec171
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e88463ac0cd6182d9b1cd0ac83ec16eba34109ec5568773d1ab337bb23a66942
3
  size 4920