MohamedAhmedAE commited on
Commit
7a4e731
·
verified ·
1 Parent(s): 3a6395a

Training in progress, step 74400, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -24,11 +24,11 @@
24
  "revision": null,
25
  "target_modules": [
26
  "k_proj",
27
- "v_proj",
28
- "o_proj",
29
- "up_proj",
30
  "down_proj",
 
 
31
  "gate_proj",
 
32
  "q_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
 
24
  "revision": null,
25
  "target_modules": [
26
  "k_proj",
 
 
 
27
  "down_proj",
28
+ "up_proj",
29
+ "v_proj",
30
  "gate_proj",
31
+ "o_proj",
32
  "q_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:575f61e221017f2179441559ba498b0a30f681ccc1e377951330eaee49c39822
3
  size 1342238560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adade87829d537ad6739250c483b6706bdd890e1f0c9de83cdf06f44b32e5090
3
  size 1342238560
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0d6c756dbba2ef1c1b74580a28ab2dcc48755bf2857720a1da48a545c30bb169
3
  size 683268498
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc5615a410c858c99e3cf7362945f36a6e9c191706d0a029ed4b1e854e10c57b
3
  size 683268498
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:362b5dd6eca4f983f206304be27a2b716fdb6e0e11a8b449b3be57f73f075b0d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dccc9cbc468284f92a4012f3a3803f3d2cbb7abb4eeb9a6c35e73e04a7df0c1b
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ad2a07100542f920d814c50556d1f596c833109e04c422ef0d43afdfc163dcec
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e73dccec2330458d52f8108cf9299eca128bc179d6d432f80a29c728273abc45
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.05146569428960303,
5
  "eval_steps": 500,
6
- "global_step": 74000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2597,6 +2597,20 @@
2597
  "learning_rate": 1.997713838369215e-05,
2598
  "loss": 1.6841,
2599
  "step": 74000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2600
  }
2601
  ],
2602
  "logging_steps": 200,
@@ -2616,7 +2630,7 @@
2616
  "attributes": {}
2617
  }
2618
  },
2619
- "total_flos": 9.85329379279872e+17,
2620
  "train_batch_size": 1,
2621
  "trial_name": null,
2622
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.05174388723170899,
5
  "eval_steps": 500,
6
+ "global_step": 74400,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2597
  "learning_rate": 1.997713838369215e-05,
2598
  "loss": 1.6841,
2599
  "step": 74000
2600
+ },
2601
+ {
2602
+ "epoch": 0.05160479076065601,
2603
+ "grad_norm": 3.7980597019195557,
2604
+ "learning_rate": 1.9977014708158027e-05,
2605
+ "loss": 1.727,
2606
+ "step": 74200
2607
+ },
2608
+ {
2609
+ "epoch": 0.05174388723170899,
2610
+ "grad_norm": 4.492558479309082,
2611
+ "learning_rate": 1.997689069954923e-05,
2612
+ "loss": 1.6307,
2613
+ "step": 74400
2614
  }
2615
  ],
2616
  "logging_steps": 200,
 
2630
  "attributes": {}
2631
  }
2632
  },
2633
+ "total_flos": 9.906764066420982e+17,
2634
  "train_batch_size": 1,
2635
  "trial_name": null,
2636
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:acabefbc48b0af9c7f4952fd51a56e04992b5bb070918bb0e676e541c603140d
3
  size 6840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:973554a532ca404cb4db7d5fef3755d36407a35ec7c1a90c277a7941be0f4f1e
3
  size 6840