iamnguyen commited on
Commit
99d67a0
1 Parent(s): 57d6b69

Training in progress, step 2580, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dc2136cfbd00690f3783dc17e8800db2dc8ad727fd6fe2cc8c6c953b2599e207
3
  size 147770496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b4d9136a9ce0e4ccfef59d92612be6224b1453529c1e6acd828713f37075749
3
  size 147770496
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ea2b85354be8b703c4c2920b8b2fb21f3e0e76e72838162be0d20e1b9317205
3
  size 75455810
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe674b371c6564c8793475ac817ef36331813aeddb6980a9b5b85d709be533a8
3
  size 75455810
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3fd8571639a832ac3a6bdbe8b83a9d5174ee7552fcb8cdc3d2effb6c405a5866
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd0207b1251f238352e39cb342463c968aa002551f624c293b10d9e3d04f1f0b
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:83a970aca6d8492ea417337203ea2e40d5dcc9a674de898354c4c9743b9beef6
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ab51e931596bdb995a7ee5081800ce8215a3ea0018197244a214eabc916219f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.9981337664165743,
5
  "eval_steps": 500,
6
- "global_step": 2576,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -18039,6 +18039,34 @@
18039
  "learning_rate": 6.177385484029685e-11,
18040
  "loss": 1.3969,
18041
  "step": 2576
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18042
  }
18043
  ],
18044
  "logging_steps": 1.0,
@@ -18053,12 +18081,12 @@
18053
  "should_evaluate": false,
18054
  "should_log": false,
18055
  "should_save": true,
18056
- "should_training_stop": false
18057
  },
18058
  "attributes": {}
18059
  }
18060
  },
18061
- "total_flos": 2.547514766592221e+18,
18062
  "train_batch_size": 1,
18063
  "trial_name": null,
18064
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.999683663569395,
5
  "eval_steps": 500,
6
+ "global_step": 2580,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
18039
  "learning_rate": 6.177385484029685e-11,
18040
  "loss": 1.3969,
18041
  "step": 2576
18042
+ },
18043
+ {
18044
+ "epoch": 0.9985212407047794,
18045
+ "grad_norm": 0.19256563484668732,
18046
+ "learning_rate": 3.474782465096027e-11,
18047
+ "loss": 1.3432,
18048
+ "step": 2577
18049
+ },
18050
+ {
18051
+ "epoch": 0.9989087149929846,
18052
+ "grad_norm": 0.19377000629901886,
18053
+ "learning_rate": 1.544348756044034e-11,
18054
+ "loss": 1.4043,
18055
+ "step": 2578
18056
+ },
18057
+ {
18058
+ "epoch": 0.9992961892811898,
18059
+ "grad_norm": 0.18626244366168976,
18060
+ "learning_rate": 3.860873380445718e-12,
18061
+ "loss": 1.3605,
18062
+ "step": 2579
18063
+ },
18064
+ {
18065
+ "epoch": 0.999683663569395,
18066
+ "grad_norm": 0.17782603204250336,
18067
+ "learning_rate": 0.0,
18068
+ "loss": 1.3834,
18069
+ "step": 2580
18070
  }
18071
  ],
18072
  "logging_steps": 1.0,
 
18081
  "should_evaluate": false,
18082
  "should_log": false,
18083
  "should_save": true,
18084
+ "should_training_stop": true
18085
  },
18086
  "attributes": {}
18087
  }
18088
  },
18089
+ "total_flos": 2.55147909551824e+18,
18090
  "train_batch_size": 1,
18091
  "trial_name": null,
18092
  "trial_params": null