ardaspear commited on
Commit
8a8733e
·
verified ·
1 Parent(s): ad56bea

Training in progress, step 348, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7cf77b4f1ea716a72c126f9f52ff71d162d7064e459728574ac51ac8863ce784
3
  size 1329145928
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7da7958fb3316598cd7bd31f14962a4b9f95cac05c29e0e6fb364dd2eb01f50b
3
  size 1329145928
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d4cdd23b21c04acae9cd565d08c279f300ff00d504f453e44013debf830e0c9
3
  size 841979032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98fd910ed301d669c3dfc10e6103c6d4a82dee980f2ba7401e4a834f14c2b5f9
3
  size 841979032
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a7e55b1d5bb15e7a21053569831fe746e36b83ade98888145c8c664203412af
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1683e4c2451132a87e57889d6dcd25b40ca1af97d0d6c414788ddde9fa5bd447
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8f85d6d8690a59509a6114c21bc6a09f49525d54d07da8e164b75d78e5e653b
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24f9209ec94f86ca8913e01c292863444e950b743a761710e89589aaf1adbcb1
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 5.301859378814697,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-300",
4
- "epoch": 2.591792656587473,
5
  "eval_steps": 50,
6
- "global_step": 300,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -273,6 +273,34 @@
273
  "eval_samples_per_second": 5.949,
274
  "eval_steps_per_second": 1.495,
275
  "step": 300
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
276
  }
277
  ],
278
  "logging_steps": 10,
@@ -296,12 +324,12 @@
296
  "should_evaluate": false,
297
  "should_log": false,
298
  "should_save": true,
299
- "should_training_stop": false
300
  },
301
  "attributes": {}
302
  }
303
  },
304
- "total_flos": 1.0933430207800934e+18,
305
  "train_batch_size": 8,
306
  "trial_name": null,
307
  "trial_params": null
 
1
  {
2
  "best_metric": 5.301859378814697,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-300",
4
+ "epoch": 3.0064794816414686,
5
  "eval_steps": 50,
6
+ "global_step": 348,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
273
  "eval_samples_per_second": 5.949,
274
  "eval_steps_per_second": 1.495,
275
  "step": 300
276
+ },
277
+ {
278
+ "epoch": 2.6781857451403885,
279
+ "grad_norm": 12.293045997619629,
280
+ "learning_rate": 6.1728263459614796e-06,
281
+ "loss": 20.8827,
282
+ "step": 310
283
+ },
284
+ {
285
+ "epoch": 2.7645788336933044,
286
+ "grad_norm": 18.166112899780273,
287
+ "learning_rate": 3.367439399426087e-06,
288
+ "loss": 21.0383,
289
+ "step": 320
290
+ },
291
+ {
292
+ "epoch": 2.85097192224622,
293
+ "grad_norm": 17.80660629272461,
294
+ "learning_rate": 1.396265834701982e-06,
295
+ "loss": 21.3134,
296
+ "step": 330
297
+ },
298
+ {
299
+ "epoch": 2.937365010799136,
300
+ "grad_norm": 16.692588806152344,
301
+ "learning_rate": 2.7632247671177667e-07,
302
+ "loss": 20.1229,
303
+ "step": 340
304
  }
305
  ],
306
  "logging_steps": 10,
 
324
  "should_evaluate": false,
325
  "should_log": false,
326
  "should_save": true,
327
+ "should_training_stop": true
328
  },
329
  "attributes": {}
330
  }
331
  },
332
+ "total_flos": 1.2504081766652314e+18,
333
  "train_batch_size": 8,
334
  "trial_name": null,
335
  "trial_params": null