fats-fme commited on
Commit
6169896
·
verified ·
1 Parent(s): fef5f7a

Training in progress, step 46, checkpoint

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c3161d89e2ad1b44a18bf418bfb17aa903d17835c252c09465b64747372bac3f
3
  size 70667778
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71f9ad67d7da082ffc7e6bcb5da67c989ae4179d2e6742fcb02dcddde436b548
3
  size 70667778
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:06ab5129ab155330a39a9371aacd2f73f3677218096c1c34b3766d995f841a91
3
  size 14512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:380920090fa298dc1c0b364af9064413da5ecdbe3aeb596471d663387915393e
3
  size 14512
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:81435bd0a6a3789b553c69a79f8c384151f9688d0cd7d745e20e5aadeece4761
3
  size 14512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c7c9e22c75a13a9902d17f3192f927c82d7333c0b4d285610b63194d12e2883
3
  size 14512
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:844a9e8b65ea1c19a7e6553433121a4f8c699c309e7900990529efa7ada6321b
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca03a7964455ac25dc387068890b645c2e3e0fcd3245a993a5c6b7145edfe622
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.768,
5
  "eval_steps": 12,
6
- "global_step": 36,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -291,6 +291,76 @@
291
  "eval_samples_per_second": 21.637,
292
  "eval_steps_per_second": 5.478,
293
  "step": 36
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
294
  }
295
  ],
296
  "logging_steps": 1,
@@ -305,12 +375,12 @@
305
  "should_evaluate": false,
306
  "should_log": false,
307
  "should_save": true,
308
- "should_training_stop": false
309
  },
310
  "attributes": {}
311
  }
312
  },
313
- "total_flos": 5190870965944320.0,
314
  "train_batch_size": 2,
315
  "trial_name": null,
316
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.9813333333333333,
5
  "eval_steps": 12,
6
+ "global_step": 46,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
291
  "eval_samples_per_second": 21.637,
292
  "eval_steps_per_second": 5.478,
293
  "step": 36
294
+ },
295
+ {
296
+ "epoch": 0.7893333333333333,
297
+ "grad_norm": NaN,
298
+ "learning_rate": 7.4e-05,
299
+ "loss": 0.0,
300
+ "step": 37
301
+ },
302
+ {
303
+ "epoch": 0.8106666666666666,
304
+ "grad_norm": NaN,
305
+ "learning_rate": 7.6e-05,
306
+ "loss": 0.0,
307
+ "step": 38
308
+ },
309
+ {
310
+ "epoch": 0.832,
311
+ "grad_norm": NaN,
312
+ "learning_rate": 7.800000000000001e-05,
313
+ "loss": 0.0,
314
+ "step": 39
315
+ },
316
+ {
317
+ "epoch": 0.8533333333333334,
318
+ "grad_norm": NaN,
319
+ "learning_rate": 8e-05,
320
+ "loss": 0.0,
321
+ "step": 40
322
+ },
323
+ {
324
+ "epoch": 0.8746666666666667,
325
+ "grad_norm": NaN,
326
+ "learning_rate": 8.2e-05,
327
+ "loss": 0.0,
328
+ "step": 41
329
+ },
330
+ {
331
+ "epoch": 0.896,
332
+ "grad_norm": NaN,
333
+ "learning_rate": 8.4e-05,
334
+ "loss": 0.0,
335
+ "step": 42
336
+ },
337
+ {
338
+ "epoch": 0.9173333333333333,
339
+ "grad_norm": NaN,
340
+ "learning_rate": 8.6e-05,
341
+ "loss": 0.0,
342
+ "step": 43
343
+ },
344
+ {
345
+ "epoch": 0.9386666666666666,
346
+ "grad_norm": NaN,
347
+ "learning_rate": 8.800000000000001e-05,
348
+ "loss": 0.0,
349
+ "step": 44
350
+ },
351
+ {
352
+ "epoch": 0.96,
353
+ "grad_norm": NaN,
354
+ "learning_rate": 9e-05,
355
+ "loss": 0.0,
356
+ "step": 45
357
+ },
358
+ {
359
+ "epoch": 0.9813333333333333,
360
+ "grad_norm": NaN,
361
+ "learning_rate": 9.200000000000001e-05,
362
+ "loss": 0.0,
363
+ "step": 46
364
  }
365
  ],
366
  "logging_steps": 1,
 
375
  "should_evaluate": false,
376
  "should_log": false,
377
  "should_save": true,
378
+ "should_training_stop": true
379
  },
380
  "attributes": {}
381
  }
382
  },
383
+ "total_flos": 6632779567595520.0,
384
  "train_batch_size": 2,
385
  "trial_name": null,
386
  "trial_params": null