mgh6 commited on
Commit
2928e6f
1 Parent(s): 5d63a5a

Training in progress, step 10240, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ca37dad208975487ae890dd2447d9fd111a4d500fddec1a394f9efeef88557e
3
  size 4725595416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3326816602b76848a6a8b29dd3a51168076de7c214927edd394c75013a454acf
3
  size 4725595416
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0b8169b03af949ee4c5ddf91eed20622e053b6dde7d912b4c3311f20f6495a7
3
  size 9179193343
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69eff366bb4160a16c74c582bd7f1c525455e190d832ad429ff154102b6ca5ef
3
  size 9179193343
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cb732be4c200b0b68d66033f545d69b981d9d12cd9b1fae529a2b5f11bc8689a
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b60dc5f90831a3cb044b4ac92382bb3956ae1e5aaac38365110e8ca62c577d1
3
  size 14503
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7802dd5061761c471fafcb314e7fa5bea1fb541e8e1ce0821e89ee84f8e88b84
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85e731ed421ccadf919d8023a4de52dddc9be17926d4166cc53e8083d7604c53
3
  size 623
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 3752.7509765625,
3
  "best_model_checkpoint": "mgh6/TCS_Pairing_VAE/checkpoint-7680",
4
- "epoch": 0.5674335269724873,
5
  "eval_steps": 512,
6
- "global_step": 7680,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -307,6 +307,106 @@
307
  "eval_samples_per_second": 66.508,
308
  "eval_steps_per_second": 66.508,
309
  "step": 7680
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
310
  }
311
  ],
312
  "logging_steps": 256,
 
1
  {
2
  "best_metric": 3752.7509765625,
3
  "best_model_checkpoint": "mgh6/TCS_Pairing_VAE/checkpoint-7680",
4
+ "epoch": 0.7565780359633163,
5
  "eval_steps": 512,
6
+ "global_step": 10240,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
307
  "eval_samples_per_second": 66.508,
308
  "eval_steps_per_second": 66.508,
309
  "step": 7680
310
+ },
311
+ {
312
+ "epoch": 0.59,
313
+ "learning_rate": 8.827249889168022e-05,
314
+ "loss": 4972.3188,
315
+ "step": 7936
316
+ },
317
+ {
318
+ "epoch": 0.61,
319
+ "learning_rate": 8.789419240431506e-05,
320
+ "loss": 5409.0205,
321
+ "step": 8192
322
+ },
323
+ {
324
+ "epoch": 0.61,
325
+ "eval_loss": 4419.3115234375,
326
+ "eval_runtime": 56.0194,
327
+ "eval_samples_per_second": 60.658,
328
+ "eval_steps_per_second": 60.658,
329
+ "step": 8192
330
+ },
331
+ {
332
+ "epoch": 0.62,
333
+ "learning_rate": 8.751588591694991e-05,
334
+ "loss": 4755.2881,
335
+ "step": 8448
336
+ },
337
+ {
338
+ "epoch": 0.64,
339
+ "learning_rate": 8.713757942958476e-05,
340
+ "loss": 4503.3687,
341
+ "step": 8704
342
+ },
343
+ {
344
+ "epoch": 0.64,
345
+ "eval_loss": 4440.9599609375,
346
+ "eval_runtime": 50.1462,
347
+ "eval_samples_per_second": 67.762,
348
+ "eval_steps_per_second": 67.762,
349
+ "step": 8704
350
+ },
351
+ {
352
+ "epoch": 0.66,
353
+ "learning_rate": 8.67592729422196e-05,
354
+ "loss": 4803.3394,
355
+ "step": 8960
356
+ },
357
+ {
358
+ "epoch": 0.68,
359
+ "learning_rate": 8.638096645485444e-05,
360
+ "loss": 5031.4937,
361
+ "step": 9216
362
+ },
363
+ {
364
+ "epoch": 0.68,
365
+ "eval_loss": 5361.60546875,
366
+ "eval_runtime": 49.6714,
367
+ "eval_samples_per_second": 68.41,
368
+ "eval_steps_per_second": 68.41,
369
+ "step": 9216
370
+ },
371
+ {
372
+ "epoch": 0.7,
373
+ "learning_rate": 8.600265996748929e-05,
374
+ "loss": 4789.9038,
375
+ "step": 9472
376
+ },
377
+ {
378
+ "epoch": 0.72,
379
+ "learning_rate": 8.562435348012414e-05,
380
+ "loss": 5079.5186,
381
+ "step": 9728
382
+ },
383
+ {
384
+ "epoch": 0.72,
385
+ "eval_loss": 4070.673828125,
386
+ "eval_runtime": 49.4243,
387
+ "eval_samples_per_second": 68.752,
388
+ "eval_steps_per_second": 68.752,
389
+ "step": 9728
390
+ },
391
+ {
392
+ "epoch": 0.74,
393
+ "learning_rate": 8.524604699275897e-05,
394
+ "loss": 5474.73,
395
+ "step": 9984
396
+ },
397
+ {
398
+ "epoch": 0.76,
399
+ "learning_rate": 8.486774050539382e-05,
400
+ "loss": 4787.0361,
401
+ "step": 10240
402
+ },
403
+ {
404
+ "epoch": 0.76,
405
+ "eval_loss": 4277.46337890625,
406
+ "eval_runtime": 49.554,
407
+ "eval_samples_per_second": 68.572,
408
+ "eval_steps_per_second": 68.572,
409
+ "step": 10240
410
  }
411
  ],
412
  "logging_steps": 256,