cruiser commited on
Commit
d6d170e
1 Parent(s): 84fe8db

Training in progress, epoch 41

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd048f33688c4ea0e91078361721d0a1a838d1b3ecd5a89ee8a8cdaa92753ae5
3
  size 535707205
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94c6a4fd00ca81806b474a775bbe09a92a26b172751515cb979fd4f4ac7a26c8
3
  size 535707205
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:03e618138b7245bfd7581c728c4a0da296de2da08ca617e72e1c1e8ca89ab529
3
  size 267858605
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:586ebd437ff9ab9ff636107d4e140898931477277f699700b7ae4559250ac36e
3
  size 267858605
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0fc5cb68573008ff83962111e44f2cc9863cf41e64dff7053e3c4218462cf257
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dccbecb658b80e3a5361f0d5179c9d4d4504aed644c1eb08c8b8834f23a53dfd
3
  size 14575
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:915bc0ed0c7e5de215d6c27a59333bfda9c22ef35bc3ee1e9b3a72c381fa3944
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f3bc848ccd1cf01fec715361fb0dbcdfb69c00930c1757363207e7656620b81
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 1.606531023979187,
3
  "best_model_checkpoint": "distilbert_final_tweet_dataset_200_epoch/checkpoint-230",
4
- "epoch": 39.0,
5
- "global_step": 4485,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -444,11 +444,37 @@
444
  "eval_samples_per_second": 440.237,
445
  "eval_steps_per_second": 55.709,
446
  "step": 4485
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
447
  }
448
  ],
449
  "max_steps": 23000,
450
  "num_train_epochs": 200,
451
- "total_flos": 1447397208928476.0,
452
  "trial_name": null,
453
  "trial_params": null
454
  }
 
1
  {
2
  "best_metric": 1.606531023979187,
3
  "best_model_checkpoint": "distilbert_final_tweet_dataset_200_epoch/checkpoint-230",
4
+ "epoch": 41.0,
5
+ "global_step": 4715,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
444
  "eval_samples_per_second": 440.237,
445
  "eval_steps_per_second": 55.709,
446
  "step": 4485
447
+ },
448
+ {
449
+ "epoch": 39.13,
450
+ "learning_rate": 4.021739130434783e-05,
451
+ "loss": 0.0057,
452
+ "step": 4500
453
+ },
454
+ {
455
+ "epoch": 40.0,
456
+ "eval_accuracy": 0.6697530864197531,
457
+ "eval_f1": 0.6710645630658144,
458
+ "eval_loss": 3.224386215209961,
459
+ "eval_runtime": 0.7446,
460
+ "eval_samples_per_second": 435.153,
461
+ "eval_steps_per_second": 55.066,
462
+ "step": 4600
463
+ },
464
+ {
465
+ "epoch": 41.0,
466
+ "eval_accuracy": 0.6697530864197531,
467
+ "eval_f1": 0.6710645630658144,
468
+ "eval_loss": 3.2694830894470215,
469
+ "eval_runtime": 0.7256,
470
+ "eval_samples_per_second": 446.497,
471
+ "eval_steps_per_second": 56.501,
472
+ "step": 4715
473
  }
474
  ],
475
  "max_steps": 23000,
476
  "num_train_epochs": 200,
477
+ "total_flos": 1521622706822244.0,
478
  "trial_name": null,
479
  "trial_params": null
480
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:03e618138b7245bfd7581c728c4a0da296de2da08ca617e72e1c1e8ca89ab529
3
  size 267858605
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:586ebd437ff9ab9ff636107d4e140898931477277f699700b7ae4559250ac36e
3
  size 267858605
runs/May01_10-03-26_b04f269e142e/events.out.tfevents.1682935412.b04f269e142e.1987.2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f57c1bfdbb915fe67b73f68e91d1cce29da332c0aec97bfe0470c9ab498cd5d4
3
- size 19810
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc9957b9276e34c6f8ac79887f48b328cfaaa82cc3acee7f0e5b05038b44f8e3
3
+ size 20705