Check commited on
Commit
c955500
Β·
1 Parent(s): 1ac48d6

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629788442.126761/events.out.tfevents.1629788442.c435e1c5ee04.920.71 +3 -0
  11. model-bin/finetune/base/log/1629789087.6950493/events.out.tfevents.1629789087.c435e1c5ee04.920.73 +3 -0
  12. model-bin/finetune/base/log/1629789729.6492243/events.out.tfevents.1629789729.c435e1c5ee04.920.75 +3 -0
  13. model-bin/finetune/base/log/1629790377.8418694/events.out.tfevents.1629790377.c435e1c5ee04.920.77 +3 -0
  14. model-bin/finetune/base/log/1629791034.1306834/events.out.tfevents.1629791034.c435e1c5ee04.920.79 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629788442.c435e1c5ee04.920.70 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629789087.c435e1c5ee04.920.72 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629789729.c435e1c5ee04.920.74 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629790377.c435e1c5ee04.920.76 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629791034.c435e1c5ee04.920.78 +3 -0
model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e15aae07094715fa3940581269509a90f02f966e57838d740a66a09d2497d9b1
3
  size 722165009
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abb032a26cbde4b62d89539fc2e8e5881b35423dcd963a58f4158155e7e2bf88
3
  size 722165009
model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e48222865b2d6155a6f64120f77f66c5f33da659043bc8746f79e1f0a465e1bd
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2c85302de9ded7dd6c08545970e38ead0397e1e232dc1cc0397a355e678b399
3
  size 377909911
model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3240d1383b02806c1cba54298be39d75b6ab2e708f15feb5955feff6869092ad
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd52303b56dc5bf8bee3e5b43d2bf8c89b3afc6ef4263289063a2c641f525b38
3
  size 14503
model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:60a77c8dd462fc3b4be2ff5da7bab717e71264b6dc82fc18d24d6930b1315c4f
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d9cd3fedc5e7eb14c9de984a70f18c3445b012f7a0208035cf4a96e7423bac5
3
  size 559
model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a9d5be5ad540e619a04127837aff42d25861054b44f9bd177ae0dc86a6ff1bfa
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a378270aa4ce26218f87463f54f775fff54af5666c2db0fab97c9dec6dc51967
3
  size 623
model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.18631571186315712,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-55501",
4
- "epoch": 465.0,
5
- "global_step": 58240,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -173466,11 +173466,800 @@
173466
  "eval_steps_per_second": 0.676,
173467
  "eval_wer": 0.1968630900029317,
173468
  "step": 58240
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
173469
  }
173470
  ],
173471
  "max_steps": 625000,
173472
  "num_train_epochs": 5000,
173473
- "total_flos": 1.6390003887200823e+20,
173474
  "trial_name": null,
173475
  "trial_params": null
173476
  }
 
1
  {
2
  "best_metric": 0.18631571186315712,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-55501",
4
+ "epoch": 470.0,
5
+ "global_step": 58863,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
173466
  "eval_steps_per_second": 0.676,
173467
  "eval_wer": 0.1968630900029317,
173468
  "step": 58240
173469
+ },
173470
+ {
173471
+ "epoch": 465.04,
173472
+ "learning_rate": 9.082964743589744e-06,
173473
+ "loss": 0.3577,
173474
+ "step": 58245
173475
+ },
173476
+ {
173477
+ "epoch": 465.08,
173478
+ "learning_rate": 9.082884615384617e-06,
173479
+ "loss": 0.3515,
173480
+ "step": 58250
173481
+ },
173482
+ {
173483
+ "epoch": 465.12,
173484
+ "learning_rate": 9.082804487179487e-06,
173485
+ "loss": 0.5173,
173486
+ "step": 58255
173487
+ },
173488
+ {
173489
+ "epoch": 465.16,
173490
+ "learning_rate": 9.08272435897436e-06,
173491
+ "loss": 0.5225,
173492
+ "step": 58260
173493
+ },
173494
+ {
173495
+ "epoch": 465.2,
173496
+ "learning_rate": 9.082644230769233e-06,
173497
+ "loss": 1.2218,
173498
+ "step": 58265
173499
+ },
173500
+ {
173501
+ "epoch": 465.24,
173502
+ "learning_rate": 9.082564102564102e-06,
173503
+ "loss": 0.3957,
173504
+ "step": 58270
173505
+ },
173506
+ {
173507
+ "epoch": 465.28,
173508
+ "learning_rate": 9.082483974358975e-06,
173509
+ "loss": 0.4104,
173510
+ "step": 58275
173511
+ },
173512
+ {
173513
+ "epoch": 465.32,
173514
+ "learning_rate": 9.082403846153847e-06,
173515
+ "loss": 0.3723,
173516
+ "step": 58280
173517
+ },
173518
+ {
173519
+ "epoch": 465.36,
173520
+ "learning_rate": 9.082323717948718e-06,
173521
+ "loss": 0.5224,
173522
+ "step": 58285
173523
+ },
173524
+ {
173525
+ "epoch": 465.4,
173526
+ "learning_rate": 9.08224358974359e-06,
173527
+ "loss": 1.3738,
173528
+ "step": 58290
173529
+ },
173530
+ {
173531
+ "epoch": 465.44,
173532
+ "learning_rate": 9.082163461538463e-06,
173533
+ "loss": 0.4635,
173534
+ "step": 58295
173535
+ },
173536
+ {
173537
+ "epoch": 465.48,
173538
+ "learning_rate": 9.082083333333334e-06,
173539
+ "loss": 0.3809,
173540
+ "step": 58300
173541
+ },
173542
+ {
173543
+ "epoch": 465.52,
173544
+ "learning_rate": 9.082003205128205e-06,
173545
+ "loss": 0.3892,
173546
+ "step": 58305
173547
+ },
173548
+ {
173549
+ "epoch": 465.56,
173550
+ "learning_rate": 9.081923076923078e-06,
173551
+ "loss": 0.488,
173552
+ "step": 58310
173553
+ },
173554
+ {
173555
+ "epoch": 465.6,
173556
+ "learning_rate": 9.08184294871795e-06,
173557
+ "loss": 1.1037,
173558
+ "step": 58315
173559
+ },
173560
+ {
173561
+ "epoch": 465.64,
173562
+ "learning_rate": 9.081762820512821e-06,
173563
+ "loss": 0.3588,
173564
+ "step": 58320
173565
+ },
173566
+ {
173567
+ "epoch": 465.68,
173568
+ "learning_rate": 9.081682692307692e-06,
173569
+ "loss": 0.3848,
173570
+ "step": 58325
173571
+ },
173572
+ {
173573
+ "epoch": 465.72,
173574
+ "learning_rate": 9.081602564102566e-06,
173575
+ "loss": 0.3331,
173576
+ "step": 58330
173577
+ },
173578
+ {
173579
+ "epoch": 465.76,
173580
+ "learning_rate": 9.081522435897437e-06,
173581
+ "loss": 0.871,
173582
+ "step": 58335
173583
+ },
173584
+ {
173585
+ "epoch": 465.8,
173586
+ "learning_rate": 9.081442307692308e-06,
173587
+ "loss": 1.4403,
173588
+ "step": 58340
173589
+ },
173590
+ {
173591
+ "epoch": 465.84,
173592
+ "learning_rate": 9.08136217948718e-06,
173593
+ "loss": 0.3826,
173594
+ "step": 58345
173595
+ },
173596
+ {
173597
+ "epoch": 465.88,
173598
+ "learning_rate": 9.081282051282053e-06,
173599
+ "loss": 0.416,
173600
+ "step": 58350
173601
+ },
173602
+ {
173603
+ "epoch": 465.92,
173604
+ "learning_rate": 9.081201923076924e-06,
173605
+ "loss": 0.3328,
173606
+ "step": 58355
173607
+ },
173608
+ {
173609
+ "epoch": 465.96,
173610
+ "learning_rate": 9.081121794871795e-06,
173611
+ "loss": 0.5354,
173612
+ "step": 58360
173613
+ },
173614
+ {
173615
+ "epoch": 466.0,
173616
+ "learning_rate": 9.081041666666668e-06,
173617
+ "loss": 1.529,
173618
+ "step": 58365
173619
+ },
173620
+ {
173621
+ "epoch": 466.0,
173622
+ "eval_loss": 0.405769407749176,
173623
+ "eval_runtime": 40.2672,
173624
+ "eval_samples_per_second": 20.811,
173625
+ "eval_steps_per_second": 0.671,
173626
+ "eval_wer": 0.2038402569905819,
173627
+ "step": 58365
173628
+ },
173629
+ {
173630
+ "epoch": 470.04,
173631
+ "learning_rate": 9.08096153846154e-06,
173632
+ "loss": 0.3582,
173633
+ "step": 58370
173634
+ },
173635
+ {
173636
+ "epoch": 470.08,
173637
+ "learning_rate": 9.080881410256411e-06,
173638
+ "loss": 0.351,
173639
+ "step": 58375
173640
+ },
173641
+ {
173642
+ "epoch": 470.12,
173643
+ "learning_rate": 9.080801282051282e-06,
173644
+ "loss": 0.3805,
173645
+ "step": 58380
173646
+ },
173647
+ {
173648
+ "epoch": 470.16,
173649
+ "learning_rate": 9.080721153846156e-06,
173650
+ "loss": 0.456,
173651
+ "step": 58385
173652
+ },
173653
+ {
173654
+ "epoch": 470.2,
173655
+ "learning_rate": 9.080641025641025e-06,
173656
+ "loss": 1.3478,
173657
+ "step": 58390
173658
+ },
173659
+ {
173660
+ "epoch": 470.24,
173661
+ "learning_rate": 9.080560897435898e-06,
173662
+ "loss": 0.3678,
173663
+ "step": 58395
173664
+ },
173665
+ {
173666
+ "epoch": 470.28,
173667
+ "learning_rate": 9.08048076923077e-06,
173668
+ "loss": 0.3497,
173669
+ "step": 58400
173670
+ },
173671
+ {
173672
+ "epoch": 470.32,
173673
+ "learning_rate": 9.080400641025641e-06,
173674
+ "loss": 0.3635,
173675
+ "step": 58405
173676
+ },
173677
+ {
173678
+ "epoch": 470.36,
173679
+ "learning_rate": 9.080320512820514e-06,
173680
+ "loss": 0.5171,
173681
+ "step": 58410
173682
+ },
173683
+ {
173684
+ "epoch": 470.4,
173685
+ "learning_rate": 9.080240384615385e-06,
173686
+ "loss": 1.2932,
173687
+ "step": 58415
173688
+ },
173689
+ {
173690
+ "epoch": 470.44,
173691
+ "learning_rate": 9.080160256410257e-06,
173692
+ "loss": 0.3472,
173693
+ "step": 58420
173694
+ },
173695
+ {
173696
+ "epoch": 470.48,
173697
+ "learning_rate": 9.080080128205128e-06,
173698
+ "loss": 0.3133,
173699
+ "step": 58425
173700
+ },
173701
+ {
173702
+ "epoch": 470.52,
173703
+ "learning_rate": 9.080000000000001e-06,
173704
+ "loss": 0.3152,
173705
+ "step": 58430
173706
+ },
173707
+ {
173708
+ "epoch": 470.56,
173709
+ "learning_rate": 9.079919871794873e-06,
173710
+ "loss": 0.4895,
173711
+ "step": 58435
173712
+ },
173713
+ {
173714
+ "epoch": 470.6,
173715
+ "learning_rate": 9.079839743589744e-06,
173716
+ "loss": 1.096,
173717
+ "step": 58440
173718
+ },
173719
+ {
173720
+ "epoch": 470.65,
173721
+ "learning_rate": 9.079759615384615e-06,
173722
+ "loss": 0.4131,
173723
+ "step": 58445
173724
+ },
173725
+ {
173726
+ "epoch": 470.69,
173727
+ "learning_rate": 9.079679487179488e-06,
173728
+ "loss": 0.3759,
173729
+ "step": 58450
173730
+ },
173731
+ {
173732
+ "epoch": 470.73,
173733
+ "learning_rate": 9.07959935897436e-06,
173734
+ "loss": 0.3681,
173735
+ "step": 58455
173736
+ },
173737
+ {
173738
+ "epoch": 470.77,
173739
+ "learning_rate": 9.079519230769231e-06,
173740
+ "loss": 0.5257,
173741
+ "step": 58460
173742
+ },
173743
+ {
173744
+ "epoch": 470.81,
173745
+ "learning_rate": 9.079439102564104e-06,
173746
+ "loss": 1.3265,
173747
+ "step": 58465
173748
+ },
173749
+ {
173750
+ "epoch": 470.85,
173751
+ "learning_rate": 9.079358974358975e-06,
173752
+ "loss": 0.3764,
173753
+ "step": 58470
173754
+ },
173755
+ {
173756
+ "epoch": 470.89,
173757
+ "learning_rate": 9.079278846153847e-06,
173758
+ "loss": 0.31,
173759
+ "step": 58475
173760
+ },
173761
+ {
173762
+ "epoch": 470.93,
173763
+ "learning_rate": 9.079198717948718e-06,
173764
+ "loss": 0.3444,
173765
+ "step": 58480
173766
+ },
173767
+ {
173768
+ "epoch": 470.97,
173769
+ "learning_rate": 9.079118589743591e-06,
173770
+ "loss": 0.6568,
173771
+ "step": 58485
173772
+ },
173773
+ {
173774
+ "epoch": 471.0,
173775
+ "eval_loss": 0.4264553189277649,
173776
+ "eval_runtime": 39.6989,
173777
+ "eval_samples_per_second": 21.109,
173778
+ "eval_steps_per_second": 0.68,
173779
+ "eval_wer": 0.19773474607234198,
173780
+ "step": 58489
173781
+ },
173782
+ {
173783
+ "epoch": 467.01,
173784
+ "learning_rate": 9.079038461538463e-06,
173785
+ "loss": 0.5071,
173786
+ "step": 58490
173787
+ },
173788
+ {
173789
+ "epoch": 467.05,
173790
+ "learning_rate": 9.078958333333334e-06,
173791
+ "loss": 0.3601,
173792
+ "step": 58495
173793
+ },
173794
+ {
173795
+ "epoch": 467.09,
173796
+ "learning_rate": 9.078878205128205e-06,
173797
+ "loss": 0.378,
173798
+ "step": 58500
173799
+ },
173800
+ {
173801
+ "epoch": 467.13,
173802
+ "learning_rate": 9.078798076923078e-06,
173803
+ "loss": 0.427,
173804
+ "step": 58505
173805
+ },
173806
+ {
173807
+ "epoch": 467.17,
173808
+ "learning_rate": 9.07871794871795e-06,
173809
+ "loss": 0.6693,
173810
+ "step": 58510
173811
+ },
173812
+ {
173813
+ "epoch": 467.21,
173814
+ "learning_rate": 9.078637820512821e-06,
173815
+ "loss": 1.5666,
173816
+ "step": 58515
173817
+ },
173818
+ {
173819
+ "epoch": 467.25,
173820
+ "learning_rate": 9.078557692307694e-06,
173821
+ "loss": 0.3476,
173822
+ "step": 58520
173823
+ },
173824
+ {
173825
+ "epoch": 467.29,
173826
+ "learning_rate": 9.078477564102565e-06,
173827
+ "loss": 0.3344,
173828
+ "step": 58525
173829
+ },
173830
+ {
173831
+ "epoch": 467.33,
173832
+ "learning_rate": 9.078397435897437e-06,
173833
+ "loss": 0.3901,
173834
+ "step": 58530
173835
+ },
173836
+ {
173837
+ "epoch": 467.37,
173838
+ "learning_rate": 9.078317307692308e-06,
173839
+ "loss": 0.5791,
173840
+ "step": 58535
173841
+ },
173842
+ {
173843
+ "epoch": 467.41,
173844
+ "learning_rate": 9.078237179487181e-06,
173845
+ "loss": 1.095,
173846
+ "step": 58540
173847
+ },
173848
+ {
173849
+ "epoch": 467.45,
173850
+ "learning_rate": 9.078157051282051e-06,
173851
+ "loss": 0.32,
173852
+ "step": 58545
173853
+ },
173854
+ {
173855
+ "epoch": 467.49,
173856
+ "learning_rate": 9.078076923076924e-06,
173857
+ "loss": 0.2828,
173858
+ "step": 58550
173859
+ },
173860
+ {
173861
+ "epoch": 467.53,
173862
+ "learning_rate": 9.077996794871795e-06,
173863
+ "loss": 0.3248,
173864
+ "step": 58555
173865
+ },
173866
+ {
173867
+ "epoch": 467.57,
173868
+ "learning_rate": 9.077916666666667e-06,
173869
+ "loss": 0.5747,
173870
+ "step": 58560
173871
+ },
173872
+ {
173873
+ "epoch": 467.61,
173874
+ "learning_rate": 9.07783653846154e-06,
173875
+ "loss": 1.1822,
173876
+ "step": 58565
173877
+ },
173878
+ {
173879
+ "epoch": 467.65,
173880
+ "learning_rate": 9.077756410256411e-06,
173881
+ "loss": 0.3319,
173882
+ "step": 58570
173883
+ },
173884
+ {
173885
+ "epoch": 467.69,
173886
+ "learning_rate": 9.077676282051282e-06,
173887
+ "loss": 0.3283,
173888
+ "step": 58575
173889
+ },
173890
+ {
173891
+ "epoch": 467.73,
173892
+ "learning_rate": 9.077596153846154e-06,
173893
+ "loss": 0.4325,
173894
+ "step": 58580
173895
+ },
173896
+ {
173897
+ "epoch": 467.77,
173898
+ "learning_rate": 9.077516025641027e-06,
173899
+ "loss": 0.5353,
173900
+ "step": 58585
173901
+ },
173902
+ {
173903
+ "epoch": 467.81,
173904
+ "learning_rate": 9.077435897435898e-06,
173905
+ "loss": 1.3401,
173906
+ "step": 58590
173907
+ },
173908
+ {
173909
+ "epoch": 467.85,
173910
+ "learning_rate": 9.07735576923077e-06,
173911
+ "loss": 0.3492,
173912
+ "step": 58595
173913
+ },
173914
+ {
173915
+ "epoch": 467.89,
173916
+ "learning_rate": 9.077275641025641e-06,
173917
+ "loss": 0.363,
173918
+ "step": 58600
173919
+ },
173920
+ {
173921
+ "epoch": 467.93,
173922
+ "learning_rate": 9.077195512820514e-06,
173923
+ "loss": 0.3184,
173924
+ "step": 58605
173925
+ },
173926
+ {
173927
+ "epoch": 467.97,
173928
+ "learning_rate": 9.077115384615385e-06,
173929
+ "loss": 0.5845,
173930
+ "step": 58610
173931
+ },
173932
+ {
173933
+ "epoch": 468.0,
173934
+ "eval_loss": 0.3841117322444916,
173935
+ "eval_runtime": 40.8509,
173936
+ "eval_samples_per_second": 20.514,
173937
+ "eval_steps_per_second": 0.661,
173938
+ "eval_wer": 0.20130728554641597,
173939
+ "step": 58614
173940
+ },
173941
+ {
173942
+ "epoch": 472.01,
173943
+ "learning_rate": 9.077035256410257e-06,
173944
+ "loss": 0.395,
173945
+ "step": 58615
173946
+ },
173947
+ {
173948
+ "epoch": 472.05,
173949
+ "learning_rate": 9.07695512820513e-06,
173950
+ "loss": 0.3279,
173951
+ "step": 58620
173952
+ },
173953
+ {
173954
+ "epoch": 472.09,
173955
+ "learning_rate": 9.076875000000001e-06,
173956
+ "loss": 0.3135,
173957
+ "step": 58625
173958
+ },
173959
+ {
173960
+ "epoch": 472.13,
173961
+ "learning_rate": 9.076794871794872e-06,
173962
+ "loss": 0.4003,
173963
+ "step": 58630
173964
+ },
173965
+ {
173966
+ "epoch": 472.17,
173967
+ "learning_rate": 9.076714743589744e-06,
173968
+ "loss": 0.6409,
173969
+ "step": 58635
173970
+ },
173971
+ {
173972
+ "epoch": 472.21,
173973
+ "learning_rate": 9.076634615384617e-06,
173974
+ "loss": 1.1497,
173975
+ "step": 58640
173976
+ },
173977
+ {
173978
+ "epoch": 472.25,
173979
+ "learning_rate": 9.076554487179488e-06,
173980
+ "loss": 0.347,
173981
+ "step": 58645
173982
+ },
173983
+ {
173984
+ "epoch": 472.29,
173985
+ "learning_rate": 9.07647435897436e-06,
173986
+ "loss": 0.372,
173987
+ "step": 58650
173988
+ },
173989
+ {
173990
+ "epoch": 472.33,
173991
+ "learning_rate": 9.076394230769231e-06,
173992
+ "loss": 0.3361,
173993
+ "step": 58655
173994
+ },
173995
+ {
173996
+ "epoch": 472.37,
173997
+ "learning_rate": 9.076314102564104e-06,
173998
+ "loss": 0.538,
173999
+ "step": 58660
174000
+ },
174001
+ {
174002
+ "epoch": 472.41,
174003
+ "learning_rate": 9.076233974358975e-06,
174004
+ "loss": 1.161,
174005
+ "step": 58665
174006
+ },
174007
+ {
174008
+ "epoch": 472.45,
174009
+ "learning_rate": 9.076153846153847e-06,
174010
+ "loss": 0.3147,
174011
+ "step": 58670
174012
+ },
174013
+ {
174014
+ "epoch": 472.49,
174015
+ "learning_rate": 9.07607371794872e-06,
174016
+ "loss": 0.3359,
174017
+ "step": 58675
174018
+ },
174019
+ {
174020
+ "epoch": 472.53,
174021
+ "learning_rate": 9.07599358974359e-06,
174022
+ "loss": 0.426,
174023
+ "step": 58680
174024
+ },
174025
+ {
174026
+ "epoch": 472.57,
174027
+ "learning_rate": 9.075913461538462e-06,
174028
+ "loss": 0.6531,
174029
+ "step": 58685
174030
+ },
174031
+ {
174032
+ "epoch": 472.61,
174033
+ "learning_rate": 9.075833333333334e-06,
174034
+ "loss": 1.508,
174035
+ "step": 58690
174036
+ },
174037
+ {
174038
+ "epoch": 472.65,
174039
+ "learning_rate": 9.075753205128205e-06,
174040
+ "loss": 0.3804,
174041
+ "step": 58695
174042
+ },
174043
+ {
174044
+ "epoch": 472.69,
174045
+ "learning_rate": 9.075673076923077e-06,
174046
+ "loss": 0.3454,
174047
+ "step": 58700
174048
+ },
174049
+ {
174050
+ "epoch": 472.73,
174051
+ "learning_rate": 9.07559294871795e-06,
174052
+ "loss": 0.3783,
174053
+ "step": 58705
174054
+ },
174055
+ {
174056
+ "epoch": 472.77,
174057
+ "learning_rate": 9.075512820512821e-06,
174058
+ "loss": 0.613,
174059
+ "step": 58710
174060
+ },
174061
+ {
174062
+ "epoch": 472.81,
174063
+ "learning_rate": 9.075432692307692e-06,
174064
+ "loss": 1.2604,
174065
+ "step": 58715
174066
+ },
174067
+ {
174068
+ "epoch": 472.85,
174069
+ "learning_rate": 9.075352564102565e-06,
174070
+ "loss": 0.3477,
174071
+ "step": 58720
174072
+ },
174073
+ {
174074
+ "epoch": 472.89,
174075
+ "learning_rate": 9.075272435897437e-06,
174076
+ "loss": 0.3303,
174077
+ "step": 58725
174078
+ },
174079
+ {
174080
+ "epoch": 472.93,
174081
+ "learning_rate": 9.075192307692308e-06,
174082
+ "loss": 0.3996,
174083
+ "step": 58730
174084
+ },
174085
+ {
174086
+ "epoch": 472.97,
174087
+ "learning_rate": 9.07511217948718e-06,
174088
+ "loss": 0.6128,
174089
+ "step": 58735
174090
+ },
174091
+ {
174092
+ "epoch": 473.0,
174093
+ "eval_loss": 0.39583051204681396,
174094
+ "eval_runtime": 38.5651,
174095
+ "eval_samples_per_second": 21.729,
174096
+ "eval_steps_per_second": 0.7,
174097
+ "eval_wer": 0.18955640423031728,
174098
+ "step": 58738
174099
+ },
174100
+ {
174101
+ "epoch": 469.02,
174102
+ "learning_rate": 9.075032051282052e-06,
174103
+ "loss": 0.3805,
174104
+ "step": 58740
174105
+ },
174106
+ {
174107
+ "epoch": 469.06,
174108
+ "learning_rate": 9.074951923076924e-06,
174109
+ "loss": 0.2926,
174110
+ "step": 58745
174111
+ },
174112
+ {
174113
+ "epoch": 469.1,
174114
+ "learning_rate": 9.074871794871795e-06,
174115
+ "loss": 0.4002,
174116
+ "step": 58750
174117
+ },
174118
+ {
174119
+ "epoch": 469.14,
174120
+ "learning_rate": 9.074791666666667e-06,
174121
+ "loss": 0.4061,
174122
+ "step": 58755
174123
+ },
174124
+ {
174125
+ "epoch": 469.18,
174126
+ "learning_rate": 9.07471153846154e-06,
174127
+ "loss": 0.6746,
174128
+ "step": 58760
174129
+ },
174130
+ {
174131
+ "epoch": 469.22,
174132
+ "learning_rate": 9.074631410256411e-06,
174133
+ "loss": 1.0583,
174134
+ "step": 58765
174135
+ },
174136
+ {
174137
+ "epoch": 469.26,
174138
+ "learning_rate": 9.074551282051282e-06,
174139
+ "loss": 0.3149,
174140
+ "step": 58770
174141
+ },
174142
+ {
174143
+ "epoch": 469.3,
174144
+ "learning_rate": 9.074471153846155e-06,
174145
+ "loss": 0.3262,
174146
+ "step": 58775
174147
+ },
174148
+ {
174149
+ "epoch": 469.34,
174150
+ "learning_rate": 9.074391025641027e-06,
174151
+ "loss": 0.4636,
174152
+ "step": 58780
174153
+ },
174154
+ {
174155
+ "epoch": 469.38,
174156
+ "learning_rate": 9.074310897435898e-06,
174157
+ "loss": 0.7496,
174158
+ "step": 58785
174159
+ },
174160
+ {
174161
+ "epoch": 469.42,
174162
+ "learning_rate": 9.07423076923077e-06,
174163
+ "loss": 0.9887,
174164
+ "step": 58790
174165
+ },
174166
+ {
174167
+ "epoch": 469.46,
174168
+ "learning_rate": 9.074150641025643e-06,
174169
+ "loss": 0.3095,
174170
+ "step": 58795
174171
+ },
174172
+ {
174173
+ "epoch": 469.5,
174174
+ "learning_rate": 9.074070512820514e-06,
174175
+ "loss": 0.3382,
174176
+ "step": 58800
174177
+ },
174178
+ {
174179
+ "epoch": 469.54,
174180
+ "learning_rate": 9.073990384615385e-06,
174181
+ "loss": 0.3929,
174182
+ "step": 58805
174183
+ },
174184
+ {
174185
+ "epoch": 469.58,
174186
+ "learning_rate": 9.073910256410258e-06,
174187
+ "loss": 0.6524,
174188
+ "step": 58810
174189
+ },
174190
+ {
174191
+ "epoch": 469.62,
174192
+ "learning_rate": 9.07383012820513e-06,
174193
+ "loss": 1.1419,
174194
+ "step": 58815
174195
+ },
174196
+ {
174197
+ "epoch": 469.66,
174198
+ "learning_rate": 9.073750000000001e-06,
174199
+ "loss": 0.3383,
174200
+ "step": 58820
174201
+ },
174202
+ {
174203
+ "epoch": 469.7,
174204
+ "learning_rate": 9.073669871794872e-06,
174205
+ "loss": 0.398,
174206
+ "step": 58825
174207
+ },
174208
+ {
174209
+ "epoch": 469.74,
174210
+ "learning_rate": 9.073589743589745e-06,
174211
+ "loss": 0.3604,
174212
+ "step": 58830
174213
+ },
174214
+ {
174215
+ "epoch": 469.78,
174216
+ "learning_rate": 9.073509615384615e-06,
174217
+ "loss": 0.7135,
174218
+ "step": 58835
174219
+ },
174220
+ {
174221
+ "epoch": 469.82,
174222
+ "learning_rate": 9.073429487179488e-06,
174223
+ "loss": 1.2744,
174224
+ "step": 58840
174225
+ },
174226
+ {
174227
+ "epoch": 469.86,
174228
+ "learning_rate": 9.07334935897436e-06,
174229
+ "loss": 0.351,
174230
+ "step": 58845
174231
+ },
174232
+ {
174233
+ "epoch": 469.9,
174234
+ "learning_rate": 9.073269230769231e-06,
174235
+ "loss": 0.3801,
174236
+ "step": 58850
174237
+ },
174238
+ {
174239
+ "epoch": 469.94,
174240
+ "learning_rate": 9.073189102564102e-06,
174241
+ "loss": 0.372,
174242
+ "step": 58855
174243
+ },
174244
+ {
174245
+ "epoch": 469.98,
174246
+ "learning_rate": 9.073108974358975e-06,
174247
+ "loss": 0.7963,
174248
+ "step": 58860
174249
+ },
174250
+ {
174251
+ "epoch": 470.0,
174252
+ "eval_loss": 0.41615360975265503,
174253
+ "eval_runtime": 39.8592,
174254
+ "eval_samples_per_second": 21.024,
174255
+ "eval_steps_per_second": 0.677,
174256
+ "eval_wer": 0.19737984944991313,
174257
+ "step": 58863
174258
  }
174259
  ],
174260
  "max_steps": 625000,
174261
  "num_train_epochs": 5000,
174262
+ "total_flos": 1.656479841453496e+20,
174263
  "trial_name": null,
174264
  "trial_params": null
174265
  }
model-bin/finetune/base/{checkpoint-58240 β†’ checkpoint-58863}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629788442.126761/events.out.tfevents.1629788442.c435e1c5ee04.920.71 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:604c72456103c048ed1693e5c48b46865ca4fea2d95ce2a32c3d17509cf2ba26
3
+ size 4194
model-bin/finetune/base/log/1629789087.6950493/events.out.tfevents.1629789087.c435e1c5ee04.920.73 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5180d2b36f470c893631930ac5adca1c645d82258244fe95434f4e498c26e4db
3
+ size 4194
model-bin/finetune/base/log/1629789729.6492243/events.out.tfevents.1629789729.c435e1c5ee04.920.75 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99ea0f398e53dd81171f7d57f07e05be7476c85d3ca419e3f98e954337368f48
3
+ size 4194
model-bin/finetune/base/log/1629790377.8418694/events.out.tfevents.1629790377.c435e1c5ee04.920.77 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b596940341ed38c06524c37906bbb076cfc0432abecb3799d62914a952caadb5
3
+ size 4194
model-bin/finetune/base/log/1629791034.1306834/events.out.tfevents.1629791034.c435e1c5ee04.920.79 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:955ccd10408e0fdcfecfab845910d4dd18e96bb30e648211e0d888e69c852e7c
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629788442.c435e1c5ee04.920.70 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22499354f72e96e46369d0e780671621e3b5e3f983268f0d379381a43c72d156
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629789087.c435e1c5ee04.920.72 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4c486c4efc6b61b3f00d9db9cbff1c2cbe905fe172ebb850ac5f2fd27d2b850
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1629789729.c435e1c5ee04.920.74 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b15cf706c251a6c6f585b6a14b9dc4b5a93669409ff6f686376fa453db3a85b2
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629790377.c435e1c5ee04.920.76 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3defc3aac724851bdd10b9cd6305b5aa096a37939112160e5446f442fafa6d6c
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629791034.c435e1c5ee04.920.78 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:587fb13106e4ac50c2068eb3c57e604e6fbe32a14a832b92beabcfdfa87c4c6c
3
+ size 8622