Check commited on
Commit
ab60e4f
Β·
1 Parent(s): 096355b

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629941989.4511316/events.out.tfevents.1629941989.8e89bd551565.924.11 +3 -0
  11. model-bin/finetune/base/log/1629942431.5525227/events.out.tfevents.1629942431.8e89bd551565.924.13 +3 -0
  12. model-bin/finetune/base/log/1629942858.3612747/events.out.tfevents.1629942858.8e89bd551565.924.15 +3 -0
  13. model-bin/finetune/base/log/1629943284.2321923/events.out.tfevents.1629943284.8e89bd551565.924.17 +3 -0
  14. model-bin/finetune/base/log/1629943702.5924923/events.out.tfevents.1629943702.8e89bd551565.924.19 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629941989.8e89bd551565.924.10 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629942431.8e89bd551565.924.12 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629942858.8e89bd551565.924.14 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629943284.8e89bd551565.924.16 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629943702.8e89bd551565.924.18 +3 -0
model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:85223ba53941b432c78346aab6f9972c74939487ac260fa93b1d6b1f037f2caa
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2443b8ea1d53aac7f389153148ffabe5ce7712b63b8d05b786b70b8a85200a25
3
  size 722165393
model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a7d0ee63362b6103e3b14d409796c9f7d0158647e1d489337c6a63ebb1785554
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d42f5b531eb03f0e314693855366c0745885141031e2c6fb93cab7ad1340bbb1
3
  size 377909911
model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa2d10ac6adbf42dc45f693ff91266acd0ee130d5f99d16aabf290292d077da1
3
  size 14567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91ed9200715cb38ca621fab9af9f5470bcc041ea0ae387481d03723f806e19db
3
  size 14567
model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9da1248dae22f22d368d318bd65aa7f85b60c95535b8a8cd70eeacd40df98fbe
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f396b66cec689ed47334ef06d8b2f62cc0b8112d2b9e35e8e685b8ace420b29
3
  size 559
model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca8f1a5975684ec4e0bf01320f1c3c47ae2829cdaa18b8d590d05e5f125b3b39
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee56de8be16189110ea7ac83a11716df996819bb79cfacd43f9aafb201925200
3
  size 623
model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.17731766728757703,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-88857",
4
- "epoch": 717.0,
5
- "global_step": 88981,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -212577,11 +212577,800 @@
212577
  "eval_steps_per_second": 0.703,
212578
  "eval_wer": 0.1821153566758681,
212579
  "step": 88981
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
212580
  }
212581
  ],
212582
- "max_steps": 620000,
212583
  "num_train_epochs": 5000,
212584
- "total_flos": 2.5040338514743958e+20,
212585
  "trial_name": null,
212586
  "trial_params": null
212587
  }
 
1
  {
2
  "best_metric": 0.17731766728757703,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-88857",
4
+ "epoch": 716.0,
5
+ "global_step": 89603,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
212577
  "eval_steps_per_second": 0.703,
212578
  "eval_wer": 0.1821153566758681,
212579
  "step": 88981
212580
+ },
212581
+ {
212582
+ "epoch": 711.03,
212583
+ "learning_rate": 8.590560897435898e-06,
212584
+ "loss": 0.3179,
212585
+ "step": 88985
212586
+ },
212587
+ {
212588
+ "epoch": 711.07,
212589
+ "learning_rate": 8.590480769230771e-06,
212590
+ "loss": 0.2987,
212591
+ "step": 88990
212592
+ },
212593
+ {
212594
+ "epoch": 711.11,
212595
+ "learning_rate": 8.59040064102564e-06,
212596
+ "loss": 0.3108,
212597
+ "step": 88995
212598
+ },
212599
+ {
212600
+ "epoch": 711.15,
212601
+ "learning_rate": 8.590320512820514e-06,
212602
+ "loss": 0.392,
212603
+ "step": 89000
212604
+ },
212605
+ {
212606
+ "epoch": 711.19,
212607
+ "learning_rate": 8.590240384615385e-06,
212608
+ "loss": 0.8451,
212609
+ "step": 89005
212610
+ },
212611
+ {
212612
+ "epoch": 711.23,
212613
+ "learning_rate": 8.590160256410256e-06,
212614
+ "loss": 0.613,
212615
+ "step": 89010
212616
+ },
212617
+ {
212618
+ "epoch": 711.27,
212619
+ "learning_rate": 8.59008012820513e-06,
212620
+ "loss": 0.3105,
212621
+ "step": 89015
212622
+ },
212623
+ {
212624
+ "epoch": 711.31,
212625
+ "learning_rate": 8.59e-06,
212626
+ "loss": 0.2981,
212627
+ "step": 89020
212628
+ },
212629
+ {
212630
+ "epoch": 711.35,
212631
+ "learning_rate": 8.589919871794872e-06,
212632
+ "loss": 0.4424,
212633
+ "step": 89025
212634
+ },
212635
+ {
212636
+ "epoch": 711.39,
212637
+ "learning_rate": 8.589839743589743e-06,
212638
+ "loss": 0.9266,
212639
+ "step": 89030
212640
+ },
212641
+ {
212642
+ "epoch": 711.43,
212643
+ "learning_rate": 8.589759615384617e-06,
212644
+ "loss": 0.6477,
212645
+ "step": 89035
212646
+ },
212647
+ {
212648
+ "epoch": 711.47,
212649
+ "learning_rate": 8.589679487179488e-06,
212650
+ "loss": 0.3203,
212651
+ "step": 89040
212652
+ },
212653
+ {
212654
+ "epoch": 711.51,
212655
+ "learning_rate": 8.58959935897436e-06,
212656
+ "loss": 0.361,
212657
+ "step": 89045
212658
+ },
212659
+ {
212660
+ "epoch": 711.55,
212661
+ "learning_rate": 8.58951923076923e-06,
212662
+ "loss": 0.4658,
212663
+ "step": 89050
212664
+ },
212665
+ {
212666
+ "epoch": 711.59,
212667
+ "learning_rate": 8.589439102564104e-06,
212668
+ "loss": 0.9329,
212669
+ "step": 89055
212670
+ },
212671
+ {
212672
+ "epoch": 711.63,
212673
+ "learning_rate": 8.589358974358975e-06,
212674
+ "loss": 0.6051,
212675
+ "step": 89060
212676
+ },
212677
+ {
212678
+ "epoch": 711.67,
212679
+ "learning_rate": 8.589278846153846e-06,
212680
+ "loss": 0.2512,
212681
+ "step": 89065
212682
+ },
212683
+ {
212684
+ "epoch": 711.71,
212685
+ "learning_rate": 8.58919871794872e-06,
212686
+ "loss": 0.2798,
212687
+ "step": 89070
212688
+ },
212689
+ {
212690
+ "epoch": 711.75,
212691
+ "learning_rate": 8.58911858974359e-06,
212692
+ "loss": 0.4504,
212693
+ "step": 89075
212694
+ },
212695
+ {
212696
+ "epoch": 711.79,
212697
+ "learning_rate": 8.589038461538462e-06,
212698
+ "loss": 0.9256,
212699
+ "step": 89080
212700
+ },
212701
+ {
212702
+ "epoch": 711.83,
212703
+ "learning_rate": 8.588958333333334e-06,
212704
+ "loss": 0.7481,
212705
+ "step": 89085
212706
+ },
212707
+ {
212708
+ "epoch": 711.87,
212709
+ "learning_rate": 8.588878205128207e-06,
212710
+ "loss": 0.2726,
212711
+ "step": 89090
212712
+ },
212713
+ {
212714
+ "epoch": 711.91,
212715
+ "learning_rate": 8.588798076923078e-06,
212716
+ "loss": 0.313,
212717
+ "step": 89095
212718
+ },
212719
+ {
212720
+ "epoch": 711.95,
212721
+ "learning_rate": 8.58871794871795e-06,
212722
+ "loss": 0.3964,
212723
+ "step": 89100
212724
+ },
212725
+ {
212726
+ "epoch": 711.99,
212727
+ "learning_rate": 8.58863782051282e-06,
212728
+ "loss": 0.9138,
212729
+ "step": 89105
212730
+ },
212731
+ {
212732
+ "epoch": 712.0,
212733
+ "eval_loss": 0.3886844217777252,
212734
+ "eval_runtime": 39.4344,
212735
+ "eval_samples_per_second": 21.327,
212736
+ "eval_steps_per_second": 0.685,
212737
+ "eval_wer": 0.19554527046572173,
212738
+ "step": 89106
212739
+ },
212740
+ {
212741
+ "epoch": 718.03,
212742
+ "learning_rate": 8.588557692307694e-06,
212743
+ "loss": 0.3705,
212744
+ "step": 89110
212745
+ },
212746
+ {
212747
+ "epoch": 718.07,
212748
+ "learning_rate": 8.588477564102565e-06,
212749
+ "loss": 0.2904,
212750
+ "step": 89115
212751
+ },
212752
+ {
212753
+ "epoch": 718.11,
212754
+ "learning_rate": 8.588397435897436e-06,
212755
+ "loss": 0.3454,
212756
+ "step": 89120
212757
+ },
212758
+ {
212759
+ "epoch": 718.15,
212760
+ "learning_rate": 8.58831730769231e-06,
212761
+ "loss": 0.3967,
212762
+ "step": 89125
212763
+ },
212764
+ {
212765
+ "epoch": 718.19,
212766
+ "learning_rate": 8.588237179487179e-06,
212767
+ "loss": 0.915,
212768
+ "step": 89130
212769
+ },
212770
+ {
212771
+ "epoch": 718.23,
212772
+ "learning_rate": 8.588157051282052e-06,
212773
+ "loss": 0.7099,
212774
+ "step": 89135
212775
+ },
212776
+ {
212777
+ "epoch": 718.27,
212778
+ "learning_rate": 8.588076923076924e-06,
212779
+ "loss": 0.2788,
212780
+ "step": 89140
212781
+ },
212782
+ {
212783
+ "epoch": 718.31,
212784
+ "learning_rate": 8.587996794871795e-06,
212785
+ "loss": 0.2902,
212786
+ "step": 89145
212787
+ },
212788
+ {
212789
+ "epoch": 718.35,
212790
+ "learning_rate": 8.587916666666666e-06,
212791
+ "loss": 0.3752,
212792
+ "step": 89150
212793
+ },
212794
+ {
212795
+ "epoch": 718.39,
212796
+ "learning_rate": 8.58783653846154e-06,
212797
+ "loss": 0.8795,
212798
+ "step": 89155
212799
+ },
212800
+ {
212801
+ "epoch": 718.43,
212802
+ "learning_rate": 8.58775641025641e-06,
212803
+ "loss": 0.824,
212804
+ "step": 89160
212805
+ },
212806
+ {
212807
+ "epoch": 718.47,
212808
+ "learning_rate": 8.587676282051282e-06,
212809
+ "loss": 0.3274,
212810
+ "step": 89165
212811
+ },
212812
+ {
212813
+ "epoch": 718.51,
212814
+ "learning_rate": 8.587596153846155e-06,
212815
+ "loss": 0.3495,
212816
+ "step": 89170
212817
+ },
212818
+ {
212819
+ "epoch": 718.55,
212820
+ "learning_rate": 8.587516025641026e-06,
212821
+ "loss": 0.41,
212822
+ "step": 89175
212823
+ },
212824
+ {
212825
+ "epoch": 718.59,
212826
+ "learning_rate": 8.587435897435898e-06,
212827
+ "loss": 1.0472,
212828
+ "step": 89180
212829
+ },
212830
+ {
212831
+ "epoch": 718.63,
212832
+ "learning_rate": 8.587355769230769e-06,
212833
+ "loss": 0.6203,
212834
+ "step": 89185
212835
+ },
212836
+ {
212837
+ "epoch": 718.67,
212838
+ "learning_rate": 8.587275641025642e-06,
212839
+ "loss": 0.3644,
212840
+ "step": 89190
212841
+ },
212842
+ {
212843
+ "epoch": 718.71,
212844
+ "learning_rate": 8.587195512820514e-06,
212845
+ "loss": 0.2647,
212846
+ "step": 89195
212847
+ },
212848
+ {
212849
+ "epoch": 718.76,
212850
+ "learning_rate": 8.587115384615385e-06,
212851
+ "loss": 0.4371,
212852
+ "step": 89200
212853
+ },
212854
+ {
212855
+ "epoch": 718.8,
212856
+ "learning_rate": 8.587035256410256e-06,
212857
+ "loss": 1.1203,
212858
+ "step": 89205
212859
+ },
212860
+ {
212861
+ "epoch": 718.84,
212862
+ "learning_rate": 8.58695512820513e-06,
212863
+ "loss": 0.652,
212864
+ "step": 89210
212865
+ },
212866
+ {
212867
+ "epoch": 718.88,
212868
+ "learning_rate": 8.586875e-06,
212869
+ "loss": 0.3579,
212870
+ "step": 89215
212871
+ },
212872
+ {
212873
+ "epoch": 718.92,
212874
+ "learning_rate": 8.586794871794872e-06,
212875
+ "loss": 0.3215,
212876
+ "step": 89220
212877
+ },
212878
+ {
212879
+ "epoch": 718.96,
212880
+ "learning_rate": 8.586714743589745e-06,
212881
+ "loss": 0.4431,
212882
+ "step": 89225
212883
+ },
212884
+ {
212885
+ "epoch": 719.0,
212886
+ "learning_rate": 8.586634615384616e-06,
212887
+ "loss": 1.1711,
212888
+ "step": 89230
212889
+ },
212890
+ {
212891
+ "epoch": 719.0,
212892
+ "eval_loss": 0.44025877118110657,
212893
+ "eval_runtime": 38.1224,
212894
+ "eval_samples_per_second": 22.061,
212895
+ "eval_steps_per_second": 0.708,
212896
+ "eval_wer": 0.19066512768720242,
212897
+ "step": 89230
212898
+ },
212899
+ {
212900
+ "epoch": 713.04,
212901
+ "learning_rate": 8.586554487179488e-06,
212902
+ "loss": 0.3305,
212903
+ "step": 89235
212904
+ },
212905
+ {
212906
+ "epoch": 713.08,
212907
+ "learning_rate": 8.58647435897436e-06,
212908
+ "loss": 0.2897,
212909
+ "step": 89240
212910
+ },
212911
+ {
212912
+ "epoch": 713.12,
212913
+ "learning_rate": 8.586394230769232e-06,
212914
+ "loss": 0.3075,
212915
+ "step": 89245
212916
+ },
212917
+ {
212918
+ "epoch": 713.16,
212919
+ "learning_rate": 8.586314102564104e-06,
212920
+ "loss": 0.4379,
212921
+ "step": 89250
212922
+ },
212923
+ {
212924
+ "epoch": 713.2,
212925
+ "learning_rate": 8.586233974358975e-06,
212926
+ "loss": 1.2201,
212927
+ "step": 89255
212928
+ },
212929
+ {
212930
+ "epoch": 713.24,
212931
+ "learning_rate": 8.586153846153848e-06,
212932
+ "loss": 0.3432,
212933
+ "step": 89260
212934
+ },
212935
+ {
212936
+ "epoch": 713.28,
212937
+ "learning_rate": 8.58607371794872e-06,
212938
+ "loss": 0.2984,
212939
+ "step": 89265
212940
+ },
212941
+ {
212942
+ "epoch": 713.32,
212943
+ "learning_rate": 8.58599358974359e-06,
212944
+ "loss": 0.362,
212945
+ "step": 89270
212946
+ },
212947
+ {
212948
+ "epoch": 713.36,
212949
+ "learning_rate": 8.585913461538462e-06,
212950
+ "loss": 0.4702,
212951
+ "step": 89275
212952
+ },
212953
+ {
212954
+ "epoch": 713.4,
212955
+ "learning_rate": 8.585833333333335e-06,
212956
+ "loss": 1.2665,
212957
+ "step": 89280
212958
+ },
212959
+ {
212960
+ "epoch": 713.44,
212961
+ "learning_rate": 8.585753205128205e-06,
212962
+ "loss": 0.2856,
212963
+ "step": 89285
212964
+ },
212965
+ {
212966
+ "epoch": 713.48,
212967
+ "learning_rate": 8.585673076923078e-06,
212968
+ "loss": 0.3187,
212969
+ "step": 89290
212970
+ },
212971
+ {
212972
+ "epoch": 713.52,
212973
+ "learning_rate": 8.58559294871795e-06,
212974
+ "loss": 0.2725,
212975
+ "step": 89295
212976
+ },
212977
+ {
212978
+ "epoch": 713.56,
212979
+ "learning_rate": 8.58551282051282e-06,
212980
+ "loss": 0.4075,
212981
+ "step": 89300
212982
+ },
212983
+ {
212984
+ "epoch": 713.6,
212985
+ "learning_rate": 8.585432692307692e-06,
212986
+ "loss": 1.0961,
212987
+ "step": 89305
212988
+ },
212989
+ {
212990
+ "epoch": 713.64,
212991
+ "learning_rate": 8.585352564102565e-06,
212992
+ "loss": 0.3646,
212993
+ "step": 89310
212994
+ },
212995
+ {
212996
+ "epoch": 713.68,
212997
+ "learning_rate": 8.585272435897436e-06,
212998
+ "loss": 0.2833,
212999
+ "step": 89315
213000
+ },
213001
+ {
213002
+ "epoch": 713.72,
213003
+ "learning_rate": 8.585192307692308e-06,
213004
+ "loss": 0.3468,
213005
+ "step": 89320
213006
+ },
213007
+ {
213008
+ "epoch": 713.76,
213009
+ "learning_rate": 8.58511217948718e-06,
213010
+ "loss": 0.4216,
213011
+ "step": 89325
213012
+ },
213013
+ {
213014
+ "epoch": 713.8,
213015
+ "learning_rate": 8.585032051282052e-06,
213016
+ "loss": 1.2307,
213017
+ "step": 89330
213018
+ },
213019
+ {
213020
+ "epoch": 713.84,
213021
+ "learning_rate": 8.584951923076923e-06,
213022
+ "loss": 0.3273,
213023
+ "step": 89335
213024
+ },
213025
+ {
213026
+ "epoch": 713.88,
213027
+ "learning_rate": 8.584871794871795e-06,
213028
+ "loss": 0.2878,
213029
+ "step": 89340
213030
+ },
213031
+ {
213032
+ "epoch": 713.92,
213033
+ "learning_rate": 8.584791666666668e-06,
213034
+ "loss": 0.2904,
213035
+ "step": 89345
213036
+ },
213037
+ {
213038
+ "epoch": 713.96,
213039
+ "learning_rate": 8.58471153846154e-06,
213040
+ "loss": 0.5359,
213041
+ "step": 89350
213042
+ },
213043
+ {
213044
+ "epoch": 714.0,
213045
+ "learning_rate": 8.58463141025641e-06,
213046
+ "loss": 1.3037,
213047
+ "step": 89355
213048
+ },
213049
+ {
213050
+ "epoch": 714.0,
213051
+ "eval_loss": 0.38318589329719543,
213052
+ "eval_runtime": 37.6593,
213053
+ "eval_samples_per_second": 22.332,
213054
+ "eval_steps_per_second": 0.717,
213055
+ "eval_wer": 0.18946828768126556,
213056
+ "step": 89355
213057
+ },
213058
+ {
213059
+ "epoch": 726.04,
213060
+ "learning_rate": 8.584551282051284e-06,
213061
+ "loss": 0.3571,
213062
+ "step": 89360
213063
+ },
213064
+ {
213065
+ "epoch": 726.08,
213066
+ "learning_rate": 8.584471153846155e-06,
213067
+ "loss": 0.3214,
213068
+ "step": 89365
213069
+ },
213070
+ {
213071
+ "epoch": 726.12,
213072
+ "learning_rate": 8.584391025641026e-06,
213073
+ "loss": 0.3781,
213074
+ "step": 89370
213075
+ },
213076
+ {
213077
+ "epoch": 726.16,
213078
+ "learning_rate": 8.584310897435898e-06,
213079
+ "loss": 0.4707,
213080
+ "step": 89375
213081
+ },
213082
+ {
213083
+ "epoch": 726.2,
213084
+ "learning_rate": 8.58423076923077e-06,
213085
+ "loss": 1.139,
213086
+ "step": 89380
213087
+ },
213088
+ {
213089
+ "epoch": 726.24,
213090
+ "learning_rate": 8.584150641025642e-06,
213091
+ "loss": 0.3475,
213092
+ "step": 89385
213093
+ },
213094
+ {
213095
+ "epoch": 726.28,
213096
+ "learning_rate": 8.584070512820513e-06,
213097
+ "loss": 0.3237,
213098
+ "step": 89390
213099
+ },
213100
+ {
213101
+ "epoch": 726.32,
213102
+ "learning_rate": 8.583990384615385e-06,
213103
+ "loss": 0.3205,
213104
+ "step": 89395
213105
+ },
213106
+ {
213107
+ "epoch": 726.36,
213108
+ "learning_rate": 8.583910256410258e-06,
213109
+ "loss": 0.4546,
213110
+ "step": 89400
213111
+ },
213112
+ {
213113
+ "epoch": 726.4,
213114
+ "learning_rate": 8.583830128205128e-06,
213115
+ "loss": 1.338,
213116
+ "step": 89405
213117
+ },
213118
+ {
213119
+ "epoch": 726.45,
213120
+ "learning_rate": 8.58375e-06,
213121
+ "loss": 0.3786,
213122
+ "step": 89410
213123
+ },
213124
+ {
213125
+ "epoch": 726.49,
213126
+ "learning_rate": 8.583669871794874e-06,
213127
+ "loss": 0.2822,
213128
+ "step": 89415
213129
+ },
213130
+ {
213131
+ "epoch": 726.53,
213132
+ "learning_rate": 8.583589743589743e-06,
213133
+ "loss": 0.3013,
213134
+ "step": 89420
213135
+ },
213136
+ {
213137
+ "epoch": 726.57,
213138
+ "learning_rate": 8.583509615384616e-06,
213139
+ "loss": 0.4148,
213140
+ "step": 89425
213141
+ },
213142
+ {
213143
+ "epoch": 726.61,
213144
+ "learning_rate": 8.583429487179488e-06,
213145
+ "loss": 1.1928,
213146
+ "step": 89430
213147
+ },
213148
+ {
213149
+ "epoch": 726.65,
213150
+ "learning_rate": 8.583349358974359e-06,
213151
+ "loss": 0.2964,
213152
+ "step": 89435
213153
+ },
213154
+ {
213155
+ "epoch": 726.69,
213156
+ "learning_rate": 8.58326923076923e-06,
213157
+ "loss": 0.3008,
213158
+ "step": 89440
213159
+ },
213160
+ {
213161
+ "epoch": 726.73,
213162
+ "learning_rate": 8.583189102564104e-06,
213163
+ "loss": 0.3086,
213164
+ "step": 89445
213165
+ },
213166
+ {
213167
+ "epoch": 726.77,
213168
+ "learning_rate": 8.583108974358975e-06,
213169
+ "loss": 0.4667,
213170
+ "step": 89450
213171
+ },
213172
+ {
213173
+ "epoch": 726.81,
213174
+ "learning_rate": 8.583028846153846e-06,
213175
+ "loss": 1.1449,
213176
+ "step": 89455
213177
+ },
213178
+ {
213179
+ "epoch": 726.85,
213180
+ "learning_rate": 8.58294871794872e-06,
213181
+ "loss": 0.3306,
213182
+ "step": 89460
213183
+ },
213184
+ {
213185
+ "epoch": 726.89,
213186
+ "learning_rate": 8.58286858974359e-06,
213187
+ "loss": 0.2821,
213188
+ "step": 89465
213189
+ },
213190
+ {
213191
+ "epoch": 726.93,
213192
+ "learning_rate": 8.582788461538462e-06,
213193
+ "loss": 0.4959,
213194
+ "step": 89470
213195
+ },
213196
+ {
213197
+ "epoch": 726.97,
213198
+ "learning_rate": 8.582708333333333e-06,
213199
+ "loss": 0.522,
213200
+ "step": 89475
213201
+ },
213202
+ {
213203
+ "epoch": 727.0,
213204
+ "eval_loss": 0.37962403893470764,
213205
+ "eval_runtime": 37.8556,
213206
+ "eval_samples_per_second": 22.216,
213207
+ "eval_steps_per_second": 0.713,
213208
+ "eval_wer": 0.18697681309726646,
213209
+ "step": 89478
213210
+ },
213211
+ {
213212
+ "epoch": 715.02,
213213
+ "learning_rate": 8.582628205128206e-06,
213214
+ "loss": 0.3868,
213215
+ "step": 89480
213216
+ },
213217
+ {
213218
+ "epoch": 715.06,
213219
+ "learning_rate": 8.582548076923078e-06,
213220
+ "loss": 0.3219,
213221
+ "step": 89485
213222
+ },
213223
+ {
213224
+ "epoch": 715.1,
213225
+ "learning_rate": 8.582467948717949e-06,
213226
+ "loss": 0.2923,
213227
+ "step": 89490
213228
+ },
213229
+ {
213230
+ "epoch": 715.14,
213231
+ "learning_rate": 8.58238782051282e-06,
213232
+ "loss": 0.3483,
213233
+ "step": 89495
213234
+ },
213235
+ {
213236
+ "epoch": 715.18,
213237
+ "learning_rate": 8.582307692307694e-06,
213238
+ "loss": 0.6396,
213239
+ "step": 89500
213240
+ },
213241
+ {
213242
+ "epoch": 715.22,
213243
+ "learning_rate": 8.582227564102565e-06,
213244
+ "loss": 0.9591,
213245
+ "step": 89505
213246
+ },
213247
+ {
213248
+ "epoch": 715.26,
213249
+ "learning_rate": 8.582147435897436e-06,
213250
+ "loss": 0.2791,
213251
+ "step": 89510
213252
+ },
213253
+ {
213254
+ "epoch": 715.3,
213255
+ "learning_rate": 8.58206730769231e-06,
213256
+ "loss": 0.3353,
213257
+ "step": 89515
213258
+ },
213259
+ {
213260
+ "epoch": 715.34,
213261
+ "learning_rate": 8.58198717948718e-06,
213262
+ "loss": 0.4082,
213263
+ "step": 89520
213264
+ },
213265
+ {
213266
+ "epoch": 715.38,
213267
+ "learning_rate": 8.581907051282052e-06,
213268
+ "loss": 0.6496,
213269
+ "step": 89525
213270
+ },
213271
+ {
213272
+ "epoch": 715.42,
213273
+ "learning_rate": 8.581826923076923e-06,
213274
+ "loss": 1.0216,
213275
+ "step": 89530
213276
+ },
213277
+ {
213278
+ "epoch": 715.46,
213279
+ "learning_rate": 8.581746794871796e-06,
213280
+ "loss": 0.2943,
213281
+ "step": 89535
213282
+ },
213283
+ {
213284
+ "epoch": 715.5,
213285
+ "learning_rate": 8.581666666666668e-06,
213286
+ "loss": 0.2755,
213287
+ "step": 89540
213288
+ },
213289
+ {
213290
+ "epoch": 715.54,
213291
+ "learning_rate": 8.581586538461539e-06,
213292
+ "loss": 0.3416,
213293
+ "step": 89545
213294
+ },
213295
+ {
213296
+ "epoch": 715.58,
213297
+ "learning_rate": 8.58150641025641e-06,
213298
+ "loss": 0.618,
213299
+ "step": 89550
213300
+ },
213301
+ {
213302
+ "epoch": 715.62,
213303
+ "learning_rate": 8.581426282051284e-06,
213304
+ "loss": 1.0209,
213305
+ "step": 89555
213306
+ },
213307
+ {
213308
+ "epoch": 715.66,
213309
+ "learning_rate": 8.581346153846155e-06,
213310
+ "loss": 0.2874,
213311
+ "step": 89560
213312
+ },
213313
+ {
213314
+ "epoch": 715.7,
213315
+ "learning_rate": 8.581266025641026e-06,
213316
+ "loss": 0.2984,
213317
+ "step": 89565
213318
+ },
213319
+ {
213320
+ "epoch": 715.74,
213321
+ "learning_rate": 8.5811858974359e-06,
213322
+ "loss": 0.3482,
213323
+ "step": 89570
213324
+ },
213325
+ {
213326
+ "epoch": 715.78,
213327
+ "learning_rate": 8.581105769230769e-06,
213328
+ "loss": 0.6483,
213329
+ "step": 89575
213330
+ },
213331
+ {
213332
+ "epoch": 715.82,
213333
+ "learning_rate": 8.581025641025642e-06,
213334
+ "loss": 1.0368,
213335
+ "step": 89580
213336
+ },
213337
+ {
213338
+ "epoch": 715.86,
213339
+ "learning_rate": 8.580945512820513e-06,
213340
+ "loss": 0.2969,
213341
+ "step": 89585
213342
+ },
213343
+ {
213344
+ "epoch": 715.9,
213345
+ "learning_rate": 8.580865384615385e-06,
213346
+ "loss": 0.3009,
213347
+ "step": 89590
213348
+ },
213349
+ {
213350
+ "epoch": 715.94,
213351
+ "learning_rate": 8.580785256410256e-06,
213352
+ "loss": 0.3862,
213353
+ "step": 89595
213354
+ },
213355
+ {
213356
+ "epoch": 715.98,
213357
+ "learning_rate": 8.58070512820513e-06,
213358
+ "loss": 0.6047,
213359
+ "step": 89600
213360
+ },
213361
+ {
213362
+ "epoch": 716.0,
213363
+ "eval_loss": 0.42192313075065613,
213364
+ "eval_runtime": 38.1277,
213365
+ "eval_samples_per_second": 22.084,
213366
+ "eval_steps_per_second": 0.708,
213367
+ "eval_wer": 0.18980540226546616,
213368
+ "step": 89603
213369
  }
213370
  ],
213371
+ "max_steps": 625000,
213372
  "num_train_epochs": 5000,
213373
+ "total_flos": 2.521538624421044e+20,
213374
  "trial_name": null,
213375
  "trial_params": null
213376
  }
model-bin/finetune/base/{checkpoint-88981 β†’ checkpoint-89603}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629941989.4511316/events.out.tfevents.1629941989.8e89bd551565.924.11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cfe52ce5123f221d5430ba4bcc21c0098e5e3ec4b6a2fca4e799fddda0c3fb8
3
+ size 4194
model-bin/finetune/base/log/1629942431.5525227/events.out.tfevents.1629942431.8e89bd551565.924.13 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec92524f1117d47d6597750660ed02d2202182c1c8a7cecc2507723ce6f435f9
3
+ size 4194
model-bin/finetune/base/log/1629942858.3612747/events.out.tfevents.1629942858.8e89bd551565.924.15 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6e16c69a0b188bb907d6b765efb936fd01ad8b86c55782750896e7eb44b1f9e
3
+ size 4194
model-bin/finetune/base/log/1629943284.2321923/events.out.tfevents.1629943284.8e89bd551565.924.17 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df7d94a62b0805c1896bb2704d09db0ab4a2e6177763c3befc9f0f59d1494743
3
+ size 4194
model-bin/finetune/base/log/1629943702.5924923/events.out.tfevents.1629943702.8e89bd551565.924.19 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2f80f84df64b7af186cd8a8c7a1d7c4baa7bb66c57c8305efcb236d680881c4
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629941989.8e89bd551565.924.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5da17eb9052a1bd3e736635e8a2e22936429b3faedc65f6050d33e34ec409525
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629942431.8e89bd551565.924.12 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d68afe14b9644be97f1bc2680ae0e318ef494c029bfa45de4bfd04d2fdcb55af
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629942858.8e89bd551565.924.14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a95517ca0d559b2325de22f8b61ff900de7468f44bb80d0aa6f95a0c64f18bf
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629943284.8e89bd551565.924.16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2de3335406e8b08ee93085cff73d46c08ea4ff85a8a6d90a7a306600ff6f7810
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1629943702.8e89bd551565.924.18 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:774bc1c6320128780f1d19b5ee725bb98c22e1270909a68ae4558c5a3839af2e
3
+ size 8622