Check commited on
Commit
ade1408
Β·
1 Parent(s): 3bac7d3

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629858970.8441613/events.out.tfevents.1629858970.7e498afd5545.905.23 +3 -0
  11. model-bin/finetune/base/log/1629859624.9463222/events.out.tfevents.1629859624.7e498afd5545.905.25 +3 -0
  12. model-bin/finetune/base/log/1629860278.0745044/events.out.tfevents.1629860278.7e498afd5545.905.27 +3 -0
  13. model-bin/finetune/base/log/1629860932.5132222/events.out.tfevents.1629860932.7e498afd5545.905.29 +3 -0
  14. model-bin/finetune/base/log/1629861583.8800406/events.out.tfevents.1629861583.7e498afd5545.905.31 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629858970.7e498afd5545.905.22 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629859624.7e498afd5545.905.24 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629860278.7e498afd5545.905.26 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629860932.7e498afd5545.905.28 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629861583.7e498afd5545.905.30 +3 -0
model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1238d3e0407c1b5eab603ddadea5d697c591b6ed7191dded1195fc755b757e8e
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c2cf74949ac6fbe57789fc93a1ba947fc6fea6315730c11480f522422c2cb32
3
  size 722165393
model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:254d4cf0f3862cae7773484aa9cfabaf779e738f449d472541eb9a091e3fb36a
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b94fe687400b28878fb9b1535714b9e18e88910b1486e085968c125fe165b608
3
  size 377909911
model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c9688cedf4af3de5be5cd9e8d216e7a95ddb6ceceb3ddf456a28ff4bf1c92c9
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a44b4a00afd33832285479c87b3d0330ccd6122276d09600764d244ca2e0d676
3
  size 14503
model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:70f0230abe0d9e7f4ba777c6d4ca204523033a231565d6e066189da086fa610e
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa5ffd5ce4a042224436871219db3b89c25b62d29b4a4605aafd1e3b9552f733
3
  size 559
model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e207074215dc1be4bcfdc4da83cf51e498c21731dc990103b72674b1b112cc7a
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34abd4dd1bca145f62df90529f947d27c07749481bac26350eaca06e62acd0fe
3
  size 623
model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.18412114350410416,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
4
- "epoch": 570.995983935743,
5
- "global_step": 70809,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -189453,11 +189453,800 @@
189453
  "eval_steps_per_second": 0.673,
189454
  "eval_wer": 0.18833597008485545,
189455
  "step": 70809
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
189456
  }
189457
  ],
189458
  "max_steps": 620000,
189459
  "num_train_epochs": 5000,
189460
- "total_flos": 1.992660946951832e+20,
189461
  "trial_name": null,
189462
  "trial_params": null
189463
  }
 
1
  {
2
  "best_metric": 0.18412114350410416,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
4
+ "epoch": 575.995983935743,
5
+ "global_step": 71429,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
189453
  "eval_steps_per_second": 0.673,
189454
  "eval_wer": 0.18833597008485545,
189455
  "step": 70809
189456
+ },
189457
+ {
189458
+ "epoch": 571.01,
189459
+ "learning_rate": 8.881666666666667e-06,
189460
+ "loss": 0.3513,
189461
+ "step": 70810
189462
+ },
189463
+ {
189464
+ "epoch": 571.05,
189465
+ "learning_rate": 8.88158653846154e-06,
189466
+ "loss": 0.2739,
189467
+ "step": 70815
189468
+ },
189469
+ {
189470
+ "epoch": 571.09,
189471
+ "learning_rate": 8.881506410256411e-06,
189472
+ "loss": 0.3412,
189473
+ "step": 70820
189474
+ },
189475
+ {
189476
+ "epoch": 571.13,
189477
+ "learning_rate": 8.881426282051282e-06,
189478
+ "loss": 0.3873,
189479
+ "step": 70825
189480
+ },
189481
+ {
189482
+ "epoch": 571.17,
189483
+ "learning_rate": 8.881346153846154e-06,
189484
+ "loss": 0.5661,
189485
+ "step": 70830
189486
+ },
189487
+ {
189488
+ "epoch": 571.21,
189489
+ "learning_rate": 8.881266025641027e-06,
189490
+ "loss": 1.1271,
189491
+ "step": 70835
189492
+ },
189493
+ {
189494
+ "epoch": 571.25,
189495
+ "learning_rate": 8.881185897435898e-06,
189496
+ "loss": 0.3545,
189497
+ "step": 70840
189498
+ },
189499
+ {
189500
+ "epoch": 571.29,
189501
+ "learning_rate": 8.88110576923077e-06,
189502
+ "loss": 0.2882,
189503
+ "step": 70845
189504
+ },
189505
+ {
189506
+ "epoch": 571.33,
189507
+ "learning_rate": 8.881025641025641e-06,
189508
+ "loss": 0.352,
189509
+ "step": 70850
189510
+ },
189511
+ {
189512
+ "epoch": 571.37,
189513
+ "learning_rate": 8.880945512820514e-06,
189514
+ "loss": 0.5543,
189515
+ "step": 70855
189516
+ },
189517
+ {
189518
+ "epoch": 571.41,
189519
+ "learning_rate": 8.880865384615385e-06,
189520
+ "loss": 1.203,
189521
+ "step": 70860
189522
+ },
189523
+ {
189524
+ "epoch": 571.45,
189525
+ "learning_rate": 8.880785256410257e-06,
189526
+ "loss": 0.3559,
189527
+ "step": 70865
189528
+ },
189529
+ {
189530
+ "epoch": 571.49,
189531
+ "learning_rate": 8.88070512820513e-06,
189532
+ "loss": 0.3581,
189533
+ "step": 70870
189534
+ },
189535
+ {
189536
+ "epoch": 571.53,
189537
+ "learning_rate": 8.880625000000001e-06,
189538
+ "loss": 0.3475,
189539
+ "step": 70875
189540
+ },
189541
+ {
189542
+ "epoch": 571.57,
189543
+ "learning_rate": 8.880544871794872e-06,
189544
+ "loss": 0.6108,
189545
+ "step": 70880
189546
+ },
189547
+ {
189548
+ "epoch": 571.61,
189549
+ "learning_rate": 8.880464743589744e-06,
189550
+ "loss": 1.0645,
189551
+ "step": 70885
189552
+ },
189553
+ {
189554
+ "epoch": 571.65,
189555
+ "learning_rate": 8.880384615384617e-06,
189556
+ "loss": 0.4175,
189557
+ "step": 70890
189558
+ },
189559
+ {
189560
+ "epoch": 571.69,
189561
+ "learning_rate": 8.880304487179488e-06,
189562
+ "loss": 0.3068,
189563
+ "step": 70895
189564
+ },
189565
+ {
189566
+ "epoch": 571.73,
189567
+ "learning_rate": 8.88022435897436e-06,
189568
+ "loss": 0.3434,
189569
+ "step": 70900
189570
+ },
189571
+ {
189572
+ "epoch": 571.77,
189573
+ "learning_rate": 8.880144230769233e-06,
189574
+ "loss": 0.5246,
189575
+ "step": 70905
189576
+ },
189577
+ {
189578
+ "epoch": 571.81,
189579
+ "learning_rate": 8.880064102564104e-06,
189580
+ "loss": 1.2047,
189581
+ "step": 70910
189582
+ },
189583
+ {
189584
+ "epoch": 571.85,
189585
+ "learning_rate": 8.879983974358975e-06,
189586
+ "loss": 0.3039,
189587
+ "step": 70915
189588
+ },
189589
+ {
189590
+ "epoch": 571.9,
189591
+ "learning_rate": 8.879903846153847e-06,
189592
+ "loss": 0.3366,
189593
+ "step": 70920
189594
+ },
189595
+ {
189596
+ "epoch": 571.94,
189597
+ "learning_rate": 8.87982371794872e-06,
189598
+ "loss": 0.412,
189599
+ "step": 70925
189600
+ },
189601
+ {
189602
+ "epoch": 571.98,
189603
+ "learning_rate": 8.87974358974359e-06,
189604
+ "loss": 0.6262,
189605
+ "step": 70930
189606
+ },
189607
+ {
189608
+ "epoch": 572.0,
189609
+ "eval_loss": 0.37808266282081604,
189610
+ "eval_runtime": 39.9883,
189611
+ "eval_samples_per_second": 21.006,
189612
+ "eval_steps_per_second": 0.675,
189613
+ "eval_wer": 0.1896881435716387,
189614
+ "step": 70933
189615
+ },
189616
+ {
189617
+ "epoch": 572.02,
189618
+ "learning_rate": 8.879663461538462e-06,
189619
+ "loss": 0.4067,
189620
+ "step": 70935
189621
+ },
189622
+ {
189623
+ "epoch": 572.06,
189624
+ "learning_rate": 8.879583333333334e-06,
189625
+ "loss": 0.3122,
189626
+ "step": 70940
189627
+ },
189628
+ {
189629
+ "epoch": 572.1,
189630
+ "learning_rate": 8.879503205128205e-06,
189631
+ "loss": 0.3144,
189632
+ "step": 70945
189633
+ },
189634
+ {
189635
+ "epoch": 572.14,
189636
+ "learning_rate": 8.879423076923077e-06,
189637
+ "loss": 0.3966,
189638
+ "step": 70950
189639
+ },
189640
+ {
189641
+ "epoch": 572.18,
189642
+ "learning_rate": 8.87934294871795e-06,
189643
+ "loss": 0.7134,
189644
+ "step": 70955
189645
+ },
189646
+ {
189647
+ "epoch": 572.22,
189648
+ "learning_rate": 8.879262820512821e-06,
189649
+ "loss": 1.0855,
189650
+ "step": 70960
189651
+ },
189652
+ {
189653
+ "epoch": 572.26,
189654
+ "learning_rate": 8.879182692307692e-06,
189655
+ "loss": 0.2845,
189656
+ "step": 70965
189657
+ },
189658
+ {
189659
+ "epoch": 572.3,
189660
+ "learning_rate": 8.879102564102565e-06,
189661
+ "loss": 0.3495,
189662
+ "step": 70970
189663
+ },
189664
+ {
189665
+ "epoch": 572.34,
189666
+ "learning_rate": 8.879022435897437e-06,
189667
+ "loss": 0.325,
189668
+ "step": 70975
189669
+ },
189670
+ {
189671
+ "epoch": 572.38,
189672
+ "learning_rate": 8.878942307692308e-06,
189673
+ "loss": 0.6465,
189674
+ "step": 70980
189675
+ },
189676
+ {
189677
+ "epoch": 572.42,
189678
+ "learning_rate": 8.87886217948718e-06,
189679
+ "loss": 1.0453,
189680
+ "step": 70985
189681
+ },
189682
+ {
189683
+ "epoch": 572.46,
189684
+ "learning_rate": 8.878782051282052e-06,
189685
+ "loss": 0.3642,
189686
+ "step": 70990
189687
+ },
189688
+ {
189689
+ "epoch": 572.5,
189690
+ "learning_rate": 8.878701923076924e-06,
189691
+ "loss": 0.3005,
189692
+ "step": 70995
189693
+ },
189694
+ {
189695
+ "epoch": 572.54,
189696
+ "learning_rate": 8.878621794871795e-06,
189697
+ "loss": 0.3992,
189698
+ "step": 71000
189699
+ },
189700
+ {
189701
+ "epoch": 572.58,
189702
+ "learning_rate": 8.878541666666668e-06,
189703
+ "loss": 0.6974,
189704
+ "step": 71005
189705
+ },
189706
+ {
189707
+ "epoch": 572.62,
189708
+ "learning_rate": 8.87846153846154e-06,
189709
+ "loss": 0.9939,
189710
+ "step": 71010
189711
+ },
189712
+ {
189713
+ "epoch": 572.66,
189714
+ "learning_rate": 8.878381410256411e-06,
189715
+ "loss": 0.3028,
189716
+ "step": 71015
189717
+ },
189718
+ {
189719
+ "epoch": 572.7,
189720
+ "learning_rate": 8.878301282051282e-06,
189721
+ "loss": 0.3952,
189722
+ "step": 71020
189723
+ },
189724
+ {
189725
+ "epoch": 572.74,
189726
+ "learning_rate": 8.878221153846155e-06,
189727
+ "loss": 0.3883,
189728
+ "step": 71025
189729
+ },
189730
+ {
189731
+ "epoch": 572.78,
189732
+ "learning_rate": 8.878141025641027e-06,
189733
+ "loss": 0.6341,
189734
+ "step": 71030
189735
+ },
189736
+ {
189737
+ "epoch": 572.82,
189738
+ "learning_rate": 8.878060897435898e-06,
189739
+ "loss": 1.0706,
189740
+ "step": 71035
189741
+ },
189742
+ {
189743
+ "epoch": 572.86,
189744
+ "learning_rate": 8.87798076923077e-06,
189745
+ "loss": 0.2861,
189746
+ "step": 71040
189747
+ },
189748
+ {
189749
+ "epoch": 572.9,
189750
+ "learning_rate": 8.877900641025642e-06,
189751
+ "loss": 0.4219,
189752
+ "step": 71045
189753
+ },
189754
+ {
189755
+ "epoch": 572.94,
189756
+ "learning_rate": 8.877820512820514e-06,
189757
+ "loss": 0.3828,
189758
+ "step": 71050
189759
+ },
189760
+ {
189761
+ "epoch": 572.98,
189762
+ "learning_rate": 8.877740384615385e-06,
189763
+ "loss": 0.6617,
189764
+ "step": 71055
189765
+ },
189766
+ {
189767
+ "epoch": 573.0,
189768
+ "eval_loss": 0.37939906120300293,
189769
+ "eval_runtime": 39.3467,
189770
+ "eval_samples_per_second": 21.349,
189771
+ "eval_steps_per_second": 0.686,
189772
+ "eval_wer": 0.19251534649804009,
189773
+ "step": 71057
189774
+ },
189775
+ {
189776
+ "epoch": 573.02,
189777
+ "learning_rate": 8.877660256410258e-06,
189778
+ "loss": 0.3238,
189779
+ "step": 71060
189780
+ },
189781
+ {
189782
+ "epoch": 573.06,
189783
+ "learning_rate": 8.87758012820513e-06,
189784
+ "loss": 0.3552,
189785
+ "step": 71065
189786
+ },
189787
+ {
189788
+ "epoch": 573.1,
189789
+ "learning_rate": 8.877500000000001e-06,
189790
+ "loss": 0.3396,
189791
+ "step": 71070
189792
+ },
189793
+ {
189794
+ "epoch": 573.14,
189795
+ "learning_rate": 8.877419871794872e-06,
189796
+ "loss": 0.3658,
189797
+ "step": 71075
189798
+ },
189799
+ {
189800
+ "epoch": 573.18,
189801
+ "learning_rate": 8.877339743589745e-06,
189802
+ "loss": 0.8553,
189803
+ "step": 71080
189804
+ },
189805
+ {
189806
+ "epoch": 573.22,
189807
+ "learning_rate": 8.877259615384615e-06,
189808
+ "loss": 0.8966,
189809
+ "step": 71085
189810
+ },
189811
+ {
189812
+ "epoch": 573.27,
189813
+ "learning_rate": 8.877179487179488e-06,
189814
+ "loss": 0.4122,
189815
+ "step": 71090
189816
+ },
189817
+ {
189818
+ "epoch": 573.31,
189819
+ "learning_rate": 8.87709935897436e-06,
189820
+ "loss": 0.3168,
189821
+ "step": 71095
189822
+ },
189823
+ {
189824
+ "epoch": 573.35,
189825
+ "learning_rate": 8.87701923076923e-06,
189826
+ "loss": 0.3781,
189827
+ "step": 71100
189828
+ },
189829
+ {
189830
+ "epoch": 573.39,
189831
+ "learning_rate": 8.876939102564102e-06,
189832
+ "loss": 0.7079,
189833
+ "step": 71105
189834
+ },
189835
+ {
189836
+ "epoch": 573.43,
189837
+ "learning_rate": 8.876858974358975e-06,
189838
+ "loss": 0.9013,
189839
+ "step": 71110
189840
+ },
189841
+ {
189842
+ "epoch": 573.47,
189843
+ "learning_rate": 8.876778846153847e-06,
189844
+ "loss": 0.3044,
189845
+ "step": 71115
189846
+ },
189847
+ {
189848
+ "epoch": 573.51,
189849
+ "learning_rate": 8.876698717948718e-06,
189850
+ "loss": 0.3062,
189851
+ "step": 71120
189852
+ },
189853
+ {
189854
+ "epoch": 573.55,
189855
+ "learning_rate": 8.876618589743591e-06,
189856
+ "loss": 0.4105,
189857
+ "step": 71125
189858
+ },
189859
+ {
189860
+ "epoch": 573.59,
189861
+ "learning_rate": 8.876538461538462e-06,
189862
+ "loss": 0.8214,
189863
+ "step": 71130
189864
+ },
189865
+ {
189866
+ "epoch": 573.63,
189867
+ "learning_rate": 8.876458333333334e-06,
189868
+ "loss": 0.739,
189869
+ "step": 71135
189870
+ },
189871
+ {
189872
+ "epoch": 573.67,
189873
+ "learning_rate": 8.876378205128205e-06,
189874
+ "loss": 0.2696,
189875
+ "step": 71140
189876
+ },
189877
+ {
189878
+ "epoch": 573.71,
189879
+ "learning_rate": 8.876298076923078e-06,
189880
+ "loss": 0.2784,
189881
+ "step": 71145
189882
+ },
189883
+ {
189884
+ "epoch": 573.75,
189885
+ "learning_rate": 8.87621794871795e-06,
189886
+ "loss": 0.3571,
189887
+ "step": 71150
189888
+ },
189889
+ {
189890
+ "epoch": 573.79,
189891
+ "learning_rate": 8.876137820512821e-06,
189892
+ "loss": 0.9113,
189893
+ "step": 71155
189894
+ },
189895
+ {
189896
+ "epoch": 573.83,
189897
+ "learning_rate": 8.876057692307694e-06,
189898
+ "loss": 0.9159,
189899
+ "step": 71160
189900
+ },
189901
+ {
189902
+ "epoch": 573.87,
189903
+ "learning_rate": 8.875977564102565e-06,
189904
+ "loss": 0.2806,
189905
+ "step": 71165
189906
+ },
189907
+ {
189908
+ "epoch": 573.91,
189909
+ "learning_rate": 8.875897435897437e-06,
189910
+ "loss": 0.3812,
189911
+ "step": 71170
189912
+ },
189913
+ {
189914
+ "epoch": 573.95,
189915
+ "learning_rate": 8.875817307692308e-06,
189916
+ "loss": 0.4332,
189917
+ "step": 71175
189918
+ },
189919
+ {
189920
+ "epoch": 573.99,
189921
+ "learning_rate": 8.875737179487181e-06,
189922
+ "loss": 0.8548,
189923
+ "step": 71180
189924
+ },
189925
+ {
189926
+ "epoch": 574.0,
189927
+ "eval_loss": 0.3877381682395935,
189928
+ "eval_runtime": 39.9108,
189929
+ "eval_samples_per_second": 21.047,
189930
+ "eval_steps_per_second": 0.677,
189931
+ "eval_wer": 0.19365056511410267,
189932
+ "step": 71181
189933
+ },
189934
+ {
189935
+ "epoch": 574.03,
189936
+ "learning_rate": 8.875657051282052e-06,
189937
+ "loss": 0.3677,
189938
+ "step": 71185
189939
+ },
189940
+ {
189941
+ "epoch": 574.07,
189942
+ "learning_rate": 8.875576923076924e-06,
189943
+ "loss": 0.3396,
189944
+ "step": 71190
189945
+ },
189946
+ {
189947
+ "epoch": 574.11,
189948
+ "learning_rate": 8.875496794871795e-06,
189949
+ "loss": 0.3444,
189950
+ "step": 71195
189951
+ },
189952
+ {
189953
+ "epoch": 574.15,
189954
+ "learning_rate": 8.875416666666668e-06,
189955
+ "loss": 0.3975,
189956
+ "step": 71200
189957
+ },
189958
+ {
189959
+ "epoch": 574.19,
189960
+ "learning_rate": 8.875336538461538e-06,
189961
+ "loss": 0.9141,
189962
+ "step": 71205
189963
+ },
189964
+ {
189965
+ "epoch": 574.23,
189966
+ "learning_rate": 8.875256410256411e-06,
189967
+ "loss": 0.7714,
189968
+ "step": 71210
189969
+ },
189970
+ {
189971
+ "epoch": 574.27,
189972
+ "learning_rate": 8.875176282051284e-06,
189973
+ "loss": 0.3316,
189974
+ "step": 71215
189975
+ },
189976
+ {
189977
+ "epoch": 574.31,
189978
+ "learning_rate": 8.875096153846154e-06,
189979
+ "loss": 0.3371,
189980
+ "step": 71220
189981
+ },
189982
+ {
189983
+ "epoch": 574.35,
189984
+ "learning_rate": 8.875016025641027e-06,
189985
+ "loss": 0.4133,
189986
+ "step": 71225
189987
+ },
189988
+ {
189989
+ "epoch": 574.39,
189990
+ "learning_rate": 8.874935897435898e-06,
189991
+ "loss": 1.0848,
189992
+ "step": 71230
189993
+ },
189994
+ {
189995
+ "epoch": 574.43,
189996
+ "learning_rate": 8.87485576923077e-06,
189997
+ "loss": 0.6723,
189998
+ "step": 71235
189999
+ },
190000
+ {
190001
+ "epoch": 574.47,
190002
+ "learning_rate": 8.87477564102564e-06,
190003
+ "loss": 0.2706,
190004
+ "step": 71240
190005
+ },
190006
+ {
190007
+ "epoch": 574.51,
190008
+ "learning_rate": 8.874695512820514e-06,
190009
+ "loss": 0.3278,
190010
+ "step": 71245
190011
+ },
190012
+ {
190013
+ "epoch": 574.55,
190014
+ "learning_rate": 8.874615384615385e-06,
190015
+ "loss": 0.3897,
190016
+ "step": 71250
190017
+ },
190018
+ {
190019
+ "epoch": 574.59,
190020
+ "learning_rate": 8.874535256410256e-06,
190021
+ "loss": 1.1266,
190022
+ "step": 71255
190023
+ },
190024
+ {
190025
+ "epoch": 574.63,
190026
+ "learning_rate": 8.87445512820513e-06,
190027
+ "loss": 0.8757,
190028
+ "step": 71260
190029
+ },
190030
+ {
190031
+ "epoch": 574.67,
190032
+ "learning_rate": 8.874375000000001e-06,
190033
+ "loss": 0.3173,
190034
+ "step": 71265
190035
+ },
190036
+ {
190037
+ "epoch": 574.71,
190038
+ "learning_rate": 8.874294871794872e-06,
190039
+ "loss": 0.2944,
190040
+ "step": 71270
190041
+ },
190042
+ {
190043
+ "epoch": 574.76,
190044
+ "learning_rate": 8.874214743589744e-06,
190045
+ "loss": 0.4019,
190046
+ "step": 71275
190047
+ },
190048
+ {
190049
+ "epoch": 574.8,
190050
+ "learning_rate": 8.874134615384617e-06,
190051
+ "loss": 1.0625,
190052
+ "step": 71280
190053
+ },
190054
+ {
190055
+ "epoch": 574.84,
190056
+ "learning_rate": 8.874054487179488e-06,
190057
+ "loss": 0.7614,
190058
+ "step": 71285
190059
+ },
190060
+ {
190061
+ "epoch": 574.88,
190062
+ "learning_rate": 8.87397435897436e-06,
190063
+ "loss": 0.2756,
190064
+ "step": 71290
190065
+ },
190066
+ {
190067
+ "epoch": 574.92,
190068
+ "learning_rate": 8.87389423076923e-06,
190069
+ "loss": 0.3798,
190070
+ "step": 71295
190071
+ },
190072
+ {
190073
+ "epoch": 574.96,
190074
+ "learning_rate": 8.873814102564104e-06,
190075
+ "loss": 0.4301,
190076
+ "step": 71300
190077
+ },
190078
+ {
190079
+ "epoch": 575.0,
190080
+ "learning_rate": 8.873733974358975e-06,
190081
+ "loss": 1.2421,
190082
+ "step": 71305
190083
+ },
190084
+ {
190085
+ "epoch": 575.0,
190086
+ "eval_loss": 0.4396674633026123,
190087
+ "eval_runtime": 40.1793,
190088
+ "eval_samples_per_second": 20.906,
190089
+ "eval_steps_per_second": 0.672,
190090
+ "eval_wer": 0.19375594063025517,
190091
+ "step": 71305
190092
+ },
190093
+ {
190094
+ "epoch": 575.04,
190095
+ "learning_rate": 8.873653846153847e-06,
190096
+ "loss": 0.3501,
190097
+ "step": 71310
190098
+ },
190099
+ {
190100
+ "epoch": 575.08,
190101
+ "learning_rate": 8.87357371794872e-06,
190102
+ "loss": 0.3542,
190103
+ "step": 71315
190104
+ },
190105
+ {
190106
+ "epoch": 575.12,
190107
+ "learning_rate": 8.873493589743591e-06,
190108
+ "loss": 0.3445,
190109
+ "step": 71320
190110
+ },
190111
+ {
190112
+ "epoch": 575.16,
190113
+ "learning_rate": 8.873413461538462e-06,
190114
+ "loss": 0.5054,
190115
+ "step": 71325
190116
+ },
190117
+ {
190118
+ "epoch": 575.2,
190119
+ "learning_rate": 8.873333333333334e-06,
190120
+ "loss": 1.209,
190121
+ "step": 71330
190122
+ },
190123
+ {
190124
+ "epoch": 575.24,
190125
+ "learning_rate": 8.873253205128207e-06,
190126
+ "loss": 0.3995,
190127
+ "step": 71335
190128
+ },
190129
+ {
190130
+ "epoch": 575.28,
190131
+ "learning_rate": 8.873173076923078e-06,
190132
+ "loss": 0.3047,
190133
+ "step": 71340
190134
+ },
190135
+ {
190136
+ "epoch": 575.32,
190137
+ "learning_rate": 8.87309294871795e-06,
190138
+ "loss": 0.3208,
190139
+ "step": 71345
190140
+ },
190141
+ {
190142
+ "epoch": 575.36,
190143
+ "learning_rate": 8.87301282051282e-06,
190144
+ "loss": 0.4865,
190145
+ "step": 71350
190146
+ },
190147
+ {
190148
+ "epoch": 575.4,
190149
+ "learning_rate": 8.872932692307694e-06,
190150
+ "loss": 1.2983,
190151
+ "step": 71355
190152
+ },
190153
+ {
190154
+ "epoch": 575.44,
190155
+ "learning_rate": 8.872852564102565e-06,
190156
+ "loss": 0.3453,
190157
+ "step": 71360
190158
+ },
190159
+ {
190160
+ "epoch": 575.48,
190161
+ "learning_rate": 8.872772435897437e-06,
190162
+ "loss": 0.3389,
190163
+ "step": 71365
190164
+ },
190165
+ {
190166
+ "epoch": 575.52,
190167
+ "learning_rate": 8.87269230769231e-06,
190168
+ "loss": 0.3255,
190169
+ "step": 71370
190170
+ },
190171
+ {
190172
+ "epoch": 575.56,
190173
+ "learning_rate": 8.87261217948718e-06,
190174
+ "loss": 0.4804,
190175
+ "step": 71375
190176
+ },
190177
+ {
190178
+ "epoch": 575.6,
190179
+ "learning_rate": 8.872532051282052e-06,
190180
+ "loss": 1.1065,
190181
+ "step": 71380
190182
+ },
190183
+ {
190184
+ "epoch": 575.64,
190185
+ "learning_rate": 8.872451923076924e-06,
190186
+ "loss": 0.3769,
190187
+ "step": 71385
190188
+ },
190189
+ {
190190
+ "epoch": 575.68,
190191
+ "learning_rate": 8.872371794871795e-06,
190192
+ "loss": 0.2841,
190193
+ "step": 71390
190194
+ },
190195
+ {
190196
+ "epoch": 575.72,
190197
+ "learning_rate": 8.872291666666666e-06,
190198
+ "loss": 0.3687,
190199
+ "step": 71395
190200
+ },
190201
+ {
190202
+ "epoch": 575.76,
190203
+ "learning_rate": 8.87221153846154e-06,
190204
+ "loss": 0.4895,
190205
+ "step": 71400
190206
+ },
190207
+ {
190208
+ "epoch": 575.8,
190209
+ "learning_rate": 8.87213141025641e-06,
190210
+ "loss": 1.1831,
190211
+ "step": 71405
190212
+ },
190213
+ {
190214
+ "epoch": 575.84,
190215
+ "learning_rate": 8.872051282051282e-06,
190216
+ "loss": 0.3517,
190217
+ "step": 71410
190218
+ },
190219
+ {
190220
+ "epoch": 575.88,
190221
+ "learning_rate": 8.871971153846155e-06,
190222
+ "loss": 0.3023,
190223
+ "step": 71415
190224
+ },
190225
+ {
190226
+ "epoch": 575.92,
190227
+ "learning_rate": 8.871891025641027e-06,
190228
+ "loss": 0.3109,
190229
+ "step": 71420
190230
+ },
190231
+ {
190232
+ "epoch": 575.96,
190233
+ "learning_rate": 8.871810897435898e-06,
190234
+ "loss": 0.5157,
190235
+ "step": 71425
190236
+ },
190237
+ {
190238
+ "epoch": 576.0,
190239
+ "eval_loss": 0.370493620634079,
190240
+ "eval_runtime": 40.7235,
190241
+ "eval_samples_per_second": 20.627,
190242
+ "eval_steps_per_second": 0.663,
190243
+ "eval_wer": 0.18538826318909307,
190244
+ "step": 71429
190245
  }
190246
  ],
190247
  "max_steps": 620000,
190248
  "num_train_epochs": 5000,
190249
+ "total_flos": 2.0101646863500657e+20,
190250
  "trial_name": null,
190251
  "trial_params": null
190252
  }
model-bin/finetune/base/{checkpoint-70809 β†’ checkpoint-71429}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629858970.8441613/events.out.tfevents.1629858970.7e498afd5545.905.23 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdda49e59d41b85957a8cccc5577fcf0f644c5611ecd244a897f22720490ca05
3
+ size 4194
model-bin/finetune/base/log/1629859624.9463222/events.out.tfevents.1629859624.7e498afd5545.905.25 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0de7580033efd4f6523ea6ea4d985866eb98165f7ce6cef08b659fd4f77df42
3
+ size 4194
model-bin/finetune/base/log/1629860278.0745044/events.out.tfevents.1629860278.7e498afd5545.905.27 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f3764d5f0e85be9f251324fca32644721d02680b360ebd90ac23c0742e0a6f7
3
+ size 4194
model-bin/finetune/base/log/1629860932.5132222/events.out.tfevents.1629860932.7e498afd5545.905.29 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e0087eb4fb9ddf06503c1a7b2305e2eb97a4422095bef32e827069366286069
3
+ size 4194
model-bin/finetune/base/log/1629861583.8800406/events.out.tfevents.1629861583.7e498afd5545.905.31 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13c2e93be9000158ca59580c8a3592bd64fa86a07859832ca99e6402955c4981
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629858970.7e498afd5545.905.22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95341d235f3561e1cc248e9f54713ad5d5c383fc0af5c13f55763c13de9e32d3
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629859624.7e498afd5545.905.24 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c66a50c891088080220870c0899a1d489514ffd01056aad55cd7b982b2316b1
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629860278.7e498afd5545.905.26 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49a2cad8d2fc8657e15e9a3c1596b3a90b0411c1665554e0fd6049efa13743ec
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629860932.7e498afd5545.905.28 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e23d6fb0b7e0651fee49bbb66a50c47a62bbea9c783fb1b5f187de94c1bccca
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629861583.7e498afd5545.905.30 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:013cda62abdeb0bba2da058ab6dbf1e2e2fb143163b9e461ac2f41c2a39b07f1
3
+ size 8462