elsayedissa commited on
Commit
f8064d0
1 Parent(s): e5e85ab

Training in progress, step 11000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:666bd295da5110fc8ed72c097e043b17b68a4f46ec1f780894808ac7d810702a
3
  size 12347192855
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b15938f7058276e6bdfc2fddcc7a120da5e6a0c7c9c19c71e16cb873fc6886b3
3
  size 12347192855
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:492b321ad12ae8744ac23f006fa941f8bd7881d2b5e9aecb269dbdaf98b5ee88
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bdd501f6153777369e4dc961faa1bf60f1f7535ded5d73c87f2ff06b5aa8064
3
  size 6173655480
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aad3fc937f2c6f56a2a6ddd2d655b320e20d5cf3c0672d854ed64624351b0488
3
- size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:069bc79c0b97b8bebabb6777e867a3617901ff0e79b431f2ba4a791a90815e7f
3
+ size 14639
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ff19fa2317539f8bfb616fcdefc3caa2efd60d151a696a55bb38626512e9ba9
3
  size 557
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b4b38bb567a4c3bd8aafad6c492b89fde494be62ffddc238da23107b19a9945
3
  size 557
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1196c99a263cfa6a64b1d73b220e535cc282c309df0a28be120a2a37266e02ba
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e9a74567933ed8c342d4d89869156924da4e6db37efcc7105e4c34e754e42a4
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
- "epoch": 0.34711374917560484,
5
- "global_step": 10000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -2496,11 +2496,260 @@
2496
  "eval_steps_per_second": 0.062,
2497
  "eval_wer": 0.10322861869805133,
2498
  "step": 10000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2499
  }
2500
  ],
2501
  "max_steps": 25000,
2502
  "num_train_epochs": 1,
2503
- "total_flos": 3.39707363328e+20,
2504
  "trial_name": null,
2505
  "trial_params": null
2506
  }
 
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
+ "epoch": 0.3818251240931653,
5
+ "global_step": 11000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
2496
  "eval_steps_per_second": 0.062,
2497
  "eval_wer": 0.10322861869805133,
2498
  "step": 10000
2499
+ },
2500
+ {
2501
+ "epoch": 0.35,
2502
+ "learning_rate": 6.115510204081633e-06,
2503
+ "loss": 0.1827,
2504
+ "step": 10025
2505
+ },
2506
+ {
2507
+ "epoch": 0.35,
2508
+ "learning_rate": 6.10530612244898e-06,
2509
+ "loss": 0.1653,
2510
+ "step": 10050
2511
+ },
2512
+ {
2513
+ "epoch": 0.35,
2514
+ "learning_rate": 6.095102040816327e-06,
2515
+ "loss": 0.1726,
2516
+ "step": 10075
2517
+ },
2518
+ {
2519
+ "epoch": 0.35,
2520
+ "learning_rate": 6.084897959183675e-06,
2521
+ "loss": 0.1569,
2522
+ "step": 10100
2523
+ },
2524
+ {
2525
+ "epoch": 0.35,
2526
+ "learning_rate": 6.074693877551021e-06,
2527
+ "loss": 0.1796,
2528
+ "step": 10125
2529
+ },
2530
+ {
2531
+ "epoch": 0.35,
2532
+ "learning_rate": 6.064489795918368e-06,
2533
+ "loss": 0.1457,
2534
+ "step": 10150
2535
+ },
2536
+ {
2537
+ "epoch": 0.35,
2538
+ "learning_rate": 6.054285714285715e-06,
2539
+ "loss": 0.1797,
2540
+ "step": 10175
2541
+ },
2542
+ {
2543
+ "epoch": 0.35,
2544
+ "learning_rate": 6.044081632653061e-06,
2545
+ "loss": 0.1596,
2546
+ "step": 10200
2547
+ },
2548
+ {
2549
+ "epoch": 0.35,
2550
+ "learning_rate": 6.033877551020409e-06,
2551
+ "loss": 0.1855,
2552
+ "step": 10225
2553
+ },
2554
+ {
2555
+ "epoch": 0.36,
2556
+ "learning_rate": 6.023673469387755e-06,
2557
+ "loss": 0.1495,
2558
+ "step": 10250
2559
+ },
2560
+ {
2561
+ "epoch": 0.36,
2562
+ "learning_rate": 6.013469387755102e-06,
2563
+ "loss": 0.1741,
2564
+ "step": 10275
2565
+ },
2566
+ {
2567
+ "epoch": 0.36,
2568
+ "learning_rate": 6.0032653061224495e-06,
2569
+ "loss": 0.1859,
2570
+ "step": 10300
2571
+ },
2572
+ {
2573
+ "epoch": 0.36,
2574
+ "learning_rate": 5.993061224489797e-06,
2575
+ "loss": 0.1874,
2576
+ "step": 10325
2577
+ },
2578
+ {
2579
+ "epoch": 0.36,
2580
+ "learning_rate": 5.982857142857144e-06,
2581
+ "loss": 0.159,
2582
+ "step": 10350
2583
+ },
2584
+ {
2585
+ "epoch": 0.36,
2586
+ "learning_rate": 5.97265306122449e-06,
2587
+ "loss": 0.1653,
2588
+ "step": 10375
2589
+ },
2590
+ {
2591
+ "epoch": 0.36,
2592
+ "learning_rate": 5.962448979591837e-06,
2593
+ "loss": 0.1417,
2594
+ "step": 10400
2595
+ },
2596
+ {
2597
+ "epoch": 0.36,
2598
+ "learning_rate": 5.952244897959184e-06,
2599
+ "loss": 0.1819,
2600
+ "step": 10425
2601
+ },
2602
+ {
2603
+ "epoch": 0.36,
2604
+ "learning_rate": 5.942040816326531e-06,
2605
+ "loss": 0.1605,
2606
+ "step": 10450
2607
+ },
2608
+ {
2609
+ "epoch": 0.36,
2610
+ "learning_rate": 5.931836734693878e-06,
2611
+ "loss": 0.1886,
2612
+ "step": 10475
2613
+ },
2614
+ {
2615
+ "epoch": 0.36,
2616
+ "learning_rate": 5.921632653061224e-06,
2617
+ "loss": 0.1667,
2618
+ "step": 10500
2619
+ },
2620
+ {
2621
+ "epoch": 0.37,
2622
+ "learning_rate": 5.911428571428573e-06,
2623
+ "loss": 0.1765,
2624
+ "step": 10525
2625
+ },
2626
+ {
2627
+ "epoch": 0.37,
2628
+ "learning_rate": 5.901224489795919e-06,
2629
+ "loss": 0.1484,
2630
+ "step": 10550
2631
+ },
2632
+ {
2633
+ "epoch": 0.37,
2634
+ "learning_rate": 5.891020408163266e-06,
2635
+ "loss": 0.1748,
2636
+ "step": 10575
2637
+ },
2638
+ {
2639
+ "epoch": 0.37,
2640
+ "learning_rate": 5.880816326530613e-06,
2641
+ "loss": 0.1514,
2642
+ "step": 10600
2643
+ },
2644
+ {
2645
+ "epoch": 0.37,
2646
+ "learning_rate": 5.87061224489796e-06,
2647
+ "loss": 0.191,
2648
+ "step": 10625
2649
+ },
2650
+ {
2651
+ "epoch": 0.37,
2652
+ "learning_rate": 5.860408163265307e-06,
2653
+ "loss": 0.154,
2654
+ "step": 10650
2655
+ },
2656
+ {
2657
+ "epoch": 0.37,
2658
+ "learning_rate": 5.850204081632653e-06,
2659
+ "loss": 0.1758,
2660
+ "step": 10675
2661
+ },
2662
+ {
2663
+ "epoch": 0.37,
2664
+ "learning_rate": 5.84e-06,
2665
+ "loss": 0.1675,
2666
+ "step": 10700
2667
+ },
2668
+ {
2669
+ "epoch": 0.37,
2670
+ "learning_rate": 5.829795918367347e-06,
2671
+ "loss": 0.1946,
2672
+ "step": 10725
2673
+ },
2674
+ {
2675
+ "epoch": 0.37,
2676
+ "learning_rate": 5.819591836734695e-06,
2677
+ "loss": 0.1563,
2678
+ "step": 10750
2679
+ },
2680
+ {
2681
+ "epoch": 0.37,
2682
+ "learning_rate": 5.809387755102042e-06,
2683
+ "loss": 0.177,
2684
+ "step": 10775
2685
+ },
2686
+ {
2687
+ "epoch": 0.37,
2688
+ "learning_rate": 5.799183673469388e-06,
2689
+ "loss": 0.1467,
2690
+ "step": 10800
2691
+ },
2692
+ {
2693
+ "epoch": 0.38,
2694
+ "learning_rate": 5.788979591836735e-06,
2695
+ "loss": 0.1677,
2696
+ "step": 10825
2697
+ },
2698
+ {
2699
+ "epoch": 0.38,
2700
+ "learning_rate": 5.778775510204082e-06,
2701
+ "loss": 0.1638,
2702
+ "step": 10850
2703
+ },
2704
+ {
2705
+ "epoch": 0.38,
2706
+ "learning_rate": 5.768571428571429e-06,
2707
+ "loss": 0.2078,
2708
+ "step": 10875
2709
+ },
2710
+ {
2711
+ "epoch": 0.38,
2712
+ "learning_rate": 5.758367346938776e-06,
2713
+ "loss": 0.1474,
2714
+ "step": 10900
2715
+ },
2716
+ {
2717
+ "epoch": 0.38,
2718
+ "learning_rate": 5.748163265306122e-06,
2719
+ "loss": 0.1758,
2720
+ "step": 10925
2721
+ },
2722
+ {
2723
+ "epoch": 0.38,
2724
+ "learning_rate": 5.73795918367347e-06,
2725
+ "loss": 0.1453,
2726
+ "step": 10950
2727
+ },
2728
+ {
2729
+ "epoch": 0.38,
2730
+ "learning_rate": 5.727755102040817e-06,
2731
+ "loss": 0.1806,
2732
+ "step": 10975
2733
+ },
2734
+ {
2735
+ "epoch": 0.38,
2736
+ "learning_rate": 5.717551020408164e-06,
2737
+ "loss": 0.1457,
2738
+ "step": 11000
2739
+ },
2740
+ {
2741
+ "epoch": 0.38,
2742
+ "eval_loss": 0.17772968113422394,
2743
+ "eval_runtime": 31110.5279,
2744
+ "eval_samples_per_second": 0.998,
2745
+ "eval_steps_per_second": 0.062,
2746
+ "eval_wer": 0.09984380314573199,
2747
+ "step": 11000
2748
  }
2749
  ],
2750
  "max_steps": 25000,
2751
  "num_train_epochs": 1,
2752
+ "total_flos": 3.736780996608e+20,
2753
  "trial_name": null,
2754
  "trial_params": null
2755
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:492b321ad12ae8744ac23f006fa941f8bd7881d2b5e9aecb269dbdaf98b5ee88
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bdd501f6153777369e4dc961faa1bf60f1f7535ded5d73c87f2ff06b5aa8064
3
  size 6173655480
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3649b56169acd2abda9b0ee07a52cc5ff3d90c7e405842df1439d1a6f490683e
3
- size 70369
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c954f05ee34517b3eb4c4f0da7bbc163f490b1b0ac3836bb3e4f0baf08b45f5
3
+ size 76967