elsayedissa commited on
Commit
a40d74c
1 Parent(s): 39d68ac

Training in progress, step 7000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0460400a279e7766ba1a5c0ca78380731e0774cf4d291edfbc84654ee4a6aca
3
  size 12347192855
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6193d4dcef6d888921a1166b7db92e1165f3686f778873a319aa454e17ea622
3
  size 12347192855
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e0ff690a614ad4feb1584be1b22e3316419b9b45d45121399a24f7b7785e65c
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e91ae1bbae769e9ce17ed73669048b57cd296ecd5ff7d7b0071e74ca1c28a4f
3
  size 6173655480
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3788ae40196f57221f45ac4d182ab0f36e1ff61517502c4d51687db9b9b71e1f
3
- size 14511
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49ab9d2b8499c20583cc80db00868f9477a33e6dd8363c0d4ff3b9c6bb471b6b
3
+ size 14575
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da3defc4121d1071efa04d9a3c24a084fc57cd835b7f2d21ed3bc283ad291901
3
  size 557
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cafc80e01cebde114db040b18403e6c790dd6f4096e4412ec1d94153944f97a8
3
  size 557
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6c16fd3410fdd3660983e7e486d0b41995f7675ed13e0a940e3532e9d1a7db51
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d0b2313199b9b45d740a2bffd20986f78b2b72606b98ad7849fa54f34c2f5de
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
- "epoch": 0.2082682495053629,
5
- "global_step": 6000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -1500,11 +1500,260 @@
1500
  "eval_steps_per_second": 0.062,
1501
  "eval_wer": 0.10863441680453598,
1502
  "step": 6000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1503
  }
1504
  ],
1505
  "max_steps": 25000,
1506
  "num_train_epochs": 1,
1507
- "total_flos": 2.038244179968e+20,
1508
  "trial_name": null,
1509
  "trial_params": null
1510
  }
 
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
+ "epoch": 0.2429796244229234,
5
+ "global_step": 7000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
1500
  "eval_steps_per_second": 0.062,
1501
  "eval_wer": 0.10863441680453598,
1502
  "step": 6000
1503
+ },
1504
+ {
1505
+ "epoch": 0.21,
1506
+ "learning_rate": 7.747755102040817e-06,
1507
+ "loss": 0.1818,
1508
+ "step": 6025
1509
+ },
1510
+ {
1511
+ "epoch": 0.21,
1512
+ "learning_rate": 7.737551020408164e-06,
1513
+ "loss": 0.1595,
1514
+ "step": 6050
1515
+ },
1516
+ {
1517
+ "epoch": 0.21,
1518
+ "learning_rate": 7.727346938775512e-06,
1519
+ "loss": 0.199,
1520
+ "step": 6075
1521
+ },
1522
+ {
1523
+ "epoch": 0.21,
1524
+ "learning_rate": 7.717142857142857e-06,
1525
+ "loss": 0.1747,
1526
+ "step": 6100
1527
+ },
1528
+ {
1529
+ "epoch": 0.21,
1530
+ "learning_rate": 7.706938775510205e-06,
1531
+ "loss": 0.1888,
1532
+ "step": 6125
1533
+ },
1534
+ {
1535
+ "epoch": 0.21,
1536
+ "learning_rate": 7.69673469387755e-06,
1537
+ "loss": 0.1553,
1538
+ "step": 6150
1539
+ },
1540
+ {
1541
+ "epoch": 0.21,
1542
+ "learning_rate": 7.6865306122449e-06,
1543
+ "loss": 0.1924,
1544
+ "step": 6175
1545
+ },
1546
+ {
1547
+ "epoch": 0.22,
1548
+ "learning_rate": 7.676326530612246e-06,
1549
+ "loss": 0.1832,
1550
+ "step": 6200
1551
+ },
1552
+ {
1553
+ "epoch": 0.22,
1554
+ "learning_rate": 7.666122448979593e-06,
1555
+ "loss": 0.2015,
1556
+ "step": 6225
1557
+ },
1558
+ {
1559
+ "epoch": 0.22,
1560
+ "learning_rate": 7.655918367346939e-06,
1561
+ "loss": 0.1587,
1562
+ "step": 6250
1563
+ },
1564
+ {
1565
+ "epoch": 0.22,
1566
+ "learning_rate": 7.645714285714286e-06,
1567
+ "loss": 0.1847,
1568
+ "step": 6275
1569
+ },
1570
+ {
1571
+ "epoch": 0.22,
1572
+ "learning_rate": 7.635510204081634e-06,
1573
+ "loss": 0.1667,
1574
+ "step": 6300
1575
+ },
1576
+ {
1577
+ "epoch": 0.22,
1578
+ "learning_rate": 7.62530612244898e-06,
1579
+ "loss": 0.1848,
1580
+ "step": 6325
1581
+ },
1582
+ {
1583
+ "epoch": 0.22,
1584
+ "learning_rate": 7.615102040816327e-06,
1585
+ "loss": 0.1763,
1586
+ "step": 6350
1587
+ },
1588
+ {
1589
+ "epoch": 0.22,
1590
+ "learning_rate": 7.604897959183674e-06,
1591
+ "loss": 0.1918,
1592
+ "step": 6375
1593
+ },
1594
+ {
1595
+ "epoch": 0.22,
1596
+ "learning_rate": 7.594693877551021e-06,
1597
+ "loss": 0.1782,
1598
+ "step": 6400
1599
+ },
1600
+ {
1601
+ "epoch": 0.22,
1602
+ "learning_rate": 7.584489795918369e-06,
1603
+ "loss": 0.2043,
1604
+ "step": 6425
1605
+ },
1606
+ {
1607
+ "epoch": 0.22,
1608
+ "learning_rate": 7.574285714285715e-06,
1609
+ "loss": 0.1687,
1610
+ "step": 6450
1611
+ },
1612
+ {
1613
+ "epoch": 0.22,
1614
+ "learning_rate": 7.564081632653062e-06,
1615
+ "loss": 0.1986,
1616
+ "step": 6475
1617
+ },
1618
+ {
1619
+ "epoch": 0.23,
1620
+ "learning_rate": 7.553877551020409e-06,
1621
+ "loss": 0.1545,
1622
+ "step": 6500
1623
+ },
1624
+ {
1625
+ "epoch": 0.23,
1626
+ "learning_rate": 7.543673469387755e-06,
1627
+ "loss": 0.1946,
1628
+ "step": 6525
1629
+ },
1630
+ {
1631
+ "epoch": 0.23,
1632
+ "learning_rate": 7.533469387755103e-06,
1633
+ "loss": 0.1792,
1634
+ "step": 6550
1635
+ },
1636
+ {
1637
+ "epoch": 0.23,
1638
+ "learning_rate": 7.523265306122449e-06,
1639
+ "loss": 0.206,
1640
+ "step": 6575
1641
+ },
1642
+ {
1643
+ "epoch": 0.23,
1644
+ "learning_rate": 7.513061224489796e-06,
1645
+ "loss": 0.1653,
1646
+ "step": 6600
1647
+ },
1648
+ {
1649
+ "epoch": 0.23,
1650
+ "learning_rate": 7.502857142857144e-06,
1651
+ "loss": 0.2081,
1652
+ "step": 6625
1653
+ },
1654
+ {
1655
+ "epoch": 0.23,
1656
+ "learning_rate": 7.492653061224491e-06,
1657
+ "loss": 0.1745,
1658
+ "step": 6650
1659
+ },
1660
+ {
1661
+ "epoch": 0.23,
1662
+ "learning_rate": 7.482448979591838e-06,
1663
+ "loss": 0.212,
1664
+ "step": 6675
1665
+ },
1666
+ {
1667
+ "epoch": 0.23,
1668
+ "learning_rate": 7.472244897959184e-06,
1669
+ "loss": 0.1625,
1670
+ "step": 6700
1671
+ },
1672
+ {
1673
+ "epoch": 0.23,
1674
+ "learning_rate": 7.462040816326531e-06,
1675
+ "loss": 0.2008,
1676
+ "step": 6725
1677
+ },
1678
+ {
1679
+ "epoch": 0.23,
1680
+ "learning_rate": 7.451836734693878e-06,
1681
+ "loss": 0.1892,
1682
+ "step": 6750
1683
+ },
1684
+ {
1685
+ "epoch": 0.24,
1686
+ "learning_rate": 7.441632653061225e-06,
1687
+ "loss": 0.1979,
1688
+ "step": 6775
1689
+ },
1690
+ {
1691
+ "epoch": 0.24,
1692
+ "learning_rate": 7.431428571428572e-06,
1693
+ "loss": 0.1695,
1694
+ "step": 6800
1695
+ },
1696
+ {
1697
+ "epoch": 0.24,
1698
+ "learning_rate": 7.421224489795918e-06,
1699
+ "loss": 0.1914,
1700
+ "step": 6825
1701
+ },
1702
+ {
1703
+ "epoch": 0.24,
1704
+ "learning_rate": 7.411020408163267e-06,
1705
+ "loss": 0.1687,
1706
+ "step": 6850
1707
+ },
1708
+ {
1709
+ "epoch": 0.24,
1710
+ "learning_rate": 7.400816326530613e-06,
1711
+ "loss": 0.1805,
1712
+ "step": 6875
1713
+ },
1714
+ {
1715
+ "epoch": 0.24,
1716
+ "learning_rate": 7.39061224489796e-06,
1717
+ "loss": 0.1902,
1718
+ "step": 6900
1719
+ },
1720
+ {
1721
+ "epoch": 0.24,
1722
+ "learning_rate": 7.3804081632653066e-06,
1723
+ "loss": 0.212,
1724
+ "step": 6925
1725
+ },
1726
+ {
1727
+ "epoch": 0.24,
1728
+ "learning_rate": 7.370204081632654e-06,
1729
+ "loss": 0.1943,
1730
+ "step": 6950
1731
+ },
1732
+ {
1733
+ "epoch": 0.24,
1734
+ "learning_rate": 7.360000000000001e-06,
1735
+ "loss": 0.1934,
1736
+ "step": 6975
1737
+ },
1738
+ {
1739
+ "epoch": 0.24,
1740
+ "learning_rate": 7.349795918367347e-06,
1741
+ "loss": 0.1854,
1742
+ "step": 7000
1743
+ },
1744
+ {
1745
+ "epoch": 0.24,
1746
+ "eval_loss": 0.19266490638256073,
1747
+ "eval_runtime": 31023.933,
1748
+ "eval_samples_per_second": 1.0,
1749
+ "eval_steps_per_second": 0.063,
1750
+ "eval_wer": 0.10483351660871203,
1751
+ "step": 7000
1752
  }
1753
  ],
1754
  "max_steps": 25000,
1755
  "num_train_epochs": 1,
1756
+ "total_flos": 2.377951543296e+20,
1757
  "trial_name": null,
1758
  "trial_params": null
1759
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e0ff690a614ad4feb1584be1b22e3316419b9b45d45121399a24f7b7785e65c
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e91ae1bbae769e9ce17ed73669048b57cd296ecd5ff7d7b0071e74ca1c28a4f
3
  size 6173655480
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5c471ec8b7d00e44f926ec8e72e6951f01dad13c670f4fcd66f750928299dbb9
3
- size 43977
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e2bff7f5195a10fc570bcd185faa14302e6ca8f8cd84b4a9fa8001f74a0166e
3
+ size 50575