elsayedissa commited on
Commit
2685bd3
1 Parent(s): d171f03

Training in progress, step 21000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3070aeb72d29088489c0fb6e92c23b06f3de39671f0b9c078f18da734e95bf6b
3
  size 12347192855
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c1e93fa65fde9ba67832b5fde4ad330506d95a23edfec7be98020d37f17559b
3
  size 12347192855
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:47bc5d383a893937307b6252269bc95121112ddd7c888bf951e25cc0f6233b96
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f07b5f25c323d9a36c85d57f748183f9d3d14494dbfa6b195ed15224d9deafeb
3
  size 6173655480
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4979e2334c039bb7332b1f1ca1ec119b85c014a960ba0dbab629d4dd618f0ad2
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f36f42cb84b389111430939ff75a7277c877ebb6534c5db685d643483c1e7ae
3
  size 14575
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e95ea4d51ce3dbab01e59ff8912c2f7044fb7bd804fa1afc7ea77264e8c7d84
3
  size 557
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75b2d5c8b12724e300963a719034bfe44737705d2de06deafe54683c78564862
3
  size 557
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e664770f258aac914190f0cd016b820f86cc722eb7db2dbd9de5aff987606154
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7394c13a986a89cb46fa0877cb8091fc4bf0e4f28ce1608b81abde4cc024dfc
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
- "epoch": 0.6942274983512097,
5
- "global_step": 20000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -4986,11 +4986,260 @@
4986
  "eval_steps_per_second": 0.063,
4987
  "eval_wer": 0.08736786835874422,
4988
  "step": 20000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4989
  }
4990
  ],
4991
  "max_steps": 25000,
4992
  "num_train_epochs": 1,
4993
- "total_flos": 6.79414726656e+20,
4994
  "trial_name": null,
4995
  "trial_params": null
4996
  }
 
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
+ "epoch": 0.7289388732687702,
5
+ "global_step": 21000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
4986
  "eval_steps_per_second": 0.063,
4987
  "eval_wer": 0.08736786835874422,
4988
  "step": 20000
4989
+ },
4990
+ {
4991
+ "epoch": 0.7,
4992
+ "learning_rate": 2.0351020408163267e-06,
4993
+ "loss": 0.1463,
4994
+ "step": 20025
4995
+ },
4996
+ {
4997
+ "epoch": 0.7,
4998
+ "learning_rate": 2.024897959183674e-06,
4999
+ "loss": 0.1206,
5000
+ "step": 20050
5001
+ },
5002
+ {
5003
+ "epoch": 0.7,
5004
+ "learning_rate": 2.0146938775510204e-06,
5005
+ "loss": 0.1498,
5006
+ "step": 20075
5007
+ },
5008
+ {
5009
+ "epoch": 0.7,
5010
+ "learning_rate": 2.0044897959183675e-06,
5011
+ "loss": 0.1191,
5012
+ "step": 20100
5013
+ },
5014
+ {
5015
+ "epoch": 0.7,
5016
+ "learning_rate": 1.9942857142857146e-06,
5017
+ "loss": 0.1674,
5018
+ "step": 20125
5019
+ },
5020
+ {
5021
+ "epoch": 0.7,
5022
+ "learning_rate": 1.984081632653061e-06,
5023
+ "loss": 0.1326,
5024
+ "step": 20150
5025
+ },
5026
+ {
5027
+ "epoch": 0.7,
5028
+ "learning_rate": 1.9738775510204083e-06,
5029
+ "loss": 0.1602,
5030
+ "step": 20175
5031
+ },
5032
+ {
5033
+ "epoch": 0.7,
5034
+ "learning_rate": 1.9636734693877553e-06,
5035
+ "loss": 0.1298,
5036
+ "step": 20200
5037
+ },
5038
+ {
5039
+ "epoch": 0.7,
5040
+ "learning_rate": 1.9534693877551024e-06,
5041
+ "loss": 0.1644,
5042
+ "step": 20225
5043
+ },
5044
+ {
5045
+ "epoch": 0.7,
5046
+ "learning_rate": 1.943265306122449e-06,
5047
+ "loss": 0.1277,
5048
+ "step": 20250
5049
+ },
5050
+ {
5051
+ "epoch": 0.7,
5052
+ "learning_rate": 1.933061224489796e-06,
5053
+ "loss": 0.1593,
5054
+ "step": 20275
5055
+ },
5056
+ {
5057
+ "epoch": 0.7,
5058
+ "learning_rate": 1.922857142857143e-06,
5059
+ "loss": 0.113,
5060
+ "step": 20300
5061
+ },
5062
+ {
5063
+ "epoch": 0.71,
5064
+ "learning_rate": 1.91265306122449e-06,
5065
+ "loss": 0.142,
5066
+ "step": 20325
5067
+ },
5068
+ {
5069
+ "epoch": 0.71,
5070
+ "learning_rate": 1.9024489795918368e-06,
5071
+ "loss": 0.1197,
5072
+ "step": 20350
5073
+ },
5074
+ {
5075
+ "epoch": 0.71,
5076
+ "learning_rate": 1.8922448979591837e-06,
5077
+ "loss": 0.1487,
5078
+ "step": 20375
5079
+ },
5080
+ {
5081
+ "epoch": 0.71,
5082
+ "learning_rate": 1.8820408163265308e-06,
5083
+ "loss": 0.1309,
5084
+ "step": 20400
5085
+ },
5086
+ {
5087
+ "epoch": 0.71,
5088
+ "learning_rate": 1.8718367346938776e-06,
5089
+ "loss": 0.1531,
5090
+ "step": 20425
5091
+ },
5092
+ {
5093
+ "epoch": 0.71,
5094
+ "learning_rate": 1.8616326530612247e-06,
5095
+ "loss": 0.135,
5096
+ "step": 20450
5097
+ },
5098
+ {
5099
+ "epoch": 0.71,
5100
+ "learning_rate": 1.8514285714285715e-06,
5101
+ "loss": 0.1591,
5102
+ "step": 20475
5103
+ },
5104
+ {
5105
+ "epoch": 0.71,
5106
+ "learning_rate": 1.8412244897959186e-06,
5107
+ "loss": 0.1402,
5108
+ "step": 20500
5109
+ },
5110
+ {
5111
+ "epoch": 0.71,
5112
+ "learning_rate": 1.8310204081632654e-06,
5113
+ "loss": 0.1577,
5114
+ "step": 20525
5115
+ },
5116
+ {
5117
+ "epoch": 0.71,
5118
+ "learning_rate": 1.8208163265306125e-06,
5119
+ "loss": 0.1218,
5120
+ "step": 20550
5121
+ },
5122
+ {
5123
+ "epoch": 0.71,
5124
+ "learning_rate": 1.8106122448979593e-06,
5125
+ "loss": 0.1638,
5126
+ "step": 20575
5127
+ },
5128
+ {
5129
+ "epoch": 0.72,
5130
+ "learning_rate": 1.8004081632653062e-06,
5131
+ "loss": 0.1376,
5132
+ "step": 20600
5133
+ },
5134
+ {
5135
+ "epoch": 0.72,
5136
+ "learning_rate": 1.7902040816326533e-06,
5137
+ "loss": 0.1609,
5138
+ "step": 20625
5139
+ },
5140
+ {
5141
+ "epoch": 0.72,
5142
+ "learning_rate": 1.7800000000000001e-06,
5143
+ "loss": 0.1099,
5144
+ "step": 20650
5145
+ },
5146
+ {
5147
+ "epoch": 0.72,
5148
+ "learning_rate": 1.7697959183673472e-06,
5149
+ "loss": 0.1863,
5150
+ "step": 20675
5151
+ },
5152
+ {
5153
+ "epoch": 0.72,
5154
+ "learning_rate": 1.759591836734694e-06,
5155
+ "loss": 0.1185,
5156
+ "step": 20700
5157
+ },
5158
+ {
5159
+ "epoch": 0.72,
5160
+ "learning_rate": 1.749387755102041e-06,
5161
+ "loss": 0.1616,
5162
+ "step": 20725
5163
+ },
5164
+ {
5165
+ "epoch": 0.72,
5166
+ "learning_rate": 1.739183673469388e-06,
5167
+ "loss": 0.1208,
5168
+ "step": 20750
5169
+ },
5170
+ {
5171
+ "epoch": 0.72,
5172
+ "learning_rate": 1.728979591836735e-06,
5173
+ "loss": 0.1588,
5174
+ "step": 20775
5175
+ },
5176
+ {
5177
+ "epoch": 0.72,
5178
+ "learning_rate": 1.7187755102040818e-06,
5179
+ "loss": 0.1268,
5180
+ "step": 20800
5181
+ },
5182
+ {
5183
+ "epoch": 0.72,
5184
+ "learning_rate": 1.7085714285714287e-06,
5185
+ "loss": 0.1506,
5186
+ "step": 20825
5187
+ },
5188
+ {
5189
+ "epoch": 0.72,
5190
+ "learning_rate": 1.6983673469387758e-06,
5191
+ "loss": 0.1412,
5192
+ "step": 20850
5193
+ },
5194
+ {
5195
+ "epoch": 0.72,
5196
+ "learning_rate": 1.6881632653061226e-06,
5197
+ "loss": 0.1547,
5198
+ "step": 20875
5199
+ },
5200
+ {
5201
+ "epoch": 0.73,
5202
+ "learning_rate": 1.6779591836734697e-06,
5203
+ "loss": 0.1386,
5204
+ "step": 20900
5205
+ },
5206
+ {
5207
+ "epoch": 0.73,
5208
+ "learning_rate": 1.6677551020408165e-06,
5209
+ "loss": 0.1597,
5210
+ "step": 20925
5211
+ },
5212
+ {
5213
+ "epoch": 0.73,
5214
+ "learning_rate": 1.6575510204081636e-06,
5215
+ "loss": 0.103,
5216
+ "step": 20950
5217
+ },
5218
+ {
5219
+ "epoch": 0.73,
5220
+ "learning_rate": 1.6473469387755102e-06,
5221
+ "loss": 0.1609,
5222
+ "step": 20975
5223
+ },
5224
+ {
5225
+ "epoch": 0.73,
5226
+ "learning_rate": 1.637142857142857e-06,
5227
+ "loss": 0.1479,
5228
+ "step": 21000
5229
+ },
5230
+ {
5231
+ "epoch": 0.73,
5232
+ "eval_loss": 0.15253788232803345,
5233
+ "eval_runtime": 31232.4021,
5234
+ "eval_samples_per_second": 0.994,
5235
+ "eval_steps_per_second": 0.062,
5236
+ "eval_wer": 0.08848403192624074,
5237
+ "step": 21000
5238
  }
5239
  ],
5240
  "max_steps": 25000,
5241
  "num_train_epochs": 1,
5242
+ "total_flos": 7.133854629888e+20,
5243
  "trial_name": null,
5244
  "trial_params": null
5245
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:47bc5d383a893937307b6252269bc95121112ddd7c888bf951e25cc0f6233b96
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f07b5f25c323d9a36c85d57f748183f9d3d14494dbfa6b195ed15224d9deafeb
3
  size 6173655480
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4222f5599f934e25008ca9c74695648d0516086fca35c66323f38f875ddff30c
3
- size 136808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a05d2b060ee61e97fa7da66f905940a11d0a5b14104c16f577979a5e29d6fa2d
3
+ size 143532