elsayedissa commited on
Commit
6b4d9e2
1 Parent(s): 9d744d9

Training in progress, step 5000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eb55c91b103628f5be95896e1614ad2bc546605cf6236721f0622634adcd689e
3
  size 12347192855
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c171984694c48427b034af7134433824c6ffd4f9f28188dda965bf3de0c15fb0
3
  size 12347192855
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:43e47f260dd1b8c823dbd2ec8239c2bf713fcf544791f3f0b436803f94dc5f60
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d172c474a443c85f3cb4d2a327f379678abab48c79f8e28e462b950711cc5684
3
  size 6173655480
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:640548085210c8490898cfba017d8bd1baf3553a19e27f9821a34801ec2b3a88
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0a9edbae95b39e125cb13cf9656d66377898a7639a1540109f9acf0f641de62
3
  size 14575
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f245e4bee87e5915781db8b8b7a3b00fd2ef949a525de0dd949823efc21b2f4
3
  size 557
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02bfa0a41730a41e7e9ac0b2cafb5ad11cb816439654260867a463c2e192b332
3
  size 557
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:455fa7fbc6657e0c30aabe935c34b61e115426f0bbb7f1106461aac75d23221b
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:183d1da7e0c968851cacd55db8383570bdd99f0154f18e1b72c4b435effad44d
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.4207764409383577,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-arabic-24h/checkpoint-3000",
4
- "epoch": 7.1301247771836005,
5
- "global_step": 4000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -1002,11 +1002,260 @@
1002
  "eval_steps_per_second": 0.149,
1003
  "eval_wer": 0.4251609673577187,
1004
  "step": 4000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1005
  }
1006
  ],
1007
  "max_steps": 5000,
1008
  "num_train_epochs": 9,
1009
- "total_flos": 4.245301687799808e+20,
1010
  "trial_name": null,
1011
  "trial_params": null
1012
  }
 
1
  {
2
  "best_metric": 0.4207764409383577,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-arabic-24h/checkpoint-3000",
4
+ "epoch": 8.9126559714795,
5
+ "global_step": 5000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
1002
  "eval_steps_per_second": 0.149,
1003
  "eval_wer": 0.4251609673577187,
1004
  "step": 4000
1005
+ },
1006
+ {
1007
+ "epoch": 7.17,
1008
+ "learning_rate": 2.1800000000000003e-06,
1009
+ "loss": 0.002,
1010
+ "step": 4025
1011
+ },
1012
+ {
1013
+ "epoch": 7.22,
1014
+ "learning_rate": 2.1244444444444443e-06,
1015
+ "loss": 0.0042,
1016
+ "step": 4050
1017
+ },
1018
+ {
1019
+ "epoch": 7.26,
1020
+ "learning_rate": 2.0688888888888892e-06,
1021
+ "loss": 0.0021,
1022
+ "step": 4075
1023
+ },
1024
+ {
1025
+ "epoch": 7.31,
1026
+ "learning_rate": 2.0133333333333337e-06,
1027
+ "loss": 0.0035,
1028
+ "step": 4100
1029
+ },
1030
+ {
1031
+ "epoch": 7.35,
1032
+ "learning_rate": 1.9577777777777777e-06,
1033
+ "loss": 0.0026,
1034
+ "step": 4125
1035
+ },
1036
+ {
1037
+ "epoch": 7.4,
1038
+ "learning_rate": 1.9022222222222222e-06,
1039
+ "loss": 0.0037,
1040
+ "step": 4150
1041
+ },
1042
+ {
1043
+ "epoch": 7.44,
1044
+ "learning_rate": 1.8466666666666668e-06,
1045
+ "loss": 0.0024,
1046
+ "step": 4175
1047
+ },
1048
+ {
1049
+ "epoch": 7.49,
1050
+ "learning_rate": 1.7911111111111113e-06,
1051
+ "loss": 0.0038,
1052
+ "step": 4200
1053
+ },
1054
+ {
1055
+ "epoch": 7.53,
1056
+ "learning_rate": 1.7355555555555555e-06,
1057
+ "loss": 0.0024,
1058
+ "step": 4225
1059
+ },
1060
+ {
1061
+ "epoch": 7.58,
1062
+ "learning_rate": 1.6800000000000002e-06,
1063
+ "loss": 0.0036,
1064
+ "step": 4250
1065
+ },
1066
+ {
1067
+ "epoch": 7.62,
1068
+ "learning_rate": 1.6244444444444447e-06,
1069
+ "loss": 0.0028,
1070
+ "step": 4275
1071
+ },
1072
+ {
1073
+ "epoch": 7.66,
1074
+ "learning_rate": 1.568888888888889e-06,
1075
+ "loss": 0.0043,
1076
+ "step": 4300
1077
+ },
1078
+ {
1079
+ "epoch": 7.71,
1080
+ "learning_rate": 1.5133333333333334e-06,
1081
+ "loss": 0.0017,
1082
+ "step": 4325
1083
+ },
1084
+ {
1085
+ "epoch": 7.75,
1086
+ "learning_rate": 1.457777777777778e-06,
1087
+ "loss": 0.0035,
1088
+ "step": 4350
1089
+ },
1090
+ {
1091
+ "epoch": 7.8,
1092
+ "learning_rate": 1.4022222222222223e-06,
1093
+ "loss": 0.003,
1094
+ "step": 4375
1095
+ },
1096
+ {
1097
+ "epoch": 7.84,
1098
+ "learning_rate": 1.3466666666666668e-06,
1099
+ "loss": 0.0037,
1100
+ "step": 4400
1101
+ },
1102
+ {
1103
+ "epoch": 7.89,
1104
+ "learning_rate": 1.2911111111111112e-06,
1105
+ "loss": 0.0018,
1106
+ "step": 4425
1107
+ },
1108
+ {
1109
+ "epoch": 7.93,
1110
+ "learning_rate": 1.2355555555555557e-06,
1111
+ "loss": 0.004,
1112
+ "step": 4450
1113
+ },
1114
+ {
1115
+ "epoch": 7.98,
1116
+ "learning_rate": 1.1800000000000001e-06,
1117
+ "loss": 0.002,
1118
+ "step": 4475
1119
+ },
1120
+ {
1121
+ "epoch": 8.02,
1122
+ "learning_rate": 1.1244444444444446e-06,
1123
+ "loss": 0.0025,
1124
+ "step": 4500
1125
+ },
1126
+ {
1127
+ "epoch": 8.07,
1128
+ "learning_rate": 1.068888888888889e-06,
1129
+ "loss": 0.0014,
1130
+ "step": 4525
1131
+ },
1132
+ {
1133
+ "epoch": 8.11,
1134
+ "learning_rate": 1.0133333333333333e-06,
1135
+ "loss": 0.0018,
1136
+ "step": 4550
1137
+ },
1138
+ {
1139
+ "epoch": 8.16,
1140
+ "learning_rate": 9.57777777777778e-07,
1141
+ "loss": 0.0012,
1142
+ "step": 4575
1143
+ },
1144
+ {
1145
+ "epoch": 8.2,
1146
+ "learning_rate": 9.022222222222222e-07,
1147
+ "loss": 0.0022,
1148
+ "step": 4600
1149
+ },
1150
+ {
1151
+ "epoch": 8.24,
1152
+ "learning_rate": 8.466666666666668e-07,
1153
+ "loss": 0.0011,
1154
+ "step": 4625
1155
+ },
1156
+ {
1157
+ "epoch": 8.29,
1158
+ "learning_rate": 7.911111111111111e-07,
1159
+ "loss": 0.0018,
1160
+ "step": 4650
1161
+ },
1162
+ {
1163
+ "epoch": 8.33,
1164
+ "learning_rate": 7.355555555555556e-07,
1165
+ "loss": 0.0013,
1166
+ "step": 4675
1167
+ },
1168
+ {
1169
+ "epoch": 8.38,
1170
+ "learning_rate": 6.800000000000001e-07,
1171
+ "loss": 0.0016,
1172
+ "step": 4700
1173
+ },
1174
+ {
1175
+ "epoch": 8.42,
1176
+ "learning_rate": 6.244444444444445e-07,
1177
+ "loss": 0.0013,
1178
+ "step": 4725
1179
+ },
1180
+ {
1181
+ "epoch": 8.47,
1182
+ "learning_rate": 5.68888888888889e-07,
1183
+ "loss": 0.0019,
1184
+ "step": 4750
1185
+ },
1186
+ {
1187
+ "epoch": 8.51,
1188
+ "learning_rate": 5.133333333333334e-07,
1189
+ "loss": 0.0013,
1190
+ "step": 4775
1191
+ },
1192
+ {
1193
+ "epoch": 8.56,
1194
+ "learning_rate": 4.5777777777777784e-07,
1195
+ "loss": 0.0017,
1196
+ "step": 4800
1197
+ },
1198
+ {
1199
+ "epoch": 8.6,
1200
+ "learning_rate": 4.0222222222222224e-07,
1201
+ "loss": 0.0013,
1202
+ "step": 4825
1203
+ },
1204
+ {
1205
+ "epoch": 8.65,
1206
+ "learning_rate": 3.466666666666667e-07,
1207
+ "loss": 0.0021,
1208
+ "step": 4850
1209
+ },
1210
+ {
1211
+ "epoch": 8.69,
1212
+ "learning_rate": 2.9111111111111116e-07,
1213
+ "loss": 0.0012,
1214
+ "step": 4875
1215
+ },
1216
+ {
1217
+ "epoch": 8.73,
1218
+ "learning_rate": 2.3555555555555556e-07,
1219
+ "loss": 0.0026,
1220
+ "step": 4900
1221
+ },
1222
+ {
1223
+ "epoch": 8.78,
1224
+ "learning_rate": 1.8e-07,
1225
+ "loss": 0.0012,
1226
+ "step": 4925
1227
+ },
1228
+ {
1229
+ "epoch": 8.82,
1230
+ "learning_rate": 1.2444444444444446e-07,
1231
+ "loss": 0.0017,
1232
+ "step": 4950
1233
+ },
1234
+ {
1235
+ "epoch": 8.87,
1236
+ "learning_rate": 6.888888888888889e-08,
1237
+ "loss": 0.0012,
1238
+ "step": 4975
1239
+ },
1240
+ {
1241
+ "epoch": 8.91,
1242
+ "learning_rate": 1.3333333333333334e-08,
1243
+ "loss": 0.0016,
1244
+ "step": 5000
1245
+ },
1246
+ {
1247
+ "epoch": 8.91,
1248
+ "eval_loss": 0.3434337377548218,
1249
+ "eval_runtime": 4386.6592,
1250
+ "eval_samples_per_second": 2.379,
1251
+ "eval_steps_per_second": 0.149,
1252
+ "eval_wer": 0.4239008475600283,
1253
+ "step": 5000
1254
  }
1255
  ],
1256
  "max_steps": 5000,
1257
  "num_train_epochs": 9,
1258
+ "total_flos": 5.306738576228352e+20,
1259
  "trial_name": null,
1260
  "trial_params": null
1261
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:43e47f260dd1b8c823dbd2ec8239c2bf713fcf544791f3f0b436803f94dc5f60
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d172c474a443c85f3cb4d2a327f379678abab48c79f8e28e462b950711cc5684
3
  size 6173655480
runs/Jan28_23-19-38_gpu07.cyverse.org/events.out.tfevents.1674973221.gpu07.cyverse.org.79624.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c20785d93f929c6776fb55f65c13ef81d9ad4f46d8a0e0fdba027f22601ec447
3
- size 30790
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22f944f653d930146c1944127978e6c5f4f499b2926b2d171485fe46cc8860c4
3
+ size 37388