jraramhoej commited on
Commit
7f8e62e
1 Parent(s): 4381de3

End of training

Browse files
fine-tune-whisper-streaming.ipynb CHANGED
@@ -847,7 +847,7 @@
847
  },
848
  {
849
  "cell_type": "code",
850
- "execution_count": null,
851
  "id": "ee8b7b8e-1c9a-4d77-9137-1778a629e6de",
852
  "metadata": {},
853
  "outputs": [
@@ -876,8 +876,8 @@
876
  "\n",
877
  " <div>\n",
878
  " \n",
879
- " <progress value='2001' max='2000' style='width:300px; height:20px; vertical-align: middle;'></progress>\n",
880
- " [2000/2000 3:36:26, Epoch 99.01/9223372036854775807]\n",
881
  " </div>\n",
882
  " <table border=\"1\" class=\"dataframe\">\n",
883
  " <thead>\n",
@@ -895,6 +895,12 @@
895
  " <td>0.603500</td>\n",
896
  " <td>26.790773</td>\n",
897
  " </tr>\n",
 
 
 
 
 
 
898
  " </tbody>\n",
899
  "</table><p>"
900
  ],
@@ -1127,8 +1133,35 @@
1127
  " Num examples: Unknown\n",
1128
  " Batch size = 8\n",
1129
  "Reading metadata...: 677it [00:00, 6344.98it/s]\n",
1130
- "The following columns in the evaluation set don't have a corresponding argument in `WhisperForConditionalGeneration.forward` and have been ignored: input_length. If input_length are not expected by `WhisperForConditionalGeneration.forward`, you can safely ignore this message.\n"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1131
  ]
 
 
 
 
 
 
 
 
 
 
1132
  }
1133
  ],
1134
  "source": [
@@ -1157,7 +1190,7 @@
1157
  },
1158
  {
1159
  "cell_type": "code",
1160
- "execution_count": null,
1161
  "id": "6dd0e310-9b07-4133-ac14-2ed2d7524e22",
1162
  "metadata": {},
1163
  "outputs": [],
@@ -1185,7 +1218,21 @@
1185
  "execution_count": null,
1186
  "id": "95737cda-c5dd-4887-a4d0-dfcb0d61d977",
1187
  "metadata": {},
1188
- "outputs": [],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1189
  "source": [
1190
  "trainer.push_to_hub(**kwargs)"
1191
  ]
 
847
  },
848
  {
849
  "cell_type": "code",
850
+ "execution_count": 22,
851
  "id": "ee8b7b8e-1c9a-4d77-9137-1778a629e6de",
852
  "metadata": {},
853
  "outputs": [
 
876
  "\n",
877
  " <div>\n",
878
  " \n",
879
+ " <progress value='2000' max='2000' style='width:300px; height:20px; vertical-align: middle;'></progress>\n",
880
+ " [2000/2000 3:40:19, Epoch 99/9223372036854775807]\n",
881
  " </div>\n",
882
  " <table border=\"1\" class=\"dataframe\">\n",
883
  " <thead>\n",
 
895
  " <td>0.603500</td>\n",
896
  " <td>26.790773</td>\n",
897
  " </tr>\n",
898
+ " <tr>\n",
899
+ " <td>2000</td>\n",
900
+ " <td>0.000200</td>\n",
901
+ " <td>0.631077</td>\n",
902
+ " <td>35.613112</td>\n",
903
+ " </tr>\n",
904
  " </tbody>\n",
905
  "</table><p>"
906
  ],
 
1133
  " Num examples: Unknown\n",
1134
  " Batch size = 8\n",
1135
  "Reading metadata...: 677it [00:00, 6344.98it/s]\n",
1136
+ "The following columns in the evaluation set don't have a corresponding argument in `WhisperForConditionalGeneration.forward` and have been ignored: input_length. If input_length are not expected by `WhisperForConditionalGeneration.forward`, you can safely ignore this message.\n",
1137
+ "Saving model checkpoint to ./checkpoint-2000\n",
1138
+ "Configuration saved in ./checkpoint-2000/config.json\n",
1139
+ "Model weights saved in ./checkpoint-2000/pytorch_model.bin\n",
1140
+ "Feature extractor saved in ./checkpoint-2000/preprocessor_config.json\n",
1141
+ "tokenizer config file saved in ./checkpoint-2000/tokenizer_config.json\n",
1142
+ "Special tokens file saved in ./checkpoint-2000/special_tokens_map.json\n",
1143
+ "added tokens file saved in ./checkpoint-2000/added_tokens.json\n",
1144
+ "Feature extractor saved in ./preprocessor_config.json\n",
1145
+ "tokenizer config file saved in ./tokenizer_config.json\n",
1146
+ "Special tokens file saved in ./special_tokens_map.json\n",
1147
+ "added tokens file saved in ./added_tokens.json\n",
1148
+ "\n",
1149
+ "\n",
1150
+ "Training completed. Do not forget to share your model on huggingface.co/models =)\n",
1151
+ "\n",
1152
+ "\n",
1153
+ "Loading best model from ./checkpoint-1000 (score: 26.79077296641036).\n"
1154
  ]
1155
+ },
1156
+ {
1157
+ "data": {
1158
+ "text/plain": [
1159
+ "TrainOutput(global_step=2000, training_loss=0.14686792829236947, metrics={'train_runtime': 13254.7147, 'train_samples_per_second': 9.657, 'train_steps_per_second': 0.151, 'total_flos': 3.5957740879872e+19, 'train_loss': 0.14686792829236947, 'epoch': 99.01})"
1160
+ ]
1161
+ },
1162
+ "execution_count": 22,
1163
+ "metadata": {},
1164
+ "output_type": "execute_result"
1165
  }
1166
  ],
1167
  "source": [
 
1190
  },
1191
  {
1192
  "cell_type": "code",
1193
+ "execution_count": 23,
1194
  "id": "6dd0e310-9b07-4133-ac14-2ed2d7524e22",
1195
  "metadata": {},
1196
  "outputs": [],
 
1218
  "execution_count": null,
1219
  "id": "95737cda-c5dd-4887-a4d0-dfcb0d61d977",
1220
  "metadata": {},
1221
+ "outputs": [
1222
+ {
1223
+ "name": "stderr",
1224
+ "output_type": "stream",
1225
+ "text": [
1226
+ "Saving model checkpoint to ./\n",
1227
+ "Configuration saved in ./config.json\n",
1228
+ "Model weights saved in ./pytorch_model.bin\n",
1229
+ "Feature extractor saved in ./preprocessor_config.json\n",
1230
+ "tokenizer config file saved in ./tokenizer_config.json\n",
1231
+ "Special tokens file saved in ./special_tokens_map.json\n",
1232
+ "added tokens file saved in ./added_tokens.json\n"
1233
+ ]
1234
+ }
1235
+ ],
1236
  "source": [
1237
  "trainer.push_to_hub(**kwargs)"
1238
  ]
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e6ff01a3ce8803bfef56cf5cca68c92577d50423a3d1edf2a66a1338a93cd7d
3
  size 967102601
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c102a56223c9d4c0073bf307faa42881b324cda17921f66be674baeaac81536
3
  size 967102601
runs/Dec15_14-18-11_129-213-18-196/events.out.tfevents.1671113929.129-213-18-196.4128231.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c42467d082604e0422febf3b84ff084d42c13e1839cdcd4976014a09b2a401e8
3
- size 17463
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:351504b6426b282a939b5cfea71bc5f9352a2d936195d040bccab3dd4c44ccc7
3
+ size 17817