Training in progress, step 400
Browse files
fine-tune-whisper-streaming.ipynb
CHANGED
@@ -878,8 +878,8 @@
|
|
878 |
"\n",
|
879 |
" <div>\n",
|
880 |
" \n",
|
881 |
-
" <progress value='
|
882 |
-
" [
|
883 |
" </div>\n",
|
884 |
" <table border=\"1\" class=\"dataframe\">\n",
|
885 |
" <thead>\n",
|
@@ -909,6 +909,12 @@
|
|
909 |
" <td>0.306301</td>\n",
|
910 |
" <td>14.119548</td>\n",
|
911 |
" </tr>\n",
|
|
|
|
|
|
|
|
|
|
|
|
|
912 |
" </tbody>\n",
|
913 |
"</table><p>"
|
914 |
],
|
@@ -974,6 +980,25 @@
|
|
974 |
"Feature extractor saved in ./preprocessor_config.json\n",
|
975 |
"tokenizer config file saved in ./tokenizer_config.json\n",
|
976 |
"Special tokens file saved in ./special_tokens_map.json\n",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
977 |
"added tokens file saved in ./added_tokens.json\n"
|
978 |
]
|
979 |
}
|
|
|
878 |
"\n",
|
879 |
" <div>\n",
|
880 |
" \n",
|
881 |
+
" <progress value='401' max='3000' style='width:300px; height:20px; vertical-align: middle;'></progress>\n",
|
882 |
+
" [ 401/3000 2:12:24 < 14:22:26, 0.05 it/s, Epoch 1.01/9223372036854775807]\n",
|
883 |
" </div>\n",
|
884 |
" <table border=\"1\" class=\"dataframe\">\n",
|
885 |
" <thead>\n",
|
|
|
909 |
" <td>0.306301</td>\n",
|
910 |
" <td>14.119548</td>\n",
|
911 |
" </tr>\n",
|
912 |
+
" <tr>\n",
|
913 |
+
" <td>400</td>\n",
|
914 |
+
" <td>0.250100</td>\n",
|
915 |
+
" <td>0.292077</td>\n",
|
916 |
+
" <td>13.570275</td>\n",
|
917 |
+
" </tr>\n",
|
918 |
" </tbody>\n",
|
919 |
"</table><p>"
|
920 |
],
|
|
|
980 |
"Feature extractor saved in ./preprocessor_config.json\n",
|
981 |
"tokenizer config file saved in ./tokenizer_config.json\n",
|
982 |
"Special tokens file saved in ./special_tokens_map.json\n",
|
983 |
+
"added tokens file saved in ./added_tokens.json\n",
|
984 |
+
"/opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\n",
|
985 |
+
" warnings.warn('Was asked to gather along dimension 0, but all '\n",
|
986 |
+
"Reading metadata...: 18211it [00:02, 8543.60it/s] \n",
|
987 |
+
"***** Running Evaluation *****\n",
|
988 |
+
" Num examples: Unknown\n",
|
989 |
+
" Batch size = 20\n",
|
990 |
+
"Reading metadata...: 8693it [00:01, 7662.16it/s]\n",
|
991 |
+
"The following columns in the evaluation set don't have a corresponding argument in `WhisperForConditionalGeneration.forward` and have been ignored: input_length. If input_length are not expected by `WhisperForConditionalGeneration.forward`, you can safely ignore this message.\n",
|
992 |
+
"Saving model checkpoint to ./checkpoint-400\n",
|
993 |
+
"Configuration saved in ./checkpoint-400/config.json\n",
|
994 |
+
"Model weights saved in ./checkpoint-400/pytorch_model.bin\n",
|
995 |
+
"Feature extractor saved in ./checkpoint-400/preprocessor_config.json\n",
|
996 |
+
"tokenizer config file saved in ./checkpoint-400/tokenizer_config.json\n",
|
997 |
+
"Special tokens file saved in ./checkpoint-400/special_tokens_map.json\n",
|
998 |
+
"added tokens file saved in ./checkpoint-400/added_tokens.json\n",
|
999 |
+
"Feature extractor saved in ./preprocessor_config.json\n",
|
1000 |
+
"tokenizer config file saved in ./tokenizer_config.json\n",
|
1001 |
+
"Special tokens file saved in ./special_tokens_map.json\n",
|
1002 |
"added tokens file saved in ./added_tokens.json\n"
|
1003 |
]
|
1004 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 967099139
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:15e4408ab96bffb767b7d9d777f8dddfc3d5df21def73f3acab022e3daedd8c9
|
3 |
size 967099139
|
runs/Dec06_07-40-22_04812eda968b/events.out.tfevents.1670312431.04812eda968b.2650899.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5ca753c6716890873e38e7b667c4ba5606d24f74521b943b5300e4409054c24a
|
3 |
+
size 67878
|