Training in progress, step 200
Browse files- .ipynb_checkpoints/run-checkpoint.sh +7 -5
- pytorch_model.bin +1 -1
- run.sh +7 -5
- special_tokens_map.json +1 -1
- training_args.bin +1 -1
.ipynb_checkpoints/run-checkpoint.sh
CHANGED
@@ -4,16 +4,18 @@ python run_speech_recognition_ctc.py \
|
|
4 |
--dataset_config_name="ur" \
|
5 |
--output_dir="./" \
|
6 |
--overwrite_output_dir \
|
7 |
-
--max_steps="
|
8 |
-
--
|
|
|
|
|
9 |
--learning_rate="4e-4" \
|
10 |
-
--save_total_limit="
|
11 |
--evaluation_strategy="steps" \
|
12 |
--text_column_name="sentence" \
|
13 |
--length_column_name="input_length" \
|
14 |
--chars_to_ignore ,\?\.\!\-\;\:\"\โ\%\โ\โ\๏ฟฝ\ู\ู\ูู,ู\
|
15 |
-
--save_steps="
|
16 |
-
--
|
17 |
--freeze_feature_encoder \
|
18 |
--gradient_checkpointing \
|
19 |
--fp16 \
|
|
|
4 |
--dataset_config_name="ur" \
|
5 |
--output_dir="./" \
|
6 |
--overwrite_output_dir \
|
7 |
+
--max_steps="3000" \
|
8 |
+
--gradient_accumulation_steps="2" \
|
9 |
+
--lr_scheduler_type="linear" \
|
10 |
+
--per_device_train_batch_size="64" \
|
11 |
--learning_rate="4e-4" \
|
12 |
+
--save_total_limit="5" \
|
13 |
--evaluation_strategy="steps" \
|
14 |
--text_column_name="sentence" \
|
15 |
--length_column_name="input_length" \
|
16 |
--chars_to_ignore ,\?\.\!\-\;\:\"\โ\%\โ\โ\๏ฟฝ\ู\ู\ูู,ู\
|
17 |
+
--save_steps="200" \
|
18 |
+
--num_train_epochs="1000.0" \
|
19 |
--freeze_feature_encoder \
|
20 |
--gradient_checkpointing \
|
21 |
--fp16 \
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1262149169
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ab3ab630c986b23e9b31761e48867b23d33b4838a573ee4407440d45bf8bcdcc
|
3 |
size 1262149169
|
run.sh
CHANGED
@@ -4,16 +4,18 @@ python run_speech_recognition_ctc.py \
|
|
4 |
--dataset_config_name="ur" \
|
5 |
--output_dir="./" \
|
6 |
--overwrite_output_dir \
|
7 |
-
--max_steps="
|
8 |
-
--
|
|
|
|
|
9 |
--learning_rate="4e-4" \
|
10 |
-
--save_total_limit="
|
11 |
--evaluation_strategy="steps" \
|
12 |
--text_column_name="sentence" \
|
13 |
--length_column_name="input_length" \
|
14 |
--chars_to_ignore ,\?\.\!\-\;\:\"\โ\%\โ\โ\๏ฟฝ\ู\ู\ูู,ู\
|
15 |
-
--save_steps="
|
16 |
-
--
|
17 |
--freeze_feature_encoder \
|
18 |
--gradient_checkpointing \
|
19 |
--fp16 \
|
|
|
4 |
--dataset_config_name="ur" \
|
5 |
--output_dir="./" \
|
6 |
--overwrite_output_dir \
|
7 |
+
--max_steps="3000" \
|
8 |
+
--gradient_accumulation_steps="2" \
|
9 |
+
--lr_scheduler_type="linear" \
|
10 |
+
--per_device_train_batch_size="64" \
|
11 |
--learning_rate="4e-4" \
|
12 |
+
--save_total_limit="5" \
|
13 |
--evaluation_strategy="steps" \
|
14 |
--text_column_name="sentence" \
|
15 |
--length_column_name="input_length" \
|
16 |
--chars_to_ignore ,\?\.\!\-\;\:\"\โ\%\โ\โ\๏ฟฝ\ู\ู\ูู,ู\
|
17 |
+
--save_steps="200" \
|
18 |
+
--num_train_epochs="1000.0" \
|
19 |
--freeze_feature_encoder \
|
20 |
--gradient_checkpointing \
|
21 |
--fp16 \
|
special_tokens_map.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
|
|
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2991
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ff8e9fbda78cdc66136dc476d2cef558c305ffcb23a4b3428f386f79718fa61f
|
3 |
size 2991
|