Training in progress, step 100
Browse files- .ipynb_checkpoints/run-checkpoint.sh +2 -3
- pytorch_model.bin +1 -1
- run.sh +2 -3
- special_tokens_map.json +1 -1
- training_args.bin +1 -1
.ipynb_checkpoints/run-checkpoint.sh
CHANGED
@@ -5,16 +5,15 @@ python run_speech_recognition_ctc.py \
|
|
5 |
--output_dir="./" \
|
6 |
--overwrite_output_dir \
|
7 |
--max_steps="1000" \
|
8 |
-
--per_device_train_batch_size="
|
9 |
--learning_rate="3e-4" \
|
10 |
--warmup_steps="200" \
|
11 |
-
--save_steps="
|
12 |
--eval_steps="100" \
|
13 |
--save_total_limit="1" \
|
14 |
--evaluation_strategy="steps" \
|
15 |
--text_column_name="sentence" \
|
16 |
--length_column_name="input_length" \
|
17 |
-
--save_steps="5" \
|
18 |
--layerdrop="0.0" \
|
19 |
--freeze_feature_encoder \
|
20 |
--gradient_checkpointing \
|
|
|
5 |
--output_dir="./" \
|
6 |
--overwrite_output_dir \
|
7 |
--max_steps="1000" \
|
8 |
+
--per_device_train_batch_size="16" \
|
9 |
--learning_rate="3e-4" \
|
10 |
--warmup_steps="200" \
|
11 |
+
--save_steps="100" \
|
12 |
--eval_steps="100" \
|
13 |
--save_total_limit="1" \
|
14 |
--evaluation_strategy="steps" \
|
15 |
--text_column_name="sentence" \
|
16 |
--length_column_name="input_length" \
|
|
|
17 |
--layerdrop="0.0" \
|
18 |
--freeze_feature_encoder \
|
19 |
--gradient_checkpointing \
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1262321393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bb41f3394c8a1e55afa7d71f33cc3e59efbb1c8e6f16e6e78f62e44d75e697b9
|
3 |
size 1262321393
|
run.sh
CHANGED
@@ -5,16 +5,15 @@ python run_speech_recognition_ctc.py \
|
|
5 |
--output_dir="./" \
|
6 |
--overwrite_output_dir \
|
7 |
--max_steps="1000" \
|
8 |
-
--per_device_train_batch_size="
|
9 |
--learning_rate="3e-4" \
|
10 |
--warmup_steps="200" \
|
11 |
-
--save_steps="
|
12 |
--eval_steps="100" \
|
13 |
--save_total_limit="1" \
|
14 |
--evaluation_strategy="steps" \
|
15 |
--text_column_name="sentence" \
|
16 |
--length_column_name="input_length" \
|
17 |
-
--save_steps="20" \
|
18 |
--layerdrop="0.0" \
|
19 |
--freeze_feature_encoder \
|
20 |
--gradient_checkpointing \
|
|
|
5 |
--output_dir="./" \
|
6 |
--overwrite_output_dir \
|
7 |
--max_steps="1000" \
|
8 |
+
--per_device_train_batch_size="16" \
|
9 |
--learning_rate="3e-4" \
|
10 |
--warmup_steps="200" \
|
11 |
+
--save_steps="100" \
|
12 |
--eval_steps="100" \
|
13 |
--save_total_limit="1" \
|
14 |
--evaluation_strategy="steps" \
|
15 |
--text_column_name="sentence" \
|
16 |
--length_column_name="input_length" \
|
|
|
17 |
--layerdrop="0.0" \
|
18 |
--freeze_feature_encoder \
|
19 |
--gradient_checkpointing \
|
special_tokens_map.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
|
|
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2991
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a7d08cb66391e6d626ffe2f58a80bdf02f5d63b07abb13079ef02ce75c032bb5
|
3 |
size 2991
|