xla stuff
Browse files- run_xla_test.sh +11 -10
run_xla_test.sh
CHANGED
@@ -1,36 +1,37 @@
|
|
1 |
# Whisper Finetuning script for the NST dataset
|
2 |
# This is a test script for XLA on TPU
|
3 |
|
4 |
-
PT_XLA_DEBUG=1
|
|
|
5 |
--model_name_or_path="openai/whisper-small" \
|
6 |
-
--output_dir="../whisper-NST-TPU-test" \
|
7 |
--overwrite_output_dir=True \
|
8 |
--language="Norwegian" \
|
9 |
--task="transcribe" \
|
10 |
--dataset_name="NbAiLab/NST" \
|
11 |
--dataset_config="no-close" \
|
12 |
--do_train=True \
|
13 |
-
--do_eval=
|
14 |
--audio_column_name="audio" \
|
15 |
--text_column_name="text" \
|
16 |
--per_device_train_batch_size=16 \
|
17 |
--per_device_train_batch_size=16 \
|
18 |
--learning_rate=2e-5 \
|
19 |
--warmup_steps=0 \
|
20 |
-
--max_steps=
|
21 |
--gradient_checkpointing=True \
|
22 |
--gradient_accumulation_steps=1 \
|
23 |
--group_by_length=False \
|
24 |
--evaluation_strategy="steps" \
|
25 |
-
--save_steps=
|
26 |
-
--eval_steps=
|
27 |
-
--max_eval_samples=
|
28 |
-
--logging_steps=
|
29 |
-
--load_best_model_at_end=
|
30 |
--metric_for_best_model="wer" \
|
31 |
--greater_is_better=False \
|
32 |
--report_to="tensorboard" \
|
33 |
-
--predict_with_generate=
|
34 |
--generation_max_length=225 \
|
35 |
--print_training_arguments=True \
|
36 |
--xla=True \
|
|
|
1 |
# Whisper Finetuning script for the NST dataset
|
2 |
# This is a test script for XLA on TPU
|
3 |
|
4 |
+
#PT_XLA_DEBUG=1
|
5 |
+
python xla_spawn.py --num_cores=4 run_whisper_finetuning.py\
|
6 |
--model_name_or_path="openai/whisper-small" \
|
7 |
+
--output_dir="../whisper-NST-TPU-test-2" \
|
8 |
--overwrite_output_dir=True \
|
9 |
--language="Norwegian" \
|
10 |
--task="transcribe" \
|
11 |
--dataset_name="NbAiLab/NST" \
|
12 |
--dataset_config="no-close" \
|
13 |
--do_train=True \
|
14 |
+
--do_eval=False \
|
15 |
--audio_column_name="audio" \
|
16 |
--text_column_name="text" \
|
17 |
--per_device_train_batch_size=16 \
|
18 |
--per_device_train_batch_size=16 \
|
19 |
--learning_rate=2e-5 \
|
20 |
--warmup_steps=0 \
|
21 |
+
--max_steps=2000 \
|
22 |
--gradient_checkpointing=True \
|
23 |
--gradient_accumulation_steps=1 \
|
24 |
--group_by_length=False \
|
25 |
--evaluation_strategy="steps" \
|
26 |
+
--save_steps=200 \
|
27 |
+
--eval_steps=0 \
|
28 |
+
--max_eval_samples=1 \
|
29 |
+
--logging_steps=100 \
|
30 |
+
--load_best_model_at_end=False \
|
31 |
--metric_for_best_model="wer" \
|
32 |
--greater_is_better=False \
|
33 |
--report_to="tensorboard" \
|
34 |
+
--predict_with_generate=False \
|
35 |
--generation_max_length=225 \
|
36 |
--print_training_arguments=True \
|
37 |
--xla=True \
|