pere commited on
Commit
112f924
1 Parent(s): c26c107
Files changed (1) hide show
  1. run_xla_test.sh +11 -10
run_xla_test.sh CHANGED
@@ -1,36 +1,37 @@
1
  # Whisper Finetuning script for the NST dataset
2
  # This is a test script for XLA on TPU
3
 
4
- PT_XLA_DEBUG=1 python xla_spawn.py --num_cores=4 run_whisper_finetuning.py\
 
5
  --model_name_or_path="openai/whisper-small" \
6
- --output_dir="../whisper-NST-TPU-test" \
7
  --overwrite_output_dir=True \
8
  --language="Norwegian" \
9
  --task="transcribe" \
10
  --dataset_name="NbAiLab/NST" \
11
  --dataset_config="no-close" \
12
  --do_train=True \
13
- --do_eval=True \
14
  --audio_column_name="audio" \
15
  --text_column_name="text" \
16
  --per_device_train_batch_size=16 \
17
  --per_device_train_batch_size=16 \
18
  --learning_rate=2e-5 \
19
  --warmup_steps=0 \
20
- --max_steps=10 \
21
  --gradient_checkpointing=True \
22
  --gradient_accumulation_steps=1 \
23
  --group_by_length=False \
24
  --evaluation_strategy="steps" \
25
- --save_steps=10 \
26
- --eval_steps=10 \
27
- --max_eval_samples=2 \
28
- --logging_steps=10 \
29
- --load_best_model_at_end=True \
30
  --metric_for_best_model="wer" \
31
  --greater_is_better=False \
32
  --report_to="tensorboard" \
33
- --predict_with_generate=True \
34
  --generation_max_length=225 \
35
  --print_training_arguments=True \
36
  --xla=True \
 
1
  # Whisper Finetuning script for the NST dataset
2
  # This is a test script for XLA on TPU
3
 
4
+ #PT_XLA_DEBUG=1
5
+ python xla_spawn.py --num_cores=4 run_whisper_finetuning.py\
6
  --model_name_or_path="openai/whisper-small" \
7
+ --output_dir="../whisper-NST-TPU-test-2" \
8
  --overwrite_output_dir=True \
9
  --language="Norwegian" \
10
  --task="transcribe" \
11
  --dataset_name="NbAiLab/NST" \
12
  --dataset_config="no-close" \
13
  --do_train=True \
14
+ --do_eval=False \
15
  --audio_column_name="audio" \
16
  --text_column_name="text" \
17
  --per_device_train_batch_size=16 \
18
  --per_device_train_batch_size=16 \
19
  --learning_rate=2e-5 \
20
  --warmup_steps=0 \
21
+ --max_steps=2000 \
22
  --gradient_checkpointing=True \
23
  --gradient_accumulation_steps=1 \
24
  --group_by_length=False \
25
  --evaluation_strategy="steps" \
26
+ --save_steps=200 \
27
+ --eval_steps=0 \
28
+ --max_eval_samples=1 \
29
+ --logging_steps=100 \
30
+ --load_best_model_at_end=False \
31
  --metric_for_best_model="wer" \
32
  --greater_is_better=False \
33
  --report_to="tensorboard" \
34
+ --predict_with_generate=False \
35
  --generation_max_length=225 \
36
  --print_training_arguments=True \
37
  --xla=True \