flax-wav2vec2-2-bart-large / wer-sweep.yaml
sanchit-gandhi's picture
up
572827c
command:
- python3
- ${program}
- --overwrite_output_dir
- --freeze_feature_encoder
- --predict_with_generate
- --do_lower_case
- --do_train
- --do_eval
- ${args}
method: random
metric:
goal: minimize
name: eval/wer
parameters:
dataset_cache_dir:
value: /home/sanchitgandhi/cache/huggingface/datasets
dataset_config_name:
value: clean
dataset_name:
value: librispeech_asr
eval_split_name:
value: validation
eval_steps:
value: 500
generation_max_length:
value: 40
generation_num_beams:
value: 1
gradient_accumulation_steps:
value: 1
gradient_checkpointing:
value: True
learning_rate:
distribution: log_uniform
max: -6.9
min: -9.2
length_column_name:
value: input_length
logging_steps:
value: 25
max_duration_in_seconds:
value: 20
max_target_length:
value: 128
model_name_or_path:
value: ./
num_train_epochs:
value: 3
output_dir:
value: ./output_dir
per_device_eval_batch_size:
value: 2
per_device_train_batch_size:
value: 1
preprocessing_num_workers:
value: 16
text_column_name:
value: text
train_split_name:
value: train.100
warmup_steps:
value: 500
program: run_flax_speech_recognition_seq2seq.py
project: flax-wav2vec2-2-bart-large-checkpointing-scan