allow_cache: true batch_size: 8 config: ./TensorFlowTTS/examples/lightspeech/conf/lightspeech_swahiliipa.yaml dataset_config: TensorFlowTTS/preprocess/swahiliipa_preprocess.yaml dataset_mapping: dump/swahiliipa_mapper.json dataset_stats: dump/stats.npy delay_f0_energy_steps: 3 dev_dir: ./dump/valid/ energy_stat: ./dump/stats_energy.npy eval_batch_size: 16 eval_interval_steps: 5000 f0_stat: ./dump/stats_f0.npy format: npy gradient_accumulation_steps: 2 hop_size: 512 is_shuffle: true lightspeech_params: attention_probs_dropout_prob: 0.1 dataset: swahiliipa decoder_attention_head_size: 16 decoder_hidden_act: mish decoder_hidden_size: 256 decoder_intermediate_kernel_size: - 17 - 21 - 9 - 13 decoder_intermediate_size: 1024 decoder_num_attention_heads: 2 decoder_num_hidden_layers: 3 encoder_attention_head_size: 16 encoder_hidden_act: mish encoder_hidden_size: 256 encoder_intermediate_kernel_size: - 5 - 25 - 13 - 9 encoder_intermediate_size: 1024 encoder_num_attention_heads: 2 encoder_num_hidden_layers: 3 hidden_dropout_prob: 0.2 initializer_range: 0.02 max_position_embeddings: 2048 n_speakers: 1 num_mels: 80 output_attentions: false output_hidden_states: false variant_prediction_num_conv_layers: 2 variant_predictor_dropout_rate: 0.5 variant_predictor_filter: 256 variant_predictor_kernel_size: 3 log_interval_steps: 200 mel_length_threshold: 32 mixed_precision: true model_type: lightspeech num_save_intermediate_results: 1 optimizer_params: decay_steps: 150000 end_learning_rate: 5.0e-05 initial_learning_rate: 0.001 warmup_proportion: 0.02 weight_decay: 0.001 outdir: ./lightspeech-openbible pretrained: '' remove_short_samples: true resume: '' save_interval_steps: 5000 train_dir: ./dump/train/ train_max_steps: 200000 use_norm: true var_train_expr: null verbose: 1 version: '0.0'