train: seed: 1234 epochs: 100 batch_size: 6 gradient_accumulation: 4 save_every_n_epoch: 1 precision: 32 gradient_clip: 1.0 optimizer: lr: 0.01 lr_init: 0.00001 lr_end: 0.0001 warmup_steps: 2000 decay_steps: 40000 data: max_eval_sample: 8 max_sec: 40 num_workers: 1 pad_val: 1024 # same with EOS in model model: saving_path: "ckpt/" resume_checkpoint: null vocoder_config_path: "quantizer/new_ckpt/config.json" vocoder_ckpt_path: "quantizer/new_ckpt/g_00600000" datadir: "/home/liweiche/GigaSpeech/wavs" metapath: "/home/liweiche/GigaSpeech/train2.json" val_metapath: "/home/liweiche/GigaSpeech/dev2.json" sampledir: "logs/" pretrained_path: null lr: 0.0001 batch_size: 200.0 train_bucket_size: 8192 training_step: 800000 optim_flat_percent: 0.0 warmup_step: 50 adam_beta1: 0.9 adam_beta2: 0.98 ffd_size: 3072 hidden_size: 768 enc_nlayers: 6 dec_nlayers: 6 nheads: 12 ar_layer: 4 ar_ffd_size: 1024 ar_hidden_size: 256 ar_nheads: 4 aligner_softmax_temp: 1.0 layer_norm_eps: 0.00001 speaker_embed_dropout: 0.05 label_smoothing: 0.0 val_check_interval: 5000 check_val_every_n_epoch: 1 precision: "fp16" nworkers: 16 distributed: true accelerator: "ddp" version: null accumulate_grad_batches: 1 use_repetition_token: true use_repetition_gating: false repetition_penalty: 1.0 sampling_temperature: 1.0 top_k: -1 min_top_k: 3 top_p: 0.8 sample_num: 4 length_penalty_max_length: 15000 length_penalty_max_prob: 0.95 max_input_length: 2048 max_output_length: 2000 sample_rate: 16000 n_codes: 1024 n_cluster_groups: 1 phone_context_window: 4 phoneset_size: 1000 inference: top_k: 5