litagin's picture
Init
26be912
raw
history blame
1.72 kB
train:
seed: 1234
epochs: 100
batch_size: 6
gradient_accumulation: 4
save_every_n_epoch: 1
precision: 32
gradient_clip: 1.0
optimizer:
lr: 0.01
lr_init: 0.00001
lr_end: 0.0001
warmup_steps: 2000
decay_steps: 40000
data:
max_eval_sample: 8
max_sec: 40
num_workers: 1
pad_val: 1024 # same with EOS in model
model:
saving_path: "ckpt/"
resume_checkpoint: null
vocoder_config_path: "quantizer/new_ckpt/config.json"
vocoder_ckpt_path: "quantizer/new_ckpt/g_00600000"
datadir: "/home/liweiche/GigaSpeech/wavs"
metapath: "/home/liweiche/GigaSpeech/train2.json"
val_metapath: "/home/liweiche/GigaSpeech/dev2.json"
sampledir: "logs/"
pretrained_path: null
lr: 0.0001
batch_size: 200.0
train_bucket_size: 8192
training_step: 800000
optim_flat_percent: 0.0
warmup_step: 50
adam_beta1: 0.9
adam_beta2: 0.98
ffd_size: 3072
hidden_size: 768
enc_nlayers: 6
dec_nlayers: 6
nheads: 12
ar_layer: 4
ar_ffd_size: 1024
ar_hidden_size: 256
ar_nheads: 4
aligner_softmax_temp: 1.0
layer_norm_eps: 0.00001
speaker_embed_dropout: 0.05
label_smoothing: 0.0
val_check_interval: 5000
check_val_every_n_epoch: 1
precision: "fp16"
nworkers: 16
distributed: true
accelerator: "ddp"
version: null
accumulate_grad_batches: 1
use_repetition_token: true
use_repetition_gating: false
repetition_penalty: 1.0
sampling_temperature: 1.0
top_k: -1
min_top_k: 3
top_p: 0.8
sample_num: 4
length_penalty_max_length: 15000
length_penalty_max_prob: 0.95
max_input_length: 2048
max_output_length: 2000
sample_rate: 16000
n_codes: 1024
n_cluster_groups: 1
phone_context_window: 4
phoneset_size: 1000
inference:
top_k: 5