# Generated 2023-06-24 from: # /netscratch/sagar/thesis/speechbrain/recipes/RescueSpeech/Enhancement/joint-training/transformers/hparams/robust_asr_16k.yaml # yamllint disable # Model: wav2vec2 + DNN + CTC # Augmentation: SpecAugment # Authors: Sangeet Sagar 2023 # ################################ ## Model parameters sample_rate: 16000 # Decoding parameters min_decode_ratio: 0.0 max_decode_ratio: 1.0 test_beam_size: 8 num_spks: 1 # Enhancement model Encoder: &id004 !new:speechbrain.lobes.models.dual_path.Encoder kernel_size: 16 out_channels: 256 SBtfintra: &id001 !new:speechbrain.lobes.models.dual_path.SBTransformerBlock num_layers: 8 d_model: 256 nhead: 8 d_ffn: 1024 dropout: 0 use_positional_encoding: true norm_before: true SBtfinter: &id002 !new:speechbrain.lobes.models.dual_path.SBTransformerBlock num_layers: 8 d_model: 256 nhead: 8 d_ffn: 1024 dropout: 0 use_positional_encoding: true norm_before: true MaskNet: &id005 !new:speechbrain.lobes.models.dual_path.Dual_Path_Model num_spks: 1 in_channels: 256 out_channels: 256 num_layers: 2 K: 250 intra_model: *id001 inter_model: *id002 norm: ln linear_layer_after_inter_intra: false skip_around_intra: true # Whisper ASR and its decoder Decoder: &id006 !new:speechbrain.lobes.models.dual_path.Decoder in_channels: 256 out_channels: 1 kernel_size: 16 stride: 8 bias: false # Change the path to use a local model instead of the remote one pretrainer: !new:speechbrain.utils.parameter_transfer.Pretrainer loadables: encoder: !ref masknet: !ref decoder: !ref modules: encoder: *id004 masknet: *id005 decoder: *id006