cemsubakan commited on
Commit
eb43c5b
1 Parent(s): 2ef3a73

Update hyperparams.yaml

Browse files
Files changed (1) hide show
  1. hyperparams.yaml +10 -10
hyperparams.yaml CHANGED
@@ -15,11 +15,11 @@ kernel_size: 16
15
  kernel_stride: 8
16
 
17
  # Specifying the network
18
- Encoder: &id003 !new:speechbrain.lobes.models.dual_path.Encoder
19
  kernel_size: 16
20
  out_channels: 256
21
 
22
- SBtfintra: &id001 !new:speechbrain.lobes.models.dual_path.SBTransformerBlock
23
  num_layers: 8
24
  d_model: 256
25
  nhead: 8
@@ -28,7 +28,7 @@ SBtfintra: &id001 !new:speechbrain.lobes.models.dual_path.SBTransformerBlock
28
  use_positional_encoding: true
29
  norm_before: true
30
 
31
- SBtfinter: &id002 !new:speechbrain.lobes.models.dual_path.SBTransformerBlock
32
  num_layers: 8
33
  d_model: 256
34
  nhead: 8
@@ -37,19 +37,19 @@ SBtfinter: &id002 !new:speechbrain.lobes.models.dual_path.SBTransformerBlock
37
  use_positional_encoding: true
38
  norm_before: true
39
 
40
- MaskNet: &id005 !new:speechbrain.lobes.models.dual_path.Dual_Path_Model
41
  num_spks: 2
42
  in_channels: 256
43
  out_channels: 256
44
  num_layers: 2
45
  K: 250
46
- intra_model: *id001
47
- inter_model: *id002
48
  norm: ln
49
  linear_layer_after_inter_intra: false
50
  skip_around_intra: true
51
 
52
- Decoder: &id004 !new:speechbrain.lobes.models.dual_path.Decoder
53
  in_channels: 256
54
  out_channels: 1
55
  kernel_size: 16
@@ -57,9 +57,9 @@ Decoder: &id004 !new:speechbrain.lobes.models.dual_path.Decoder
57
  bias: false
58
 
59
  modules:
60
- encoder: *id003
61
- decoder: *id004
62
- masknet: *id005
63
 
64
  pretrainer: !new:speechbrain.utils.parameter_transfer.Pretrainer
65
  loadables:
15
  kernel_stride: 8
16
 
17
  # Specifying the network
18
+ Encoder: !new:speechbrain.lobes.models.dual_path.Encoder
19
  kernel_size: 16
20
  out_channels: 256
21
 
22
+ SBtfintra: !new:speechbrain.lobes.models.dual_path.SBTransformerBlock
23
  num_layers: 8
24
  d_model: 256
25
  nhead: 8
28
  use_positional_encoding: true
29
  norm_before: true
30
 
31
+ SBtfinter: !new:speechbrain.lobes.models.dual_path.SBTransformerBlock
32
  num_layers: 8
33
  d_model: 256
34
  nhead: 8
37
  use_positional_encoding: true
38
  norm_before: true
39
 
40
+ MaskNet: !new:speechbrain.lobes.models.dual_path.Dual_Path_Model
41
  num_spks: 2
42
  in_channels: 256
43
  out_channels: 256
44
  num_layers: 2
45
  K: 250
46
+ intra_model: !ref <SBtfintra>
47
+ inter_model: !ref <SBtfinter>
48
  norm: ln
49
  linear_layer_after_inter_intra: false
50
  skip_around_intra: true
51
 
52
+ Decoder: !new:speechbrain.lobes.models.dual_path.Decoder
53
  in_channels: 256
54
  out_channels: 1
55
  kernel_size: 16
57
  bias: false
58
 
59
  modules:
60
+ encoder: !ref <Encoder>
61
+ decoder: !ref <Decoder>
62
+ masknet: !ref <MaskNet>
63
 
64
  pretrainer: !new:speechbrain.utils.parameter_transfer.Pretrainer
65
  loadables: