sangeet2020 commited on
Commit
720f1ce
1 Parent(s): 602c4eb

remove old hyperparam file

Browse files
Files changed (1) hide show
  1. hyperparams.yaml +0 -105
hyperparams.yaml DELETED
@@ -1,105 +0,0 @@
1
- # Generated 2023-06-24 from:
2
- # /netscratch/sagar/thesis/speechbrain/recipes/RescueSpeech/Enhancement/joint-training/transformers/hparams/robust_asr_16k.yaml
3
- # yamllint disable
4
- # Model: wav2vec2 + DNN + CTC
5
- # Augmentation: SpecAugment
6
- # Authors: Sangeet Sagar 2023
7
- # ################################
8
-
9
- # URL for the biggest whisper model.
10
- # URL for the biggest Fairseq english whisper model.
11
- whisper_hub: openai/whisper-large-v2
12
- language: german
13
-
14
- ## Model parameters
15
- sample_rate: 16000
16
- freeze_whisper: false
17
- freeze_encoder_only: false
18
- freeze_encoder: true
19
-
20
- # These values are only used for the searchers.
21
- # They needs to be hardcoded and should not be changed with Whisper.
22
- # They are used as part of the searching process.
23
- # The bos token of the searcher will be timestamp_index
24
- # and will be concatenated with the bos, language and task tokens.
25
- timestamp_index: 50363
26
- eos_index: 50257
27
- bos_index: 50258
28
-
29
- # Decoding parameters
30
- min_decode_ratio: 0.0
31
- max_decode_ratio: 1.0
32
- test_beam_size: 8
33
-
34
- num_spks: 1
35
-
36
- # Enhancement model
37
- Encoder: &id004 !new:speechbrain.lobes.models.dual_path.Encoder
38
- kernel_size: 16
39
- out_channels: 256
40
-
41
- SBtfintra: &id001 !new:speechbrain.lobes.models.dual_path.SBTransformerBlock
42
- num_layers: 8
43
- d_model: 256
44
- nhead: 8
45
- d_ffn: 1024
46
- dropout: 0
47
- use_positional_encoding: true
48
- norm_before: true
49
-
50
- SBtfinter: &id002 !new:speechbrain.lobes.models.dual_path.SBTransformerBlock
51
- num_layers: 8
52
- d_model: 256
53
- nhead: 8
54
- d_ffn: 1024
55
- dropout: 0
56
- use_positional_encoding: true
57
- norm_before: true
58
-
59
- MaskNet: &id005 !new:speechbrain.lobes.models.dual_path.Dual_Path_Model
60
- num_spks: 1
61
- in_channels: 256
62
- out_channels: 256
63
- num_layers: 2
64
- K: 250
65
- intra_model: *id001
66
- inter_model: *id002
67
- norm: ln
68
- linear_layer_after_inter_intra: false
69
- skip_around_intra: true
70
-
71
- # Whisper ASR and its decoder
72
- Decoder: &id006 !new:speechbrain.lobes.models.dual_path.Decoder
73
- in_channels: 256
74
- out_channels: 1
75
- kernel_size: 16
76
- stride: 8
77
- bias: false
78
-
79
- whisper: &id003 !new:speechbrain.lobes.models.huggingface_whisper.HuggingFaceWhisper
80
- source: !ref <whisper_hub>
81
- freeze: !ref <freeze_whisper>
82
- freeze_encoder: !ref <freeze_encoder>
83
- save_path: whisper_checkpoints
84
- encoder_only: False
85
-
86
- decoder: !new:speechbrain.decoders.seq2seq.S2SWhisperGreedySearch
87
- model: *id003
88
- bos_index: 50363
89
- eos_index: 50257
90
- min_decode_ratio: 0.0
91
- max_decode_ratio: 1.0
92
-
93
- # Change the path to use a local model instead of the remote one
94
- pretrainer: !new:speechbrain.utils.parameter_transfer.Pretrainer
95
- loadables:
96
- encoder: !ref <Encoder>
97
- masknet: !ref <MaskNet>
98
- decoder: !ref <Decoder>
99
- whisper: !ref <whisper>
100
-
101
- modules:
102
- encoder: *id004
103
- masknet: *id005
104
- decoder: *id006
105
- whisper: *id003