File size: 7,329 Bytes
81d6d86
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
{
    "run_name": "Wav2Vec-fine-tuning-TEDx",
    "run_description": "Fine tuning TEDx",
    "seed": 42,
    // AUDIO PARAMS
    "sampling_rate": 16000,

    // VOCABULARY PARAMETERS
    "vocab":{
        "vocab_path": "example/vocab_example_ru.json", // generic vocab for Portuguese
        "blank": "<pad>", // blank token for padding
        "silence": "|", // token between words
        "unk": "<unk>" // unk token
    },

    // TRAINING
    "batch_size": 8,       // Batch size for training.
    "mixed_precision": true,     // level of optimization with NVIDIA's apex feature for automatic mixed FP16/FP32 precision (AMP), NOTE: currently only O1 is supported, and use "O1" to activate.
    "early_stop_epochs": 10, // If 0 disabled else Number of epochs for stop training with validation loss dont decrease 
    "preprocess_dataset": false, // if true, the dataset will be pre-processed and saved in disk, otherwise the audio files will be loaded in each step. Preprocessing makes training faster, but requires much more disk space.

    // OPTIMIZER
    "epochs": 100,                // total number of epochs to train.
    "lr": 0.00003,                  // Initial learning rate.
    "gradient_accumulation_steps": 24, 

    // LOGGING
    "logging_steps": 100,    // Number of steps to plot.
    "load_best_model_at_end": true,
    "save_total_limit": 3,
    "warmup_ratio": 0.04761904762142857, // 0 disable Ratio of total training steps used for a linear warmup from 0 to learning_rate
    "warmup_steps": 0, // 0 disable  Number of steps used for a linear warmup from 0 to learning_rate

    // DATA LOADING
    "num_loader_workers": 8,        // number of training data loader processes. Don't set it too big. 4-8 are goo
    
    // MODEL
    "freeze_feature_extractor": true, // Whether to freeze the feature extractor layers of the model.
    "attention_dropout": 0.1, // The dropout ratio for the attention probabilities.
    "activation_dropout": 0.1, // The dropout ratio for activations inside the fully connected layer.
    "hidden_dropout": 0.1, // The dropout probabilitiy for all fully connected layers in the embeddings, encoder, and pooler.
    "feat_proj_dropout": 0.1, // The dropout probabilitiy for all 1D convolutional layers in feature extractor.
    "mask_time_prob": 0.05, //  Propability of each feature vector along the time axis to be chosen as the start of the vector span to be masked.
    "layerdrop": 0.0,  // The LayerDrop probability. 
    "gradient_checkpointing": true,  // If True, use gradient checkpointing to save memory at the expense of slower backward pass.

    // ToDo: Implement Time mask and Frequency Mask
    "audio_augmentation":[
        // additive noise and room impulse response (RIR) simulation similar to: https://arxiv.org/pdf/2009.14153.pdf
        {
            "name": "additive",
            "sounds_path":"../../datasets/musan/speech/", // download: https://www.openslr.org/17/
            "lru_cache_size": 32, // Maximum size of the LRU cache for storing noise files in memory
            "min_snr_in_db": 13.0,
            "max_snr_in_db": 20.0,
            // "sample_rate": 16000,
            "p": 0.25
        },
        {
            "name": "additive",
            "sounds_path":"../../datasets/musan/music/", // download: https://www.openslr.org/17/
            "lru_cache_size": 32, // Maximum size of the LRU cache for storing noise files in memory
            "min_snr_in_db": 5.0,
            "max_snr_in_db": 15.0,
            // "sample_rate": 16000,
            "p": 0.25
        },
        {
            "name": "additive",
            "sounds_path":"../../datasets/musan/noise/", // download: https://www.openslr.org/17/
            "lru_cache_size": 32, // Maximum size of the LRU cache for storing noise files in memory
            "min_snr_in_db": 0.0,
            "max_snr_in_db": 15.0,
            // "sample_rate": 16000,
            "p": 0.25
        },
        // rir filter proposed by: https://ieeexplore.ieee.org/document/7953152
        {
            "name": "rir",
            "ir_path": "../../datasets/RIRS_NOISES/simulated_rirs/", // download: https://www.openslr.org/28/
            "lru_cache_size": 128, // Maximum size of the LRU cache for storing noise files in memory
            // "sample_rate": 16000,
            "p": 0.25
        }
        , 
        // {
        //     "name": "gain",
        //     "min_gain_in_db": -18.0,
        //     "max_gain_in_db": 6,
        //     "p": 0.25 // propability of apply this method, 0 is disable
        // },
        {
            "name": "pitch_shift",
            "min_semitones": -4,
            "max_semitones": 4,
            "p": 0.25 // propability of apply this method, 0 is disable
        },
        {
            "name": "gaussian",
            "min_amplitude": 0.0001,
            "max_amplitude": 0.001,
            "p": 0.25 // propability of apply this method, 0 is disable
        }
    ],

    // PATHS
    "output_path": "../checkpoints/Wav2Vec-voxpopuli/one-speaker/Final-paper/GT/RU/100-epoch/",
    // CACHE
    "dataset_cache": "../datasets/",

    // DATASETS
    "datasets":{ 
        
        "files_path": "/workspace/edresson/datasets/Common_Voice/cv-corpus-7.0-2021-07-21/ru/", // relative path for audios It's will be join with the CS
        "train":
            [
               // this dicts is pass directly for the load dataset see the documentation: https://huggingface.co/docs/datasets/package_reference/loading_methods.html#datasets.load_dataset 
            {
                "name": "csv",
                "path": "csv",
                "data_files": ["/workspace/edresson/datasets/Common_Voice/cv-corpus-7.0-2021-07-21/ru/train_converted.csv"], // csv files
                "text_column": "text",
                "path_column": "file_path"
            }, 
            {                
                "name": "csv",
                "path": "csv",
                "data_files": ["/workspace/edresson/datasets/M-AILABS/ru_RU/train_converted.csv"], // csv files
                "text_column": "text",
                "path_column": "file_path"
            }
            ]
        ,
        "devel":
            [ 
                {
                    "name": "csv",
                    "path": "csv",
                    "data_files": ["/workspace/edresson/datasets/Common_Voice/cv-corpus-7.0-2021-07-21/ru/dev_converted.csv"], // csv files
                    "text_column": "text",
                    "path_column": "file_path"
                }
        
            ] 
            ,
        "test":
            {
                "name": "csv",
                "path": "csv",
                "data_files": ["/workspace/edresson/datasets/Common_Voice/cv-corpus-7.0-2021-07-21/ru/test_converted.csv"], // csv files
                "text_column": "text",
                "path_column": "file_path"
            }
    
    }//,
    // used only for test 
    // "KenLM":{
    // "kenlm_model_path": "../../kenLM/binaries/subtitle/4-gram/lm.binary", // Path for KenLM model
    // "lexicon_path": "example/lexicon.lst", // file with all words for limit the decoder search
    // "beam": 2048, 
    // "nbest": 1,
    // "beam_threshold": 25, 
    // "lm_weight": 1, 
    // "word_score": -1,
    // "sil_weight": 0
    // }



}