Kamtera commited on
Commit
a6fba01
1 Parent(s): a2264fb

Upload train_wavernn.py with huggingface_hub

Browse files
Files changed (1) hide show
  1. train_wavernn.py +72 -0
train_wavernn.py ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+
3
+ from trainer import Trainer, TrainerArgs
4
+
5
+ from TTS.utils.audio import AudioProcessor
6
+ from TTS.vocoder.configs import WavernnConfig
7
+ from TTS.vocoder.datasets.preprocess import load_wav_data
8
+ from TTS.vocoder.models.wavernn import Wavernn
9
+ from TTS.config.shared_configs import BaseAudioConfig
10
+ from TTS.tts.configs.shared_configs import BaseDatasetConfig , CharactersConfig
11
+ from TTS.tts.datasets import load_tts_samples
12
+
13
+
14
+
15
+
16
+ output_path = os.path.dirname(os.path.abspath(__file__))
17
+ dataset_config = BaseDatasetConfig(
18
+ formatter="mozilla", meta_file_train="metadata.csv", path="/kaggle/input/persian-tts-dataset-famale"
19
+ )
20
+ audio_config = BaseAudioConfig(
21
+ sample_rate=24000,
22
+ do_trim_silence=True,
23
+ resample=False,
24
+ mel_fmin=95,
25
+ mel_fmax=8000.0,
26
+
27
+
28
+ )
29
+
30
+ config = WavernnConfig(
31
+ batch_size=64,#
32
+ eval_batch_size=16,#
33
+ num_loader_workers=1,
34
+ num_eval_loader_workers=1,
35
+ run_eval=True,
36
+ test_delay_epochs=-1,
37
+ epochs=1000,
38
+ seq_len=1280,
39
+ pad_short=2000,
40
+ use_noise_augment=False,
41
+ save_step=1000,
42
+ eval_split_size=10,
43
+ print_step=25,
44
+ print_eval=True,
45
+ mixed_precision=False,
46
+ lr=1e-4,
47
+ data_path="/kaggle/input/persian-tts-dataset-famale/wavs/",
48
+ output_path=output_path,
49
+ audio=audio_config,
50
+
51
+ )
52
+
53
+ # init audio processor
54
+ ap = AudioProcessor(**config.audio.to_dict())
55
+
56
+ # load training samples
57
+ eval_samples, train_samples = load_wav_data(config.data_path, config.eval_split_size)
58
+
59
+ # init model
60
+ model = Wavernn(config)
61
+
62
+ # init the trainer and 🚀
63
+ trainer = Trainer(
64
+ TrainerArgs(),
65
+ config,
66
+ output_path,
67
+ model=model,
68
+ train_samples=train_samples,
69
+ eval_samples=eval_samples,
70
+ training_assets={"audio_processor": ap},
71
+ )
72
+ trainer.fit()