Nithu commited on
Commit
6489817
1 Parent(s): 641d75b

Upload 7 files

Browse files
config.yaml ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ features:
2
+ energy_max: 3.2244551181793213
3
+ energy_min: -4.9544901847839355
4
+ eps: 1.0e-05
5
+ f_max: 8000
6
+ f_min: 0
7
+ hop_len_t: 0.011609977324263039
8
+ hop_length: 256
9
+ n_fft: 1024
10
+ n_mels: 80
11
+ n_stft: 513
12
+ pitch_max: 5.733940816898645
13
+ pitch_min: -4.660287183665281
14
+ sample_rate: 22050
15
+ type: spectrogram+melscale+log
16
+ win_len_t: 0.046439909297052155
17
+ win_length: 1024
18
+ window_fn: hann
19
+ global_cmvn:
20
+ stats_npz_path: fbank_mfa_gcmvn_stats.npz
21
+ transforms:
22
+ '*':
23
+ - global_cmvn
24
+ vocab_filename: vocab.txt
25
+ vocoder:
26
+ type: hifigan
27
+ config: hifigan.json
28
+ checkpoint: hifigan.bin
29
+ hub:
30
+ phonemizer: g2p
fbank_mfa_gcmvn_stats.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6735b35875c2614cee80bf861c6a604aba35671887e6f04b4449dc257bb15d34
3
+ size 1140
hifigan.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4f4f016c791fd9ca9859a9e25e7eb0a823fee2ea997c1e5ae8e1a9ea5f99b1f
3
+ size 55825897
hifigan.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "resblock": "1",
3
+ "num_gpus": 0,
4
+ "batch_size": 16,
5
+ "learning_rate": 0.0002,
6
+ "adam_b1": 0.8,
7
+ "adam_b2": 0.99,
8
+ "lr_decay": 0.999,
9
+ "seed": 1234,
10
+
11
+ "upsample_rates": [8,8,2,2],
12
+ "upsample_kernel_sizes": [16,16,4,4],
13
+ "upsample_initial_channel": 512,
14
+ "resblock_kernel_sizes": [3,7,11],
15
+ "resblock_dilation_sizes": [[1,3,5], [1,3,5], [1,3,5]],
16
+
17
+ "segment_size": 8192,
18
+ "num_mels": 80,
19
+ "num_freq": 1025,
20
+ "n_fft": 1024,
21
+ "hop_size": 256,
22
+ "win_size": 1024,
23
+
24
+ "sampling_rate": 22050,
25
+
26
+ "fmin": 0,
27
+ "fmax": 8000,
28
+ "fmax_for_loss": null,
29
+
30
+ "num_workers": 4,
31
+
32
+ "dist_config": {
33
+ "dist_backend": "nccl",
34
+ "dist_url": "tcp://localhost:54321",
35
+ "world_size": 1
36
+ }
37
+ }
pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a48d454fe66939079d0ddb70f1c062ec669f521a7cfadc608968746e312986ab
3
+ size 494816801
run_fast_speech_2.py ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ #!/usr/bin/env python3
2
+ from fairseq.checkpoint_utils import load_model_ensemble_and_task
3
+
4
+ # model = load_model_ensemble_and_task(["./pytorch_model.pt"], arg_overrides={"config_yaml": "./config.yaml", "data": "./"})
5
+ model = load_model_ensemble_and_task(["./pytorch_model.pt"], arg_overrides={"data": "./"})
vocab.txt ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ AH0 71007
2
+ N 63410
3
+ T 60842
4
+ S 40263
5
+ D 39886
6
+ R 35965
7
+ L 30358
8
+ sp 27584
9
+ IH0 27113
10
+ DH 26584
11
+ K 25851
12
+ IH1 25683
13
+ Z 25387
14
+ EH1 21690
15
+ AE1 21648
16
+ M 21537
17
+ W 18760
18
+ P 18458
19
+ ER0 18446
20
+ V 18169
21
+ IY0 17832
22
+ AH1 16995
23
+ F 15549
24
+ B 14227
25
+ HH 13468
26
+ IY1 12751
27
+ EY1 12141
28
+ AO1 11595
29
+ AA1 10589
30
+ AY1 9624
31
+ UW1 8865
32
+ SH 7449
33
+ OW1 7441
34
+ NG 6705
35
+ G 5472
36
+ ER1 4898
37
+ Y 4548
38
+ JH 4486
39
+ CH 4355
40
+ TH 3980
41
+ AW1 3607
42
+ UH1 2469
43
+ EH2 1881
44
+ spn 1774
45
+ AO0 1357
46
+ OW0 1328
47
+ EY2 1258
48
+ IH2 1251
49
+ AE2 1104
50
+ UW0 1077
51
+ AY2 1062
52
+ AA2 774
53
+ OY1 771
54
+ AO2 622
55
+ ZH 587
56
+ EH0 568
57
+ OW2 557
58
+ EY0 443
59
+ IY2 435
60
+ UW2 431
61
+ AY0 390
62
+ AE0 374
63
+ AH2 316
64
+ AW2 290
65
+ AA0 259
66
+ ER2 136
67
+ UH2 127
68
+ OY2 44
69
+ UH0 36
70
+ AW0 35
71
+ OY0 4