Jobanpreet
commited on
Commit
•
21629ac
1
Parent(s):
0f080f5
Upload 12 files
Browse files- .gitattributes +4 -0
- checkpoints/glow/.gitkeep +0 -0
- checkpoints/glow/female/G_171.pth +3 -0
- checkpoints/glow/female/config.json +1 -0
- checkpoints/glow/male/G_285.pth +3 -0
- checkpoints/glow/male/config.json +1 -0
- checkpoints/hifi/.gitkeep +0 -0
- checkpoints/hifi/female/config.json +37 -0
- checkpoints/hifi/female/do_00040000 +3 -0
- checkpoints/hifi/female/g_00040000 +3 -0
- checkpoints/hifi/male/config.json +37 -0
- checkpoints/hifi/male/do_00060000 +3 -0
- checkpoints/hifi/male/g_00060000 +3 -0
.gitattributes
CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
checkpoints/hifi/female/do_00040000 filter=lfs diff=lfs merge=lfs -text
|
37 |
+
checkpoints/hifi/female/g_00040000 filter=lfs diff=lfs merge=lfs -text
|
38 |
+
checkpoints/hifi/male/do_00060000 filter=lfs diff=lfs merge=lfs -text
|
39 |
+
checkpoints/hifi/male/g_00060000 filter=lfs diff=lfs merge=lfs -text
|
checkpoints/glow/.gitkeep
ADDED
File without changes
|
checkpoints/glow/female/G_171.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:782873243045c21ccfa5181e35cc4407073c4f07de27bfd1c2ac7d59e7d0207f
|
3 |
+
size 343838316
|
checkpoints/glow/female/config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"train": {"use_cuda": true, "log_interval": 20, "seed": 1234, "epochs": 10000, "learning_rate": 1.0, "betas": [0.9, 0.98], "eps": 1e-09, "warmup_steps": 4000, "scheduler": "noam", "batch_size": 32, "ddi": true, "fp16_run": true, "save_epoch": 1}, "data": {"load_mel_from_disk": false, "training_files": "../../data/glow/female/train.txt", "validation_files": "../../data/glow/female/valid.txt", "chars": "\u0a02\u2013\u0a59\u0a28\u0a1c\u0a16\u0a25\u0a42\u0a40\u0a10\u0a07\u0a1d\u2026n\u0a5a\u0a5c\u0a3f\u0a30\u0a15\u0a14\u0a18\u0a08\u0a70\u0a1a\u0a1b\u0a13\u0a39\u0a06\u0a38b\u0a09\u0a2e\u0a5b\u0a47\u0a2c\u0a2d\u0a4c\u0a36\u0a2f\u0a24\u0a05\u0a32\u0a27\u0a41e\u0a2a\u0a20\u0a26\u0a17\u0a4d\u0a48\u0a4b\u0a0f\u0a35\u0a3e\u0a2b\u0a21\u0a3cd\u0a1f\u0a0a\u0a71\u0a22\u0a5e\u0a23", "punc": "-:. ", "text_cleaners": ["basic_indic_cleaners"], "max_wav_value": 32768.0, "sampling_rate": 22050, "filter_length": 1024, "hop_length": 256, "win_length": 1024, "n_mel_channels": 80, "mel_fmin": 80.0, "mel_fmax": 7600.0, "add_noise": true, "add_blank": true}, "model": {"hidden_channels": 192, "filter_channels": 768, "filter_channels_dp": 256, "kernel_size": 3, "p_dropout": 0.1, "n_blocks_dec": 12, "n_layers_enc": 6, "n_heads": 2, "p_dropout_dec": 0.05, "dilation_rate": 1, "kernel_size_dec": 5, "n_block_layers": 4, "n_sqz": 2, "prenet": true, "mean_only": true, "hidden_channels_enc": 192, "hidden_channels_dec": 192, "window_size": 4}}
|
checkpoints/glow/male/G_285.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:28d96b895f25f48e5e3d418d476a93a52a67a4ed406bbaa3d30972d8094bbd62
|
3 |
+
size 343880814
|
checkpoints/glow/male/config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"train": {"use_cuda": true, "log_interval": 20, "seed": 1234, "epochs": 10000, "learning_rate": 1.0, "betas": [0.9, 0.98], "eps": 1e-09, "warmup_steps": 4000, "scheduler": "noam", "batch_size": 16, "ddi": true, "fp16_run": true, "save_epoch": 1}, "data": {"load_mel_from_disk": false, "training_files": "../../data/glow/male/train.txt", "validation_files": "../../data/glow/male/valid.txt", "chars": "\u0a73\u0a26\u0a23\u0a1c\u0a71\u0a18n\u0a1f\u0a2ed\u0a2f\u0a2b\u2026\u0a5a\u0a25\u0a47\u0a5e\u0a1a\u0a27\u0a3c\u0a4d\u0a177\u0a41\u0a24e0\u20b9\u0a32\u0a2858\u0a70\u0a4b\u0a3f\u0a08\u0a42\u0a386\u0a1d3\u0a16\u0a20\u0a0a\u0a36\u0a212\u0a30\u0a05\u0a401\u0a2c\u0a1b\u0a09\u0a5b94\u0a0f\u0a2a\u0a39\u0a02\u0a14\u0a06\u0a4c\u0a15\u0a35\u0a13b\u0964\u0a59\u0a5c\u0a2d\u0a72\u0a10\u0a48\u0a07\u2013\u0a22\u0a3e", "punc": "?.!-:/, ", "text_cleaners": ["basic_indic_cleaners"], "max_wav_value": 32768.0, "sampling_rate": 22050, "filter_length": 1024, "hop_length": 256, "win_length": 1024, "n_mel_channels": 80, "mel_fmin": 80.0, "mel_fmax": 7600.0, "add_noise": true, "add_blank": true}, "model": {"hidden_channels": 192, "filter_channels": 768, "filter_channels_dp": 256, "kernel_size": 3, "p_dropout": 0.1, "n_blocks_dec": 12, "n_layers_enc": 6, "n_heads": 2, "p_dropout_dec": 0.05, "dilation_rate": 1, "kernel_size_dec": 5, "n_block_layers": 4, "n_sqz": 2, "prenet": true, "mean_only": true, "hidden_channels_enc": 192, "hidden_channels_dec": 192, "window_size": 4}}
|
checkpoints/hifi/.gitkeep
ADDED
File without changes
|
checkpoints/hifi/female/config.json
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"resblock": "1",
|
3 |
+
"num_gpus": 0,
|
4 |
+
"batch_size": 24,
|
5 |
+
"learning_rate": 0.0002,
|
6 |
+
"adam_b1": 0.8,
|
7 |
+
"adam_b2": 0.99,
|
8 |
+
"lr_decay": 0.999,
|
9 |
+
"seed": 1234,
|
10 |
+
|
11 |
+
"upsample_rates": [8,8,2,2],
|
12 |
+
"upsample_kernel_sizes": [16,16,4,4],
|
13 |
+
"upsample_initial_channel": 512,
|
14 |
+
"resblock_kernel_sizes": [3,7,11],
|
15 |
+
"resblock_dilation_sizes": [[1,3,5], [1,3,5], [1,3,5]],
|
16 |
+
|
17 |
+
"segment_size": 8192,
|
18 |
+
"num_mels": 80,
|
19 |
+
"num_freq": 1025,
|
20 |
+
"n_fft": 1024,
|
21 |
+
"hop_size": 256,
|
22 |
+
"win_size": 1024,
|
23 |
+
|
24 |
+
"sampling_rate": 22050,
|
25 |
+
|
26 |
+
"fmin": 80,
|
27 |
+
"fmax": 7600,
|
28 |
+
"fmax_for_loss": null,
|
29 |
+
|
30 |
+
"num_workers": 4,
|
31 |
+
|
32 |
+
"dist_config": {
|
33 |
+
"dist_backend": "nccl",
|
34 |
+
"dist_url": "tcp://localhost:54321",
|
35 |
+
"world_size": 1
|
36 |
+
}
|
37 |
+
}
|
checkpoints/hifi/female/do_00040000
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:12f9f70bdcfffd97667324d3741fa8716e981d664815814b09b8bb881c3acedd
|
3 |
+
size 960570609
|
checkpoints/hifi/female/g_00040000
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5c80a4bfc8ba6c8a95773ad8591250099860f8938437ae17f059f4a0db03a008
|
3 |
+
size 55829161
|
checkpoints/hifi/male/config.json
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"resblock": "1",
|
3 |
+
"num_gpus": 0,
|
4 |
+
"batch_size": 24,
|
5 |
+
"learning_rate": 0.0002,
|
6 |
+
"adam_b1": 0.8,
|
7 |
+
"adam_b2": 0.99,
|
8 |
+
"lr_decay": 0.999,
|
9 |
+
"seed": 1234,
|
10 |
+
|
11 |
+
"upsample_rates": [8,8,2,2],
|
12 |
+
"upsample_kernel_sizes": [16,16,4,4],
|
13 |
+
"upsample_initial_channel": 512,
|
14 |
+
"resblock_kernel_sizes": [3,7,11],
|
15 |
+
"resblock_dilation_sizes": [[1,3,5], [1,3,5], [1,3,5]],
|
16 |
+
|
17 |
+
"segment_size": 8192,
|
18 |
+
"num_mels": 80,
|
19 |
+
"num_freq": 1025,
|
20 |
+
"n_fft": 1024,
|
21 |
+
"hop_size": 256,
|
22 |
+
"win_size": 1024,
|
23 |
+
|
24 |
+
"sampling_rate": 22050,
|
25 |
+
|
26 |
+
"fmin": 80,
|
27 |
+
"fmax": 7600,
|
28 |
+
"fmax_for_loss": null,
|
29 |
+
|
30 |
+
"num_workers": 4,
|
31 |
+
|
32 |
+
"dist_config": {
|
33 |
+
"dist_backend": "nccl",
|
34 |
+
"dist_url": "tcp://localhost:54321",
|
35 |
+
"world_size": 1
|
36 |
+
}
|
37 |
+
}
|
checkpoints/hifi/male/do_00060000
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f4608a77b2d962dc4e2c937541790d49e93097239ef1688291b73edec030e1ff
|
3 |
+
size 960570609
|
checkpoints/hifi/male/g_00060000
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a31c67d611f2da234318edd7d22d7446679694a80ec2aab40f25fcf723358e91
|
3 |
+
size 55829161
|