CooperElektrik commited on
Commit
742afce
1 Parent(s): d4251af

Add Elaine's model

Browse files
app.py CHANGED
@@ -78,12 +78,6 @@ if __name__ == '__main__':
78
  args = parser.parse_args()
79
  hubert_model = utils.get_hubert_model().to(args.device)
80
  models = []
81
- others = {
82
- "rudolf": "https://huggingface.co/spaces/sayashi/sovits-rudolf",
83
- "teio": "https://huggingface.co/spaces/sayashi/sovits-teio",
84
- "goldship": "https://huggingface.co/spaces/sayashi/sovits-goldship",
85
- "tannhauser": "https://huggingface.co/spaces/sayashi/sovits-tannhauser"
86
- }
87
  voices = []
88
  tts_voice_list = asyncio.get_event_loop().run_until_complete(edge_tts.list_voices())
89
  for r in tts_voice_list:
@@ -95,12 +89,8 @@ if __name__ == '__main__':
95
  models.append((name, cover, create_vc_fn(model, name)))
96
  with gr.Blocks() as app:
97
  gr.Markdown(
98
- "# <center> Sovits Models\n"
99
  "## <center> The input audio should be clean and pure voice without background music.\n"
100
- "![visitor badge](https://visitor-badge.glitch.me/badge?page_id=sayashi.Sovits-Umamusume)\n\n"
101
- "[![image](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1wfsBbMzmtLflOJeqc5ZnJiLY7L239hJW?usp=share_link)\n\n"
102
- "[![Duplicate this Space](https://huggingface.co/datasets/huggingface/badges/raw/main/duplicate-this-space-sm-dark.svg)](https://huggingface.co/spaces/sayashi/sovits-models?duplicate=true)\n\n"
103
- "[![Original Repo](https://badgen.net/badge/icon/github?icon=github&label=Original%20Repo)](https://github.com/svc-develop-team/so-vits-svc)"
104
 
105
  )
106
  with gr.Tabs():
 
78
  args = parser.parse_args()
79
  hubert_model = utils.get_hubert_model().to(args.device)
80
  models = []
 
 
 
 
 
 
81
  voices = []
82
  tts_voice_list = asyncio.get_event_loop().run_until_complete(edge_tts.list_voices())
83
  for r in tts_voice_list:
 
89
  models.append((name, cover, create_vc_fn(model, name)))
90
  with gr.Blocks() as app:
91
  gr.Markdown(
92
+ "# <center> KoMETA AI Demo\n"
93
  "## <center> The input audio should be clean and pure voice without background music.\n"
 
 
 
 
94
 
95
  )
96
  with gr.Tabs():
models/alice/cover.png DELETED
Binary file (743 kB)
 
models/{alice → elaine}/config.json RENAMED
@@ -1,93 +1,98 @@
1
- {
2
- "train": {
3
- "log_interval": 200,
4
- "eval_interval": 800,
5
- "seed": 1234,
6
- "epochs": 10000,
7
- "learning_rate": 0.0001,
8
- "betas": [
9
- 0.8,
10
- 0.99
11
- ],
12
- "eps": 1e-09,
13
- "batch_size": 32,
14
- "fp16_run": false,
15
- "lr_decay": 0.999875,
16
- "segment_size": 10240,
17
- "init_lr_ratio": 1,
18
- "warmup_epochs": 0,
19
- "c_mel": 45,
20
- "c_kl": 1.0,
21
- "use_sr": true,
22
- "max_speclen": 512,
23
- "port": "8001",
24
- "keep_ckpts": 99
25
- },
26
- "data": {
27
- "training_files": "filelists/train.txt",
28
- "validation_files": "filelists/val.txt",
29
- "max_wav_value": 32768.0,
30
- "sampling_rate": 44100,
31
- "filter_length": 2048,
32
- "hop_length": 512,
33
- "win_length": 2048,
34
- "n_mel_channels": 80,
35
- "mel_fmin": 0.0,
36
- "mel_fmax": 22050
37
- },
38
- "model": {
39
- "inter_channels": 192,
40
- "hidden_channels": 192,
41
- "filter_channels": 768,
42
- "n_heads": 2,
43
- "n_layers": 6,
44
- "kernel_size": 3,
45
- "p_dropout": 0.1,
46
- "resblock": "1",
47
- "resblock_kernel_sizes": [
48
- 3,
49
- 7,
50
- 11
51
- ],
52
- "resblock_dilation_sizes": [
53
- [
54
- 1,
55
- 3,
56
- 5
57
- ],
58
- [
59
- 1,
60
- 3,
61
- 5
62
- ],
63
- [
64
- 1,
65
- 3,
66
- 5
67
- ]
68
- ],
69
- "upsample_rates": [
70
- 8,
71
- 8,
72
- 2,
73
- 2,
74
- 2
75
- ],
76
- "upsample_initial_channel": 512,
77
- "upsample_kernel_sizes": [
78
- 16,
79
- 16,
80
- 4,
81
- 4,
82
- 4
83
- ],
84
- "n_layers_q": 3,
85
- "use_spectral_norm": false,
86
- "gin_channels": 256,
87
- "ssl_dim": 256,
88
- "n_speakers": 200
89
- },
90
- "spk": {
91
- "alice": 0
92
- }
 
 
 
 
 
93
  }
 
1
+ {
2
+ "train": {
3
+ "log_interval": 200,
4
+ "eval_interval": 8000,
5
+ "seed": 70235,
6
+ "epochs": 500,
7
+ "learning_rate": 0.0002,
8
+ "betas": [
9
+ 0.8,
10
+ 0.99
11
+ ],
12
+ "eps": 1e-09,
13
+ "batch_size": 8,
14
+ "fp16_run": false,
15
+ "bf16_run": true,
16
+ "lr_decay": 0.999875,
17
+ "segment_size": 10240,
18
+ "init_lr_ratio": 1,
19
+ "warmup_epochs": 0,
20
+ "c_mel": 45,
21
+ "c_kl": 1.0,
22
+ "use_sr": true,
23
+ "max_speclen": 512,
24
+ "port": "8001",
25
+ "keep_ckpts": 3,
26
+ "num_workers": 4,
27
+ "log_version": 0,
28
+ "ckpt_name_by_step": false,
29
+ "accumulate_grad_batches": 1
30
+ },
31
+ "data": {
32
+ "training_files": "filelists/44k/train.txt",
33
+ "validation_files": "filelists/44k/val.txt",
34
+ "max_wav_value": 32768.0,
35
+ "sampling_rate": 44100,
36
+ "filter_length": 2048,
37
+ "hop_length": 512,
38
+ "win_length": 2048,
39
+ "n_mel_channels": 80,
40
+ "mel_fmin": 0.0,
41
+ "mel_fmax": 22050
42
+ },
43
+ "model": {
44
+ "inter_channels": 192,
45
+ "hidden_channels": 192,
46
+ "filter_channels": 768,
47
+ "n_heads": 2,
48
+ "n_layers": 6,
49
+ "kernel_size": 3,
50
+ "p_dropout": 0.1,
51
+ "resblock": "1",
52
+ "resblock_kernel_sizes": [
53
+ 3,
54
+ 7,
55
+ 11
56
+ ],
57
+ "resblock_dilation_sizes": [
58
+ [
59
+ 1,
60
+ 3,
61
+ 5
62
+ ],
63
+ [
64
+ 1,
65
+ 3,
66
+ 5
67
+ ],
68
+ [
69
+ 1,
70
+ 3,
71
+ 5
72
+ ]
73
+ ],
74
+ "upsample_rates": [
75
+ 8,
76
+ 8,
77
+ 2,
78
+ 2,
79
+ 2
80
+ ],
81
+ "upsample_initial_channel": 512,
82
+ "upsample_kernel_sizes": [
83
+ 16,
84
+ 16,
85
+ 4,
86
+ 4,
87
+ 4
88
+ ],
89
+ "n_layers_q": 3,
90
+ "use_spectral_norm": false,
91
+ "gin_channels": 256,
92
+ "ssl_dim": 256,
93
+ "n_speakers": 200
94
+ },
95
+ "spk": {
96
+ "mommylaine": 0
97
+ }
98
  }
models/{alice/alice.pth → elaine/elaine.pth} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6d1bef76e26beeadcae5f716cc0b60abb2aac4aae1316cac709cc439726cf533
3
- size 180883747
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:656c2ea75e261d37496e75e93598a019e09a1bbf70d9d6f9de8f1550ff5a7c3e
3
+ size 542789469