File size: 7,222 Bytes
9a85c9a
 
 
 
 
 
 
 
 
 
 
 
 
b49028b
b92e1c5
9a85c9a
 
ff3b347
f16ce36
ff3b347
 
d532b1b
0d7d0b8
 
 
 
ff3b347
94d1049
ff3b347
 
 
9a85c9a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b92e1c5
a3654b1
 
b92e1c5
 
9a85c9a
 
 
 
 
 
 
 
5478222
9a85c9a
 
88c2b88
b92e1c5
 
9a85c9a
 
9b9ccec
db45ded
 
 
 
 
 
 
 
9a85c9a
ff8b33b
9a85c9a
db45ded
9a85c9a
 
 
 
 
 
 
 
 
 
 
e9073c2
26acab0
e9073c2
9e67411
6343e6d
3bfffe0
 
 
 
 
 
 
 
 
 
9a85c9a
 
05f7bfe
ff3b347
05f7bfe
0b2fffd
db45ded
ff3b347
ee5528e
3acd203
ff3b347
4593ac7
7a7857a
 
 
 
c468b59
3acd203
714116a
ff3b347
 
db45ded
9a85c9a
 
db45ded
 
e9073c2
db45ded
 
428be2c
eea2265
fdcc4cf
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
import sys, os
import torch
import argparse
import commons
import utils
from models import SynthesizerTrn
from text.symbols import symbols
from text import cleaned_text_to_sequence, get_bert
from text.cleaner import clean_text
import gradio as gr
import soundfile as sf
from datetime import datetime
import pytz

tz = pytz.timezone('Asia/Shanghai')
net_g = None
models = {
    "MistyNikki": "./MODELS/nikki.pth",
    "AlluWin": "./MODELS/AW.pth",
    "VVV":"./MODELS/v3.pth",
    "StayTerra": "./MODELS/TERRA.pth",
    "Rrabbitt": "./MODELS/rabbit4900.pth",
    "Xelo": "./MODELS/HER_1100.pth",
    "Hypnosia": "./MODELS/hypno.pth",
    "Appelsin":"./MODELS/orange.pth",
    "Kitsune": "./MODELS/fox2100.pth",
    "Silverleg": "./MODELS/J8900.pth",
    "PremJ": "./MODELS/premj.pth",
    "Robo!Girl": "./MODELS/BG1300.pth",
    "HeavyHammer": "./MODELS/hammer.pth",
    "TaxiAI": "./MODELS/DLM.pth",
}

def get_text(text, language_str, hps):
    norm_text, phone, tone, word2ph = clean_text(text, language_str)
    phone, tone, language = cleaned_text_to_sequence(phone, tone, language_str)

    if hps.data.add_blank:
        phone = commons.intersperse(phone, 0)
        tone = commons.intersperse(tone, 0)
        language = commons.intersperse(language, 0)
        for i in range(len(word2ph)):
            word2ph[i] = word2ph[i] * 2
        word2ph[0] += 1
    bert = get_bert(norm_text, word2ph, language_str)
    del word2ph

    assert bert.shape[-1] == len(phone)

    phone = torch.LongTensor(phone)
    tone = torch.LongTensor(tone)
    language = torch.LongTensor(language)

    return bert, phone, tone, language

def infer(text, sdp_ratio, noise_scale, noise_scale_w, length_scale, sid, model_dir):
    global net_g
    bert, phones, tones, lang_ids = get_text(text, "ZH", hps)
    with torch.no_grad():
        x_tst=phones.to(device).unsqueeze(0)
        tones=tones.to(device).unsqueeze(0)
        lang_ids=lang_ids.to(device).unsqueeze(0)
        bert = bert.to(device).unsqueeze(0)
        x_tst_lengths = torch.LongTensor([phones.size(0)]).to(device)
        del phones
        speakers = torch.LongTensor([hps.data.spk2id[sid]]).to(device)
        audio = net_g.infer(x_tst, x_tst_lengths, speakers, tones, lang_ids, bert, sdp_ratio=sdp_ratio
                           , noise_scale=noise_scale, noise_scale_w=noise_scale_w, length_scale=length_scale)[0][0,0].data.cpu().float().numpy()
        del x_tst, tones, lang_ids, bert, x_tst_lengths, speakers
        sf.write("tmp.wav", audio, 44100)
        return audio



def convert_wav_to_mp3(wav_file):
    global tz
    now = datetime.now(tz).strftime('%m%d%H%M%S')    
    os.makedirs('out', exist_ok=True)  
    output_path_mp3 = os.path.join('out', f"{now}.mp3")

    renamed_input_path = os.path.join('in', f"in.wav")
    os.makedirs('in', exist_ok=True)
    os.rename(wav_file.name, renamed_input_path)
    command = ["ffmpeg", "-i", renamed_input_path, "-acodec", "libmp3lame", "-y", output_path_mp3]
    os.system(" ".join(command))
    print(str(output_path_mp3))
    return output_path_mp3
    
def tts_generator(text, sdp_ratio, noise_scale, noise_scale_w, length_scale, model):
    global net_g,speakers,tz
    now = datetime.now(tz).strftime('%m-%d %H:%M:%S')
    model_path = models[model]
    net_g, _, _, _ = utils.load_checkpoint(model_path, net_g, None, skip_optimizer=True)
    print(now+text)
    try:
        with torch.no_grad():
            audio = infer(text, sdp_ratio=sdp_ratio, noise_scale=noise_scale, noise_scale_w=noise_scale_w, length_scale=length_scale, sid=speaker,model_dir=model)
        with open('tmp.wav', 'rb') as wav_file:
            mp3 = convert_wav_to_mp3(wav_file)  
        return "生成语音成功", (hps.data.sampling_rate, audio), mp3
    except Exception as e:
        return "生成语音失败:" + str(e), None, None


if __name__ == "__main__":
    hps = utils.get_hparams_from_file("./configs/config.json")
    device = "cuda:0" if torch.cuda.is_available() else "cpu"
   
    net_g = SynthesizerTrn(
        len(symbols),
        hps.data.filter_length // 2 + 1,
        hps.train.segment_size // hps.data.hop_length,
        n_speakers=hps.data.n_speakers,
        **hps.model).to(device)
    _ = net_g.eval()

    speaker_ids = hps.data.spk2id
    speaker = list(speaker_ids.keys())[0]
    theme='remilia/Ghostly'
    
    with gr.Blocks(theme=theme) as app:
        with gr.Column():
            gr.HTML('''<br><br>
    <p style="margin-bottom: 10px; font-size: 110%">
    本空间仅支持中文生成<br>
    Currently, this space only supports Chinese generation.
    現在、このスペースは中国語の生成のみをサポートしています。
   Use <b>English</b> to generate, please go to this <a href="https://huggingface.co/spaces/Ailyth/Multi-voice-TTS-GPT-SoVITS" target="_blank">SPACE</a>
   </p>
    <p style="margin-bottom: 10px; font-size: 100%">
   <b>日本語</b>で生成するために、<a href="https://huggingface.co/spaces/Ailyth/Multi-voice-TTS-GPT-SoVITS" target="_blank">こちら</a>へ進んでください。
    </p><hr>''')
            with gr.Column():

                gr.Markdown('''
                **仅供测试用** These models only speak Chinese for now.
                ''')
                text = gr.TextArea(label="输入需要生成语音的文字", placeholder="输入文字",
                                value="在不在?能不能借给我三百块钱买可乐",
                                info="使用huggingface的免费CPU进行推理,因此速度不快,一次性不要输入超过500字。字数越多,生成速度越慢,请耐心等待,只会说中文。"
,
                                  )
                model = gr.Radio(choices=list(models.keys()), value=list(models.keys())[0], label='选择音声模型')
                with gr.Accordion(label="展开设置生成参数", open=False):
                    sdp_ratio = gr.Slider(minimum=0, maximum=1, value=0.2, step=0.01, label='SDP/DP混合比',info='可控制一定程度的语调变化')
                    noise_scale = gr.Slider(minimum=0.1, maximum=1.5, value=0.5, step=0.01, label='感情变化')
                    noise_scale_w = gr.Slider(minimum=0.1, maximum=1.4, value=0.9, step=0.01, label='音节长度')
                    length_scale = gr.Slider(minimum=0.1, maximum=2, value=1, step=0.01, label='生成语音总长度',info='数值越大,语速越慢')
                btn = gr.Button("✨生成", variant="primary")
            with gr.Column():
                audio_output = gr.Audio(label="试听")
                MP3_output = gr.File(label="💾下载")
                text_output = gr.Textbox(label="❗调试信息")
                gr.Markdown("""
                
                """)
        btn.click(
                tts_generator,
                inputs=[text, sdp_ratio, noise_scale, noise_scale_w, length_scale, model],
                outputs=[text_output, audio_output,MP3_output]
                )
       
        gr.HTML('''<div align=center><img id="visitor-badge" alt="visitor badge" src="https://visitor-badge.laobi.icu/badge?page_id=nepheTTS" /></div>''')
    app.launch(show_error=True)