File size: 2,879 Bytes
abaceb0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
import os

import gradio as gr
import subprocess
from subprocess import call

basePath = os.path.dirname(os.path.realpath(__file__))

outputPath = os.path.join(basePath, 'outputs')
inputAudioPath = basePath + '/inputs/kim_audio.mp3'
inputVideoPath = basePath + '/inputs/kimk_7s_raw.mp4'
lipSyncedOutputPath = basePath + '/outputs/result.mp4'

with gr.Blocks() as ui:
    with gr.Row():
        video = gr.File(label="Video or Image", info="Filepath of video/image that contains faces to use")
        audio = gr.File(label="Audio", info="Filepath of video/audio file to use as raw audio source")
        with gr.Column():
            checkpoint = gr.Radio(["wav2lip", "wav2lip_gan"], label="Checkpoint",
                                  info="Name of saved checkpoint to load weights from")
            no_smooth = gr.Checkbox(label="No Smooth",
                                    info="Prevent smoothing face detections over a short temporal window")
            resize_factor = gr.Slider(minimum=1, maximum=4, step=1, label="Resize Factor",
                                      info="Reduce the resolution by this factor. Sometimes, best results are obtained at 480p or 720p")
    with gr.Row():
        with gr.Column():
            pad_top = gr.Slider(minimum=0, maximum=50, step=1, value=0, label="Pad Top", info="Padding above")
            pad_bottom = gr.Slider(minimum=0, maximum=50, step=1, value=10,
                                   label="Pad Bottom (Often increasing this to 20 allows chin to be included)",
                                   info="Padding below lips")
            pad_left = gr.Slider(minimum=0, maximum=50, step=1, value=0, label="Pad Left",
                                 info="Padding to the left of lips")
            pad_right = gr.Slider(minimum=0, maximum=50, step=1, value=0, label="Pad Right",
                                  info="Padding to the right of lips")
            generate_btn = gr.Button("Generate")
        with gr.Column():
            result = gr.Video()


    def generate(video, audio, checkpoint, no_smooth, resize_factor, pad_top, pad_bottom, pad_left, pad_right):
        if video is None or audio is None or checkpoint is None:
            return

        smooth = "--nosmooth" if no_smooth else ""

        cmd = [
            "python",
            "inference.py",
            "--checkpoint_path", f"checkpoints/{checkpoint}.pth",
            "--segmentation_path", "checkpoints/face_segmentation.pth",
            "--enhance_face", "gfpgan",
            "--face", video.name,
            "--audio", audio.name,
            "--outfile", "results/output.mp4",
        ]

        call(cmd)
        return "results/output.mp4"


    generate_btn.click(
        generate,
        [video, audio, checkpoint, pad_top, pad_bottom, pad_left, pad_right, resize_factor],
        result)

ui.queue().launch(debug=True)