import os import gradio as gr from PIL import Image as im from scipy.io.wavfile import write def generateVideo(input_img, input_audio): data = im.fromarray(input_img) # saving the final output # as a PNG file data.save('MakeItTalk/examples/in_image.jpg') write('MakeItTalk/examples/in_audio.wav', input_audio[0], input_audio[1]) input_img = 'in_image.jpg' input_audio = 'in_audio.wav' os.system(f"python3 MakeItTalk/main_end2end.py --jpg {input_img}") #add image video_name = 'MakeItTalk/examples/in_image_pred_fls_in_audio_audio_embed.mp4' return video_name demo = gr.Interface( fn=generateVideo, inputs=[gr.Image(shape=(256, 256)), gr.Audio(), ], outputs= gr.Video().style(height=256, width=256), title='Audio Driven Animation', description='This is my app', examples =[['example_image.jpg', 'marlene_example.wav']] ) demo.launch()