Byte-Beats-demo / app.py
qichenhuang's picture
Update app.py
5c23d25 verified
import gradio as gr
import torch
from transformers import pipeline
from transformers import AutoProcessor, MusicgenForConditionalGeneration
import scipy
import wave
import io
def generate_music(text):
processor = AutoProcessor.from_pretrained("facebook/musicgen-small")
model = MusicgenForConditionalGeneration.from_pretrained("facebook/musicgen-small")
inputs = processor(
text=text,
padding=True,
return_tensors="pt",
)
audio_values = model.generate(**inputs, do_sample=True
, guidance_scale=3
, max_new_tokens=1350)
sampling_rate = model.config.audio_encoder.sampling_rate
scipy.io.wavfile.write("musicgen_out.wav", rate=sampling_rate, data=audio_values[0, 0].numpy())
file_path= "musicgen_out.wav"
try:
Audio = gr.Audio(file_path)
return Audio
except Exception as e:
return str(e)
# theme='Taithrah/Minimal'
# theme = 'freddyaboulton/test-blue'
theme = gr.themes.Monochrome(
primary_hue="indigo",
secondary_hue="violet",
neutral_hue="purple",
).set(
body_background_fill='*background_fill_secondary',
link_text_color='*secondary_700',
link_text_color_dark='*secondary_600'
)
# 创建 Gradio Blocks 界面
iface = gr.Blocks(theme = theme)
with iface as demo:
# 创建一个列容器
with gr.Row():
with gr.Column(scale=1):
gr.Image("logo.jpg", height = 138, width = 108, show_download_button = False, show_label = False, show_share_button = False, min_width = 108)
with gr.Column(scale=5):
company_intro = gr.Markdown("## Byte beats\n\n We simplify soundtrack creation, empowering filmmakers globally, thereby reducing production expenses and shortening the process.\n\n\n")
with gr.Row():
# 在左侧列中添加文本输入框
with gr.Column(scale=1):
input_text = gr.Textbox(lines=6, placeholder="Please enter your script or upload video below...")
video = gr.UploadButton("upload your video")
example_demo = gr.Markdown("## Example Demo\n\n")
# 在右侧列中添加音频输出组件
with gr.Column(scale=3):
output_audio = gr.Audio(label="Generated Music")
generate_button = gr.Button("Generate Music")
generate_button.click(fn=generate_music, inputs=input_text, outputs=output_audio)
# 启动界面
demo.queue(max_size=4).launch(share = True)
# demo = gr.Interface(
# fn=generate_music,
# inputs='text',
# outputs='audio',
# )
# demo.queue(max_size=4).launch(share = True)