--- datasets: - google/MusicCaps language: - en library_name: diffusers tags: - music pipeline_tag: text-to-audio --- # AudioLDM 2 Music for Zalo AI Challenge 2023 This checkpoint is the result of finetuning AudioLDM 2 Music (https://huggingface.co/cvssp/audioldm2-music) on the challenge dataset + MusicCaps (https://www.kaggle.com/datasets/googleai/musiccaps) ## Uses First, install the required packages: ``` pip install --upgrade diffusers transformers accelerate ``` ### Text-to-Audio ```python from diffusers import AudioLDM2Pipeline import torch repo_id = "vtrungnhan9/audioldm2-music-zac2023" pipe = AudioLDM2Pipeline.from_pretrained(repo_id, torch_dtype=torch.float16) pipe = pipe.to("cuda") prompt = "This music is instrumental. The tempo is medium with synthesiser arrangements, digital drums and electronic music. The music is upbeat, pulsating, youthful, buoyant, exciting, punchy, psychedelic and has propulsive beats with a dance groove. This music is Techno Pop/EDM." neg_prompt = "bad quality" audio = pipe(prompt, negative_prompt=neg_prompt, num_inference_steps=200, audio_length_in_s=10.0, guidance_scale=10).audios[0] ``` The resulting audio output can be saved as a .wav file: ```python import scipy scipy.io.wavfile.write("techno.wav", rate=16000, data=audio) ``` Or displayed in a Jupyter Notebook / Google Colab: ```python from IPython.display import Audio Audio(audio, rate=16000) ``` ## Training Details ### Training Data * You can download the challenge dataset from link: https://challenge.zalo.ai/portal/background-music-generation * You can download MusicCaps from link: https://www.kaggle.com/datasets/googleai/musiccaps [More Information Needed] ### Training Procedure Please refer at https://github.com/declare-lab/tango/blob/master/train.py for training procedure ## Citation **BibTeX:** ``` @article{liu2023audioldm2, title={"AudioLDM 2: Learning Holistic Audio Generation with Self-supervised Pretraining"}, author={Haohe Liu and Qiao Tian and Yi Yuan and Xubo Liu and Xinhao Mei and Qiuqiang Kong and Yuping Wang and Wenwu Wang and Yuxuan Wang and Mark D. Plumbley}, journal={arXiv preprint arXiv:2308.05734}, year={2023} } ``` ## Model Card Contact vtrungnhan16@gmail.com