Spaces:
Running
title: Music Descriptor
emoji: π
colorFrom: blue
colorTo: indigo
sdk: gradio
sdk_version: 3.29.0
app_file: app.py
pinned: true
license: cc-by-nc-4.0
Demo Introduction
This is an example of using the MERT-v1-95M model as backbone to conduct multiple music understanding tasks with the universal represenation.
The tasks include EMO, GS, MTGInstrument, MTGGenre, MTGTop50, MTGMood, NSynthI, NSynthP, VocalSetS, VocalSetT.
More models can be referred at the map organization page.
Known Issues
Audio Format Support
Theorectically, all the audio formats supported by torchaudio.load() can be used in the demo. Theese should include but not limited to WAV, AMB, MP3, FLAC
.
Error Output
Due the hardware limitation of the machine hosting our demospecification (2 CPU and 16GB RAM), there might be Error
output when uploading long audios.
Unfortunately, we couldn't fix this in a short time since our team are all volunteer researchers.
We recommend to test audios less than 30 seconds or using the live mode if you are trying the Music Descriptor demo hosted online at HuggingFace Space.
This issue is expected to solve in the future by applying more community-support GPU resources or using other audio encoding strategy.
In the current stage, if you want to directly run the demo with longer audios, you could:
- clone this space
git clone https://huggingface.co/spaces/m-a-p/Music-Descriptor
and deploy the demo on your own machine with higher performance following the official instruction. The code will automatically use GPU for inference if there is GPU that can be detected bytorch.cuda.is_available()
. - develop your own application with the MERT models if you have the experience of machine learning.