File size: 1,371 Bytes
ba28442
 
 
 
 
 
 
 
 
 
 
 
 
 
0609db6
ba28442
bc13a8a
0609db6
 
 
2cebf50
 
 
 
0609db6
 
2cebf50
0609db6
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
---
dataset_info:
  features:
  - name: image
    dtype: image
  - name: text
    dtype: string
  splits:
  - name: train
    num_bytes: 378108023.375
    num_examples: 1581
  download_size: 373552088
  dataset_size: 378108023.375
---
# MDCT-1k

Over 1000 audio clips from the [Google music captions dataset](https://huggingface.co/datasets/google/MusicCaps) represented as 512x512 time-frequency images. More information is provided in the [report](MP3_diffusion.pdf).

The time-frequency images are created from the MDCT coefficients of the 0-12kHz frequency band for 20 second audio clips.

Other audio diffusion models operate in the space of the magnitude spectrogram or mel magnitude spectrogram. Since the phase is discarded, this requires the use of a vocoder for audio generation. When operating in the space of the mel-spectrogram, high frequencies are represented with insufficient time resolution, leading to a noticable loss of quality.

Operating in the MDCT space does not require a vocoder, nor does it oversample or undersample any range of frequencies.

Please see [this notebook showing how to load the dataset and convert from the MDCT images back to audio](load_dataset.ipynb)

Additionally, [this notebook includes an example of the audio generated by fine tuning on this dataset and shows how to use the inference pipeline](music_inference.ipynb)