|
--- |
|
dataset_info: |
|
features: |
|
- name: image |
|
dtype: image |
|
- name: text |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 378108023.375 |
|
num_examples: 1581 |
|
download_size: 373552088 |
|
dataset_size: 378108023.375 |
|
--- |
|
# MDCT-1k |
|
|
|
Over 1000 audio clips from the [Google music captions dataset](https://huggingface.co/datasets/google/MusicCaps) represented as 512x512 time-frequency images. |
|
|
|
The time-frequency images are created from the MDCT coefficients of the 0-12kHz frequency band for 20 second audio clips. |
|
|
|
Please see [this notebook showing how to load the dataset and convert from the MDCT images back to audio](load_dataset.ipynb) |
|
|
|
Most other audio diffusion models operate in the space of the magnitude spectrogram or mel magnitude spectrogram. Since the phase is discarded, this requires the use of a vocoder for audio generation. When operating in the space of the mel-spectrogram, high frequencies are represented with insufficient time resolution, leading to a noticable loss of quality. |
|
|
|
Operating in the MDCT space does not require a vocoder, nor does it oversample or undersample any range of frequencies. |