--- license: mit --- # Amphion Vocoder Pretrained Models We provide a [DiffWave](https://github.com/open-mmlab/Amphion/tree/main/egs/vocoder/diffusion) pretrained checkpoint, which is trained on 125 hours of speech data and 80 hours of singing voice data. ## Quick Start To utilize these pretrained vocoders, just run the following commands: ### Step1: Download the checkpoint ```bash git lfs install git clone https://huggingface.co/amphion/diffwave ``` ### Step2: Clone the Amphion's Source Code of GitHub ```bash git clone https://github.com/open-mmlab/Amphion.git ``` ### Step3: Specify the checkpoint's path Use the soft link to specify the downloaded checkpoint in the first step: ```bash cd Amphion mkdir -p ckpts/vocoder ln -s "$(realpath ../diffwave/diffwave)" pretrained/diffwave ``` ### Step4: Inference For analysis synthesis on the processed dataset, raw waveform, or predicted mel spectrograms, you can follow the inference part of [this recipe](https://github.com/open-mmlab/Amphion/tree/main/egs/vocoder/diffusion). ```bash sh egs/vocoder/diffusion/diffwave/run.sh --stage 3 \ --infer_mode [Your chosen inference mode] \ --infer_datasets [Datasets you want to inference, needed when infer_from_dataset] \ --infer_feature_dir [Your path to your predicted acoustic features, needed when infer_from_feature] \ --infer_audio_dir [Your path to your audio files, needed when infer_form_audio] \ --infer_expt_dir Amphion/ckpts/vocoder/[YourExptName] \ --infer_output_dir Amphion/ckpts/vocoder/[YourExptName]/result \ ```