Setsugesuka commited on
Commit
060ad1f
1 Parent(s): 5316f67

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -0
README.md CHANGED
@@ -1,3 +1,44 @@
1
  ---
2
  license: mit
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
  ---
4
+
5
+ # Amphion Vocoder Pretrained Models
6
+
7
+ We provide a [DiffWave](https://github.com/open-mmlab/Amphion/tree/main/egs/vocoder/diffusion) pretrained checkpoint, which is trained on 125 hours of speech data and 80 hours of singing voice data.
8
+
9
+ ## Quick Start
10
+
11
+ To utilize these pretrained vocoders, just run the following commands:
12
+
13
+ ### Step1: Download the checkpoint
14
+ ```bash
15
+ git lfs install
16
+ git clone https://huggingface.co/amphion/diffwave
17
+ ```
18
+
19
+ ### Step2: Clone the Amphion's Source Code of GitHub
20
+ ```bash
21
+ git clone https://github.com/open-mmlab/Amphion.git
22
+ ```
23
+
24
+ ### Step3: Specify the checkpoint's path
25
+ Use the soft link to specify the downloaded checkpoint in the first step:
26
+
27
+ ```bash
28
+ cd Amphion
29
+ mkdir -p ckpts/vocoder
30
+ ln -s "$(realpath ../diffwave/diffwave)" pretrained/diffwave
31
+ ```
32
+
33
+ ### Step4: Inference
34
+ For analysis synthesis on the processed dataset, raw waveform, or predicted mel spectrograms, you can follow the inference part of [this recipe](https://github.com/open-mmlab/Amphion/tree/main/egs/vocoder/diffusion).
35
+
36
+ ```bash
37
+ sh egs/vocoder/diffusion/diffwave/run.sh --stage 3 \
38
+ --infer_mode [Your chosen inference mode] \
39
+ --infer_datasets [Datasets you want to inference, needed when infer_from_dataset] \
40
+ --infer_feature_dir [Your path to your predicted acoustic features, needed when infer_from_feature] \
41
+ --infer_audio_dir [Your path to your audio files, needed when infer_form_audio] \
42
+ --infer_expt_dir Amphion/ckpts/vocoder/[YourExptName] \
43
+ --infer_output_dir Amphion/ckpts/vocoder/[YourExptName]/result \
44
+ ```