# Copyright (c) 2023 Amphion. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. ######## Build Experiment Environment ########### exp_dir=$(cd `dirname $0`; pwd) work_dir=$(dirname $(dirname $(dirname $exp_dir))) export WORK_DIR=$work_dir export PYTHONPATH=$work_dir export PYTHONIOENCODING=UTF-8 cd $work_dir/modules/monotonic_align mkdir -p monotonic_align python setup.py build_ext --inplace cd $work_dir ######## Parse the Given Parameters from the Commond ########### # options=$(getopt -o c:n:s --long gpu:,config:,infer_expt_dir:,infer_output_dir:,infer_source_file:,infer_source_audio_dir:,infer_target_speaker:,infer_key_shift:,infer_vocoder_dir:,name:,stage: -- "$@") options=$(getopt -o c:n:s --long gpu:,config:,infer_expt_dir:,infer_output_dir:,infer_mode:,infer_dataset:,infer_testing_set:,infer_text:,infer_speaker_name:,name:,stage: -- "$@") eval set -- "$options" while true; do case $1 in # Experimental Configuration File -c | --config) shift; exp_config=$1 ; shift ;; # Experimental Name -n | --name) shift; exp_name=$1 ; shift ;; # Running Stage -s | --stage) shift; running_stage=$1 ; shift ;; # Visible GPU machines. The default value is "0". --gpu) shift; gpu=$1 ; shift ;; # [Only for Inference] The experiment dir. The value is like "[Your path to save logs and checkpoints]/[YourExptName]" --infer_expt_dir) shift; infer_expt_dir=$1 ; shift ;; # [Only for Inference] The output dir to save inferred audios. Its default value is "$expt_dir/result" --infer_output_dir) shift; infer_output_dir=$1 ; shift ;; # [Only for Inference] The inference mode. It can be "batch" to generate speech by batch, or "single" to generage a single clip of speech. --infer_mode) shift; infer_mode=$1 ; shift ;; # [Only for Inference] The inference dataset. It is only used when the inference model is "batch". --infer_dataset) shift; infer_dataset=$1 ; shift ;; # [Only for Inference] The inference testing set. It is only used when the inference model is "batch". It can be "test" set split from the dataset, or "golden_test" carefully selected from the testing set. --infer_testing_set) shift; infer_testing_set=$1 ; shift ;; # [Only for Inference] The text to be synthesized from. It is only used when the inference model is "single". --infer_text) shift; infer_text=$1 ; shift ;; # [Only for Inference] The speaker voice to be delivered in the synthesized speech. It is only used when the inference model is "single". --infer_speaker_name) shift; infer_speaker_name=$1 ; shift ;; --) shift ; break ;; *) echo "Invalid option: $1" exit 1 ;; esac done ### Value check ### if [ -z "$running_stage" ]; then echo "[Error] Please specify the running stage" exit 1 fi if [ -z "$exp_config" ]; then exp_config="${exp_dir}"/exp_config.json fi echo "Exprimental Configuration File: $exp_config" if [ -z "$gpu" ]; then gpu="0" fi ######## Features Extraction ########### if [ $running_stage -eq 1 ]; then CUDA_VISIBLE_DEVICES=$gpu /home/pai/envs/amphion/bin/python "${work_dir}"/bins/tts/preprocess.py \ --config=$exp_config \ --num_workers=4 fi ######## Training ########### if [ $running_stage -eq 2 ]; then if [ -z "$exp_name" ]; then echo "[Error] Please specify the experiments name" exit 1 fi echo "Exprimental Name: $exp_name" CUDA_VISIBLE_DEVICES=$gpu accelerate launch "${work_dir}"/bins/tts/train.py \ --config $exp_config \ --exp_name $exp_name \ --log_level debug fi ######## Inference ########### if [ $running_stage -eq 3 ]; then if [ -z "$infer_expt_dir" ]; then echo "[Error] Please specify the experimental directionary. The value is like [Your path to save logs and checkpoints]/[YourExptName]" exit 1 fi if [ -z "$infer_output_dir" ]; then infer_output_dir="$expt_dir/result" fi if [ -z "$infer_mode" ]; then echo "[Error] Please specify the inference mode, e.g., "batch", "single"" exit 1 fi if [ "$infer_mode" = "batch" ] && [ -z "$infer_dataset" ]; then echo "[Error] Please specify the dataset used in inference when the inference mode is batch" exit 1 fi if [ "$infer_mode" = "batch" ] && [ -z "$infer_testing_set" ]; then echo "[Error] Please specify the testing set used in inference when the inference mode is batch" exit 1 fi if [ "$infer_mode" = "single" ] && [ -z "$infer_text" ]; then echo "[Error] Please specify the text to be synthesized when the inference mode is single" exit 1 fi if [ "$infer_mode" = "single" ]; then echo 'Text: ' ${infer_text} infer_dataset=None infer_testing_set=None elif [ "$infer_mode" = "batch" ]; then infer_text='' infer_speaker_name=None fi CUDA_VISIBLE_DEVICES=$gpu accelerate launch "$work_dir"/bins/tts/inference.py \ --config $exp_config \ --acoustics_dir $infer_expt_dir \ --output_dir $infer_output_dir \ --mode $infer_mode \ --dataset $infer_dataset \ --testing_set $infer_testing_set \ --text "$infer_text" \ --speaker_name $infer_speaker_name \ --log_level debug # CUDA_VISIBLE_DEVICES=$gpu accelerate launch "$work_dir"/bins/tts/inference.py \ # --config $exp_config \ # --acoustics_dir $infer_expt_dir \ # --output_dir $infer_output_dir \ # --mode $infer_mode \ # --dataset $infer_dataset \ # --test_list_file $infer_testing_set \ # --text "$infer_text" \ # --speaker_name $infer_speaker_name \ # --log_level debug fi