zyingt's picture
Upload 685 files
0d80816
# Copyright (c) 2023 Amphion.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
######## Build Experiment Environment ###########
exp_dir=$(cd `dirname $0`; pwd)
work_dir=$(dirname $(dirname $(dirname $exp_dir)))
export WORK_DIR=$work_dir
export PYTHONPATH=$work_dir
export PYTHONIOENCODING=UTF-8
cd $work_dir/modules/monotonic_align
mkdir -p monotonic_align
python setup.py build_ext --inplace
cd $work_dir
######## Parse the Given Parameters from the Commond ###########
# options=$(getopt -o c:n:s --long gpu:,config:,infer_expt_dir:,infer_output_dir:,infer_source_file:,infer_source_audio_dir:,infer_target_speaker:,infer_key_shift:,infer_vocoder_dir:,name:,stage: -- "$@")
options=$(getopt -o c:n:s --long gpu:,config:,infer_expt_dir:,infer_output_dir:,infer_mode:,infer_dataset:,infer_testing_set:,infer_text:,infer_speaker_name:,name:,stage: -- "$@")
eval set -- "$options"
while true; do
case $1 in
# Experimental Configuration File
-c | --config) shift; exp_config=$1 ; shift ;;
# Experimental Name
-n | --name) shift; exp_name=$1 ; shift ;;
# Running Stage
-s | --stage) shift; running_stage=$1 ; shift ;;
# Visible GPU machines. The default value is "0".
--gpu) shift; gpu=$1 ; shift ;;
# [Only for Inference] The experiment dir. The value is like "[Your path to save logs and checkpoints]/[YourExptName]"
--infer_expt_dir) shift; infer_expt_dir=$1 ; shift ;;
# [Only for Inference] The output dir to save inferred audios. Its default value is "$expt_dir/result"
--infer_output_dir) shift; infer_output_dir=$1 ; shift ;;
# [Only for Inference] The inference mode. It can be "batch" to generate speech by batch, or "single" to generage a single clip of speech.
--infer_mode) shift; infer_mode=$1 ; shift ;;
# [Only for Inference] The inference dataset. It is only used when the inference model is "batch".
--infer_dataset) shift; infer_dataset=$1 ; shift ;;
# [Only for Inference] The inference testing set. It is only used when the inference model is "batch". It can be "test" set split from the dataset, or "golden_test" carefully selected from the testing set.
--infer_testing_set) shift; infer_testing_set=$1 ; shift ;;
# [Only for Inference] The text to be synthesized from. It is only used when the inference model is "single".
--infer_text) shift; infer_text=$1 ; shift ;;
# [Only for Inference] The speaker voice to be delivered in the synthesized speech. It is only used when the inference model is "single".
--infer_speaker_name) shift; infer_speaker_name=$1 ; shift ;;
--) shift ; break ;;
*) echo "Invalid option: $1" exit 1 ;;
esac
done
### Value check ###
if [ -z "$running_stage" ]; then
echo "[Error] Please specify the running stage"
exit 1
fi
if [ -z "$exp_config" ]; then
exp_config="${exp_dir}"/exp_config.json
fi
echo "Exprimental Configuration File: $exp_config"
if [ -z "$gpu" ]; then
gpu="0"
fi
######## Features Extraction ###########
if [ $running_stage -eq 1 ]; then
CUDA_VISIBLE_DEVICES=$gpu /home/pai/envs/amphion/bin/python "${work_dir}"/bins/tts/preprocess.py \
--config=$exp_config \
--num_workers=4
fi
######## Training ###########
if [ $running_stage -eq 2 ]; then
if [ -z "$exp_name" ]; then
echo "[Error] Please specify the experiments name"
exit 1
fi
echo "Exprimental Name: $exp_name"
CUDA_VISIBLE_DEVICES=$gpu accelerate launch "${work_dir}"/bins/tts/train.py \
--config $exp_config \
--exp_name $exp_name \
--resume \
--log_level debug \
--checkpoint_path "/mnt/workspace/tzeying/data/vits_on_libritts_hifitts/logs/HifiTTS_all/checkpoint/epoch-0030_step-0312356_loss-38.448391"
fi
######## Inference ###########
if [ $running_stage -eq 3 ]; then
if [ -z "$infer_expt_dir" ]; then
echo "[Error] Please specify the experimental directionary. The value is like [Your path to save logs and checkpoints]/[YourExptName]"
exit 1
fi
if [ -z "$infer_output_dir" ]; then
infer_output_dir="$expt_dir/result"
fi
if [ -z "$infer_mode" ]; then
echo "[Error] Please specify the inference mode, e.g., "batch", "single""
exit 1
fi
if [ "$infer_mode" = "batch" ] && [ -z "$infer_dataset" ]; then
echo "[Error] Please specify the dataset used in inference when the inference mode is batch"
exit 1
fi
if [ "$infer_mode" = "batch" ] && [ -z "$infer_testing_set" ]; then
echo "[Error] Please specify the testing set used in inference when the inference mode is batch"
exit 1
fi
if [ "$infer_mode" = "single" ] && [ -z "$infer_text" ]; then
echo "[Error] Please specify the text to be synthesized when the inference mode is single"
exit 1
fi
if [ "$infer_mode" = "single" ]; then
echo 'Text: ' ${infer_text}
infer_dataset=None
infer_testing_set=None
elif [ "$infer_mode" = "batch" ]; then
infer_text=''
infer_speaker_name=None
fi
CUDA_VISIBLE_DEVICES=$gpu accelerate launch "$work_dir"/bins/tts/inference.py \
--config $exp_config \
--checkpoint_path $infer_expt_dir \
--output_dir $infer_output_dir \
--mode $infer_mode \
--dataset $infer_dataset \
--testing_set $infer_testing_set \
--text "$infer_text" \
--speaker_name $infer_speaker_name \
--log_level debug
fi