Yurii Paniv
Mit model (#28)
0ea31c1 unverified
# Setup environment
Link: https://espnet.github.io/espnet/installation.html
```sh
sudo apt-get install cmake sox libsndfile1-dev ffmpeg
git clone --branch v.202301 https://github.com/espnet/espnet
cd ./espnet/tools
./setup_anaconda.sh anaconda espnet 3.8
. ./activate_python.sh
make
pip install --upgrade torch torchaudio # or setup same versions
make
. ./activate_python.sh; python3 check_install.py
```
# Run training
ESPNET is a dynamic framework. For the latest guide, please refer to https://github.com/espnet/espnet/tree/master/egs2/TEMPLATE/tts1
This page provides general launching steps on how training was performed for reference, and this doesn't cover data preparation.
NOTE: before running the script below, copy [./train_vits.yaml](./train_vits.yaml) to your `<espnet_root>/egs2/ljspeech/tts1/conf/tuning/train_vits.yaml`
```sh
cd ../egs2/ljspeech/tts1
pip install torchvision # to save figures
pip install speechbrain
./run.sh \
--stage 2 \
--use_xvector true \
--xvector_tool speechbrain \
--fs 22050 \
--n_fft 1024 \
--n_shift 256 \
--win_length null \
--dumpdir dump/22k \
--expdir exp/22k \
--tts_task gan_tts \
--feats_extract linear_spectrogram \
--feats_normalize none \
--train_config ./conf/tuning/train_vits.yaml \
--inference_config ./conf/tuning/decode_vits.yaml
```