Bert-VITS2-2.2-CLAP / README.md
OedoSoldier's picture
Create README.md
ef4d549
|
raw
history blame
303 Bytes
Pre-trained Bert-VITS2-2.2-CLAP model.
This model is trained with Chinese / Japanese / English speech data, the speaker embedding is removed due to copyright issues.
You can use auxiliary data to fine-tune the model to avoid catastrophic forgetting of multi-language ability and CLAP control ability.