--- license: mpl-2.0 datasets: - CommonVoice - LibriSpeech - LibriTTS - VCTK - VoxCeleb tags: - speaker-encoding - speaker-embedding - speaker-verification - speaker - speech library_name: PyTorch --- This repository contains released Mozilla SpeakerEncoder models. Due to original codebase migration and SpeakerEncoder depracation, I decided to extract inference code into my own repository. For ease of integration, I also release relevant models: - `model_mozilla_128.pth.tar` - which is [Speaker-Encoder-iter25k](https://drive.google.com/open?id=1CwXxwm0qKk2lcPL_4EQDirD8XMyvHIqR) that was trained on LibriSpeech, - `model_mozilla_256.pth.tar` - which is [Speaker-Encoder](https://drive.google.com/drive/folders/15oeBYf6Qn1edONkVLXe82MzdIi3O_9m3?usp=sharing) by [@mueller91](https://github.com/mueller91) that was trained on LibriTTS, VCTK, VoxCeleb and CommonVoice. Original SpeakerEncoder implementation is available [here](https://github.com/coqui-ai/TTS/tree/f9001a4bdd8a169da8b4eefd481461f370def8f9/TTS/speaker_encoder) and originally released models are available [here](https://github.com/mozilla/TTS/wiki/Released-Models). If you are interested in the current version of Coqui.ai TTS go [here](https://github.com/coqui-ai/TTS).