--- license: mit task_categories: - text-to-speech language: - en pretty_name: WhisperSpeech --- # The WhisperSpeech Dataset This dataset contains data to train SPEAR TTS-like text-to-speech models that utilized semantic tokens derived from the OpenAI Whisper speech recognition model. We currently provide semantic and acoustic tokens for the LibriLight and LibriTTS datasets (English only). Acoustic tokens: - 24kHz EnCodec 6kbps (8 quantizers) Semantic tokens: - Whisper tiny VQ bottleneck trained on a subset of LibriLight Available LibriLight subsets: - `small`/`medium`/`large` (following the original dataset division but with `large` excluding the speaker `6454`) - a separate ≈1300hr single-speaker subset based on the `6454` speaker from the `large` subset for training single-speaker TTS models We plan to add more acoustic tokens from other codecs in the future.