Edit model card

Fine-tuned SpeechT5 TTS Model for Haitian Creole

This model is a fine-tuned version of microsoft/speecht5-tts for the Haitian Creole language. It was fine-tuned on the CMU Haitian dataset.

Model Description

The model is based on the SpeechT5 architecture, which is a variant of the T5 (Text-to-Text Transfer Transformer) model designed specifically for text-to-speech tasks. The model is capable of converting input text in Haitian Creole into corresponding speech.

Intended Uses & Limitations

The model is intended for text-to-speech (TTS) applications in Haitian Creole language processing. It can be used for generating speech from written text, enabling applications such as audiobook narration, voice assistants, and more.

However, there are some limitations to be aware of:

  • The model's performance heavily depends on the quality and diversity of the training data. Fine-tuning on more diverse and specific datasets might improve its performance.
  • Like all machine learning models, this model may produce inaccuracies or errors in speech synthesis, especially for complex sentences or domain-specific jargon.

Training and Evaluation Data

The model was fine-tuned on the CMU Haitian dataset, which contains text and corresponding audio samples in Haitian Creole. The dataset was split into training and evaluation sets to assess the model's performance.

Training Procedure

Training Hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • per_device_train_batch_size: 16
  • gradient_accumulation_steps: 2
  • warmup_steps: 500
  • max_steps: 4000
  • gradient_checkpointing: True
  • fp16: True
  • evaluation_strategy: no
  • per_device_eval_batch_size: 8
  • save_steps: 1000
  • logging_steps: 25
  • report_to: ["tensorboard"]
  • greater_is_better: False

Training Results

The training progress and evaluation results are as follows:

Training Loss Epoch Step Validation Loss
0.5147 2.42 1000 0.4753
0.4932 4.84 2000 0.4629
0.4926 7.26 3000 0.4566
0.4907 9.69 4000 0.4542
0.4839 12.11 5000 0.4532

Training Output

The training was completed with the following output:

  • Global Step: 4000
  • Training Loss: 0.3344
  • Training Runtime: 7123.63 seconds
  • Training Samples per Second: 17.97
  • Training Steps per Second: 0.562
  • Total FLOPs: 1.1690e+16

Framework Versions

  • Transformers 4.31.0
  • PyTorch 2.0.1+cu118
  • Datasets 2.13.1
  • Tokenizers 0.13.3
Downloads last month
23
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.