YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Pre-trained TDNN-LSTM-CTC models for the TIMIT dataset with icefall.

The model was trained on full TIMIT with the scripts in icefall.
See (https://github.com/k2-fsa/icefall/tree/master/egs/timit/ASR/tdnn_lstm_ctc) for more details of this model.

How to use

See (https://github.com/k2-fsa/icefall/blob/master/egs/timit/ASR/tdnn_lstm_ctc/Pre-trained.md)

Training procedure

The main repositories are list below, we will update the training and decoding scripts with the update of version.
k2: https://github.com/k2-fsa/k2 icefall: https://github.com/k2-fsa/icefall lhotse: https://github.com/lhotse-speech/lhotse

git clone https://github.com/k2-fsa/icefall
cd icefall
  • Preparing data.
cd egs/timit/ASR
bash ./prepare.sh
  • Training
export CUDA_VISIBLE_DEVICES="0"
python tdnn_lstm_ctc/train.py --bucketing-sampler True \
                                --concatenate-cuts False \
                                --max-duration 200 \
                                --world-size 1

Evaluation results

The best decoding results (PER, equals to WER) on TIMIT TEST are listed below, we got this result by averaging models from epoch 16 to 25, the lm_scale is 0.08, the decoding method is whole-lattice-rescoring.

TEST
PER 19.71%
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.