YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
Fine-tuning of wav2vec2-large-lv60
on 100h of Librispeech training data. Results are a bit worse than those reported in the Appendix in Table 3 of the original paper.
Model was trained on librispeech-clean-train.100 with following hyper-parameters:
- 2 GPUs Titan RTX
- Total update steps 17500
- Batch size per GPU: 16 corresponding to a total batch size of ca. ~750 seconds
- Adam with linear decaying learning rate with 3000 warmup steps
- dynamic padding for batch
- fp16
- attention_mask was used during training
Result (WER) on Librispeech test:
"clean" | "other" |
---|---|
4.0 | 10.3 |
- Downloads last month
- 8
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.