YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
Hugging Face's logo
language: - yo - en datasets: - JW300 + Menyo-20k
byt5-base-eng-yor-mt
Model description
byt5-base-eng-yor-mt is a machine translation model from English language to Yorùbá language based on a fine-tuned byt5-base model. It establishes a strong baseline for automatically translating texts from English to Yorùbá.
Specifically, this model is a byt5-base model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k
Limitations and bias
This model is limited by its training dataset. This may not generalize well for all use cases in different domains.
Training data
This model was fine-tuned on on JW300 corpus and Menyo-20k dataset
Training procedure
This model was trained on NVIDIA V100 GPU
Eval results on Test set (BLEU score)
Fine-tuning byt5-base achieves 12.23 BLEU on Menyo-20k test set while mt5-base achieves 9.82
BibTeX entry and citation info
By David Adelani
- Downloads last month
- 149
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.