byt5-base-zu-en-mt is a machine translation model from isiZulu to English based on a fine-tuned google/byt5-base model. It establishes a strong baseline for automatically translating texts from isiZulu to English.
This model is limited by its training dataset. This may not generalize well for all use cases in different domains.
Specifically, this model is a byt5-base model that was fine-tuned on JW300 isiZulu corpus and LAFAND. The model was trained using isiXhosa (xh_ZA) as the language since the pre-trained model does not initially support isiZulu. Thus, you need to use the xh_ZA for language code when evaluating the model.
This model was trained on NVIDIA V100 GPU
Fine-tuning byt5-base achieves 29.32 BLEU on LAFAND test set
By David Adelani
- Downloads last month