Edit model card

Fine-tuned mBART Model for English to Urdu Translation

This repository contains a fine-tuned mBART model for English to Urdu translation. The model has been trained on a custom dataset and evaluated on test data.

Model Information

  • Model Name: abdulwaheed1/english-to-urdu-translation-mbart
  • Base Model: facebook/mbart-large-50
  • Tokenizer: facebook/mbart-large-50
  • Source Language: English (en)
  • Target Language: Urdu (ur)

Usage

python
from transformers import MBart50TokenizerFast, MBartForConditionalGeneration

# Load the fine-tuned model
model_name = "abdulwaheed1/english-to-urdu-translation-mbart"
tokenizer = MBart50TokenizerFast.from_pretrained(model_name, src_lang="en_XX", tgt_lang="ur_PK")
model = MBartForConditionalGeneration.from_pretrained(model_name)

Evaluation

The model has been evaluated on a test dataset, and the following metrics were obtained:

  • BLEU Score: 35.87
  • Generation Length: 42.56
  • Meteor Score: 0.60

Training Details

The model was trained using the transformers library with the following configuration:

  • Training Loss: 1.5697
  • Validation Loss: 1.1256

Dataset

The model was fine-tuned on a custom English-Urdu translation dataset. If you wish to use the same dataset, you can find the preprocessing script and dataset files in the data directory.

Acknowledgments

The fine-tuning process and code were inspired by the Hugging Face Transformers library.


Downloads last month
128
Safetensors
Model size
611M params
Tensor type
F32
Β·

Spaces using abdulwaheed1/english-to-urdu-translation-mbart 2