Edit model card

t5-small-finetuned-hausa-to-english

This model is a fine-tuned version of google-t5/t5-small on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.5734
  • Bleu: 71.4633
  • Gen Len: 6.984

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0008
  • train_batch_size: 32
  • eval_batch_size: 32
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_steps: 3000
  • num_epochs: 30
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Bleu Gen Len
3.4817 1.0 749 2.2949 34.5266 7.5496
2.0957 2.0 1498 1.6167 53.0797 6.8785
1.562 3.0 2247 1.3650 60.3083 7.0095
1.2877 4.0 2996 1.2149 63.8437 6.9444
1.1026 5.0 3745 1.1452 66.7883 7.13
0.9531 6.0 4494 1.1028 67.3774 6.911
0.8402 7.0 5243 1.0995 68.0354 6.8114
0.7513 8.0 5992 1.0878 69.4876 7.0216
0.6746 9.0 6741 1.1109 69.7327 7.1134
0.6073 10.0 7490 1.1167 70.1607 7.0526
0.5531 11.0 8239 1.1468 69.8006 6.8101
0.4981 12.0 8988 1.1856 70.5423 6.8789
0.4544 13.0 9737 1.2019 70.5876 6.9313
0.4095 14.0 10486 1.2347 70.7996 6.8371
0.373 15.0 11235 1.2734 71.0903 7.0274
0.3408 16.0 11984 1.2974 71.104 7.0025
0.3096 17.0 12733 1.3313 70.7308 6.925
0.2856 18.0 13482 1.3820 70.9862 6.9656
0.2601 19.0 14231 1.4016 71.1836 7.0082
0.2404 20.0 14980 1.4483 71.0219 6.9268
0.2241 21.0 15729 1.4714 71.2721 6.9552
0.2065 22.0 16478 1.4814 71.3874 6.9968
0.1942 23.0 17227 1.5090 71.4722 6.9404
0.1831 24.0 17976 1.5265 71.4556 6.9771
0.173 25.0 18725 1.5379 71.4026 6.9998
0.1662 26.0 19474 1.5530 71.4843 6.9932
0.159 27.0 20223 1.5668 71.3663 6.9784
0.1568 28.0 20972 1.5742 71.3261 6.9734
0.1566 29.0 21721 1.5739 71.4435 6.9843
0.156 30.0 22470 1.5734 71.4633 6.984

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu121
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
6
Safetensors
Model size
60.5M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for Kumshe/t5-small-finetuned-hausa-to-english

Base model

google-t5/t5-small
Finetuned
(1502)
this model