--- language: - en - ro license: apache-2.0 tags: - translation - wmt16 - Lvxue datasets: - wmt16 metrics: - sacrebleu - bleu model-index: - name: Lvxue/finetuned-mt5-small-10epoch results: - task: type: translation name: Translation dataset: name: wmt16 type: wmt16 config: ro-en split: test metrics: - type: bleu value: 6.0012 name: BLEU verified: true verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMzAwOGM0NWI2ODYxNTcyOWZmZGYxMzZmM2RhZjYzZTc4ZDk4MTBkYzJiNjY3MjYxNDI3ZTBjYmY0Mzc4NGZhOCIsInZlcnNpb24iOjF9.kx434Ckp7xt_qfxxaJ2XTXJjUUDbNUCY-gRDCSe_FPeVec69Fap3NhL4u9hZioaNbTqpXxNwMHGRnkgjJ1jHAw - type: loss value: 1.7407585382461548 name: loss verified: true verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiNGRhMjYwYTY4YzNhNjQ0ZjlhM2IxMDQ1OWZjOWRmZjNhNGRhODc1NmE3Y2M4ODg3NzRjZWU2ZDU1MWZlZGUxNSIsInZlcnNpb24iOjF9.IHuQPWQp4xxuAOVvkmFdJQGa4uqBYq5_aS0uiK3uqTgfxSnxqFawLfgE2ou5RnRHFPiEQ4h5m_DDAjZBxBSLAA - type: gen_len value: 18.2281 name: gen_len verified: true verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiY2ZiYTA5MzBkZDExY2FlNThjZDAxYmJiZjI2NjkzMDU2OTRjOGU1NjNiNjNmNGJkODk4MTg0YmI1Y2U5ZTY0YiIsInZlcnNpb24iOjF9.5XEsIvtfSnmnEvQg3GVVmVcURti_Jw3hSLQryeoAfgdLw6GEqKqIcu-WHsWnc6-9EDMfLQahrQneNlKpe8xDDw --- # finetuned-mt5-small-10epoch This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on the wmt16 ro-en dataset. It achieves the following results on the evaluation set: - Loss: 1.7274 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 48 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10.0 ### Training results ### Framework versions - Transformers 4.20.1 - Pytorch 1.12.0+cu102 - Datasets 2.3.2 - Tokenizers 0.12.1