mt5_small_wmt16_de_en
This model is a fine-tuned version of google/mt5-small on the wmt16 dataset. It achieves the following results on the evaluation set:
- Loss: 2.4612
- Rouge1: 0.3666
- Rouge2: 0.147
- Rougel: 0.3362
- Sacrebleu: 6.4622
Model description
Multilingual T5 (mT5) is a massively multilingual pretrained text-to-text transformer model, trained following a similar recipe as T5.
Intended uses & limitations
This is tried to be familiarized with the mt5 model in order to use it for the translation of English to Korean.
Training and evaluation data
This work was done as an exercise for English-Korean translation, so I trained by selecting only very small part of a very large original dataset. Therefore, the quality is not expected to be very good. ์ด ์ผ์ ์์ด ํ๊ตญ์ด ๋ฒ์ญ์ ์ํ ์ฐ์ต์ผ๋ก ํ ๊ฒ์ด๊ธฐ ๋๋ฌธ์ ๋งค์ฐ ํฐ ์ dataset์์ ์์ฃผ ์์ ํฌ๊ธฐ๋ง์ ๊ธ๋ญ์น๋ง ์ ํ์ ํด์ ํ๋ จ์ ํ๋ค. ๋ฐ๋ผ์ ์ง์ ๊ทธ๋ฆฌ ์ข์ง ์์ ๊ฒ์ผ๋ก ์์๋๋ค.
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0005
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
Training results
Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Sacrebleu |
---|---|---|---|---|---|---|---|
3.3059 | 1.6 | 500 | 2.5597 | 0.3398 | 0.1261 | 0.3068 | 5.5524 |
2.4093 | 3.2 | 1000 | 2.4996 | 0.3609 | 0.144 | 0.3304 | 6.2002 |
2.2322 | 4.8 | 1500 | 2.4612 | 0.3666 | 0.147 | 0.3362 | 6.4622 |
Framework versions
- Transformers 4.32.0
- Pytorch 2.0.1+cu118
- Datasets 2.14.4
- Tokenizers 0.13.3
- Downloads last month
- 22
Model tree for chunwoolee0/mt5_small_wmt16_de_en
Base model
google/mt5-smallDataset used to train chunwoolee0/mt5_small_wmt16_de_en
Evaluation results
- Rouge1 on wmt16validation set self-reported0.367
- Sacrebleu on wmt16validation set self-reported6.462