Edit model card

KoT5_Translate_ko_jp

This model is a fine-tuned version of KETI-AIR/ke-t5-base on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.3331
  • Bleu: 44.5463

Model description

ํ•œ๊ตญ์–ด-์ผ๋ณธ์–ด ๋ฒˆ์—ญ๊ธฐ ๋ชจ๋ธ์„ ์œ„ํ•ด์„œ ๋งŒ๋“ค์—ˆ์Šต๋‹ˆ๋‹ค. KETI-AIR๋‹˜์ด ๊ณต์œ ํ•ด์ฃผ์‹  ke-t5-base์— Text2Text Task๋กœ ํ•œ๊ตญ์–ด-์ผ๋ณธ์–ด Translate๋ฅผ ์œ„ํ•ด์„œ Fine-Tuning ์ง„ํ–‰ํ•œ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค.

Training and evaluation data

noahkim/Kor_Jpn_Translation_Dataset ์ œ๊ฐ€ AIHub์—์„œ ๋‹ค์šด ๋ฐ›์•„ ํ—ˆ๊น…ํŽ˜์ด์Šค์— ๊ณต์œ ํ•œ ํ•œ๊ตญ์–ด-์ผ๋ณธ์–ด ๋ฌธํ™” ๋ถ„์•ผ ์ด์ค‘ ๋ง๋ญ‰์น˜๋ฅผ Fine-Tuning ๋ฐ์ดํ„ฐ์…‹์œผ๋กœ ํ™œ์šฉํ–ˆ์Šต๋‹ˆ๋‹ค.

Supported Tasks and Leaderboards

Translation

Languages

Kor Jpan

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0005
  • train_batch_size: 64
  • eval_batch_size: 128
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 128
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 1

Training results

Training Loss Epoch Step Validation Loss Bleu
3.8739 0.08 500 1.7216 3.3261
1.2621 0.15 1000 0.6792 28.6184
0.7413 0.23 1500 0.5153 35.9355
0.635 0.3 2000 0.4807 38.4874
0.5643 0.38 2500 0.4322 40.7997
0.5137 0.46 3000 0.4027 41.9025
0.4806 0.53 3500 0.3862 42.5947
0.4552 0.61 4000 0.3721 42.9976
0.4395 0.69 4500 0.3585 43.5369
0.4213 0.76 5000 0.3487 44.0028
0.411 0.84 5500 0.3418 44.1845
0.3992 0.91 6000 0.3348 44.3701
0.3966 0.99 6500 0.3331 44.5463

Framework versions

  • Transformers 4.26.1
  • Pytorch 1.13.1+cu116
  • Datasets 2.10.1
  • Tokenizers 0.13.2
Downloads last month
44