tdobrxl's picture
Update README.md
959db39
|
raw
history blame
No virus
779 Bytes
# Overview
This is a fine-tuned version of the model [Helsinki-NLP/opus-mt-en-vi](https://huggingface.co/Helsinki-NLP/opus-mt-en-vi?text=My+name+is+Sarah+and+I+live+in+London) on the dataset [IWSLT'15 English-Vietnamese](https://huggingface.co/datasets/mt_eng_vietnamese).
Performance in terms of [sacrebleu](https://huggingface.co/docs/datasets/v1.5.0/using_metrics.html) on the test set is as follows:
* Original opus-mt-en-vi: 29.83
* Fine-tuned opus-mt-en-vi: 37.35
# Parameters
* learning_rate=2e-5
* batch_size: 32
* weight_decay=0.01
* num_train_epochs=1
# Thoughts
* Model `Helsinki-NLP/opus-mt-en-vi` is small (around 260MB), and can be easily deployed to a cheap server (e.g., EC2 t2.medium) without a GPU
* Easier and much faster to train compared to t5 or byt5.