Back to all models
translation mask_token:
Query this model
🔥 This model is currently loaded and running on the Inference API. ⚠️ This model could not be loaded by the inference API. ⚠️ This model can be loaded on the Inference API on-demand.
JSON Output
API endpoint  

⚡️ Upgrade your account to access the Inference API

Share Copied link to clipboard

Monthly model downloads

Helsinki-NLP/opus-mt-en-urj Helsinki-NLP/opus-mt-en-urj
N/a downloads
last 30 days

pytorch

tf

Contributed by

Language Technology Research Group at the University of Helsinki university
1 team member · 1325 models

How to use this model directly from the 🤗/transformers library:

			
Copy to clipboard
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM tokenizer = AutoTokenizer.from_pretrained("Helsinki-NLP/opus-mt-en-urj") model = AutoModelForSeq2SeqLM.from_pretrained("Helsinki-NLP/opus-mt-en-urj")
Uploaded in S3

eng-urj

  • source group: English

  • target group: Uralic languages

  • OPUS readme: eng-urj

  • model: transformer

  • source language(s): eng

  • target language(s): est fin fkv_Latn hun izh kpv krl liv_Latn mdf mhr myv sma sme udm vro

  • model: transformer

  • pre-processing: normalization + SentencePiece (spm32k,spm32k)

  • a sentence initial language token is required in the form of >>id<< (id = valid target language ID)

  • download original weights: opus2m-2020-08-02.zip

  • test set translations: opus2m-2020-08-02.test.txt

  • test set scores: opus2m-2020-08-02.eval.txt

Benchmarks

testset BLEU chr-F
newsdev2015-enfi-engfin.eng.fin 18.3 0.519
newsdev2018-enet-engest.eng.est 19.3 0.520
newssyscomb2009-enghun.eng.hun 15.4 0.471
newstest2009-enghun.eng.hun 15.7 0.468
newstest2015-enfi-engfin.eng.fin 20.2 0.534
newstest2016-enfi-engfin.eng.fin 20.7 0.541
newstest2017-enfi-engfin.eng.fin 23.6 0.566
newstest2018-enet-engest.eng.est 20.8 0.535
newstest2018-enfi-engfin.eng.fin 15.8 0.499
newstest2019-enfi-engfin.eng.fin 19.9 0.518
newstestB2016-enfi-engfin.eng.fin 16.6 0.509
newstestB2017-enfi-engfin.eng.fin 19.4 0.529
Tatoeba-test.eng-chm.eng.chm 1.3 0.127
Tatoeba-test.eng-est.eng.est 51.0 0.692
Tatoeba-test.eng-fin.eng.fin 34.6 0.597
Tatoeba-test.eng-fkv.eng.fkv 2.2 0.302
Tatoeba-test.eng-hun.eng.hun 35.6 0.591
Tatoeba-test.eng-izh.eng.izh 5.7 0.211
Tatoeba-test.eng-kom.eng.kom 3.0 0.012
Tatoeba-test.eng-krl.eng.krl 8.5 0.230
Tatoeba-test.eng-liv.eng.liv 2.7 0.077
Tatoeba-test.eng-mdf.eng.mdf 2.8 0.007
Tatoeba-test.eng.multi 35.1 0.588
Tatoeba-test.eng-myv.eng.myv 1.3 0.014
Tatoeba-test.eng-sma.eng.sma 1.8 0.095
Tatoeba-test.eng-sme.eng.sme 6.8 0.204
Tatoeba-test.eng-udm.eng.udm 1.1 0.121

System Info: