Edit model card

eng-tut

  • source group: English

  • target group: Altaic languages

  • OPUS readme: eng-tut

  • model: transformer

  • source language(s): eng

  • target language(s): aze_Latn bak chv crh crh_Latn kaz_Cyrl kaz_Latn kir_Cyrl kjh kum mon nog ota_Arab ota_Latn sah tat tat_Arab tat_Latn tuk tuk_Latn tur tyv uig_Arab uig_Cyrl uzb_Cyrl uzb_Latn xal

  • model: transformer

  • pre-processing: normalization + SentencePiece (spm32k,spm32k)

  • a sentence initial language token is required in the form of >>id<< (id = valid target language ID)

  • download original weights: opus2m-2020-08-02.zip

  • test set translations: opus2m-2020-08-02.test.txt

  • test set scores: opus2m-2020-08-02.eval.txt

Benchmarks

testset BLEU chr-F
newsdev2016-entr-engtur.eng.tur 10.4 0.438
newstest2016-entr-engtur.eng.tur 9.1 0.414
newstest2017-entr-engtur.eng.tur 9.5 0.414
newstest2018-entr-engtur.eng.tur 9.5 0.415
Tatoeba-test.eng-aze.eng.aze 27.2 0.580
Tatoeba-test.eng-bak.eng.bak 5.8 0.298
Tatoeba-test.eng-chv.eng.chv 4.6 0.301
Tatoeba-test.eng-crh.eng.crh 6.5 0.342
Tatoeba-test.eng-kaz.eng.kaz 11.8 0.360
Tatoeba-test.eng-kir.eng.kir 24.6 0.499
Tatoeba-test.eng-kjh.eng.kjh 2.2 0.052
Tatoeba-test.eng-kum.eng.kum 8.0 0.229
Tatoeba-test.eng-mon.eng.mon 10.3 0.362
Tatoeba-test.eng.multi 19.5 0.451
Tatoeba-test.eng-nog.eng.nog 1.5 0.117
Tatoeba-test.eng-ota.eng.ota 0.2 0.035
Tatoeba-test.eng-sah.eng.sah 0.7 0.080
Tatoeba-test.eng-tat.eng.tat 10.8 0.320
Tatoeba-test.eng-tuk.eng.tuk 5.6 0.323
Tatoeba-test.eng-tur.eng.tur 34.2 0.623
Tatoeba-test.eng-tyv.eng.tyv 8.1 0.192
Tatoeba-test.eng-uig.eng.uig 0.1 0.158
Tatoeba-test.eng-uzb.eng.uzb 4.2 0.298
Tatoeba-test.eng-xal.eng.xal 0.1 0.061

System Info:

Downloads last month
20
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Spaces using Helsinki-NLP/opus-mt-en-tut 7