File size: 2,205 Bytes
4c9ef03
 
 
 
bfe5caa
 
4c9ef03
 
 
 
 
 
 
 
bfe5caa
4c9ef03
 
 
 
 
bfe5caa
4c9ef03
bfe5caa
4c9ef03
 
 
 
bfe5caa
0fcb40e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
---
language:
- en
- ru
- multilingual
license: cc-by-4.0
tags:
- translation
- opus-mt-tc
model-index:
- name: opus-mt-tc-base-en-ru
  results:
  - task:
      type: translation
      name: Translation eng-rus
    dataset:
      name: tatoeba-test-v2021-08-07
      type: tatoeba_mt
      args: eng-rus
    metrics:
    - type: bleu
      value: 45.6
      name: BLEU
---

# Opus Tatoeba English-Russian

*This model was obtained by running the script [convert_marian_to_pytorch.py](https://github.com/huggingface/transformers/blob/master/src/transformers/models/marian/convert_marian_to_pytorch.py). The original models were trained by [J�rg Tiedemann](https://blogs.helsinki.fi/tiedeman/) using the [MarianNMT](https://marian-nmt.github.io/) library. See all available `MarianMTModel` models on the profile of the [Helsinki NLP](https://huggingface.co/Helsinki-NLP) group.*

* dataset: opus+bt
* model: transformer-align
* source language(s): eng
* target language(s): rus
* model: transformer-align
* pre-processing: normalization + SentencePiece (spm32k,spm32k)
* download: [opus+bt-2021-04-14.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/eng-rus/opus+bt-2021-04-14.zip)
* test set translations: [opus+bt-2021-04-14.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/eng-rus/opus+bt-2021-04-14.test.txt)
* test set scores: [opus+bt-2021-04-14.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/eng-rus/opus+bt-2021-04-14.eval.txt)

## Benchmarks

| testset | BLEU  | chr-F | #sent | #words | BP |
|---------|-------|-------|-------|--------|----|
| newstest2012.eng-rus 	| 31.2 	| 0.588 	| 3003 	| 64830 	| 0.980 |
| newstest2013.eng-rus 	| 23.2 	| 0.518 	| 3000 	| 58560 	| 0.974 |
| newstest2015-enru.eng-rus 	| 29.1 	| 0.581 	| 2818 	| 55915 	| 1.000 |
| newstest2016-enru.eng-rus 	| 27.4 	| 0.565 	| 2998 	| 62018 	| 0.993 |
| newstest2017-enru.eng-rus 	| 30.8 	| 0.592 	| 3001 	| 60255 	| 0.998 |
| newstest2018-enru.eng-rus 	| 27.4 	| 0.572 	| 3000 	| 61920 	| 1.000 |
| newstest2019-enru.eng-rus 	| 27.1 	| 0.540 	| 1997 	| 48153 	| 0.927 |
| Tatoeba-test.eng-rus 	| 45.6 	| 0.652 	| 10000 	| 66872 	| 0.987 |
| tico19-test.eng-rus 	| 27.5 	| 0.556 	| 2100 	| 55837 	| 0.927 |