tiedeman commited on
Commit
c93612f
1 Parent(s): 0a2c562

Initial commit

Browse files
.gitattributes CHANGED
@@ -29,3 +29,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
29
  *.zip filter=lfs diff=lfs merge=lfs -text
30
  *.zst filter=lfs diff=lfs merge=lfs -text
31
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
29
  *.zip filter=lfs diff=lfs merge=lfs -text
30
  *.zst filter=lfs diff=lfs merge=lfs -text
31
  *tfevents* filter=lfs diff=lfs merge=lfs -text
32
+ *.spm filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,2354 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - ast
4
+ - ca
5
+ - cbk
6
+ - es
7
+ - fr
8
+ - fro
9
+ - gcf
10
+ - gl
11
+ - ht
12
+ - it
13
+ - itc
14
+ - la
15
+ - la_Latn
16
+ - lad
17
+ - lij
18
+ - lld
19
+ - oc
20
+ - pms
21
+ - pt
22
+ - ro
23
+ - xx_Latn
24
+
25
+ tags:
26
+ - translation
27
+ - opus-mt-tc
28
+
29
+ license: cc-by-4.0
30
+ model-index:
31
+ - name: opus-mt-tc-big-itc-itc
32
+ results:
33
+ - task:
34
+ name: Translation ast-cat
35
+ type: translation
36
+ args: ast-cat
37
+ dataset:
38
+ name: flores101-devtest
39
+ type: flores_101
40
+ args: ast cat devtest
41
+ metrics:
42
+ - name: BLEU
43
+ type: bleu
44
+ value: 31.8
45
+ - name: chr-F
46
+ type: chrf
47
+ value: 0.57870
48
+ - task:
49
+ name: Translation ast-fra
50
+ type: translation
51
+ args: ast-fra
52
+ dataset:
53
+ name: flores101-devtest
54
+ type: flores_101
55
+ args: ast fra devtest
56
+ metrics:
57
+ - name: BLEU
58
+ type: bleu
59
+ value: 31.1
60
+ - name: chr-F
61
+ type: chrf
62
+ value: 0.56761
63
+ - task:
64
+ name: Translation ast-glg
65
+ type: translation
66
+ args: ast-glg
67
+ dataset:
68
+ name: flores101-devtest
69
+ type: flores_101
70
+ args: ast glg devtest
71
+ metrics:
72
+ - name: BLEU
73
+ type: bleu
74
+ value: 27.9
75
+ - name: chr-F
76
+ type: chrf
77
+ value: 0.55161
78
+ - task:
79
+ name: Translation ast-ita
80
+ type: translation
81
+ args: ast-ita
82
+ dataset:
83
+ name: flores101-devtest
84
+ type: flores_101
85
+ args: ast ita devtest
86
+ metrics:
87
+ - name: BLEU
88
+ type: bleu
89
+ value: 22.1
90
+ - name: chr-F
91
+ type: chrf
92
+ value: 0.51764
93
+ - task:
94
+ name: Translation ast-oci
95
+ type: translation
96
+ args: ast-oci
97
+ dataset:
98
+ name: flores101-devtest
99
+ type: flores_101
100
+ args: ast oci devtest
101
+ metrics:
102
+ - name: BLEU
103
+ type: bleu
104
+ value: 20.6
105
+ - name: chr-F
106
+ type: chrf
107
+ value: 0.49545
108
+ - task:
109
+ name: Translation ast-por
110
+ type: translation
111
+ args: ast-por
112
+ dataset:
113
+ name: flores101-devtest
114
+ type: flores_101
115
+ args: ast por devtest
116
+ metrics:
117
+ - name: BLEU
118
+ type: bleu
119
+ value: 31.5
120
+ - name: chr-F
121
+ type: chrf
122
+ value: 0.57347
123
+ - task:
124
+ name: Translation ast-ron
125
+ type: translation
126
+ args: ast-ron
127
+ dataset:
128
+ name: flores101-devtest
129
+ type: flores_101
130
+ args: ast ron devtest
131
+ metrics:
132
+ - name: BLEU
133
+ type: bleu
134
+ value: 24.8
135
+ - name: chr-F
136
+ type: chrf
137
+ value: 0.52317
138
+ - task:
139
+ name: Translation ast-spa
140
+ type: translation
141
+ args: ast-spa
142
+ dataset:
143
+ name: flores101-devtest
144
+ type: flores_101
145
+ args: ast spa devtest
146
+ metrics:
147
+ - name: BLEU
148
+ type: bleu
149
+ value: 21.2
150
+ - name: chr-F
151
+ type: chrf
152
+ value: 0.49741
153
+ - task:
154
+ name: Translation cat-ast
155
+ type: translation
156
+ args: cat-ast
157
+ dataset:
158
+ name: flores101-devtest
159
+ type: flores_101
160
+ args: cat ast devtest
161
+ metrics:
162
+ - name: BLEU
163
+ type: bleu
164
+ value: 24.7
165
+ - name: chr-F
166
+ type: chrf
167
+ value: 0.56754
168
+ - task:
169
+ name: Translation cat-fra
170
+ type: translation
171
+ args: cat-fra
172
+ dataset:
173
+ name: flores101-devtest
174
+ type: flores_101
175
+ args: cat fra devtest
176
+ metrics:
177
+ - name: BLEU
178
+ type: bleu
179
+ value: 38.4
180
+ - name: chr-F
181
+ type: chrf
182
+ value: 0.63368
183
+ - task:
184
+ name: Translation cat-glg
185
+ type: translation
186
+ args: cat-glg
187
+ dataset:
188
+ name: flores101-devtest
189
+ type: flores_101
190
+ args: cat glg devtest
191
+ metrics:
192
+ - name: BLEU
193
+ type: bleu
194
+ value: 32.2
195
+ - name: chr-F
196
+ type: chrf
197
+ value: 0.59596
198
+ - task:
199
+ name: Translation cat-ita
200
+ type: translation
201
+ args: cat-ita
202
+ dataset:
203
+ name: flores101-devtest
204
+ type: flores_101
205
+ args: cat ita devtest
206
+ metrics:
207
+ - name: BLEU
208
+ type: bleu
209
+ value: 26.3
210
+ - name: chr-F
211
+ type: chrf
212
+ value: 0.55886
213
+ - task:
214
+ name: Translation cat-oci
215
+ type: translation
216
+ args: cat-oci
217
+ dataset:
218
+ name: flores101-devtest
219
+ type: flores_101
220
+ args: cat oci devtest
221
+ metrics:
222
+ - name: BLEU
223
+ type: bleu
224
+ value: 24.6
225
+ - name: chr-F
226
+ type: chrf
227
+ value: 0.54285
228
+ - task:
229
+ name: Translation cat-por
230
+ type: translation
231
+ args: cat-por
232
+ dataset:
233
+ name: flores101-devtest
234
+ type: flores_101
235
+ args: cat por devtest
236
+ metrics:
237
+ - name: BLEU
238
+ type: bleu
239
+ value: 37.7
240
+ - name: chr-F
241
+ type: chrf
242
+ value: 0.62913
243
+ - task:
244
+ name: Translation cat-ron
245
+ type: translation
246
+ args: cat-ron
247
+ dataset:
248
+ name: flores101-devtest
249
+ type: flores_101
250
+ args: cat ron devtest
251
+ metrics:
252
+ - name: BLEU
253
+ type: bleu
254
+ value: 29.5
255
+ - name: chr-F
256
+ type: chrf
257
+ value: 0.56885
258
+ - task:
259
+ name: Translation cat-spa
260
+ type: translation
261
+ args: cat-spa
262
+ dataset:
263
+ name: flores101-devtest
264
+ type: flores_101
265
+ args: cat spa devtest
266
+ metrics:
267
+ - name: BLEU
268
+ type: bleu
269
+ value: 24.6
270
+ - name: chr-F
271
+ type: chrf
272
+ value: 0.53372
273
+ - task:
274
+ name: Translation fra-ast
275
+ type: translation
276
+ args: fra-ast
277
+ dataset:
278
+ name: flores101-devtest
279
+ type: flores_101
280
+ args: fra ast devtest
281
+ metrics:
282
+ - name: BLEU
283
+ type: bleu
284
+ value: 20.7
285
+ - name: chr-F
286
+ type: chrf
287
+ value: 0.52696
288
+ - task:
289
+ name: Translation fra-cat
290
+ type: translation
291
+ args: fra-cat
292
+ dataset:
293
+ name: flores101-devtest
294
+ type: flores_101
295
+ args: fra cat devtest
296
+ metrics:
297
+ - name: BLEU
298
+ type: bleu
299
+ value: 34.6
300
+ - name: chr-F
301
+ type: chrf
302
+ value: 0.60492
303
+ - task:
304
+ name: Translation fra-glg
305
+ type: translation
306
+ args: fra-glg
307
+ dataset:
308
+ name: flores101-devtest
309
+ type: flores_101
310
+ args: fra glg devtest
311
+ metrics:
312
+ - name: BLEU
313
+ type: bleu
314
+ value: 30.3
315
+ - name: chr-F
316
+ type: chrf
317
+ value: 0.57485
318
+ - task:
319
+ name: Translation fra-ita
320
+ type: translation
321
+ args: fra-ita
322
+ dataset:
323
+ name: flores101-devtest
324
+ type: flores_101
325
+ args: fra ita devtest
326
+ metrics:
327
+ - name: BLEU
328
+ type: bleu
329
+ value: 27.3
330
+ - name: chr-F
331
+ type: chrf
332
+ value: 0.56493
333
+ - task:
334
+ name: Translation fra-oci
335
+ type: translation
336
+ args: fra-oci
337
+ dataset:
338
+ name: flores101-devtest
339
+ type: flores_101
340
+ args: fra oci devtest
341
+ metrics:
342
+ - name: BLEU
343
+ type: bleu
344
+ value: 28.2
345
+ - name: chr-F
346
+ type: chrf
347
+ value: 0.57449
348
+ - task:
349
+ name: Translation fra-por
350
+ type: translation
351
+ args: fra-por
352
+ dataset:
353
+ name: flores101-devtest
354
+ type: flores_101
355
+ args: fra por devtest
356
+ metrics:
357
+ - name: BLEU
358
+ type: bleu
359
+ value: 36.9
360
+ - name: chr-F
361
+ type: chrf
362
+ value: 0.62211
363
+ - task:
364
+ name: Translation fra-ron
365
+ type: translation
366
+ args: fra-ron
367
+ dataset:
368
+ name: flores101-devtest
369
+ type: flores_101
370
+ args: fra ron devtest
371
+ metrics:
372
+ - name: BLEU
373
+ type: bleu
374
+ value: 29.4
375
+ - name: chr-F
376
+ type: chrf
377
+ value: 0.56998
378
+ - task:
379
+ name: Translation fra-spa
380
+ type: translation
381
+ args: fra-spa
382
+ dataset:
383
+ name: flores101-devtest
384
+ type: flores_101
385
+ args: fra spa devtest
386
+ metrics:
387
+ - name: BLEU
388
+ type: bleu
389
+ value: 24.2
390
+ - name: chr-F
391
+ type: chrf
392
+ value: 0.52880
393
+ - task:
394
+ name: Translation glg-ast
395
+ type: translation
396
+ args: glg-ast
397
+ dataset:
398
+ name: flores101-devtest
399
+ type: flores_101
400
+ args: glg ast devtest
401
+ metrics:
402
+ - name: BLEU
403
+ type: bleu
404
+ value: 22.4
405
+ - name: chr-F
406
+ type: chrf
407
+ value: 0.55090
408
+ - task:
409
+ name: Translation glg-cat
410
+ type: translation
411
+ args: glg-cat
412
+ dataset:
413
+ name: flores101-devtest
414
+ type: flores_101
415
+ args: glg cat devtest
416
+ metrics:
417
+ - name: BLEU
418
+ type: bleu
419
+ value: 32.6
420
+ - name: chr-F
421
+ type: chrf
422
+ value: 0.60550
423
+ - task:
424
+ name: Translation glg-fra
425
+ type: translation
426
+ args: glg-fra
427
+ dataset:
428
+ name: flores101-devtest
429
+ type: flores_101
430
+ args: glg fra devtest
431
+ metrics:
432
+ - name: BLEU
433
+ type: bleu
434
+ value: 36.0
435
+ - name: chr-F
436
+ type: chrf
437
+ value: 0.62026
438
+ - task:
439
+ name: Translation glg-ita
440
+ type: translation
441
+ args: glg-ita
442
+ dataset:
443
+ name: flores101-devtest
444
+ type: flores_101
445
+ args: glg ita devtest
446
+ metrics:
447
+ - name: BLEU
448
+ type: bleu
449
+ value: 25.9
450
+ - name: chr-F
451
+ type: chrf
452
+ value: 0.55834
453
+ - task:
454
+ name: Translation glg-oci
455
+ type: translation
456
+ args: glg-oci
457
+ dataset:
458
+ name: flores101-devtest
459
+ type: flores_101
460
+ args: glg oci devtest
461
+ metrics:
462
+ - name: BLEU
463
+ type: bleu
464
+ value: 21.9
465
+ - name: chr-F
466
+ type: chrf
467
+ value: 0.52520
468
+ - task:
469
+ name: Translation glg-por
470
+ type: translation
471
+ args: glg-por
472
+ dataset:
473
+ name: flores101-devtest
474
+ type: flores_101
475
+ args: glg por devtest
476
+ metrics:
477
+ - name: BLEU
478
+ type: bleu
479
+ value: 32.7
480
+ - name: chr-F
481
+ type: chrf
482
+ value: 0.60027
483
+ - task:
484
+ name: Translation glg-ron
485
+ type: translation
486
+ args: glg-ron
487
+ dataset:
488
+ name: flores101-devtest
489
+ type: flores_101
490
+ args: glg ron devtest
491
+ metrics:
492
+ - name: BLEU
493
+ type: bleu
494
+ value: 27.8
495
+ - name: chr-F
496
+ type: chrf
497
+ value: 0.55621
498
+ - task:
499
+ name: Translation glg-spa
500
+ type: translation
501
+ args: glg-spa
502
+ dataset:
503
+ name: flores101-devtest
504
+ type: flores_101
505
+ args: glg spa devtest
506
+ metrics:
507
+ - name: BLEU
508
+ type: bleu
509
+ value: 24.4
510
+ - name: chr-F
511
+ type: chrf
512
+ value: 0.53219
513
+ - task:
514
+ name: Translation ita-ast
515
+ type: translation
516
+ args: ita-ast
517
+ dataset:
518
+ name: flores101-devtest
519
+ type: flores_101
520
+ args: ita ast devtest
521
+ metrics:
522
+ - name: BLEU
523
+ type: bleu
524
+ value: 17.1
525
+ - name: chr-F
526
+ type: chrf
527
+ value: 0.50741
528
+ - task:
529
+ name: Translation ita-cat
530
+ type: translation
531
+ args: ita-cat
532
+ dataset:
533
+ name: flores101-devtest
534
+ type: flores_101
535
+ args: ita cat devtest
536
+ metrics:
537
+ - name: BLEU
538
+ type: bleu
539
+ value: 27.9
540
+ - name: chr-F
541
+ type: chrf
542
+ value: 0.57061
543
+ - task:
544
+ name: Translation ita-fra
545
+ type: translation
546
+ args: ita-fra
547
+ dataset:
548
+ name: flores101-devtest
549
+ type: flores_101
550
+ args: ita fra devtest
551
+ metrics:
552
+ - name: BLEU
553
+ type: bleu
554
+ value: 32.0
555
+ - name: chr-F
556
+ type: chrf
557
+ value: 0.60199
558
+ - task:
559
+ name: Translation ita-glg
560
+ type: translation
561
+ args: ita-glg
562
+ dataset:
563
+ name: flores101-devtest
564
+ type: flores_101
565
+ args: ita glg devtest
566
+ metrics:
567
+ - name: BLEU
568
+ type: bleu
569
+ value: 25.9
570
+ - name: chr-F
571
+ type: chrf
572
+ value: 0.55312
573
+ - task:
574
+ name: Translation ita-oci
575
+ type: translation
576
+ args: ita-oci
577
+ dataset:
578
+ name: flores101-devtest
579
+ type: flores_101
580
+ args: ita oci devtest
581
+ metrics:
582
+ - name: BLEU
583
+ type: bleu
584
+ value: 18.1
585
+ - name: chr-F
586
+ type: chrf
587
+ value: 0.48447
588
+ - task:
589
+ name: Translation ita-por
590
+ type: translation
591
+ args: ita-por
592
+ dataset:
593
+ name: flores101-devtest
594
+ type: flores_101
595
+ args: ita por devtest
596
+ metrics:
597
+ - name: BLEU
598
+ type: bleu
599
+ value: 29.0
600
+ - name: chr-F
601
+ type: chrf
602
+ value: 0.58162
603
+ - task:
604
+ name: Translation ita-ron
605
+ type: translation
606
+ args: ita-ron
607
+ dataset:
608
+ name: flores101-devtest
609
+ type: flores_101
610
+ args: ita ron devtest
611
+ metrics:
612
+ - name: BLEU
613
+ type: bleu
614
+ value: 24.2
615
+ - name: chr-F
616
+ type: chrf
617
+ value: 0.53703
618
+ - task:
619
+ name: Translation ita-spa
620
+ type: translation
621
+ args: ita-spa
622
+ dataset:
623
+ name: flores101-devtest
624
+ type: flores_101
625
+ args: ita spa devtest
626
+ metrics:
627
+ - name: BLEU
628
+ type: bleu
629
+ value: 23.1
630
+ - name: chr-F
631
+ type: chrf
632
+ value: 0.52238
633
+ - task:
634
+ name: Translation oci-ast
635
+ type: translation
636
+ args: oci-ast
637
+ dataset:
638
+ name: flores101-devtest
639
+ type: flores_101
640
+ args: oci ast devtest
641
+ metrics:
642
+ - name: BLEU
643
+ type: bleu
644
+ value: 20.2
645
+ - name: chr-F
646
+ type: chrf
647
+ value: 0.53010
648
+ - task:
649
+ name: Translation oci-cat
650
+ type: translation
651
+ args: oci-cat
652
+ dataset:
653
+ name: flores101-devtest
654
+ type: flores_101
655
+ args: oci cat devtest
656
+ metrics:
657
+ - name: BLEU
658
+ type: bleu
659
+ value: 32.2
660
+ - name: chr-F
661
+ type: chrf
662
+ value: 0.59946
663
+ - task:
664
+ name: Translation oci-fra
665
+ type: translation
666
+ args: oci-fra
667
+ dataset:
668
+ name: flores101-devtest
669
+ type: flores_101
670
+ args: oci fra devtest
671
+ metrics:
672
+ - name: BLEU
673
+ type: bleu
674
+ value: 39.0
675
+ - name: chr-F
676
+ type: chrf
677
+ value: 0.64290
678
+ - task:
679
+ name: Translation oci-glg
680
+ type: translation
681
+ args: oci-glg
682
+ dataset:
683
+ name: flores101-devtest
684
+ type: flores_101
685
+ args: oci glg devtest
686
+ metrics:
687
+ - name: BLEU
688
+ type: bleu
689
+ value: 28.0
690
+ - name: chr-F
691
+ type: chrf
692
+ value: 0.56737
693
+ - task:
694
+ name: Translation oci-ita
695
+ type: translation
696
+ args: oci-ita
697
+ dataset:
698
+ name: flores101-devtest
699
+ type: flores_101
700
+ args: oci ita devtest
701
+ metrics:
702
+ - name: BLEU
703
+ type: bleu
704
+ value: 24.2
705
+ - name: chr-F
706
+ type: chrf
707
+ value: 0.54220
708
+ - task:
709
+ name: Translation oci-por
710
+ type: translation
711
+ args: oci-por
712
+ dataset:
713
+ name: flores101-devtest
714
+ type: flores_101
715
+ args: oci por devtest
716
+ metrics:
717
+ - name: BLEU
718
+ type: bleu
719
+ value: 35.7
720
+ - name: chr-F
721
+ type: chrf
722
+ value: 0.62127
723
+ - task:
724
+ name: Translation oci-ron
725
+ type: translation
726
+ args: oci-ron
727
+ dataset:
728
+ name: flores101-devtest
729
+ type: flores_101
730
+ args: oci ron devtest
731
+ metrics:
732
+ - name: BLEU
733
+ type: bleu
734
+ value: 28.0
735
+ - name: chr-F
736
+ type: chrf
737
+ value: 0.55906
738
+ - task:
739
+ name: Translation oci-spa
740
+ type: translation
741
+ args: oci-spa
742
+ dataset:
743
+ name: flores101-devtest
744
+ type: flores_101
745
+ args: oci spa devtest
746
+ metrics:
747
+ - name: BLEU
748
+ type: bleu
749
+ value: 22.8
750
+ - name: chr-F
751
+ type: chrf
752
+ value: 0.52110
753
+ - task:
754
+ name: Translation por-ast
755
+ type: translation
756
+ args: por-ast
757
+ dataset:
758
+ name: flores101-devtest
759
+ type: flores_101
760
+ args: por ast devtest
761
+ metrics:
762
+ - name: BLEU
763
+ type: bleu
764
+ value: 22.5
765
+ - name: chr-F
766
+ type: chrf
767
+ value: 0.54539
768
+ - task:
769
+ name: Translation por-cat
770
+ type: translation
771
+ args: por-cat
772
+ dataset:
773
+ name: flores101-devtest
774
+ type: flores_101
775
+ args: por cat devtest
776
+ metrics:
777
+ - name: BLEU
778
+ type: bleu
779
+ value: 36.4
780
+ - name: chr-F
781
+ type: chrf
782
+ value: 0.61809
783
+ - task:
784
+ name: Translation por-fra
785
+ type: translation
786
+ args: por-fra
787
+ dataset:
788
+ name: flores101-devtest
789
+ type: flores_101
790
+ args: por fra devtest
791
+ metrics:
792
+ - name: BLEU
793
+ type: bleu
794
+ value: 39.7
795
+ - name: chr-F
796
+ type: chrf
797
+ value: 0.64343
798
+ - task:
799
+ name: Translation por-glg
800
+ type: translation
801
+ args: por-glg
802
+ dataset:
803
+ name: flores101-devtest
804
+ type: flores_101
805
+ args: por glg devtest
806
+ metrics:
807
+ - name: BLEU
808
+ type: bleu
809
+ value: 30.4
810
+ - name: chr-F
811
+ type: chrf
812
+ value: 0.57965
813
+ - task:
814
+ name: Translation por-ita
815
+ type: translation
816
+ args: por-ita
817
+ dataset:
818
+ name: flores101-devtest
819
+ type: flores_101
820
+ args: por ita devtest
821
+ metrics:
822
+ - name: BLEU
823
+ type: bleu
824
+ value: 26.3
825
+ - name: chr-F
826
+ type: chrf
827
+ value: 0.55841
828
+ - task:
829
+ name: Translation por-oci
830
+ type: translation
831
+ args: por-oci
832
+ dataset:
833
+ name: flores101-devtest
834
+ type: flores_101
835
+ args: por oci devtest
836
+ metrics:
837
+ - name: BLEU
838
+ type: bleu
839
+ value: 25.3
840
+ - name: chr-F
841
+ type: chrf
842
+ value: 0.54829
843
+ - task:
844
+ name: Translation por-ron
845
+ type: translation
846
+ args: por-ron
847
+ dataset:
848
+ name: flores101-devtest
849
+ type: flores_101
850
+ args: por ron devtest
851
+ metrics:
852
+ - name: BLEU
853
+ type: bleu
854
+ value: 29.8
855
+ - name: chr-F
856
+ type: chrf
857
+ value: 0.57283
858
+ - task:
859
+ name: Translation por-spa
860
+ type: translation
861
+ args: por-spa
862
+ dataset:
863
+ name: flores101-devtest
864
+ type: flores_101
865
+ args: por spa devtest
866
+ metrics:
867
+ - name: BLEU
868
+ type: bleu
869
+ value: 25.2
870
+ - name: chr-F
871
+ type: chrf
872
+ value: 0.53513
873
+ - task:
874
+ name: Translation ron-ast
875
+ type: translation
876
+ args: ron-ast
877
+ dataset:
878
+ name: flores101-devtest
879
+ type: flores_101
880
+ args: ron ast devtest
881
+ metrics:
882
+ - name: BLEU
883
+ type: bleu
884
+ value: 20.1
885
+ - name: chr-F
886
+ type: chrf
887
+ value: 0.52265
888
+ - task:
889
+ name: Translation ron-cat
890
+ type: translation
891
+ args: ron-cat
892
+ dataset:
893
+ name: flores101-devtest
894
+ type: flores_101
895
+ args: ron cat devtest
896
+ metrics:
897
+ - name: BLEU
898
+ type: bleu
899
+ value: 32.6
900
+ - name: chr-F
901
+ type: chrf
902
+ value: 0.59689
903
+ - task:
904
+ name: Translation ron-fra
905
+ type: translation
906
+ args: ron-fra
907
+ dataset:
908
+ name: flores101-devtest
909
+ type: flores_101
910
+ args: ron fra devtest
911
+ metrics:
912
+ - name: BLEU
913
+ type: bleu
914
+ value: 37.4
915
+ - name: chr-F
916
+ type: chrf
917
+ value: 0.63060
918
+ - task:
919
+ name: Translation ron-glg
920
+ type: translation
921
+ args: ron-glg
922
+ dataset:
923
+ name: flores101-devtest
924
+ type: flores_101
925
+ args: ron glg devtest
926
+ metrics:
927
+ - name: BLEU
928
+ type: bleu
929
+ value: 29.3
930
+ - name: chr-F
931
+ type: chrf
932
+ value: 0.56677
933
+ - task:
934
+ name: Translation ron-ita
935
+ type: translation
936
+ args: ron-ita
937
+ dataset:
938
+ name: flores101-devtest
939
+ type: flores_101
940
+ args: ron ita devtest
941
+ metrics:
942
+ - name: BLEU
943
+ type: bleu
944
+ value: 25.6
945
+ - name: chr-F
946
+ type: chrf
947
+ value: 0.55485
948
+ - task:
949
+ name: Translation ron-oci
950
+ type: translation
951
+ args: ron-oci
952
+ dataset:
953
+ name: flores101-devtest
954
+ type: flores_101
955
+ args: ron oci devtest
956
+ metrics:
957
+ - name: BLEU
958
+ type: bleu
959
+ value: 21.8
960
+ - name: chr-F
961
+ type: chrf
962
+ value: 0.52433
963
+ - task:
964
+ name: Translation ron-por
965
+ type: translation
966
+ args: ron-por
967
+ dataset:
968
+ name: flores101-devtest
969
+ type: flores_101
970
+ args: ron por devtest
971
+ metrics:
972
+ - name: BLEU
973
+ type: bleu
974
+ value: 36.1
975
+ - name: chr-F
976
+ type: chrf
977
+ value: 0.61831
978
+ - task:
979
+ name: Translation ron-spa
980
+ type: translation
981
+ args: ron-spa
982
+ dataset:
983
+ name: flores101-devtest
984
+ type: flores_101
985
+ args: ron spa devtest
986
+ metrics:
987
+ - name: BLEU
988
+ type: bleu
989
+ value: 24.1
990
+ - name: chr-F
991
+ type: chrf
992
+ value: 0.52712
993
+ - task:
994
+ name: Translation spa-ast
995
+ type: translation
996
+ args: spa-ast
997
+ dataset:
998
+ name: flores101-devtest
999
+ type: flores_101
1000
+ args: spa ast devtest
1001
+ metrics:
1002
+ - name: BLEU
1003
+ type: bleu
1004
+ value: 15.7
1005
+ - name: chr-F
1006
+ type: chrf
1007
+ value: 0.49008
1008
+ - task:
1009
+ name: Translation spa-cat
1010
+ type: translation
1011
+ args: spa-cat
1012
+ dataset:
1013
+ name: flores101-devtest
1014
+ type: flores_101
1015
+ args: spa cat devtest
1016
+ metrics:
1017
+ - name: BLEU
1018
+ type: bleu
1019
+ value: 23.2
1020
+ - name: chr-F
1021
+ type: chrf
1022
+ value: 0.53905
1023
+ - task:
1024
+ name: Translation spa-fra
1025
+ type: translation
1026
+ args: spa-fra
1027
+ dataset:
1028
+ name: flores101-devtest
1029
+ type: flores_101
1030
+ args: spa fra devtest
1031
+ metrics:
1032
+ - name: BLEU
1033
+ type: bleu
1034
+ value: 27.4
1035
+ - name: chr-F
1036
+ type: chrf
1037
+ value: 0.57078
1038
+ - task:
1039
+ name: Translation spa-glg
1040
+ type: translation
1041
+ args: spa-glg
1042
+ dataset:
1043
+ name: flores101-devtest
1044
+ type: flores_101
1045
+ args: spa glg devtest
1046
+ metrics:
1047
+ - name: BLEU
1048
+ type: bleu
1049
+ value: 22.0
1050
+ - name: chr-F
1051
+ type: chrf
1052
+ value: 0.52563
1053
+ - task:
1054
+ name: Translation spa-ita
1055
+ type: translation
1056
+ args: spa-ita
1057
+ dataset:
1058
+ name: flores101-devtest
1059
+ type: flores_101
1060
+ args: spa ita devtest
1061
+ metrics:
1062
+ - name: BLEU
1063
+ type: bleu
1064
+ value: 22.3
1065
+ - name: chr-F
1066
+ type: chrf
1067
+ value: 0.52783
1068
+ - task:
1069
+ name: Translation spa-oci
1070
+ type: translation
1071
+ args: spa-oci
1072
+ dataset:
1073
+ name: flores101-devtest
1074
+ type: flores_101
1075
+ args: spa oci devtest
1076
+ metrics:
1077
+ - name: BLEU
1078
+ type: bleu
1079
+ value: 16.3
1080
+ - name: chr-F
1081
+ type: chrf
1082
+ value: 0.48064
1083
+ - task:
1084
+ name: Translation spa-por
1085
+ type: translation
1086
+ args: spa-por
1087
+ dataset:
1088
+ name: flores101-devtest
1089
+ type: flores_101
1090
+ args: spa por devtest
1091
+ metrics:
1092
+ - name: BLEU
1093
+ type: bleu
1094
+ value: 25.8
1095
+ - name: chr-F
1096
+ type: chrf
1097
+ value: 0.55736
1098
+ - task:
1099
+ name: Translation spa-ron
1100
+ type: translation
1101
+ args: spa-ron
1102
+ dataset:
1103
+ name: flores101-devtest
1104
+ type: flores_101
1105
+ args: spa ron devtest
1106
+ metrics:
1107
+ - name: BLEU
1108
+ type: bleu
1109
+ value: 21.4
1110
+ - name: chr-F
1111
+ type: chrf
1112
+ value: 0.51623
1113
+ - task:
1114
+ name: Translation fra-spa
1115
+ type: translation
1116
+ args: fra-spa
1117
+ dataset:
1118
+ name: news-test2008
1119
+ type: news-test2008
1120
+ args: fra-spa
1121
+ metrics:
1122
+ - name: BLEU
1123
+ type: bleu
1124
+ value: 33.9
1125
+ - name: chr-F
1126
+ type: chrf
1127
+ value: 0.58939
1128
+ - task:
1129
+ name: Translation spa-fra
1130
+ type: translation
1131
+ args: spa-fra
1132
+ dataset:
1133
+ name: news-test2008
1134
+ type: news-test2008
1135
+ args: spa-fra
1136
+ metrics:
1137
+ - name: BLEU
1138
+ type: bleu
1139
+ value: 32.4
1140
+ - name: chr-F
1141
+ type: chrf
1142
+ value: 0.58695
1143
+ - task:
1144
+ name: Translation cat-fra
1145
+ type: translation
1146
+ args: cat-fra
1147
+ dataset:
1148
+ name: tatoeba-test-v2021-08-07
1149
+ type: tatoeba_mt
1150
+ args: cat-fra
1151
+ metrics:
1152
+ - name: BLEU
1153
+ type: bleu
1154
+ value: 54.6
1155
+ - name: chr-F
1156
+ type: chrf
1157
+ value: 0.71201
1158
+ - task:
1159
+ name: Translation cat-ita
1160
+ type: translation
1161
+ args: cat-ita
1162
+ dataset:
1163
+ name: tatoeba-test-v2021-08-07
1164
+ type: tatoeba_mt
1165
+ args: cat-ita
1166
+ metrics:
1167
+ - name: BLEU
1168
+ type: bleu
1169
+ value: 58.4
1170
+ - name: chr-F
1171
+ type: chrf
1172
+ value: 0.74198
1173
+ - task:
1174
+ name: Translation cat-por
1175
+ type: translation
1176
+ args: cat-por
1177
+ dataset:
1178
+ name: tatoeba-test-v2021-08-07
1179
+ type: tatoeba_mt
1180
+ args: cat-por
1181
+ metrics:
1182
+ - name: BLEU
1183
+ type: bleu
1184
+ value: 57.4
1185
+ - name: chr-F
1186
+ type: chrf
1187
+ value: 0.74930
1188
+ - task:
1189
+ name: Translation cat-spa
1190
+ type: translation
1191
+ args: cat-spa
1192
+ dataset:
1193
+ name: tatoeba-test-v2021-08-07
1194
+ type: tatoeba_mt
1195
+ args: cat-spa
1196
+ metrics:
1197
+ - name: BLEU
1198
+ type: bleu
1199
+ value: 78.1
1200
+ - name: chr-F
1201
+ type: chrf
1202
+ value: 0.87844
1203
+ - task:
1204
+ name: Translation fra-cat
1205
+ type: translation
1206
+ args: fra-cat
1207
+ dataset:
1208
+ name: tatoeba-test-v2021-08-07
1209
+ type: tatoeba_mt
1210
+ args: fra-cat
1211
+ metrics:
1212
+ - name: BLEU
1213
+ type: bleu
1214
+ value: 46.2
1215
+ - name: chr-F
1216
+ type: chrf
1217
+ value: 0.66525
1218
+ - task:
1219
+ name: Translation fra-ita
1220
+ type: translation
1221
+ args: fra-ita
1222
+ dataset:
1223
+ name: tatoeba-test-v2021-08-07
1224
+ type: tatoeba_mt
1225
+ args: fra-ita
1226
+ metrics:
1227
+ - name: BLEU
1228
+ type: bleu
1229
+ value: 53.8
1230
+ - name: chr-F
1231
+ type: chrf
1232
+ value: 0.72742
1233
+ - task:
1234
+ name: Translation fra-por
1235
+ type: translation
1236
+ args: fra-por
1237
+ dataset:
1238
+ name: tatoeba-test-v2021-08-07
1239
+ type: tatoeba_mt
1240
+ args: fra-por
1241
+ metrics:
1242
+ - name: BLEU
1243
+ type: bleu
1244
+ value: 48.6
1245
+ - name: chr-F
1246
+ type: chrf
1247
+ value: 0.68413
1248
+ - task:
1249
+ name: Translation fra-ron
1250
+ type: translation
1251
+ args: fra-ron
1252
+ dataset:
1253
+ name: tatoeba-test-v2021-08-07
1254
+ type: tatoeba_mt
1255
+ args: fra-ron
1256
+ metrics:
1257
+ - name: BLEU
1258
+ type: bleu
1259
+ value: 44.0
1260
+ - name: chr-F
1261
+ type: chrf
1262
+ value: 0.65009
1263
+ - task:
1264
+ name: Translation fra-spa
1265
+ type: translation
1266
+ args: fra-spa
1267
+ dataset:
1268
+ name: tatoeba-test-v2021-08-07
1269
+ type: tatoeba_mt
1270
+ args: fra-spa
1271
+ metrics:
1272
+ - name: BLEU
1273
+ type: bleu
1274
+ value: 54.8
1275
+ - name: chr-F
1276
+ type: chrf
1277
+ value: 0.72080
1278
+ - task:
1279
+ name: Translation glg-por
1280
+ type: translation
1281
+ args: glg-por
1282
+ dataset:
1283
+ name: tatoeba-test-v2021-08-07
1284
+ type: tatoeba_mt
1285
+ args: glg-por
1286
+ metrics:
1287
+ - name: BLEU
1288
+ type: bleu
1289
+ value: 61.1
1290
+ - name: chr-F
1291
+ type: chrf
1292
+ value: 0.76720
1293
+ - task:
1294
+ name: Translation glg-spa
1295
+ type: translation
1296
+ args: glg-spa
1297
+ dataset:
1298
+ name: tatoeba-test-v2021-08-07
1299
+ type: tatoeba_mt
1300
+ args: glg-spa
1301
+ metrics:
1302
+ - name: BLEU
1303
+ type: bleu
1304
+ value: 71.7
1305
+ - name: chr-F
1306
+ type: chrf
1307
+ value: 0.82362
1308
+ - task:
1309
+ name: Translation ita-cat
1310
+ type: translation
1311
+ args: ita-cat
1312
+ dataset:
1313
+ name: tatoeba-test-v2021-08-07
1314
+ type: tatoeba_mt
1315
+ args: ita-cat
1316
+ metrics:
1317
+ - name: BLEU
1318
+ type: bleu
1319
+ value: 56.4
1320
+ - name: chr-F
1321
+ type: chrf
1322
+ value: 0.72529
1323
+ - task:
1324
+ name: Translation ita-fra
1325
+ type: translation
1326
+ args: ita-fra
1327
+ dataset:
1328
+ name: tatoeba-test-v2021-08-07
1329
+ type: tatoeba_mt
1330
+ args: ita-fra
1331
+ metrics:
1332
+ - name: BLEU
1333
+ type: bleu
1334
+ value: 65.2
1335
+ - name: chr-F
1336
+ type: chrf
1337
+ value: 0.77932
1338
+ - task:
1339
+ name: Translation ita-por
1340
+ type: translation
1341
+ args: ita-por
1342
+ dataset:
1343
+ name: tatoeba-test-v2021-08-07
1344
+ type: tatoeba_mt
1345
+ args: ita-por
1346
+ metrics:
1347
+ - name: BLEU
1348
+ type: bleu
1349
+ value: 54.0
1350
+ - name: chr-F
1351
+ type: chrf
1352
+ value: 0.72798
1353
+ - task:
1354
+ name: Translation ita-ron
1355
+ type: translation
1356
+ args: ita-ron
1357
+ dataset:
1358
+ name: tatoeba-test-v2021-08-07
1359
+ type: tatoeba_mt
1360
+ args: ita-ron
1361
+ metrics:
1362
+ - name: BLEU
1363
+ type: bleu
1364
+ value: 51.1
1365
+ - name: chr-F
1366
+ type: chrf
1367
+ value: 0.70814
1368
+ - task:
1369
+ name: Translation ita-spa
1370
+ type: translation
1371
+ args: ita-spa
1372
+ dataset:
1373
+ name: tatoeba-test-v2021-08-07
1374
+ type: tatoeba_mt
1375
+ args: ita-spa
1376
+ metrics:
1377
+ - name: BLEU
1378
+ type: bleu
1379
+ value: 62.9
1380
+ - name: chr-F
1381
+ type: chrf
1382
+ value: 0.77455
1383
+ - task:
1384
+ name: Translation lad-spa
1385
+ type: translation
1386
+ args: lad-spa
1387
+ dataset:
1388
+ name: tatoeba-test-v2021-08-07
1389
+ type: tatoeba_mt
1390
+ args: lad-spa
1391
+ metrics:
1392
+ - name: BLEU
1393
+ type: bleu
1394
+ value: 34.7
1395
+ - name: chr-F
1396
+ type: chrf
1397
+ value: 0.52243
1398
+ - task:
1399
+ name: Translation lad_Latn-spa
1400
+ type: translation
1401
+ args: lad_Latn-spa
1402
+ dataset:
1403
+ name: tatoeba-test-v2021-08-07
1404
+ type: tatoeba_mt
1405
+ args: lad_Latn-spa
1406
+ metrics:
1407
+ - name: BLEU
1408
+ type: bleu
1409
+ value: 42.6
1410
+ - name: chr-F
1411
+ type: chrf
1412
+ value: 0.59363
1413
+ - task:
1414
+ name: Translation oci-fra
1415
+ type: translation
1416
+ args: oci-fra
1417
+ dataset:
1418
+ name: tatoeba-test-v2021-08-07
1419
+ type: tatoeba_mt
1420
+ args: oci-fra
1421
+ metrics:
1422
+ - name: BLEU
1423
+ type: bleu
1424
+ value: 29.6
1425
+ - name: chr-F
1426
+ type: chrf
1427
+ value: 0.49660
1428
+ - task:
1429
+ name: Translation pms-ita
1430
+ type: translation
1431
+ args: pms-ita
1432
+ dataset:
1433
+ name: tatoeba-test-v2021-08-07
1434
+ type: tatoeba_mt
1435
+ args: pms-ita
1436
+ metrics:
1437
+ - name: BLEU
1438
+ type: bleu
1439
+ value: 20.0
1440
+ - name: chr-F
1441
+ type: chrf
1442
+ value: 0.40221
1443
+ - task:
1444
+ name: Translation por-cat
1445
+ type: translation
1446
+ args: por-cat
1447
+ dataset:
1448
+ name: tatoeba-test-v2021-08-07
1449
+ type: tatoeba_mt
1450
+ args: por-cat
1451
+ metrics:
1452
+ - name: BLEU
1453
+ type: bleu
1454
+ value: 52.2
1455
+ - name: chr-F
1456
+ type: chrf
1457
+ value: 0.71146
1458
+ - task:
1459
+ name: Translation por-fra
1460
+ type: translation
1461
+ args: por-fra
1462
+ dataset:
1463
+ name: tatoeba-test-v2021-08-07
1464
+ type: tatoeba_mt
1465
+ args: por-fra
1466
+ metrics:
1467
+ - name: BLEU
1468
+ type: bleu
1469
+ value: 60.9
1470
+ - name: chr-F
1471
+ type: chrf
1472
+ value: 0.75565
1473
+ - task:
1474
+ name: Translation por-glg
1475
+ type: translation
1476
+ args: por-glg
1477
+ dataset:
1478
+ name: tatoeba-test-v2021-08-07
1479
+ type: tatoeba_mt
1480
+ args: por-glg
1481
+ metrics:
1482
+ - name: BLEU
1483
+ type: bleu
1484
+ value: 59.0
1485
+ - name: chr-F
1486
+ type: chrf
1487
+ value: 0.75348
1488
+ - task:
1489
+ name: Translation por-ita
1490
+ type: translation
1491
+ args: por-ita
1492
+ dataset:
1493
+ name: tatoeba-test-v2021-08-07
1494
+ type: tatoeba_mt
1495
+ args: por-ita
1496
+ metrics:
1497
+ - name: BLEU
1498
+ type: bleu
1499
+ value: 58.8
1500
+ - name: chr-F
1501
+ type: chrf
1502
+ value: 0.76883
1503
+ - task:
1504
+ name: Translation por-ron
1505
+ type: translation
1506
+ args: por-ron
1507
+ dataset:
1508
+ name: tatoeba-test-v2021-08-07
1509
+ type: tatoeba_mt
1510
+ args: por-ron
1511
+ metrics:
1512
+ - name: BLEU
1513
+ type: bleu
1514
+ value: 46.6
1515
+ - name: chr-F
1516
+ type: chrf
1517
+ value: 0.67838
1518
+ - task:
1519
+ name: Translation por-spa
1520
+ type: translation
1521
+ args: por-spa
1522
+ dataset:
1523
+ name: tatoeba-test-v2021-08-07
1524
+ type: tatoeba_mt
1525
+ args: por-spa
1526
+ metrics:
1527
+ - name: BLEU
1528
+ type: bleu
1529
+ value: 64.8
1530
+ - name: chr-F
1531
+ type: chrf
1532
+ value: 0.79336
1533
+ - task:
1534
+ name: Translation ron-fra
1535
+ type: translation
1536
+ args: ron-fra
1537
+ dataset:
1538
+ name: tatoeba-test-v2021-08-07
1539
+ type: tatoeba_mt
1540
+ args: ron-fra
1541
+ metrics:
1542
+ - name: BLEU
1543
+ type: bleu
1544
+ value: 55.0
1545
+ - name: chr-F
1546
+ type: chrf
1547
+ value: 0.70307
1548
+ - task:
1549
+ name: Translation ron-ita
1550
+ type: translation
1551
+ args: ron-ita
1552
+ dataset:
1553
+ name: tatoeba-test-v2021-08-07
1554
+ type: tatoeba_mt
1555
+ args: ron-ita
1556
+ metrics:
1557
+ - name: BLEU
1558
+ type: bleu
1559
+ value: 53.7
1560
+ - name: chr-F
1561
+ type: chrf
1562
+ value: 0.73862
1563
+ - task:
1564
+ name: Translation ron-por
1565
+ type: translation
1566
+ args: ron-por
1567
+ dataset:
1568
+ name: tatoeba-test-v2021-08-07
1569
+ type: tatoeba_mt
1570
+ args: ron-por
1571
+ metrics:
1572
+ - name: BLEU
1573
+ type: bleu
1574
+ value: 50.7
1575
+ - name: chr-F
1576
+ type: chrf
1577
+ value: 0.70889
1578
+ - task:
1579
+ name: Translation ron-spa
1580
+ type: translation
1581
+ args: ron-spa
1582
+ dataset:
1583
+ name: tatoeba-test-v2021-08-07
1584
+ type: tatoeba_mt
1585
+ args: ron-spa
1586
+ metrics:
1587
+ - name: BLEU
1588
+ type: bleu
1589
+ value: 57.2
1590
+ - name: chr-F
1591
+ type: chrf
1592
+ value: 0.73529
1593
+ - task:
1594
+ name: Translation spa-cat
1595
+ type: translation
1596
+ args: spa-cat
1597
+ dataset:
1598
+ name: tatoeba-test-v2021-08-07
1599
+ type: tatoeba_mt
1600
+ args: spa-cat
1601
+ metrics:
1602
+ - name: BLEU
1603
+ type: bleu
1604
+ value: 67.9
1605
+ - name: chr-F
1606
+ type: chrf
1607
+ value: 0.82758
1608
+ - task:
1609
+ name: Translation spa-fra
1610
+ type: translation
1611
+ args: spa-fra
1612
+ dataset:
1613
+ name: tatoeba-test-v2021-08-07
1614
+ type: tatoeba_mt
1615
+ args: spa-fra
1616
+ metrics:
1617
+ - name: BLEU
1618
+ type: bleu
1619
+ value: 57.3
1620
+ - name: chr-F
1621
+ type: chrf
1622
+ value: 0.73113
1623
+ - task:
1624
+ name: Translation spa-glg
1625
+ type: translation
1626
+ args: spa-glg
1627
+ dataset:
1628
+ name: tatoeba-test-v2021-08-07
1629
+ type: tatoeba_mt
1630
+ args: spa-glg
1631
+ metrics:
1632
+ - name: BLEU
1633
+ type: bleu
1634
+ value: 63.0
1635
+ - name: chr-F
1636
+ type: chrf
1637
+ value: 0.77332
1638
+ - task:
1639
+ name: Translation spa-ita
1640
+ type: translation
1641
+ args: spa-ita
1642
+ dataset:
1643
+ name: tatoeba-test-v2021-08-07
1644
+ type: tatoeba_mt
1645
+ args: spa-ita
1646
+ metrics:
1647
+ - name: BLEU
1648
+ type: bleu
1649
+ value: 60.3
1650
+ - name: chr-F
1651
+ type: chrf
1652
+ value: 0.77046
1653
+ - task:
1654
+ name: Translation spa-por
1655
+ type: translation
1656
+ args: spa-por
1657
+ dataset:
1658
+ name: tatoeba-test-v2021-08-07
1659
+ type: tatoeba_mt
1660
+ args: spa-por
1661
+ metrics:
1662
+ - name: BLEU
1663
+ type: bleu
1664
+ value: 59.1
1665
+ - name: chr-F
1666
+ type: chrf
1667
+ value: 0.75854
1668
+ - task:
1669
+ name: Translation spa-ron
1670
+ type: translation
1671
+ args: spa-ron
1672
+ dataset:
1673
+ name: tatoeba-test-v2021-08-07
1674
+ type: tatoeba_mt
1675
+ args: spa-ron
1676
+ metrics:
1677
+ - name: BLEU
1678
+ type: bleu
1679
+ value: 45.5
1680
+ - name: chr-F
1681
+ type: chrf
1682
+ value: 0.66679
1683
+ - task:
1684
+ name: Translation fra-ita
1685
+ type: translation
1686
+ args: fra-ita
1687
+ dataset:
1688
+ name: newstest2009
1689
+ type: wmt-2009-news
1690
+ args: fra-ita
1691
+ metrics:
1692
+ - name: BLEU
1693
+ type: bleu
1694
+ value: 31.2
1695
+ - name: chr-F
1696
+ type: chrf
1697
+ value: 0.59764
1698
+ - task:
1699
+ name: Translation fra-spa
1700
+ type: translation
1701
+ args: fra-spa
1702
+ dataset:
1703
+ name: newstest2009
1704
+ type: wmt-2009-news
1705
+ args: fra-spa
1706
+ metrics:
1707
+ - name: BLEU
1708
+ type: bleu
1709
+ value: 32.5
1710
+ - name: chr-F
1711
+ type: chrf
1712
+ value: 0.58829
1713
+ - task:
1714
+ name: Translation ita-fra
1715
+ type: translation
1716
+ args: ita-fra
1717
+ dataset:
1718
+ name: newstest2009
1719
+ type: wmt-2009-news
1720
+ args: ita-fra
1721
+ metrics:
1722
+ - name: BLEU
1723
+ type: bleu
1724
+ value: 31.6
1725
+ - name: chr-F
1726
+ type: chrf
1727
+ value: 0.59084
1728
+ - task:
1729
+ name: Translation ita-spa
1730
+ type: translation
1731
+ args: ita-spa
1732
+ dataset:
1733
+ name: newstest2009
1734
+ type: wmt-2009-news
1735
+ args: ita-spa
1736
+ metrics:
1737
+ - name: BLEU
1738
+ type: bleu
1739
+ value: 33.5
1740
+ - name: chr-F
1741
+ type: chrf
1742
+ value: 0.59669
1743
+ - task:
1744
+ name: Translation spa-fra
1745
+ type: translation
1746
+ args: spa-fra
1747
+ dataset:
1748
+ name: newstest2009
1749
+ type: wmt-2009-news
1750
+ args: spa-fra
1751
+ metrics:
1752
+ - name: BLEU
1753
+ type: bleu
1754
+ value: 32.3
1755
+ - name: chr-F
1756
+ type: chrf
1757
+ value: 0.59096
1758
+ - task:
1759
+ name: Translation spa-ita
1760
+ type: translation
1761
+ args: spa-ita
1762
+ dataset:
1763
+ name: newstest2009
1764
+ type: wmt-2009-news
1765
+ args: spa-ita
1766
+ metrics:
1767
+ - name: BLEU
1768
+ type: bleu
1769
+ value: 33.2
1770
+ - name: chr-F
1771
+ type: chrf
1772
+ value: 0.60783
1773
+ - task:
1774
+ name: Translation fra-spa
1775
+ type: translation
1776
+ args: fra-spa
1777
+ dataset:
1778
+ name: newstest2010
1779
+ type: wmt-2010-news
1780
+ args: fra-spa
1781
+ metrics:
1782
+ - name: BLEU
1783
+ type: bleu
1784
+ value: 37.8
1785
+ - name: chr-F
1786
+ type: chrf
1787
+ value: 0.62250
1788
+ - task:
1789
+ name: Translation spa-fra
1790
+ type: translation
1791
+ args: spa-fra
1792
+ dataset:
1793
+ name: newstest2010
1794
+ type: wmt-2010-news
1795
+ args: spa-fra
1796
+ metrics:
1797
+ - name: BLEU
1798
+ type: bleu
1799
+ value: 36.2
1800
+ - name: chr-F
1801
+ type: chrf
1802
+ value: 0.61953
1803
+ - task:
1804
+ name: Translation fra-spa
1805
+ type: translation
1806
+ args: fra-spa
1807
+ dataset:
1808
+ name: newstest2011
1809
+ type: wmt-2011-news
1810
+ args: fra-spa
1811
+ metrics:
1812
+ - name: BLEU
1813
+ type: bleu
1814
+ value: 39.8
1815
+ - name: chr-F
1816
+ type: chrf
1817
+ value: 0.62953
1818
+ - task:
1819
+ name: Translation spa-fra
1820
+ type: translation
1821
+ args: spa-fra
1822
+ dataset:
1823
+ name: newstest2011
1824
+ type: wmt-2011-news
1825
+ args: spa-fra
1826
+ metrics:
1827
+ - name: BLEU
1828
+ type: bleu
1829
+ value: 34.9
1830
+ - name: chr-F
1831
+ type: chrf
1832
+ value: 0.61130
1833
+ - task:
1834
+ name: Translation fra-spa
1835
+ type: translation
1836
+ args: fra-spa
1837
+ dataset:
1838
+ name: newstest2012
1839
+ type: wmt-2012-news
1840
+ args: fra-spa
1841
+ metrics:
1842
+ - name: BLEU
1843
+ type: bleu
1844
+ value: 39.0
1845
+ - name: chr-F
1846
+ type: chrf
1847
+ value: 0.62397
1848
+ - task:
1849
+ name: Translation spa-fra
1850
+ type: translation
1851
+ args: spa-fra
1852
+ dataset:
1853
+ name: newstest2012
1854
+ type: wmt-2012-news
1855
+ args: spa-fra
1856
+ metrics:
1857
+ - name: BLEU
1858
+ type: bleu
1859
+ value: 34.3
1860
+ - name: chr-F
1861
+ type: chrf
1862
+ value: 0.60927
1863
+ - task:
1864
+ name: Translation fra-spa
1865
+ type: translation
1866
+ args: fra-spa
1867
+ dataset:
1868
+ name: newstest2013
1869
+ type: wmt-2013-news
1870
+ args: fra-spa
1871
+ metrics:
1872
+ - name: BLEU
1873
+ type: bleu
1874
+ value: 34.9
1875
+ - name: chr-F
1876
+ type: chrf
1877
+ value: 0.59312
1878
+ - task:
1879
+ name: Translation spa-fra
1880
+ type: translation
1881
+ args: spa-fra
1882
+ dataset:
1883
+ name: newstest2013
1884
+ type: wmt-2013-news
1885
+ args: spa-fra
1886
+ metrics:
1887
+ - name: BLEU
1888
+ type: bleu
1889
+ value: 33.6
1890
+ - name: chr-F
1891
+ type: chrf
1892
+ value: 0.59468
1893
+ - task:
1894
+ name: Translation cat-ita
1895
+ type: translation
1896
+ args: cat-ita
1897
+ dataset:
1898
+ name: wmt21-ml-wp
1899
+ type: wmt21-ml-wp
1900
+ args: cat-ita
1901
+ metrics:
1902
+ - name: BLEU
1903
+ type: bleu
1904
+ value: 47.8
1905
+ - name: chr-F
1906
+ type: chrf
1907
+ value: 0.69968
1908
+ - task:
1909
+ name: Translation cat-oci
1910
+ type: translation
1911
+ args: cat-oci
1912
+ dataset:
1913
+ name: wmt21-ml-wp
1914
+ type: wmt21-ml-wp
1915
+ args: cat-oci
1916
+ metrics:
1917
+ - name: BLEU
1918
+ type: bleu
1919
+ value: 51.6
1920
+ - name: chr-F
1921
+ type: chrf
1922
+ value: 0.73808
1923
+ - task:
1924
+ name: Translation cat-ron
1925
+ type: translation
1926
+ args: cat-ron
1927
+ dataset:
1928
+ name: wmt21-ml-wp
1929
+ type: wmt21-ml-wp
1930
+ args: cat-ron
1931
+ metrics:
1932
+ - name: BLEU
1933
+ type: bleu
1934
+ value: 29.0
1935
+ - name: chr-F
1936
+ type: chrf
1937
+ value: 0.51178
1938
+ - task:
1939
+ name: Translation ita-cat
1940
+ type: translation
1941
+ args: ita-cat
1942
+ dataset:
1943
+ name: wmt21-ml-wp
1944
+ type: wmt21-ml-wp
1945
+ args: ita-cat
1946
+ metrics:
1947
+ - name: BLEU
1948
+ type: bleu
1949
+ value: 48.9
1950
+ - name: chr-F
1951
+ type: chrf
1952
+ value: 0.70538
1953
+ - task:
1954
+ name: Translation ita-oci
1955
+ type: translation
1956
+ args: ita-oci
1957
+ dataset:
1958
+ name: wmt21-ml-wp
1959
+ type: wmt21-ml-wp
1960
+ args: ita-oci
1961
+ metrics:
1962
+ - name: BLEU
1963
+ type: bleu
1964
+ value: 32.0
1965
+ - name: chr-F
1966
+ type: chrf
1967
+ value: 0.59025
1968
+ - task:
1969
+ name: Translation ita-ron
1970
+ type: translation
1971
+ args: ita-ron
1972
+ dataset:
1973
+ name: wmt21-ml-wp
1974
+ type: wmt21-ml-wp
1975
+ args: ita-ron
1976
+ metrics:
1977
+ - name: BLEU
1978
+ type: bleu
1979
+ value: 28.9
1980
+ - name: chr-F
1981
+ type: chrf
1982
+ value: 0.51261
1983
+ - task:
1984
+ name: Translation oci-cat
1985
+ type: translation
1986
+ args: oci-cat
1987
+ dataset:
1988
+ name: wmt21-ml-wp
1989
+ type: wmt21-ml-wp
1990
+ args: oci-cat
1991
+ metrics:
1992
+ - name: BLEU
1993
+ type: bleu
1994
+ value: 66.1
1995
+ - name: chr-F
1996
+ type: chrf
1997
+ value: 0.80908
1998
+ - task:
1999
+ name: Translation oci-ita
2000
+ type: translation
2001
+ args: oci-ita
2002
+ dataset:
2003
+ name: wmt21-ml-wp
2004
+ type: wmt21-ml-wp
2005
+ args: oci-ita
2006
+ metrics:
2007
+ - name: BLEU
2008
+ type: bleu
2009
+ value: 39.6
2010
+ - name: chr-F
2011
+ type: chrf
2012
+ value: 0.63584
2013
+ - task:
2014
+ name: Translation oci-ron
2015
+ type: translation
2016
+ args: oci-ron
2017
+ dataset:
2018
+ name: wmt21-ml-wp
2019
+ type: wmt21-ml-wp
2020
+ args: oci-ron
2021
+ metrics:
2022
+ - name: BLEU
2023
+ type: bleu
2024
+ value: 24.6
2025
+ - name: chr-F
2026
+ type: chrf
2027
+ value: 0.47384
2028
+ - task:
2029
+ name: Translation ron-cat
2030
+ type: translation
2031
+ args: ron-cat
2032
+ dataset:
2033
+ name: wmt21-ml-wp
2034
+ type: wmt21-ml-wp
2035
+ args: ron-cat
2036
+ metrics:
2037
+ - name: BLEU
2038
+ type: bleu
2039
+ value: 31.1
2040
+ - name: chr-F
2041
+ type: chrf
2042
+ value: 0.52994
2043
+ - task:
2044
+ name: Translation ron-ita
2045
+ type: translation
2046
+ args: ron-ita
2047
+ dataset:
2048
+ name: wmt21-ml-wp
2049
+ type: wmt21-ml-wp
2050
+ args: ron-ita
2051
+ metrics:
2052
+ - name: BLEU
2053
+ type: bleu
2054
+ value: 29.6
2055
+ - name: chr-F
2056
+ type: chrf
2057
+ value: 0.52714
2058
+ - task:
2059
+ name: Translation ron-oci
2060
+ type: translation
2061
+ args: ron-oci
2062
+ dataset:
2063
+ name: wmt21-ml-wp
2064
+ type: wmt21-ml-wp
2065
+ args: ron-oci
2066
+ metrics:
2067
+ - name: BLEU
2068
+ type: bleu
2069
+ value: 21.3
2070
+ - name: chr-F
2071
+ type: chrf
2072
+ value: 0.45932
2073
+ ---
2074
+ # opus-mt-tc-big-itc-itc
2075
+
2076
+ ## Table of Contents
2077
+ - [Model Details](#model-details)
2078
+ - [Uses](#uses)
2079
+ - [Risks, Limitations and Biases](#risks-limitations-and-biases)
2080
+ - [How to Get Started With the Model](#how-to-get-started-with-the-model)
2081
+ - [Training](#training)
2082
+ - [Evaluation](#evaluation)
2083
+ - [Citation Information](#citation-information)
2084
+ - [Acknowledgements](#acknowledgements)
2085
+
2086
+ ## Model Details
2087
+
2088
+ Neural machine translation model for translating from Italic languages (itc) to Italic languages (itc).
2089
+
2090
+ This model is part of the [OPUS-MT project](https://github.com/Helsinki-NLP/Opus-MT), an effort to make neural machine translation models widely available and accessible for many languages in the world. All models are originally trained using the amazing framework of [Marian NMT](https://marian-nmt.github.io/), an efficient NMT implementation written in pure C++. The models have been converted to pyTorch using the transformers library by huggingface. Training data is taken from [OPUS](https://opus.nlpl.eu/) and training pipelines use the procedures of [OPUS-MT-train](https://github.com/Helsinki-NLP/Opus-MT-train).
2091
+ **Model Description:**
2092
+ - **Developed by:** Language Technology Research Group at the University of Helsinki
2093
+ - **Model Type:** Translation (transformer-big)
2094
+ - **Release**: 2022-08-10
2095
+ - **License:** CC-BY-4.0
2096
+ - **Language(s):**
2097
+ - Source Language(s): ast cat cbk fra fro glg hat ita lad lad_Latn lat lat_Latn lij lld oci pms por ron spa
2098
+ - Target Language(s): ast cat fra gcf glg hat ita lad lad_Latn lat lat_Latn oci por ron spa
2099
+ - Valid Target Language Labels: >>ast<< >>cat<< >>fra<< >>gcf<< >>glg<< >>hat<< >>ita<< >>lad<< >>lad_Latn<< >>lat<< >>lat_Latn<< >>oci<< >>por<< >>ron<< >>spa<<
2100
+ - **Original Model**: [opusTCv20210807_transformer-big_2022-08-10.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/itc-itc/opusTCv20210807_transformer-big_2022-08-10.zip)
2101
+ - **Resources for more information:**
2102
+ - [OPUS-MT-train GitHub Repo](https://github.com/Helsinki-NLP/OPUS-MT-train)
2103
+ - More information about released models for this language pair: [OPUS-MT itc-itc README](https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/itc-itc/README.md)
2104
+ - [More information about MarianNMT models in the transformers library](https://huggingface.co/docs/transformers/model_doc/marian)
2105
+ - [Tatoeba Translation Challenge](https://github.com/Helsinki-NLP/Tatoeba-Challenge/
2106
+
2107
+ This is a multilingual translation model with multiple target languages. A sentence initial language token is required in the form of `>>id<<` (id = valid target language ID), e.g. `>>ast<<`
2108
+
2109
+ ## Uses
2110
+
2111
+ This model can be used for translation and text-to-text generation.
2112
+
2113
+ ## Risks, Limitations and Biases
2114
+
2115
+ **CONTENT WARNING: Readers should be aware that the model is trained on various public data sets that may contain content that is disturbing, offensive, and can propagate historical and current stereotypes.**
2116
+
2117
+ Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
2118
+
2119
+ ## How to Get Started With the Model
2120
+
2121
+ A short example code:
2122
+
2123
+ ```python
2124
+ from transformers import MarianMTModel, MarianTokenizer
2125
+
2126
+ src_text = [
2127
+ ">>fra<< Charras anglés?",
2128
+ ">>fra<< Vull veure't."
2129
+ ]
2130
+
2131
+ model_name = "pytorch-models/opus-mt-tc-big-itc-itc"
2132
+ tokenizer = MarianTokenizer.from_pretrained(model_name)
2133
+ model = MarianMTModel.from_pretrained(model_name)
2134
+ translated = model.generate(**tokenizer(src_text, return_tensors="pt", padding=True))
2135
+
2136
+ for t in translated:
2137
+ print( tokenizer.decode(t, skip_special_tokens=True) )
2138
+
2139
+ # expected output:
2140
+ # Conversations anglaises ?
2141
+ # Je veux te voir.
2142
+ ```
2143
+
2144
+ You can also use OPUS-MT models with the transformers pipelines, for example:
2145
+
2146
+ ```python
2147
+ from transformers import pipeline
2148
+ pipe = pipeline("translation", model="Helsinki-NLP/opus-mt-tc-big-itc-itc")
2149
+ print(pipe(">>fra<< Charras anglés?"))
2150
+
2151
+ # expected output: Conversations anglaises ?
2152
+ ```
2153
+
2154
+ ## Training
2155
+
2156
+ - **Data**: opusTCv20210807 ([source](https://github.com/Helsinki-NLP/Tatoeba-Challenge))
2157
+ - **Pre-processing**: SentencePiece (spm32k,spm32k)
2158
+ - **Model Type:** transformer-big
2159
+ - **Original MarianNMT Model**: [opusTCv20210807_transformer-big_2022-08-10.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/itc-itc/opusTCv20210807_transformer-big_2022-08-10.zip)
2160
+ - **Training Scripts**: [GitHub Repo](https://github.com/Helsinki-NLP/OPUS-MT-train)
2161
+
2162
+ ## Evaluation
2163
+
2164
+ * test set translations: [opusTCv20210807_transformer-big_2022-08-10.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/itc-itc/opusTCv20210807_transformer-big_2022-08-10.test.txt)
2165
+ * test set scores: [opusTCv20210807_transformer-big_2022-08-10.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/itc-itc/opusTCv20210807_transformer-big_2022-08-10.eval.txt)
2166
+ * benchmark results: [benchmark_results.txt](benchmark_results.txt)
2167
+ * benchmark output: [benchmark_translations.zip](benchmark_translations.zip)
2168
+
2169
+ | langpair | testset | chr-F | BLEU | #sent | #words |
2170
+ |----------|---------|-------|-------|-------|--------|
2171
+ | cat-fra | tatoeba-test-v2021-08-07 | 0.71201 | 54.6 | 700 | 5664 |
2172
+ | cat-ita | tatoeba-test-v2021-08-07 | 0.74198 | 58.4 | 298 | 2028 |
2173
+ | cat-por | tatoeba-test-v2021-08-07 | 0.74930 | 57.4 | 747 | 6119 |
2174
+ | cat-spa | tatoeba-test-v2021-08-07 | 0.87844 | 78.1 | 1534 | 12094 |
2175
+ | fra-cat | tatoeba-test-v2021-08-07 | 0.66525 | 46.2 | 700 | 5342 |
2176
+ | fra-ita | tatoeba-test-v2021-08-07 | 0.72742 | 53.8 | 10091 | 62060 |
2177
+ | fra-por | tatoeba-test-v2021-08-07 | 0.68413 | 48.6 | 10518 | 77650 |
2178
+ | fra-ron | tatoeba-test-v2021-08-07 | 0.65009 | 44.0 | 1925 | 12252 |
2179
+ | fra-spa | tatoeba-test-v2021-08-07 | 0.72080 | 54.8 | 10294 | 78406 |
2180
+ | glg-por | tatoeba-test-v2021-08-07 | 0.76720 | 61.1 | 433 | 3105 |
2181
+ | glg-spa | tatoeba-test-v2021-08-07 | 0.82362 | 71.7 | 2121 | 17443 |
2182
+ | ita-cat | tatoeba-test-v2021-08-07 | 0.72529 | 56.4 | 298 | 2109 |
2183
+ | ita-fra | tatoeba-test-v2021-08-07 | 0.77932 | 65.2 | 10091 | 66377 |
2184
+ | ita-por | tatoeba-test-v2021-08-07 | 0.72798 | 54.0 | 3066 | 25668 |
2185
+ | ita-ron | tatoeba-test-v2021-08-07 | 0.70814 | 51.1 | 1005 | 6209 |
2186
+ | ita-spa | tatoeba-test-v2021-08-07 | 0.77455 | 62.9 | 5000 | 34937 |
2187
+ | lad_Latn-spa | tatoeba-test-v2021-08-07 | 0.59363 | 42.6 | 239 | 1239 |
2188
+ | lad-spa | tatoeba-test-v2021-08-07 | 0.52243 | 34.7 | 276 | 1448 |
2189
+ | oci-fra | tatoeba-test-v2021-08-07 | 0.49660 | 29.6 | 806 | 6302 |
2190
+ | pms-ita | tatoeba-test-v2021-08-07 | 0.40221 | 20.0 | 232 | 1721 |
2191
+ | por-cat | tatoeba-test-v2021-08-07 | 0.71146 | 52.2 | 747 | 6149 |
2192
+ | por-fra | tatoeba-test-v2021-08-07 | 0.75565 | 60.9 | 10518 | 80459 |
2193
+ | por-glg | tatoeba-test-v2021-08-07 | 0.75348 | 59.0 | 433 | 3016 |
2194
+ | por-ita | tatoeba-test-v2021-08-07 | 0.76883 | 58.8 | 3066 | 24897 |
2195
+ | por-ron | tatoeba-test-v2021-08-07 | 0.67838 | 46.6 | 681 | 4521 |
2196
+ | por-spa | tatoeba-test-v2021-08-07 | 0.79336 | 64.8 | 10947 | 87335 |
2197
+ | ron-fra | tatoeba-test-v2021-08-07 | 0.70307 | 55.0 | 1925 | 13347 |
2198
+ | ron-ita | tatoeba-test-v2021-08-07 | 0.73862 | 53.7 | 1005 | 6352 |
2199
+ | ron-por | tatoeba-test-v2021-08-07 | 0.70889 | 50.7 | 681 | 4593 |
2200
+ | ron-spa | tatoeba-test-v2021-08-07 | 0.73529 | 57.2 | 1959 | 12679 |
2201
+ | spa-cat | tatoeba-test-v2021-08-07 | 0.82758 | 67.9 | 1534 | 12343 |
2202
+ | spa-fra | tatoeba-test-v2021-08-07 | 0.73113 | 57.3 | 10294 | 83501 |
2203
+ | spa-glg | tatoeba-test-v2021-08-07 | 0.77332 | 63.0 | 2121 | 16581 |
2204
+ | spa-ita | tatoeba-test-v2021-08-07 | 0.77046 | 60.3 | 5000 | 34515 |
2205
+ | spa-lad_Latn | tatoeba-test-v2021-08-07 | 0.40084 | 14.7 | 239 | 1254 |
2206
+ | spa-por | tatoeba-test-v2021-08-07 | 0.75854 | 59.1 | 10947 | 87610 |
2207
+ | spa-ron | tatoeba-test-v2021-08-07 | 0.66679 | 45.5 | 1959 | 12503 |
2208
+ | ast-cat | flores101-devtest | 0.57870 | 31.8 | 1012 | 27304 |
2209
+ | ast-fra | flores101-devtest | 0.56761 | 31.1 | 1012 | 28343 |
2210
+ | ast-glg | flores101-devtest | 0.55161 | 27.9 | 1012 | 26582 |
2211
+ | ast-ita | flores101-devtest | 0.51764 | 22.1 | 1012 | 27306 |
2212
+ | ast-oci | flores101-devtest | 0.49545 | 20.6 | 1012 | 27305 |
2213
+ | ast-por | flores101-devtest | 0.57347 | 31.5 | 1012 | 26519 |
2214
+ | ast-ron | flores101-devtest | 0.52317 | 24.8 | 1012 | 26799 |
2215
+ | ast-spa | flores101-devtest | 0.49741 | 21.2 | 1012 | 29199 |
2216
+ | cat-ast | flores101-devtest | 0.56754 | 24.7 | 1012 | 24572 |
2217
+ | cat-fra | flores101-devtest | 0.63368 | 38.4 | 1012 | 28343 |
2218
+ | cat-glg | flores101-devtest | 0.59596 | 32.2 | 1012 | 26582 |
2219
+ | cat-ita | flores101-devtest | 0.55886 | 26.3 | 1012 | 27306 |
2220
+ | cat-oci | flores101-devtest | 0.54285 | 24.6 | 1012 | 27305 |
2221
+ | cat-por | flores101-devtest | 0.62913 | 37.7 | 1012 | 26519 |
2222
+ | cat-ron | flores101-devtest | 0.56885 | 29.5 | 1012 | 26799 |
2223
+ | cat-spa | flores101-devtest | 0.53372 | 24.6 | 1012 | 29199 |
2224
+ | fra-ast | flores101-devtest | 0.52696 | 20.7 | 1012 | 24572 |
2225
+ | fra-cat | flores101-devtest | 0.60492 | 34.6 | 1012 | 27304 |
2226
+ | fra-glg | flores101-devtest | 0.57485 | 30.3 | 1012 | 26582 |
2227
+ | fra-ita | flores101-devtest | 0.56493 | 27.3 | 1012 | 27306 |
2228
+ | fra-oci | flores101-devtest | 0.57449 | 28.2 | 1012 | 27305 |
2229
+ | fra-por | flores101-devtest | 0.62211 | 36.9 | 1012 | 26519 |
2230
+ | fra-ron | flores101-devtest | 0.56998 | 29.4 | 1012 | 26799 |
2231
+ | fra-spa | flores101-devtest | 0.52880 | 24.2 | 1012 | 29199 |
2232
+ | glg-ast | flores101-devtest | 0.55090 | 22.4 | 1012 | 24572 |
2233
+ | glg-cat | flores101-devtest | 0.60550 | 32.6 | 1012 | 27304 |
2234
+ | glg-fra | flores101-devtest | 0.62026 | 36.0 | 1012 | 28343 |
2235
+ | glg-ita | flores101-devtest | 0.55834 | 25.9 | 1012 | 27306 |
2236
+ | glg-oci | flores101-devtest | 0.52520 | 21.9 | 1012 | 27305 |
2237
+ | glg-por | flores101-devtest | 0.60027 | 32.7 | 1012 | 26519 |
2238
+ | glg-ron | flores101-devtest | 0.55621 | 27.8 | 1012 | 26799 |
2239
+ | glg-spa | flores101-devtest | 0.53219 | 24.4 | 1012 | 29199 |
2240
+ | ita-ast | flores101-devtest | 0.50741 | 17.1 | 1012 | 24572 |
2241
+ | ita-cat | flores101-devtest | 0.57061 | 27.9 | 1012 | 27304 |
2242
+ | ita-fra | flores101-devtest | 0.60199 | 32.0 | 1012 | 28343 |
2243
+ | ita-glg | flores101-devtest | 0.55312 | 25.9 | 1012 | 26582 |
2244
+ | ita-oci | flores101-devtest | 0.48447 | 18.1 | 1012 | 27305 |
2245
+ | ita-por | flores101-devtest | 0.58162 | 29.0 | 1012 | 26519 |
2246
+ | ita-ron | flores101-devtest | 0.53703 | 24.2 | 1012 | 26799 |
2247
+ | ita-spa | flores101-devtest | 0.52238 | 23.1 | 1012 | 29199 |
2248
+ | oci-ast | flores101-devtest | 0.53010 | 20.2 | 1012 | 24572 |
2249
+ | oci-cat | flores101-devtest | 0.59946 | 32.2 | 1012 | 27304 |
2250
+ | oci-fra | flores101-devtest | 0.64290 | 39.0 | 1012 | 28343 |
2251
+ | oci-glg | flores101-devtest | 0.56737 | 28.0 | 1012 | 26582 |
2252
+ | oci-ita | flores101-devtest | 0.54220 | 24.2 | 1012 | 27306 |
2253
+ | oci-por | flores101-devtest | 0.62127 | 35.7 | 1012 | 26519 |
2254
+ | oci-ron | flores101-devtest | 0.55906 | 28.0 | 1012 | 26799 |
2255
+ | oci-spa | flores101-devtest | 0.52110 | 22.8 | 1012 | 29199 |
2256
+ | por-ast | flores101-devtest | 0.54539 | 22.5 | 1012 | 24572 |
2257
+ | por-cat | flores101-devtest | 0.61809 | 36.4 | 1012 | 27304 |
2258
+ | por-fra | flores101-devtest | 0.64343 | 39.7 | 1012 | 28343 |
2259
+ | por-glg | flores101-devtest | 0.57965 | 30.4 | 1012 | 26582 |
2260
+ | por-ita | flores101-devtest | 0.55841 | 26.3 | 1012 | 27306 |
2261
+ | por-oci | flores101-devtest | 0.54829 | 25.3 | 1012 | 27305 |
2262
+ | por-ron | flores101-devtest | 0.57283 | 29.8 | 1012 | 26799 |
2263
+ | por-spa | flores101-devtest | 0.53513 | 25.2 | 1012 | 29199 |
2264
+ | ron-ast | flores101-devtest | 0.52265 | 20.1 | 1012 | 24572 |
2265
+ | ron-cat | flores101-devtest | 0.59689 | 32.6 | 1012 | 27304 |
2266
+ | ron-fra | flores101-devtest | 0.63060 | 37.4 | 1012 | 28343 |
2267
+ | ron-glg | flores101-devtest | 0.56677 | 29.3 | 1012 | 26582 |
2268
+ | ron-ita | flores101-devtest | 0.55485 | 25.6 | 1012 | 27306 |
2269
+ | ron-oci | flores101-devtest | 0.52433 | 21.8 | 1012 | 27305 |
2270
+ | ron-por | flores101-devtest | 0.61831 | 36.1 | 1012 | 26519 |
2271
+ | ron-spa | flores101-devtest | 0.52712 | 24.1 | 1012 | 29199 |
2272
+ | spa-ast | flores101-devtest | 0.49008 | 15.7 | 1012 | 24572 |
2273
+ | spa-cat | flores101-devtest | 0.53905 | 23.2 | 1012 | 27304 |
2274
+ | spa-fra | flores101-devtest | 0.57078 | 27.4 | 1012 | 28343 |
2275
+ | spa-glg | flores101-devtest | 0.52563 | 22.0 | 1012 | 26582 |
2276
+ | spa-ita | flores101-devtest | 0.52783 | 22.3 | 1012 | 27306 |
2277
+ | spa-oci | flores101-devtest | 0.48064 | 16.3 | 1012 | 27305 |
2278
+ | spa-por | flores101-devtest | 0.55736 | 25.8 | 1012 | 26519 |
2279
+ | spa-ron | flores101-devtest | 0.51623 | 21.4 | 1012 | 26799 |
2280
+ | fra-ita | newssyscomb2009 | 0.60995 | 32.1 | 502 | 11551 |
2281
+ | fra-spa | newssyscomb2009 | 0.60224 | 34.2 | 502 | 12503 |
2282
+ | ita-fra | newssyscomb2009 | 0.61237 | 33.7 | 502 | 12331 |
2283
+ | ita-spa | newssyscomb2009 | 0.60706 | 35.4 | 502 | 12503 |
2284
+ | spa-fra | newssyscomb2009 | 0.61290 | 34.6 | 502 | 12331 |
2285
+ | spa-ita | newssyscomb2009 | 0.61632 | 33.3 | 502 | 11551 |
2286
+ | fra-spa | news-test2008 | 0.58939 | 33.9 | 2051 | 52586 |
2287
+ | spa-fra | news-test2008 | 0.58695 | 32.4 | 2051 | 52685 |
2288
+ | fra-ita | newstest2009 | 0.59764 | 31.2 | 2525 | 63466 |
2289
+ | fra-spa | newstest2009 | 0.58829 | 32.5 | 2525 | 68111 |
2290
+ | ita-fra | newstest2009 | 0.59084 | 31.6 | 2525 | 69263 |
2291
+ | ita-spa | newstest2009 | 0.59669 | 33.5 | 2525 | 68111 |
2292
+ | spa-fra | newstest2009 | 0.59096 | 32.3 | 2525 | 69263 |
2293
+ | spa-ita | newstest2009 | 0.60783 | 33.2 | 2525 | 63466 |
2294
+ | fra-spa | newstest2010 | 0.62250 | 37.8 | 2489 | 65480 |
2295
+ | spa-fra | newstest2010 | 0.61953 | 36.2 | 2489 | 66022 |
2296
+ | fra-spa | newstest2011 | 0.62953 | 39.8 | 3003 | 79476 |
2297
+ | spa-fra | newstest2011 | 0.61130 | 34.9 | 3003 | 80626 |
2298
+ | fra-spa | newstest2012 | 0.62397 | 39.0 | 3003 | 79006 |
2299
+ | spa-fra | newstest2012 | 0.60927 | 34.3 | 3003 | 78011 |
2300
+ | fra-spa | newstest2013 | 0.59312 | 34.9 | 3000 | 70528 |
2301
+ | spa-fra | newstest2013 | 0.59468 | 33.6 | 3000 | 70037 |
2302
+ | cat-ita | wmt21-ml-wp | 0.69968 | 47.8 | 1743 | 42735 |
2303
+ | cat-oci | wmt21-ml-wp | 0.73808 | 51.6 | 1743 | 43736 |
2304
+ | cat-ron | wmt21-ml-wp | 0.51178 | 29.0 | 1743 | 42895 |
2305
+ | ita-cat | wmt21-ml-wp | 0.70538 | 48.9 | 1743 | 43833 |
2306
+ | ita-oci | wmt21-ml-wp | 0.59025 | 32.0 | 1743 | 43736 |
2307
+ | ita-ron | wmt21-ml-wp | 0.51261 | 28.9 | 1743 | 42895 |
2308
+ | oci-cat | wmt21-ml-wp | 0.80908 | 66.1 | 1743 | 43833 |
2309
+ | oci-ita | wmt21-ml-wp | 0.63584 | 39.6 | 1743 | 42735 |
2310
+ | oci-ron | wmt21-ml-wp | 0.47384 | 24.6 | 1743 | 42895 |
2311
+ | ron-cat | wmt21-ml-wp | 0.52994 | 31.1 | 1743 | 43833 |
2312
+ | ron-ita | wmt21-ml-wp | 0.52714 | 29.6 | 1743 | 42735 |
2313
+ | ron-oci | wmt21-ml-wp | 0.45932 | 21.3 | 1743 | 43736 |
2314
+
2315
+ ## Citation Information
2316
+
2317
+ * Publications: [OPUS-MT – Building open translation services for the World](https://aclanthology.org/2020.eamt-1.61/) and [The Tatoeba Translation Challenge – Realistic Data Sets for Low Resource and Multilingual MT](https://aclanthology.org/2020.wmt-1.139/) (Please, cite if you use this model.)
2318
+
2319
+ ```
2320
+ @inproceedings{tiedemann-thottingal-2020-opus,
2321
+ title = "{OPUS}-{MT} {--} Building open translation services for the World",
2322
+ author = {Tiedemann, J{\"o}rg and Thottingal, Santhosh},
2323
+ booktitle = "Proceedings of the 22nd Annual Conference of the European Association for Machine Translation",
2324
+ month = nov,
2325
+ year = "2020",
2326
+ address = "Lisboa, Portugal",
2327
+ publisher = "European Association for Machine Translation",
2328
+ url = "https://aclanthology.org/2020.eamt-1.61",
2329
+ pages = "479--480",
2330
+ }
2331
+
2332
+ @inproceedings{tiedemann-2020-tatoeba,
2333
+ title = "The Tatoeba Translation Challenge {--} Realistic Data Sets for Low Resource and Multilingual {MT}",
2334
+ author = {Tiedemann, J{\"o}rg},
2335
+ booktitle = "Proceedings of the Fifth Conference on Machine Translation",
2336
+ month = nov,
2337
+ year = "2020",
2338
+ address = "Online",
2339
+ publisher = "Association for Computational Linguistics",
2340
+ url = "https://aclanthology.org/2020.wmt-1.139",
2341
+ pages = "1174--1182",
2342
+ }
2343
+ ```
2344
+
2345
+ ## Acknowledgements
2346
+
2347
+ The work is supported by the [European Language Grid](https://www.european-language-grid.eu/) as [pilot project 2866](https://live.european-language-grid.eu/catalogue/#/resource/projects/2866), by the [FoTran project](https://www.helsinki.fi/en/researchgroups/natural-language-understanding-with-cross-lingual-grounding), funded by the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No 771113), and the [MeMAD project](https://memad.eu/), funded by the European Union’s Horizon 2020 Research and Innovation Programme under grant agreement No 780069. We are also grateful for the generous computational resources and IT infrastructure provided by [CSC -- IT Center for Science](https://www.csc.fi/), Finland.
2348
+
2349
+ ## Model conversion info
2350
+
2351
+ * transformers version: 4.16.2
2352
+ * OPUS-MT git hash: 8b9f0b0
2353
+ * port time: Fri Aug 12 13:02:38 EEST 2022
2354
+ * port machine: LM0-400-22516.local
benchmark_results.txt ADDED
@@ -0,0 +1,294 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ast-cat flores101-dev 0.56993 31.1 997 25962
2
+ ast-fra flores101-dev 0.56952 32.2 997 26706
3
+ ast-glg flores101-dev 0.54400 28.3 997 25265
4
+ ast-ita flores101-dev 0.51441 22.6 997 25840
5
+ ast-oci flores101-dev 0.49415 21.1 997 25819
6
+ ast-por flores101-dev 0.56712 30.9 997 25287
7
+ ast-ron flores101-dev 0.52266 25.1 997 25616
8
+ ast-spa flores101-dev 0.48952 21.3 997 27793
9
+ cat-ast flores101-dev 0.55160 23.9 997 23232
10
+ cat-fra flores101-dev 0.63184 38.3 997 26706
11
+ cat-glg flores101-dev 0.59388 32.3 997 25265
12
+ cat-ita flores101-dev 0.55787 26.2 997 25840
13
+ cat-oci flores101-dev 0.54514 25.6 997 25819
14
+ cat-por flores101-dev 0.62313 36.9 997 25287
15
+ cat-ron flores101-dev 0.56799 29.5 997 25616
16
+ cat-spa flores101-dev 0.52900 24.3 997 27793
17
+ fra-ast flores101-dev 0.51966 20.6 997 23232
18
+ fra-cat flores101-dev 0.59877 33.7 997 25962
19
+ fra-glg flores101-dev 0.57259 30.1 997 25265
20
+ fra-ita flores101-dev 0.55876 26.1 997 25840
21
+ fra-oci flores101-dev 0.57502 28.5 997 25819
22
+ fra-por flores101-dev 0.61771 36.5 997 25287
23
+ fra-ron flores101-dev 0.56672 30.0 997 25616
24
+ fra-spa flores101-dev 0.51968 23.8 997 27793
25
+ glg-ast flores101-dev 0.54482 22.7 997 23232
26
+ glg-cat flores101-dev 0.60313 32.8 997 25962
27
+ glg-fra flores101-dev 0.61767 35.9 997 26706
28
+ glg-ita flores101-dev 0.55075 25.0 997 25840
29
+ glg-oci flores101-dev 0.52977 22.7 997 25819
30
+ glg-por flores101-dev 0.59659 32.3 997 25287
31
+ glg-ron flores101-dev 0.55550 28.4 997 25616
32
+ glg-spa flores101-dev 0.53044 24.5 997 27793
33
+ ita-ast flores101-dev 0.49418 16.5 997 23232
34
+ ita-cat flores101-dev 0.56351 27.4 997 25962
35
+ ita-fra flores101-dev 0.59322 31.0 997 26706
36
+ ita-glg flores101-dev 0.54433 25.1 997 25265
37
+ ita-oci flores101-dev 0.47877 18.1 997 25819
38
+ ita-por flores101-dev 0.57238 28.3 997 25287
39
+ ita-ron flores101-dev 0.52778 23.8 997 25616
40
+ ita-spa flores101-dev 0.51145 22.5 997 27793
41
+ oci-ast flores101-dev 0.52269 20.2 997 23232
42
+ oci-cat flores101-dev 0.59532 32.1 997 25962
43
+ oci-fra flores101-dev 0.64711 39.7 997 26706
44
+ oci-glg flores101-dev 0.56790 28.9 997 25265
45
+ oci-ita flores101-dev 0.53848 24.3 997 25840
46
+ oci-por flores101-dev 0.61621 35.6 997 25287
47
+ oci-ron flores101-dev 0.55893 28.8 997 25616
48
+ oci-spa flores101-dev 0.51491 22.6 997 27793
49
+ por-ast flores101-dev 0.53306 21.4 997 23232
50
+ por-cat flores101-dev 0.60956 35.1 997 25962
51
+ por-fra flores101-dev 0.63829 39.5 997 26706
52
+ por-glg flores101-dev 0.57314 29.8 997 25265
53
+ por-ita flores101-dev 0.55032 25.5 997 25840
54
+ por-oci flores101-dev 0.54152 25.0 997 25819
55
+ por-ron flores101-dev 0.57130 30.3 997 25616
56
+ por-spa flores101-dev 0.52929 24.9 997 27793
57
+ ron-ast flores101-dev 0.51473 19.8 997 23232
58
+ ron-cat flores101-dev 0.59300 32.4 997 25962
59
+ ron-fra flores101-dev 0.62557 37.4 997 26706
60
+ ron-glg flores101-dev 0.56859 29.6 997 25265
61
+ ron-ita flores101-dev 0.54677 24.4 997 25840
62
+ ron-oci flores101-dev 0.52448 22.6 997 25819
63
+ ron-por flores101-dev 0.60829 34.9 997 25287
64
+ ron-spa flores101-dev 0.51970 24.0 997 27793
65
+ spa-ast flores101-dev 0.48365 15.1 997 23232
66
+ spa-cat flores101-dev 0.53735 22.7 997 25962
67
+ spa-fra flores101-dev 0.56861 27.7 997 26706
68
+ spa-glg flores101-dev 0.52786 22.3 997 25265
69
+ spa-ita flores101-dev 0.52474 21.6 997 25840
70
+ spa-oci flores101-dev 0.48027 16.8 997 25819
71
+ spa-por flores101-dev 0.55183 25.2 997 25287
72
+ spa-ron flores101-dev 0.51403 22.0 997 25616
73
+ ast-cat flores101-devtest 0.57870 31.8 1012 27304
74
+ ast-fra flores101-devtest 0.56761 31.1 1012 28343
75
+ ast-glg flores101-devtest 0.55161 27.9 1012 26582
76
+ ast-ita flores101-devtest 0.51764 22.1 1012 27306
77
+ ast-oci flores101-devtest 0.49545 20.6 1012 27305
78
+ ast-por flores101-devtest 0.57347 31.5 1012 26519
79
+ ast-ron flores101-devtest 0.52317 24.8 1012 26799
80
+ ast-spa flores101-devtest 0.49741 21.2 1012 29199
81
+ cat-ast flores101-devtest 0.56754 24.7 1012 24572
82
+ cat-fra flores101-devtest 0.63368 38.4 1012 28343
83
+ cat-glg flores101-devtest 0.59596 32.2 1012 26582
84
+ cat-ita flores101-devtest 0.55886 26.3 1012 27306
85
+ cat-oci flores101-devtest 0.54285 24.6 1012 27305
86
+ cat-por flores101-devtest 0.62913 37.7 1012 26519
87
+ cat-ron flores101-devtest 0.56885 29.5 1012 26799
88
+ cat-spa flores101-devtest 0.53372 24.6 1012 29199
89
+ fra-ast flores101-devtest 0.52696 20.7 1012 24572
90
+ fra-cat flores101-devtest 0.60492 34.6 1012 27304
91
+ fra-glg flores101-devtest 0.57485 30.3 1012 26582
92
+ fra-ita flores101-devtest 0.56493 27.3 1012 27306
93
+ fra-oci flores101-devtest 0.57449 28.2 1012 27305
94
+ fra-por flores101-devtest 0.62211 36.9 1012 26519
95
+ fra-ron flores101-devtest 0.56998 29.4 1012 26799
96
+ fra-spa flores101-devtest 0.52880 24.2 1012 29199
97
+ glg-ast flores101-devtest 0.55090 22.4 1012 24572
98
+ glg-cat flores101-devtest 0.60550 32.6 1012 27304
99
+ glg-fra flores101-devtest 0.62026 36.0 1012 28343
100
+ glg-ita flores101-devtest 0.55834 25.9 1012 27306
101
+ glg-oci flores101-devtest 0.52520 21.9 1012 27305
102
+ glg-por flores101-devtest 0.60027 32.7 1012 26519
103
+ glg-ron flores101-devtest 0.55621 27.8 1012 26799
104
+ glg-spa flores101-devtest 0.53219 24.4 1012 29199
105
+ ita-ast flores101-devtest 0.50741 17.1 1012 24572
106
+ ita-cat flores101-devtest 0.57061 27.9 1012 27304
107
+ ita-fra flores101-devtest 0.60199 32.0 1012 28343
108
+ ita-glg flores101-devtest 0.55312 25.9 1012 26582
109
+ ita-oci flores101-devtest 0.48447 18.1 1012 27305
110
+ ita-por flores101-devtest 0.58162 29.0 1012 26519
111
+ ita-ron flores101-devtest 0.53703 24.2 1012 26799
112
+ ita-spa flores101-devtest 0.52238 23.1 1012 29199
113
+ oci-ast flores101-devtest 0.53010 20.2 1012 24572
114
+ oci-cat flores101-devtest 0.59946 32.2 1012 27304
115
+ oci-fra flores101-devtest 0.64290 39.0 1012 28343
116
+ oci-glg flores101-devtest 0.56737 28.0 1012 26582
117
+ oci-ita flores101-devtest 0.54220 24.2 1012 27306
118
+ oci-por flores101-devtest 0.62127 35.7 1012 26519
119
+ oci-ron flores101-devtest 0.55906 28.0 1012 26799
120
+ oci-spa flores101-devtest 0.52110 22.8 1012 29199
121
+ por-ast flores101-devtest 0.54539 22.5 1012 24572
122
+ por-cat flores101-devtest 0.61809 36.4 1012 27304
123
+ por-fra flores101-devtest 0.64343 39.7 1012 28343
124
+ por-glg flores101-devtest 0.57965 30.4 1012 26582
125
+ por-ita flores101-devtest 0.55841 26.3 1012 27306
126
+ por-oci flores101-devtest 0.54829 25.3 1012 27305
127
+ por-ron flores101-devtest 0.57283 29.8 1012 26799
128
+ por-spa flores101-devtest 0.53513 25.2 1012 29199
129
+ ron-ast flores101-devtest 0.52265 20.1 1012 24572
130
+ ron-cat flores101-devtest 0.59689 32.6 1012 27304
131
+ ron-fra flores101-devtest 0.63060 37.4 1012 28343
132
+ ron-glg flores101-devtest 0.56677 29.3 1012 26582
133
+ ron-ita flores101-devtest 0.55485 25.6 1012 27306
134
+ ron-oci flores101-devtest 0.52433 21.8 1012 27305
135
+ ron-por flores101-devtest 0.61831 36.1 1012 26519
136
+ ron-spa flores101-devtest 0.52712 24.1 1012 29199
137
+ spa-ast flores101-devtest 0.49008 15.7 1012 24572
138
+ spa-cat flores101-devtest 0.53905 23.2 1012 27304
139
+ spa-fra flores101-devtest 0.57078 27.4 1012 28343
140
+ spa-glg flores101-devtest 0.52563 22.0 1012 26582
141
+ spa-ita flores101-devtest 0.52783 22.3 1012 27306
142
+ spa-oci flores101-devtest 0.48064 16.3 1012 27305
143
+ spa-por flores101-devtest 0.55736 25.8 1012 26519
144
+ spa-ron flores101-devtest 0.51623 21.4 1012 26799
145
+ fra-ita newssyscomb2009 0.60995 32.1 502 11551
146
+ fra-spa newssyscomb2009 0.60224 34.2 502 12503
147
+ ita-fra newssyscomb2009 0.61237 33.7 502 12331
148
+ ita-spa newssyscomb2009 0.60706 35.4 502 12503
149
+ spa-fra newssyscomb2009 0.61290 34.6 502 12331
150
+ spa-ita newssyscomb2009 0.61632 33.3 502 11551
151
+ fra-spa news-test2008 0.58939 33.9 2051 52586
152
+ spa-fra news-test2008 0.58695 32.4 2051 52685
153
+ fra-ita newstest2009 0.59764 31.2 2525 63466
154
+ fra-spa newstest2009 0.58829 32.5 2525 68111
155
+ ita-fra newstest2009 0.59084 31.6 2525 69263
156
+ ita-spa newstest2009 0.59669 33.5 2525 68111
157
+ spa-fra newstest2009 0.59096 32.3 2525 69263
158
+ spa-ita newstest2009 0.60783 33.2 2525 63466
159
+ fra-spa newstest2010 0.62250 37.8 2489 65480
160
+ spa-fra newstest2010 0.61953 36.2 2489 66022
161
+ fra-spa newstest2011 0.62953 39.8 3003 79476
162
+ spa-fra newstest2011 0.61130 34.9 3003 80626
163
+ fra-spa newstest2012 0.62397 39.0 3003 79006
164
+ spa-fra newstest2012 0.60927 34.3 3003 78011
165
+ fra-spa newstest2013 0.59312 34.9 3000 70528
166
+ spa-fra newstest2013 0.59468 33.6 3000 70037
167
+ cat-fra tatoeba-test-v2020-07-28 0.71379 55.0 686 5517
168
+ cat-ita tatoeba-test-v2020-07-28 0.74192 58.6 294 1995
169
+ cat-por tatoeba-test-v2020-07-28 0.74959 57.6 729 5847
170
+ fra-cat tatoeba-test-v2020-07-28 0.66559 46.4 686 5214
171
+ fra-ita tatoeba-test-v2020-07-28 0.72709 53.7 10000 61282
172
+ fra-por tatoeba-test-v2020-07-28 0.68142 48.3 10000 71341
173
+ fra-ron tatoeba-test-v2020-07-28 0.65019 43.9 1952 12383
174
+ fra-spa tatoeba-test-v2020-07-28 0.71972 54.7 10000 76002
175
+ glg-por tatoeba-test-v2020-07-28 0.76871 61.2 430 3078
176
+ ita-cat tatoeba-test-v2020-07-28 0.72695 56.6 294 2074
177
+ ita-fra tatoeba-test-v2020-07-28 0.77930 65.2 10000 65549
178
+ ita-por tatoeba-test-v2020-07-28 0.72649 53.6 2500 18891
179
+ ita-ron tatoeba-test-v2020-07-28 0.70789 51.1 1004 6203
180
+ lad-spa tatoeba-test-v2020-07-28 0.56506 38.4 207 1076
181
+ lat_Latn-por tatoeba-test-v2020-07-28 0.31405 11.3 4999 64023
182
+ por-cat tatoeba-test-v2020-07-28 0.70936 51.9 729 5878
183
+ por-fra tatoeba-test-v2020-07-28 0.75256 60.6 10000 73898
184
+ por-glg tatoeba-test-v2020-07-28 0.75444 59.0 430 2989
185
+ por-ita tatoeba-test-v2020-07-28 0.76190 57.4 2500 18301
186
+ por-lat_Latn tatoeba-test-v2020-07-28 0.32111 5.6 4999 49173
187
+ por-ron tatoeba-test-v2020-07-28 0.67789 46.6 681 4529
188
+ por-spa tatoeba-test-v2020-07-28 0.79326 64.6 10000 77915
189
+ ron-fra tatoeba-test-v2020-07-28 0.70274 54.9 1952 13478
190
+ ron-ita tatoeba-test-v2020-07-28 0.73841 53.6 1004 6345
191
+ ron-por tatoeba-test-v2020-07-28 0.70987 51.0 681 4604
192
+ ron-spa tatoeba-test-v2020-07-28 0.73537 57.2 1961 12694
193
+ spa-fra tatoeba-test-v2020-07-28 0.73050 57.2 10000 80915
194
+ spa-lad tatoeba-test-v2020-07-28 0.36854 12.8 207 1090
195
+ spa-por tatoeba-test-v2020-07-28 0.76180 59.6 10000 77911
196
+ spa-ron tatoeba-test-v2020-07-28 0.66673 45.5 1961 12518
197
+ cat-fra tatoeba-test-v2021-03-30 0.71041 54.4 705 5724
198
+ cat-ita tatoeba-test-v2021-03-30 0.74355 58.9 296 2010
199
+ cat-por tatoeba-test-v2021-03-30 0.74920 57.3 753 6173
200
+ egl-ita tatoeba-test-v2021-03-30 0.25460 5.3 203 1008
201
+ fra-cat tatoeba-test-v2021-03-30 0.66514 46.2 705 5398
202
+ fra-ita tatoeba-test-v2021-03-30 0.72735 53.8 10041 61568
203
+ fra-oci tatoeba-test-v2021-03-30 0.37528 11.8 807 6057
204
+ fra-pcd tatoeba-test-v2021-03-30 0.13333 0.4 268 1578
205
+ fra-por tatoeba-test-v2021-03-30 0.68127 48.3 10151 72451
206
+ fra-ron tatoeba-test-v2021-03-30 0.65019 43.9 1952 12383
207
+ fra-spa tatoeba-test-v2021-03-30 0.72025 54.8 10122 77059
208
+ glg-por tatoeba-test-v2021-03-30 0.76744 61.3 438 3138
209
+ ita-cat tatoeba-test-v2021-03-30 0.72871 57.0 296 2089
210
+ ita-egl tatoeba-test-v2021-03-30 0.11118 1.0 203 1093
211
+ ita-fra tatoeba-test-v2021-03-30 0.77930 65.2 10041 65856
212
+ ita-pms tatoeba-test-v2021-03-30 0.17895 1.0 233 1939
213
+ ita-por tatoeba-test-v2021-03-30 0.72846 54.0 5069 38699
214
+ ita-ron tatoeba-test-v2021-03-30 0.70691 51.0 1015 6261
215
+ lad_Latn-spa tatoeba-test-v2021-03-30 0.60681 44.7 242 1294
216
+ lad-spa tatoeba-test-v2021-03-30 0.56680 39.5 260 1405
217
+ lat_Latn-por tatoeba-test-v2021-03-30 0.31424 11.4 5001 64073
218
+ oci-fra tatoeba-test-v2021-03-30 0.49707 29.6 807 6310
219
+ pcd-fra tatoeba-test-v2021-03-30 0.39112 17.7 268 1685
220
+ pms-ita tatoeba-test-v2021-03-30 0.40099 19.8 233 1734
221
+ por-cat tatoeba-test-v2021-03-30 0.71112 52.1 753 6209
222
+ por-fra tatoeba-test-v2021-03-30 0.75323 60.7 10151 75038
223
+ por-glg tatoeba-test-v2021-03-30 0.75379 59.0 438 3048
224
+ por-ita tatoeba-test-v2021-03-30 0.76475 57.9 5069 37479
225
+ por-lat_Latn tatoeba-test-v2021-03-30 0.32119 5.6 5001 49208
226
+ por-ron tatoeba-test-v2021-03-30 0.67676 46.7 715 4725
227
+ por-spa tatoeba-test-v2021-03-30 0.79310 64.7 10395 80844
228
+ ron-fra tatoeba-test-v2021-03-30 0.70274 54.9 1952 13478
229
+ ron-ita tatoeba-test-v2021-03-30 0.73780 53.5 1015 6401
230
+ ron-por tatoeba-test-v2021-03-30 0.71087 51.1 715 4807
231
+ ron-spa tatoeba-test-v2021-03-30 0.73537 57.2 1961 12694
232
+ spa-fra tatoeba-test-v2021-03-30 0.73092 57.3 10122 82050
233
+ spa-lad tatoeba-test-v2021-03-30 0.37849 14.6 260 1417
234
+ spa-lad_Latn tatoeba-test-v2021-03-30 0.41093 15.1 242 1306
235
+ spa-por tatoeba-test-v2021-03-30 0.76015 59.3 10395 81022
236
+ spa-ron tatoeba-test-v2021-03-30 0.66673 45.5 1961 12518
237
+ cat-fra tatoeba-test-v2021-08-07 0.71201 54.6 700 5664
238
+ cat-ita tatoeba-test-v2021-08-07 0.74198 58.4 298 2028
239
+ cat-por tatoeba-test-v2021-08-07 0.74930 57.4 747 6119
240
+ cat-spa tatoeba-test-v2021-08-07 0.87844 78.1 1534 12094
241
+ egl-ita tatoeba-test-v2021-08-07 0.25485 5.3 202 1004
242
+ fra-cat tatoeba-test-v2021-08-07 0.66525 46.2 700 5342
243
+ fra-ita tatoeba-test-v2021-08-07 0.72742 53.8 10091 62060
244
+ fra-oci tatoeba-test-v2021-08-07 0.37505 11.8 806 6047
245
+ fra-pcd tatoeba-test-v2021-08-07 0.13336 0.4 266 1569
246
+ fra-por tatoeba-test-v2021-08-07 0.68413 48.6 10518 77650
247
+ fra-ron tatoeba-test-v2021-08-07 0.65009 44.0 1925 12252
248
+ fra-spa tatoeba-test-v2021-08-07 0.72080 54.8 10294 78406
249
+ glg-por tatoeba-test-v2021-08-07 0.76720 61.1 433 3105
250
+ glg-spa tatoeba-test-v2021-08-07 0.82362 71.7 2121 17443
251
+ ita-cat tatoeba-test-v2021-08-07 0.72529 56.4 298 2109
252
+ ita-egl tatoeba-test-v2021-08-07 0.11133 1.0 202 1088
253
+ ita-fra tatoeba-test-v2021-08-07 0.77932 65.2 10091 66377
254
+ ita-pms tatoeba-test-v2021-08-07 0.17907 1.0 232 1926
255
+ ita-por tatoeba-test-v2021-08-07 0.72798 54.0 3066 25668
256
+ ita-ron tatoeba-test-v2021-08-07 0.70814 51.1 1005 6209
257
+ ita-spa tatoeba-test-v2021-08-07 0.77455 62.9 5000 34937
258
+ lad_Latn-spa tatoeba-test-v2021-08-07 0.59363 42.6 239 1239
259
+ lad-spa tatoeba-test-v2021-08-07 0.52243 34.7 276 1448
260
+ lat_Latn-por tatoeba-test-v2021-08-07 0.31410 11.3 5000 64031
261
+ oci-fra tatoeba-test-v2021-08-07 0.49660 29.6 806 6302
262
+ pcd-fra tatoeba-test-v2021-08-07 0.39257 17.7 266 1677
263
+ pms-ita tatoeba-test-v2021-08-07 0.40221 20.0 232 1721
264
+ por-cat tatoeba-test-v2021-08-07 0.71146 52.2 747 6149
265
+ por-fra tatoeba-test-v2021-08-07 0.75565 60.9 10518 80459
266
+ por-glg tatoeba-test-v2021-08-07 0.75348 59.0 433 3016
267
+ por-ita tatoeba-test-v2021-08-07 0.76883 58.8 3066 24897
268
+ por-lat_Latn tatoeba-test-v2021-08-07 0.32112 5.6 5000 49182
269
+ por-ron tatoeba-test-v2021-08-07 0.67838 46.6 681 4521
270
+ por-spa tatoeba-test-v2021-08-07 0.79336 64.8 10947 87335
271
+ ron-fra tatoeba-test-v2021-08-07 0.70307 55.0 1925 13347
272
+ ron-ita tatoeba-test-v2021-08-07 0.73862 53.7 1005 6352
273
+ ron-por tatoeba-test-v2021-08-07 0.70889 50.7 681 4593
274
+ ron-spa tatoeba-test-v2021-08-07 0.73529 57.2 1959 12679
275
+ spa-cat tatoeba-test-v2021-08-07 0.82758 67.9 1534 12343
276
+ spa-fra tatoeba-test-v2021-08-07 0.73113 57.3 10294 83501
277
+ spa-glg tatoeba-test-v2021-08-07 0.77332 63.0 2121 16581
278
+ spa-ita tatoeba-test-v2021-08-07 0.77046 60.3 5000 34515
279
+ spa-lad tatoeba-test-v2021-08-07 0.34217 13.4 276 1464
280
+ spa-lad_Latn tatoeba-test-v2021-08-07 0.40084 14.7 239 1254
281
+ spa-por tatoeba-test-v2021-08-07 0.75854 59.1 10947 87610
282
+ spa-ron tatoeba-test-v2021-08-07 0.66679 45.5 1959 12503
283
+ cat-ita wmt21-ml-wp 0.69968 47.8 1743 42735
284
+ cat-oci wmt21-ml-wp 0.73808 51.6 1743 43736
285
+ cat-ron wmt21-ml-wp 0.51178 29.0 1743 42895
286
+ ita-cat wmt21-ml-wp 0.70538 48.9 1743 43833
287
+ ita-oci wmt21-ml-wp 0.59025 32.0 1743 43736
288
+ ita-ron wmt21-ml-wp 0.51261 28.9 1743 42895
289
+ oci-cat wmt21-ml-wp 0.80908 66.1 1743 43833
290
+ oci-ita wmt21-ml-wp 0.63584 39.6 1743 42735
291
+ oci-ron wmt21-ml-wp 0.47384 24.6 1743 42895
292
+ ron-cat wmt21-ml-wp 0.52994 31.1 1743 43833
293
+ ron-ita wmt21-ml-wp 0.52714 29.6 1743 42735
294
+ ron-oci wmt21-ml-wp 0.45932 21.3 1743 43736
benchmark_translations.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:621c7657fe954235aa2ce27f34c19f3086b20696bd55608fa1ffd5263dce47d4
3
+ size 45914103
config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_dropout": 0.0,
3
+ "activation_function": "relu",
4
+ "architectures": [
5
+ "MarianMTModel"
6
+ ],
7
+ "attention_dropout": 0.0,
8
+ "bad_words_ids": [
9
+ [
10
+ 35282
11
+ ]
12
+ ],
13
+ "bos_token_id": 0,
14
+ "classifier_dropout": 0.0,
15
+ "d_model": 1024,
16
+ "decoder_attention_heads": 16,
17
+ "decoder_ffn_dim": 4096,
18
+ "decoder_layerdrop": 0.0,
19
+ "decoder_layers": 6,
20
+ "decoder_start_token_id": 35282,
21
+ "decoder_vocab_size": 35283,
22
+ "dropout": 0.1,
23
+ "encoder_attention_heads": 16,
24
+ "encoder_ffn_dim": 4096,
25
+ "encoder_layerdrop": 0.0,
26
+ "encoder_layers": 6,
27
+ "eos_token_id": 29158,
28
+ "forced_eos_token_id": 29158,
29
+ "init_std": 0.02,
30
+ "is_encoder_decoder": true,
31
+ "max_length": 512,
32
+ "max_position_embeddings": 1024,
33
+ "model_type": "marian",
34
+ "normalize_embedding": false,
35
+ "num_beams": 4,
36
+ "num_hidden_layers": 6,
37
+ "pad_token_id": 35282,
38
+ "scale_embedding": true,
39
+ "share_encoder_decoder_embeddings": true,
40
+ "static_position_embeddings": true,
41
+ "torch_dtype": "float16",
42
+ "transformers_version": "4.18.0.dev0",
43
+ "use_cache": true,
44
+ "vocab_size": 35283
45
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3159e3ce64dbbf65dc9521e49f4e3fcf04c9b0544282c4c7b9f332548063d70
3
+ size 497396163
source.spm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afade4a78a9e1c0853e1614f9f2f607d522fdceda4eceb510883b7954a287220
3
+ size 810819
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"}
target.spm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8452e3069d0e13d220a2cf2da7cf1bd9cda32a653721a3cab29ed3b8d7d5ed64
3
+ size 811011
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"source_lang": "itc", "target_lang": "itc", "unk_token": "<unk>", "eos_token": "</s>", "pad_token": "<pad>", "model_max_length": 512, "sp_model_kwargs": {}, "separate_vocabs": false, "special_tokens_map_file": null, "name_or_path": "marian-models/opusTCv20210807_transformer-big_2022-08-10/itc-itc", "tokenizer_class": "MarianTokenizer"}
vocab.json ADDED
The diff for this file is too large to render. See raw diff