File size: 1,978 Bytes
23c6f46
 
 
735c282
23c6f46
 
 
 
 
d03d5bb
23c6f46
d03d5bb
 
 
23c6f46
c2e98d1
 
d03d5bb
 
23c6f46
d03d5bb
 
 
 
 
 
23c6f46
d03d5bb
23c6f46
d03d5bb
23c6f46
d03d5bb
23c6f46
d03d5bb
23c6f46
d0e8555
 
d03d5bb
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
---
language:
  - nn
  - en
tags:
- translation
license: cc-by-4.0
---

## HPLT MT release v1.0

This repository contains the translation model for nn-en trained with OPUS and HPLT data. For usage instructions, evaluation scripts, and inference scripts, please refer to the [HPLT-MT-Models v1.0](https://github.com/hplt-project/HPLT-MT-Models/tree/main/v1.0) GitHub repository.

### Model Info

* Source language: Norwegian
* Target language: English
* Dataset: OPUS and HPLT data
* Model architecture: Transformer-base
* Tokenizer: SentencePiece (Unigram)
* Cleaning: We used OpusCleaner with a set of basic rules. Details can be found in the filter files in [Github](https://github.com/hplt-project/HPLT-MT-Models/tree/main/v1.0/data/en-nn/raw/v2)

You can also read our deliverable report [here](https://hplt-project.org/HPLT_D5_1___Translation_models_for_select_language_pairs.pdf) for more details.

### Usage


The model has been trained with Marian. To run inference, refer to the [Inference/Decoding/Translation](https://github.com/hplt-project/HPLT-MT-Models/tree/main/v1.0#inferencedecodingtranslation) section of our GitHub repository.

The model can be used with the Hugging Face framework if the weights are converted to the Hugging Face format. We might provide this in the future; contributions are also welcome.

### Benchmarks

| testset                                | BLEU | chrF++ | COMET22 |
| -------------------------------------- | ---- | ----- | ----- |
| flores200     | 23.1 | 47.3  | 0.7042  |
| ntrex | 21.1   | 44.6  | 0.669  |

### Acknowledgements

This project has received funding from the European Union's Horizon Europe research and innovation programme under grant agreement No 101070350 and from UK Research and Innovation (UKRI) under the UK government's Horizon Europe funding guarantee [grant number 10052546]

Brought to you by researchers from the University of Edinburgh, Charles University in Prague, and the whole HPLT consortium.