File size: 8,366 Bytes
d83f0d8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
db2caff
d190699
f75ae68
9cd8273
d83f0d8
 
b84f4c9
 
83bd757
0ea9bda
b84f4c9
 
 
 
 
 
d29ff40
b84f4c9
 
 
f94bd14
2351aec
b84f4c9
 
 
 
de751fb
9028ecb
f94bd14
9028ecb
883c1f5
 
 
 
 
 
0ed8cf2
a90c2f0
 
 
 
 
 
 
 
 
 
 
 
 
6654247
 
 
a90c2f0
 
 
 
 
 
 
 
 
 
 
 
03066f8
a90c2f0
03066f8
 
 
fb08f72
34571e8
42b91e3
34571e8
 
 
 
 
 
 
 
 
 
 
 
de751fb
03066f8
 
 
de751fb
b5f78f9
b50c66c
03066f8
 
 
 
848e23a
 
 
2351aec
 
 
 
 
8c05266
 
 
 
 
 
 
2351aec
 
f94bd14
2351aec
f94bd14
 
3812117
f94bd14
 
3812117
f94bd14
 
3812117
f94bd14
 
2351aec
f94bd14
e99062c
bc7abcb
bbbd07e
2351aec
f94bd14
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
---
license: apache-2.0
language:
- en
- gl
- de
- es
- ca
- it
- fr
- eu
- pt
metrics:
- comet
- bleu
pipeline_tag: translation
widget:
- text: <s> [spa_Latn] Ayer él se fue, tomó sus cosas y se puso a navegar. \n[cat_Latn]
inference: false
---

# Plume32k

This is the model card of Plume (**P**arallel **L**ang**u**age **M**od**e**l) with a vocabulary size of 32k.

## Table of Contents
<details>
<summary>Click to expand</summary>

- [Model description](#model-description)
- [Intended uses and limitations](#intended-uses-and-limitations)
- [Run the model](#run-the-model)
- [Training](#training)
- [Evaluation](#evaluation)
- [Citation](#citation)
- [Additional information](#additional-information)
  
</details>

## Summary

Plume is the first LLM trained for Neural Machine Translation with only parallel Catalan-Centric data from scratch. It is a language model with the same architecture as Gemma 2B. The model is trained for general translation tasks at sentence level. For more information about training, architecture and interpretability of the model check out the paper;  "Investigating the translation capabilities of Large Language Models trained on parallel data only". The preprint is available on [arXiv](https://arxiv.org/abs/2406.09140).

- **Developed by:** The Language Technologies Unit from Barcelona Supercomputing Center (BSC).
- **Languages:** Spanish, French, Italian, Portuguese, Galician, German, English, and Basque.
- **License:** Apache License, Version 2.0

## Model Description

In recent years, Large Language Models (LLMs) have demonstrated exceptional proficiency across a broad spectrum of Natural Language Processing (NLP) tasks, including Machine Translation. However, previous methodologies predominantly relied on iterative processes such as instruction fine-tuning or continual pre-training, leaving unexplored the challenges of training LLMs solely on parallel data. In this work, we introduce Plume (**P**arallel **L**ang**u**age **M**od**e**l), a collection of three 2B LLMs featuring varying vocabulary sizes (32k, 128k, and 256k) trained exclusively on  Catalan-centric parallel examples. These models perform comparable to previous encoder-decoder architectures on 16 supervised translation directions and 56 zero-shot ones.

For more details regarding the model architecture, the dataset and model interpretability take a look at the [paper](https://arxiv.org/abs/2406.09140).

## Intended Uses and Limitations

The model is proficient in 16 supervised translation directions that include Catalan and is capable of translating in other 56 zero-shot directions as well.

At the time of submission, no measures have been taken to estimate the bias and added toxicity embedded in the model. However, we are aware that our models may be biased since the corpora have been collected using crawling techniques on multiple web sources. We intend to conduct research in these areas in the future, and if completed, this model card will be updated.

## Run the model


```python
from transformers import AutoTokenizer, AutoModelForCausalLM

# language codes: spa_Latn (Spanish), cat_Latn (Catalan), eng_Latn (English), ita_Latn (Italian), 
# eus_Latn (Basque), deu_Latn (German), por_Latn (Portuguese), glg_Latn (Galician), fra_Latn (French)

model_id = "projecte-aina/Plume32k"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)

src_lang_code = 'spa_Latn'
tgt_lang_code = 'cat_Latn'
sentence = 'Ayer se fue, tomó sus cosas y se puso a navegar.'
prompt = '<s> [{}] {} \n[{}]'.format(src_lang_code, sentence, tgt_lang_code)
input_ids = tokenizer(prompt, return_tensors='pt').input_ids
output_ids = model.generate( input_ids, max_length=200, num_beams=5 )
input_length = input_ids.shape[1]
generated_text = tokenizer.decode(output_ids[0, input_length: ], skip_special_tokens=True).strip()
# Ahir se'n va anar, va agafar les seves coses i es va posar a navegar.
```

## Training

For training, the learning rate is warmed up from 1e-7 to a maximum of 3e-4 over the first 2000 steps. We apply a weight decay of 0.1 and a gradient clipping of 1.0. During training, we set an effective batch size of 81,920 tokens per gradient step distributed over 40 NVIDIA H100-64GB GPUs. We use DeepSpeed with full *float32* training. We show in the next table the training hyperparameters:

| **Hyper-Parameter** |     **Value**                     |
|---------------------|--------------------------|
| Batch size          | 40                       |
| Number of Epochs    | 1                        |
| Optimizer           | Adam                     |
| Adam-β₁             | 0.9                      |
| Adam-β₂             | 0.999                    |
| Adam-ε              | 1e-08                    |
| Learning rate       | 3e-04                    |
| LR Scheduler        | Linear                   |
| Warmup Steps        | 2000                     |


More training details are specified in the [paper](https://arxiv.org/abs/2406.09140). Code for training the model and running other experiments can be found in our [GitHub repository](https://github.com/projecte-aina/Plume).

## Evaluation

Below are the evaluation results on Flores-200 and NTREX for supervised MT directions. For more details about model evaluation check out the [paper](https://arxiv.org/abs/2406.09140).

| Model  | FLORES BLEU | FLORES COMET | NTREX BLEU | NTREX COMET |
|----------------------|-------------|--------------|------------|-------------|
| NLLB-1.3B            | 31.02       | 0.86         | 29.68      | 0.85        |
| NLLB-600M            | 29.24       | 0.85         | 28.37      | 0.84        |
| Bilinguals BSC       | 31.93       | 0.86         | 29.77      | 0.84        |
| **Plume 32k**           | 30.44       | 0.86         | 28.46      | 0.84        |
| **Plume 128k**          | 30.81       | 0.86         | 28.78      | 0.84        |
| **Plume 256k**          | 30.72       | 0.86         | 28.87      | 0.84        |


## Citation

```bibtex
@misc{gilabert2024investigating,
      title={Investigating the translation capabilities of Large Language Models trained on parallel data only}, 
      author={Javier García Gilabert and Carlos Escolano and Aleix Sant Savall and Francesca De Luca Fornaciari and Audrey Mash and Xixian Liao and Maite Melero},
      year={2024},
      eprint={2406.09140},
      archivePrefix={arXiv}
}
```

## Additional information

### Author
The Language Technologies Unit from Barcelona Supercomputing Center.

### Contact
Feel free to write us at with any questions you may have to {javier.garcia1, carlos.escolano, aleix.santsavall, francesca.delucafornaciari, audrey.mash, xixian.liao, maite.melero}@bsc.es 

### Copyright
Copyright(c) 2023 by Language Technologies Unit, Barcelona Supercomputing Center.

### License
[Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0)

### Funding

This work has been promoted and financed by the Government of Catalonia through the [Aina](https://projecteaina.cat/) project, by the Ministerio para la Transformación Digital y de la Función Pública and Plan de Recuperación, Transformación y Resiliencia - Funded by EU – NextGenerationEU within the framework of the project [ILENIA](https://proyectoilenia.es/) with reference 2022/TL22/00215337, as well as by [DeepR3](https://ixa2.si.ehu.eus/deepr3/) (TED2021-130295B-C32) founded by MCIN/AEI/10.13039/501100011033 and European Union NextGeneration EU/PRTR.


### Disclaimer

<details>
<summary>Click to expand</summary>

The model published in this repository is intended for a generalist purpose and is available to third parties under a permissive Apache License, Version 2.0. 

Be aware that the model may have biases and/or any other undesirable distortions.

When third parties deploy or provide systems and/or services to other parties using this model (or any system based on it) 
or become users of the model, they should note that it is their responsibility to mitigate the risks arising from its use and, 
in any event, to comply with applicable regulations, including regulations regarding the use of Artificial Intelligence.

In no event shall the owner and creator of the model (Barcelona Supercomputing Center) be liable for any results arising from the use made by third parties.

</details>