File size: 2,177 Bytes
4d56008
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fcd3557
4d56008
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
---
license: apache-2.0
base_model: openlm-research/open_llama_3b_v2
datasets:
- xaviviro/oasst2_euskera_gpt
tags:
- finetune
- chatml
- gpt4
- catalan
model-index:
- name: SUGARRA-3B
  results: []
library_name: transformers
widget:
- text: "<|im_start|>user\nNor zen Isaac Newton?<|im_end|>\n<|im_start|>assistant\n"
language:
- eu
- en
---

# SUGARRA: Modelo 3B experimental ChatML euskaldun.


![SUGARRA](sugarra_1.png)

SUGARRA es el resultado de finetunear el modelo [open_llama_3b_v2](/openlm-research/open_llama_3b_v2) con las instrucciones [OpenAssistant v2](/datasets/OpenAssistant/oasst2) traducidas automáticamente al euskera usando recursos de [Helsinki-NLP](/Helsinki-NLP) y tratadas en formato ChatML.


# Prompt Template

SUGARRA usa el prompt template **ChatML**:

```
<|im_start|>user
Nor zen Isaac Newton?<|im_end|>
<|im_start|>assistant\n
```

[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)

## Referències

```
@software{xaviviro2024sugarra,
  author = {xaviviro},
  title = {SUGARRA: Modelo 3B experimental ChatML euskaldun.},
  month = January,
  year = 2024,
  url = {https://huggingface.co/xaviviro/SUGARRA-3B}
}
```

```
@software{openlm2023openllama,
  author = {Geng, Xinyang and Liu, Hao},
  title = {OpenLLaMA: An Open Reproduction of LLaMA},
  month = May,
  year = 2023,
  url = {https://github.com/openlm-research/open_llama}
}
```
```
@software{together2023redpajama,
  author = {Together Computer},
  title = {RedPajama-Data: An Open Source Recipe to Reproduce LLaMA training dataset},
  month = April,
  year = 2023,
  url = {https://github.com/togethercomputer/RedPajama-Data}
}
```
```
@article{touvron2023llama,
  title={Llama: Open and efficient foundation language models},
  author={Touvron, Hugo and Lavril, Thibaut and Izacard, Gautier and Martinet, Xavier and Lachaux, Marie-Anne and Lacroix, Timoth{\'e}e and Rozi{\`e}re, Baptiste and Goyal, Naman and Hambro, Eric and Azhar, Faisal and others},
  journal={arXiv preprint arXiv:2302.13971},
  year={2023}
}
```