update model name
Browse files
README.md
CHANGED
@@ -12,9 +12,9 @@ metrics:
|
|
12 |
pipeline_tag: text-generation
|
13 |
---
|
14 |
|
15 |
-
# **Model Card for
|
16 |
|
17 |
-
|
18 |
|
19 |
|
20 |
# **Model Details**
|
@@ -22,7 +22,7 @@ Basque LLaMA is a collection of foundation models specifically tuned for Basque.
|
|
22 |
|
23 |
## **Model Description**
|
24 |
|
25 |
-
|
26 |
|
27 |
The models are released in three sizes: 7B, 13B and 70B.
|
28 |
|
@@ -44,7 +44,7 @@ Use the code below to get started with the model.
|
|
44 |
|
45 |
from transformers import pipeline
|
46 |
|
47 |
-
pipe = pipeline("text-generation", model=”HiTZ/
|
48 |
|
49 |
text = "Euskara adimen artifizialera iritsi da!"
|
50 |
|
@@ -62,12 +62,12 @@ pipe(text, max_new_tokens=50, num_beams=5)
|
|
62 |
|
63 |
# **Uses**
|
64 |
|
65 |
-
|
66 |
|
67 |
|
68 |
## **Direct Use**
|
69 |
|
70 |
-
|
71 |
|
72 |
|
73 |
## **Out-of-Scope Use**
|
@@ -77,7 +77,7 @@ The model was not fine-tuned to follow instructions or to work as a chat assista
|
|
77 |
|
78 |
# **Bias, Risks, and Limitations**
|
79 |
|
80 |
-
In an effort to alleviate the potentially disturbing or harmful content,
|
81 |
|
82 |
Please see the LLaMA’s _Ethical Considerations and Limitations _for further information.
|
83 |
|
@@ -115,7 +115,7 @@ The models were trained using the GPT-Neox library on the HPC CINECA computing c
|
|
115 |
</td>
|
116 |
</tr>
|
117 |
<tr>
|
118 |
-
<td>
|
119 |
</td>
|
120 |
<td><p style="text-align: right">
|
121 |
2000</p>
|
@@ -139,7 +139,7 @@ The models were trained using the GPT-Neox library on the HPC CINECA computing c
|
|
139 |
</td>
|
140 |
</tr>
|
141 |
<tr>
|
142 |
-
<td>
|
143 |
</td>
|
144 |
<td><p style="text-align: right">
|
145 |
1000</p>
|
@@ -163,7 +163,7 @@ The models were trained using the GPT-Neox library on the HPC CINECA computing c
|
|
163 |
</td>
|
164 |
</tr>
|
165 |
<tr>
|
166 |
-
<td>
|
167 |
</td>
|
168 |
<td><p style="text-align: right">
|
169 |
1680</p>
|
@@ -389,7 +389,7 @@ The model was evaluated using the LM Evaluation harness library from Eleuther AI
|
|
389 |
</td>
|
390 |
</tr>
|
391 |
<tr>
|
392 |
-
<td><strong>
|
393 |
</td>
|
394 |
<td>35.67
|
395 |
</td>
|
@@ -411,7 +411,7 @@ The model was evaluated using the LM Evaluation harness library from Eleuther AI
|
|
411 |
</td>
|
412 |
</tr>
|
413 |
<tr>
|
414 |
-
<td><strong>
|
415 |
</td>
|
416 |
<td>53.56
|
417 |
</td>
|
@@ -433,7 +433,7 @@ The model was evaluated using the LM Evaluation harness library from Eleuther AI
|
|
433 |
</td>
|
434 |
</tr>
|
435 |
<tr>
|
436 |
-
<td><strong>
|
437 |
</td>
|
438 |
<td><strong>71.78</strong>
|
439 |
</td>
|
|
|
12 |
pipeline_tag: text-generation
|
13 |
---
|
14 |
|
15 |
+
# **Model Card for Latxa 7b**
|
16 |
|
17 |
+
Latxa is a collection of foundation models specifically tuned for Basque. Based on Meta’s LLaMA 2 model family, these models were further trained with Euscrawl, a highly curated Basque corpora ([Artetxe et al., 2022](https://aclanthology.org/2022.emnlp-main.499/)). Ranging from 7 billion to 70 billion parameters, these models are currently the biggest and best-performing LLMs built for Basque. This is the 7b repository, links to other models can be found in the index at the bottom.
|
18 |
|
19 |
|
20 |
# **Model Details**
|
|
|
22 |
|
23 |
## **Model Description**
|
24 |
|
25 |
+
Latxa is a family of Large Language Models (LLM) based on Meta’s [LLaMA models](https://huggingface.co/meta-llama). Current LLMs exhibit incredible performance for high-resource languages such as English, but, in the case of Basque and other low-resource languages, their performance is close to a random guesser. These limitations widen the gap between high- and low-resource languages when it comes to digital development. We present Latxa to overcome these limitations and promote the development of LLM-based technology and research for the Basque language. Latxa models follow the same architecture as their original counterparts and were further trained in Euscrawl v1 ([Artetxe et al., 2022](https://aclanthology.org/2022.emnlp-main.499/)), a high-quality Basque corpora.
|
26 |
|
27 |
The models are released in three sizes: 7B, 13B and 70B.
|
28 |
|
|
|
44 |
|
45 |
from transformers import pipeline
|
46 |
|
47 |
+
pipe = pipeline("text-generation", model=”HiTZ/latxa-7b-v1”)
|
48 |
|
49 |
text = "Euskara adimen artifizialera iritsi da!"
|
50 |
|
|
|
62 |
|
63 |
# **Uses**
|
64 |
|
65 |
+
Latxa models are intended to be used with Basque data; for any other language the performance is not guaranteed. Same as the original, Latxa inherits the [LLaMA-2 License](https://ai.meta.com/llama/license/) which allows for commercial and research use.
|
66 |
|
67 |
|
68 |
## **Direct Use**
|
69 |
|
70 |
+
Latxa family models are pre-trained LLMs without any task-specific or instruction fine-tuning. That is, the model can either be prompted to perform a specific task or further fine-tuned for specific use cases.
|
71 |
|
72 |
|
73 |
## **Out-of-Scope Use**
|
|
|
77 |
|
78 |
# **Bias, Risks, and Limitations**
|
79 |
|
80 |
+
In an effort to alleviate the potentially disturbing or harmful content, Latxa has been trained on carefully selected and processed data which comes mainly from local media, national/regional newspapers, encyclopedias and blogs (see Euscrawl below). Still, the model is based on LLaMA models and can potentially carry the same bias, risk and limitations.
|
81 |
|
82 |
Please see the LLaMA’s _Ethical Considerations and Limitations _for further information.
|
83 |
|
|
|
115 |
</td>
|
116 |
</tr>
|
117 |
<tr>
|
118 |
+
<td>Latxa 7B
|
119 |
</td>
|
120 |
<td><p style="text-align: right">
|
121 |
2000</p>
|
|
|
139 |
</td>
|
140 |
</tr>
|
141 |
<tr>
|
142 |
+
<td>Latxa 13B
|
143 |
</td>
|
144 |
<td><p style="text-align: right">
|
145 |
1000</p>
|
|
|
163 |
</td>
|
164 |
</tr>
|
165 |
<tr>
|
166 |
+
<td>Latxa 70B
|
167 |
</td>
|
168 |
<td><p style="text-align: right">
|
169 |
1680</p>
|
|
|
389 |
</td>
|
390 |
</tr>
|
391 |
<tr>
|
392 |
+
<td><strong>Latxa 7B</strong>
|
393 |
</td>
|
394 |
<td>35.67
|
395 |
</td>
|
|
|
411 |
</td>
|
412 |
</tr>
|
413 |
<tr>
|
414 |
+
<td><strong>Latxa 13B</strong>
|
415 |
</td>
|
416 |
<td>53.56
|
417 |
</td>
|
|
|
433 |
</td>
|
434 |
</tr>
|
435 |
<tr>
|
436 |
+
<td><strong>Latxa 70B</strong>
|
437 |
</td>
|
438 |
<td><strong>71.78</strong>
|
439 |
</td>
|