juletxara commited on
Commit
87ebf1a
1 Parent(s): 3ee0836

update model name

Browse files
Files changed (1) hide show
  1. README.md +13 -13
README.md CHANGED
@@ -12,9 +12,9 @@ metrics:
12
  pipeline_tag: text-generation
13
  ---
14
 
15
- # **Model Card for Basque Llama 7b**
16
 
17
- Basque LLaMA is a collection of foundation models specifically tuned for Basque. Based on Meta’s LLaMA 2 model family, these models were further trained with Euscrawl, a highly curated Basque corpora ([Artetxe et al., 2022](https://aclanthology.org/2022.emnlp-main.499/)). Ranging from 7 billion to 70 billion parameters, these models are currently the biggest and best-performing LLMs built for Basque. This is the 7b repository, links to other models can be found in the index at the bottom.
18
 
19
 
20
  # **Model Details**
@@ -22,7 +22,7 @@ Basque LLaMA is a collection of foundation models specifically tuned for Basque.
22
 
23
  ## **Model Description**
24
 
25
- Basque LLaMA is a family of Large Language Models (LLM) based on Meta’s [LLaMA models](https://huggingface.co/meta-llama). Current LLMs exhibit incredible performance for high-resource languages such as English, but, in the case of Basque and other low-resource languages, their performance is close to a random guesser. These limitations widen the gap between high- and low-resource languages when it comes to digital development. We present Basque LLaMA to overcome these limitations and promote the development of LLM-based technology and research for the Basque language. Basque LLaMA models follow the same architecture as their original counterparts and were further trained in Euscrawl v1 ([Artetxe et al., 2022](https://aclanthology.org/2022.emnlp-main.499/)), a high-quality Basque corpora.
26
 
27
  The models are released in three sizes: 7B, 13B and 70B.
28
 
@@ -44,7 +44,7 @@ Use the code below to get started with the model.
44
 
45
  from transformers import pipeline
46
 
47
- pipe = pipeline("text-generation", model=”HiTZ/basque-llama-2-7b-v1”)
48
 
49
  text = "Euskara adimen artifizialera iritsi da!"
50
 
@@ -62,12 +62,12 @@ pipe(text, max_new_tokens=50, num_beams=5)
62
 
63
  # **Uses**
64
 
65
- Basque LLaMA models are intended to be used with Basque data; for any other language the performance is not guaranteed. Same as the original, Basque LLaMA inherits the [LLaMA-2 License](https://ai.meta.com/llama/license/) which allows for commercial and research use.
66
 
67
 
68
  ## **Direct Use**
69
 
70
- Basque LLaMA family models are pre-trained LLMs without any task-specific or instruction fine-tuning. That is, the model can either be prompted to perform a specific task or further fine-tuned for specific use cases.
71
 
72
 
73
  ## **Out-of-Scope Use**
@@ -77,7 +77,7 @@ The model was not fine-tuned to follow instructions or to work as a chat assista
77
 
78
  # **Bias, Risks, and Limitations**
79
 
80
- In an effort to alleviate the potentially disturbing or harmful content, Basque LLaMA has been trained on carefully selected and processed data which comes mainly from local media, national/regional newspapers, encyclopedias and blogs (see Euscrawl below). Still, the model is based on LLaMA models and can potentially carry the same bias, risk and limitations.
81
 
82
  Please see the LLaMA’s _Ethical Considerations and Limitations _for further information.
83
 
@@ -115,7 +115,7 @@ The models were trained using the GPT-Neox library on the HPC CINECA computing c
115
  </td>
116
  </tr>
117
  <tr>
118
- <td>Basque LLaMA 7B
119
  </td>
120
  <td><p style="text-align: right">
121
  2000</p>
@@ -139,7 +139,7 @@ The models were trained using the GPT-Neox library on the HPC CINECA computing c
139
  </td>
140
  </tr>
141
  <tr>
142
- <td>Basque LLaMA 13B
143
  </td>
144
  <td><p style="text-align: right">
145
  1000</p>
@@ -163,7 +163,7 @@ The models were trained using the GPT-Neox library on the HPC CINECA computing c
163
  </td>
164
  </tr>
165
  <tr>
166
- <td>Basque LLaMA 70B
167
  </td>
168
  <td><p style="text-align: right">
169
  1680</p>
@@ -389,7 +389,7 @@ The model was evaluated using the LM Evaluation harness library from Eleuther AI
389
  </td>
390
  </tr>
391
  <tr>
392
- <td><strong>Basque LLaMA 7B</strong>
393
  </td>
394
  <td>35.67
395
  </td>
@@ -411,7 +411,7 @@ The model was evaluated using the LM Evaluation harness library from Eleuther AI
411
  </td>
412
  </tr>
413
  <tr>
414
- <td><strong>Basque LLaMA 13B</strong>
415
  </td>
416
  <td>53.56
417
  </td>
@@ -433,7 +433,7 @@ The model was evaluated using the LM Evaluation harness library from Eleuther AI
433
  </td>
434
  </tr>
435
  <tr>
436
- <td><strong>Basque LLaMA 70B</strong>
437
  </td>
438
  <td><strong>71.78</strong>
439
  </td>
 
12
  pipeline_tag: text-generation
13
  ---
14
 
15
+ # **Model Card for Latxa 7b**
16
 
17
+ Latxa is a collection of foundation models specifically tuned for Basque. Based on Meta’s LLaMA 2 model family, these models were further trained with Euscrawl, a highly curated Basque corpora ([Artetxe et al., 2022](https://aclanthology.org/2022.emnlp-main.499/)). Ranging from 7 billion to 70 billion parameters, these models are currently the biggest and best-performing LLMs built for Basque. This is the 7b repository, links to other models can be found in the index at the bottom.
18
 
19
 
20
  # **Model Details**
 
22
 
23
  ## **Model Description**
24
 
25
+ Latxa is a family of Large Language Models (LLM) based on Meta’s [LLaMA models](https://huggingface.co/meta-llama). Current LLMs exhibit incredible performance for high-resource languages such as English, but, in the case of Basque and other low-resource languages, their performance is close to a random guesser. These limitations widen the gap between high- and low-resource languages when it comes to digital development. We present Latxa to overcome these limitations and promote the development of LLM-based technology and research for the Basque language. Latxa models follow the same architecture as their original counterparts and were further trained in Euscrawl v1 ([Artetxe et al., 2022](https://aclanthology.org/2022.emnlp-main.499/)), a high-quality Basque corpora.
26
 
27
  The models are released in three sizes: 7B, 13B and 70B.
28
 
 
44
 
45
  from transformers import pipeline
46
 
47
+ pipe = pipeline("text-generation", model=”HiTZ/latxa-7b-v1”)
48
 
49
  text = "Euskara adimen artifizialera iritsi da!"
50
 
 
62
 
63
  # **Uses**
64
 
65
+ Latxa models are intended to be used with Basque data; for any other language the performance is not guaranteed. Same as the original, Latxa inherits the [LLaMA-2 License](https://ai.meta.com/llama/license/) which allows for commercial and research use.
66
 
67
 
68
  ## **Direct Use**
69
 
70
+ Latxa family models are pre-trained LLMs without any task-specific or instruction fine-tuning. That is, the model can either be prompted to perform a specific task or further fine-tuned for specific use cases.
71
 
72
 
73
  ## **Out-of-Scope Use**
 
77
 
78
  # **Bias, Risks, and Limitations**
79
 
80
+ In an effort to alleviate the potentially disturbing or harmful content, Latxa has been trained on carefully selected and processed data which comes mainly from local media, national/regional newspapers, encyclopedias and blogs (see Euscrawl below). Still, the model is based on LLaMA models and can potentially carry the same bias, risk and limitations.
81
 
82
  Please see the LLaMA’s _Ethical Considerations and Limitations _for further information.
83
 
 
115
  </td>
116
  </tr>
117
  <tr>
118
+ <td>Latxa 7B
119
  </td>
120
  <td><p style="text-align: right">
121
  2000</p>
 
139
  </td>
140
  </tr>
141
  <tr>
142
+ <td>Latxa 13B
143
  </td>
144
  <td><p style="text-align: right">
145
  1000</p>
 
163
  </td>
164
  </tr>
165
  <tr>
166
+ <td>Latxa 70B
167
  </td>
168
  <td><p style="text-align: right">
169
  1680</p>
 
389
  </td>
390
  </tr>
391
  <tr>
392
+ <td><strong>Latxa 7B</strong>
393
  </td>
394
  <td>35.67
395
  </td>
 
411
  </td>
412
  </tr>
413
  <tr>
414
+ <td><strong>Latxa 13B</strong>
415
  </td>
416
  <td>53.56
417
  </td>
 
433
  </td>
434
  </tr>
435
  <tr>
436
+ <td><strong>Latxa 70B</strong>
437
  </td>
438
  <td><strong>71.78</strong>
439
  </td>