mapama247 commited on
Commit
4257f04
1 Parent(s): d917d6b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -4
README.md CHANGED
@@ -181,7 +181,8 @@ Note: A small amount of English data was kept to avoid catastrophic forgetting.
181
 
182
  The training corpus has been tokenized using a byte version of [Byte-Pair Encoding (BPE)](https://github.com/openai/gpt-2) with a vocabulary size of 50,257 tokens.
183
  After training a new tokenizer and adapting [falcon-7b](https://huggingface.co/tiiuae/falcon-7b)'s embedding layer, the model was
184
- further pre-trained in three target languages: Catalan, Spanish and English.
 
185
  The training lasted a total of 320 hours on 8 NVIDIA H100 GPUs with 80GB RAM.
186
 
187
 
@@ -233,11 +234,11 @@ This work was partially funded by:
233
  <details>
234
  <summary>Click to expand</summary>
235
 
236
- The model published in this repository is intended for a generalist purpose and is available to third parties.
237
 
238
- This model may have biases and/or any other undesirable distortions.
239
 
240
- When third parties deploy or provide systems and/or services to other parties using this model (or any systems based on it)
241
  or become users of the model, they should note that it is their responsibility to mitigate the risks arising from its use and,
242
  in any event, to comply with applicable regulations, including regulations regarding the use of Artificial Intelligence.
243
 
 
181
 
182
  The training corpus has been tokenized using a byte version of [Byte-Pair Encoding (BPE)](https://github.com/openai/gpt-2) with a vocabulary size of 50,257 tokens.
183
  After training a new tokenizer and adapting [falcon-7b](https://huggingface.co/tiiuae/falcon-7b)'s embedding layer, the model was
184
+ further pre-trained in three target languages: Catalan, Spanish and English.
185
+
186
  The training lasted a total of 320 hours on 8 NVIDIA H100 GPUs with 80GB RAM.
187
 
188
 
 
234
  <details>
235
  <summary>Click to expand</summary>
236
 
237
+ The model published in this repository is intended for a generalist purpose and is available to third parties under a permissive Apache License, Version 2.0.
238
 
239
+ Be aware that the model may have biases and/or any other undesirable distortions.
240
 
241
+ When third parties deploy or provide systems and/or services to other parties using this model (or any system based on it)
242
  or become users of the model, they should note that it is their responsibility to mitigate the risks arising from its use and,
243
  in any event, to comply with applicable regulations, including regulations regarding the use of Artificial Intelligence.
244