mmarimon commited on
Commit
e1fc578
1 Parent(s): 5c06509

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -40,6 +40,8 @@ inference:
40
  - [Citing information](#citing-information)
41
  - [Disclaimer](#disclaimer)
42
 
 
 
43
  ## Model description
44
  RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019.
45
 
@@ -97,7 +99,7 @@ Intelligence (SEDIA) within the framework of the Plan-TL.},
97
  }
98
  ```
99
 
100
- ## Disclaimer
101
 
102
  The models published in this repository are intended for a generalist purpose and are available to third parties. These models may have bias and/or any other undesirable distortions.
103
 
 
40
  - [Citing information](#citing-information)
41
  - [Disclaimer](#disclaimer)
42
 
43
+ </details>
44
+
45
  ## Model description
46
  RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019.
47
 
 
99
  }
100
  ```
101
 
102
+ ### Disclaimer
103
 
104
  The models published in this repository are intended for a generalist purpose and are available to third parties. These models may have bias and/or any other undesirable distortions.
105