mmarimon commited on
Commit
56cc8af
1 Parent(s): 54daa54

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -4
README.md CHANGED
@@ -35,6 +35,7 @@ widget:
35
  - [Model Description](#model-description)
36
  - [Intended Uses and Limitations](#intended-uses-and-limitations)
37
  - [How to Use](#how-to-use)
 
38
  - [Training](#training)
39
  - [Training Data](#training-data)
40
  - [Training Procedure](#training-procedure)
@@ -60,13 +61,9 @@ widget:
60
  ## Model Description
61
  RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019.
62
 
63
-
64
  ## Intended Uses and Limitations
65
  You can use the raw model for fill mask or fine-tune it to a downstream task.
66
 
67
- The training data used for this model has not been released as a dataset one can browse. We know it contains a lot of
68
- unfiltered content from the internet, which is far from neutral. At the time of submission, no measures have been taken to estimate the bias and toxicity embedded in the model. However, we are well aware that our models may be biased since the corpora have been collected using crawling techniques on multiple web sources. We intend to conduct research in these areas in the future, and if completed, this model card will be updated.
69
-
70
  ## How to Use
71
  You can use this model directly with a pipeline for fill mask. Since the generation relies on some randomness, we set a seed for reproducibility:
72
 
@@ -109,6 +106,10 @@ Here is how to use this model to get the features of a given text in PyTorch:
109
  torch.Size([1, 19, 1024])
110
  ```
111
 
 
 
 
 
112
  ## Training
113
 
114
  ### Training Data
 
35
  - [Model Description](#model-description)
36
  - [Intended Uses and Limitations](#intended-uses-and-limitations)
37
  - [How to Use](#how-to-use)
38
+ - [Limitations and bias](#limitations-and-bias)
39
  - [Training](#training)
40
  - [Training Data](#training-data)
41
  - [Training Procedure](#training-procedure)
 
61
  ## Model Description
62
  RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019.
63
 
 
64
  ## Intended Uses and Limitations
65
  You can use the raw model for fill mask or fine-tune it to a downstream task.
66
 
 
 
 
67
  ## How to Use
68
  You can use this model directly with a pipeline for fill mask. Since the generation relies on some randomness, we set a seed for reproducibility:
69
 
 
106
  torch.Size([1, 19, 1024])
107
  ```
108
 
109
+ ## Limitations and bias
110
+
111
+ At the time of submission, no measures have been taken to estimate the bias and toxicity embedded in the model. However, we are well aware that our models may be biased since the corpora have been collected using crawling techniques on multiple web sources. We intend to conduct research in these areas in the future, and if completed, this model card will be updated.
112
+
113
  ## Training
114
 
115
  ### Training Data