asier-gutierrez commited on
Commit
4f8deae
1 Parent(s): 8252578

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -37,7 +37,7 @@ Some of the statistics of the corpus:
37
  The training corpus has been tokenized using a byte version of Byte-Pair Encoding (BPE) used in the original [RoBERTA](https://arxiv.org/abs/1907.11692) model with a vocabulary size of 50,262 tokens. The RoBERTa-large-bne pre-training consists of a masked language model training that follows the approach employed for the RoBERTa large. The training lasted a total of 96 hours with 32 computing nodes each one with 4 NVIDIA V100 GPUs of 16GB VRAM.
38
 
39
  ## Evaluation and results
40
- For evaluation details visit our [GitHub repository](https://github.com/PlanTL-SANIDAD/lm-spanish).
41
 
42
  ## Citing
43
  Check out our paper for all the details: https://arxiv.org/abs/2107.07253
 
37
  The training corpus has been tokenized using a byte version of Byte-Pair Encoding (BPE) used in the original [RoBERTA](https://arxiv.org/abs/1907.11692) model with a vocabulary size of 50,262 tokens. The RoBERTa-large-bne pre-training consists of a masked language model training that follows the approach employed for the RoBERTa large. The training lasted a total of 96 hours with 32 computing nodes each one with 4 NVIDIA V100 GPUs of 16GB VRAM.
38
 
39
  ## Evaluation and results
40
+ For evaluation details visit our [GitHub repository](https://github.com/PlanTL-GOB-ES/lm-spanish).
41
 
42
  ## Citing
43
  Check out our paper for all the details: https://arxiv.org/abs/2107.07253