angel-poc commited on
Commit
08cbbd4
1 Parent(s): 0b91fed

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -72,7 +72,7 @@ widget:
72
 
73
  ## Model description
74
 
75
- The **roberta-large-bne-massive** is a Intent Classification model for the Catalan language fine-tuned from the roberta-large-bne-massive model, a [RoBERTa](https://arxiv.org/abs/1907.11692) based model pre-trained on a medium-size corpus collected from publicly available corpora and crawlers.
76
 
77
  ## Intended uses and limitations
78
 
@@ -99,7 +99,7 @@ At the time of submission, no measures have been taken to estimate the bias embe
99
  ## Training
100
 
101
  ### Training data
102
- We used the Catalan split of the [MASSIVE](https://huggingface.co/datasets/AmazonScience/massive) dataset for training and evaluation.
103
 
104
  ### Training procedure
105
  The model was trained with a batch size of 16 and a learning rate of 1e-5 for 20 epochs. We then selected the best checkpoint using the downstream task metric in the corresponding development set and then evaluated it on the test set.
72
 
73
  ## Model description
74
 
75
+ The **roberta-large-bne-massive** is a Intent Classification model for the Spanish language fine-tuned from the roberta-large-bne-massive model, a [RoBERTa](https://arxiv.org/abs/1907.11692) based model pre-trained on a medium-size corpus collected from publicly available corpora and crawlers.
76
 
77
  ## Intended uses and limitations
78
 
99
  ## Training
100
 
101
  ### Training data
102
+ We used the Spanish split of the [MASSIVE](https://huggingface.co/datasets/AmazonScience/massive) dataset for training and evaluation.
103
 
104
  ### Training procedure
105
  The model was trained with a batch size of 16 and a learning rate of 1e-5 for 20 epochs. We then selected the best checkpoint using the downstream task metric in the corresponding development set and then evaluated it on the test set.