Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -16,7 +16,7 @@ In [DeBERTa V3](https://arxiv.org/abs/2111.09543), we further improved the effic
16
 
17
  Please check the [official repository](https://github.com/microsoft/DeBERTa) for more implementation details and updates.
18
 
19
- The DeBERTa V3 base model comes with 12 layers and a hidden size of 768. It has only 86M backbone parameters with a vocabulary containing 128K tokens which introduces 98M parameters in the Embedding layer. This model was trained using the 160GB data as DeBERTa V2.
20
 
21
 
22
  #### Fine-tuning on NLU tasks
 
16
 
17
  Please check the [official repository](https://github.com/microsoft/DeBERTa) for more implementation details and updates.
18
 
19
+ The DeBERTa V3 base model comes with 12 layers and a hidden size of 768. It has only 86M backbone parameters with a vocabulary containing 128K tokens which introduces 98M parameters in the Embedding layer. This model was trained using the same 160GB data as DeBERTa V2.
20
 
21
 
22
  #### Fine-tuning on NLU tasks