Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -16,7 +16,7 @@ In [DeBERTa V3](https://arxiv.org/abs/2111.09543), we further improved the effic
16
 
17
  Please check the [official repository](https://github.com/microsoft/DeBERTa) for more implementation details and updates.
18
 
19
- The DeBERTa V3 large model comes with 24 layers and a hidden size of 1024. It has 304M backbone parameters with a vocabulary containing 128K tokens which introduces 131M parameters in the Embedding layer. This model was trained using the 160GB data as DeBERTa V2.
20
 
21
 
22
  #### Fine-tuning on NLU tasks
 
16
 
17
  Please check the [official repository](https://github.com/microsoft/DeBERTa) for more implementation details and updates.
18
 
19
+ The DeBERTa V3 large model comes with 24 layers and a hidden size of 1024. It has 304M backbone parameters with a vocabulary containing 128K tokens which introduces 131M parameters in the Embedding layer. This model was trained using the same 160GB data as DeBERTa V2.
20
 
21
 
22
  #### Fine-tuning on NLU tasks