DeBERTa commited on
Commit
67754fc
1 Parent(s): f008b26

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -22
README.md CHANGED
@@ -5,26 +5,5 @@ license: mit
5
 
6
  ## DeBERTa: Decoding-enhanced BERT with Disentangled Attention
7
 
8
- [DeBERTa](https://arxiv.org/abs/2006.03654) improves the BERT and RoBERTa models using disentangled attention and enhanced mask decoder. With those two improvements, DeBERTa out perform RoBERTa on a majority of NLU tasks with 80GB training data.
9
 
10
- Please check the [official repository](https://github.com/microsoft/DeBERTa) for more details and updates.
11
-
12
- This the DeBERTa V2 xlarge model fine-tuned with MNLI task, 24 layers, 1536 hidden size. Total parameters 900M.
13
-
14
- ## This model is deprecated, please use [DeBERTa-V2-XLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xlarge-mnli)
15
-
16
-
17
- ### Citation
18
-
19
- If you find DeBERTa useful for your work, please cite the following paper:
20
-
21
- ``` latex
22
- @inproceedings{
23
- he2021deberta,
24
- title={DEBERTA: DECODING-ENHANCED BERT WITH DISENTANGLED ATTENTION},
25
- author={Pengcheng He and Xiaodong Liu and Jianfeng Gao and Weizhu Chen},
26
- booktitle={International Conference on Learning Representations},
27
- year={2021},
28
- url={https://openreview.net/forum?id=XPZIaotutsD}
29
- }
30
- ```
5
 
6
  ## DeBERTa: Decoding-enhanced BERT with Disentangled Attention
7
 
8
+ ## This model is DEPRECATED, please use [DeBERTa-V2-XLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xlarge-mnli)
9