DeBERTa commited on
Commit
98da113
1 Parent(s): 8711b9e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -0
README.md CHANGED
@@ -69,6 +69,15 @@ python -m torch.distributed.launch --nproc_per_node=${num_gpus} \
69
  If you find DeBERTa useful for your work, please cite the following paper:
70
 
71
  ``` latex
 
 
 
 
 
 
 
 
 
72
  @inproceedings{
73
  he2021deberta,
74
  title={DEBERTA: DECODING-ENHANCED BERT WITH DISENTANGLED ATTENTION},
 
69
  If you find DeBERTa useful for your work, please cite the following paper:
70
 
71
  ``` latex
72
+ @misc{he2021debertav3,
73
+ title={DeBERTaV3: Improving DeBERTa using ELECTRA-Style Pre-Training with Gradient-Disentangled Embedding Sharing},
74
+ author={Pengcheng He and Jianfeng Gao and Weizhu Chen},
75
+ year={2021},
76
+ eprint={2111.09543},
77
+ archivePrefix={arXiv},
78
+ primaryClass={cs.CL}
79
+ }
80
+
81
  @inproceedings{
82
  he2021deberta,
83
  title={DEBERTA: DECODING-ENHANCED BERT WITH DISENTANGLED ATTENTION},