Decoding-enhanced BERT with Disentangled Attention
GPTKB entity
Statements (28)
| Predicate | Object |
|---|---|
| gptkbp:instanceOf |
gptkb:academic_journal
|
| gptkbp:abbreviation |
gptkb:DeBERTa
|
| gptkbp:author |
gptkb:Pengcheng_He
gptkb:Weizhu_Chen gptkb:Xiaodong_Liu gptkb:Jianfeng_Gao |
| gptkbp:basedOn |
gptkb:BERT
|
| gptkbp:citation |
1000+
|
| gptkbp:field |
gptkb:Machine_Learning
gptkb:Natural_Language_Processing |
| gptkbp:improves |
gptkb:BERT
gptkb:RoBERTa gptkb:XLNet |
| gptkbp:introduced |
disentangled attention mechanism
enhanced mask decoder |
| gptkbp:language |
English
|
| gptkbp:proposedBy |
gptkb:Microsoft_Research
|
| gptkbp:publicationYear |
2020
|
| gptkbp:relatedTo |
transformer models
pre-trained language models |
| gptkbp:repository |
https://github.com/microsoft/DeBERTa
|
| gptkbp:usedFor |
gptkb:GLUE_benchmark
gptkb:SQuAD gptkb:SuperGLUE language understanding tasks |
| gptkbp:bfsParent |
gptkb:DeBERTa
|
| gptkbp:bfsLayer |
8
|
| https://www.w3.org/2000/01/rdf-schema#label |
Decoding-enhanced BERT with Disentangled Attention
|