TinyBERT: Distilling BERT for Natural Language Understanding
GPTKB entity
Statements (28)
| Predicate | Object |
|---|---|
| gptkbp:instanceOf |
gptkb:academic_journal
|
| gptkbp:application |
mobile and edge devices
|
| gptkbp:arXivID |
1909.10351
|
| gptkbp:author |
Chen, Xiao
Jiang, Xin Jiao, Xiaoqi Li, Linlin Shang, Lifeng Wang, Fang Wang, Qun Yin, Yichun |
| gptkbp:focusesOn |
natural language understanding
knowledge distillation model compression |
| gptkbp:improves |
gptkb:BERT
|
| gptkbp:proposedBy |
gptkb:TinyBERT
embedding distillation layer-to-layer distillation prediction distillation transformer distillation two-stage learning framework |
| gptkbp:publishedIn |
gptkb:EMNLP_2020
|
| gptkbp:title |
gptkb:TinyBERT:_Distilling_BERT_for_Natural_Language_Understanding
|
| gptkbp:url |
https://arxiv.org/abs/1909.10351
|
| gptkbp:year |
2020
|
| gptkbp:bfsParent |
gptkb:TinyBERT
|
| gptkbp:bfsLayer |
9
|
| https://www.w3.org/2000/01/rdf-schema#label |
TinyBERT: Distilling BERT for Natural Language Understanding
|