TinyBERT: Distilling BERT for Natural Language Understanding

GPTKB entity

Statements (28)
Predicate Object
gptkbp:instanceOf gptkb:academic_journal
gptkbp:application mobile and edge devices
gptkbp:arXivID 1909.10351
gptkbp:author Chen, Xiao
Jiang, Xin
Jiao, Xiaoqi
Li, Linlin
Shang, Lifeng
Wang, Fang
Wang, Qun
Yin, Yichun
gptkbp:focusesOn natural language understanding
knowledge distillation
model compression
https://www.w3.org/2000/01/rdf-schema#label TinyBERT: Distilling BERT for Natural Language Understanding
gptkbp:improves gptkb:BERT
gptkbp:proposedBy gptkb:TinyBERT
embedding distillation
layer-to-layer distillation
prediction distillation
transformer distillation
two-stage learning framework
gptkbp:publishedIn gptkb:EMNLP_2020
gptkbp:title gptkb:TinyBERT:_Distilling_BERT_for_Natural_Language_Understanding
gptkbp:url https://arxiv.org/abs/1909.10351
gptkbp:year 2020
gptkbp:bfsParent gptkb:TinyBERT
gptkbp:bfsLayer 6