GLaM: Efficient Scaling of Language Models with Mixture-of-Experts (Du et al., 2022)
GPTKB entity
Statements (26)
Predicate | Object |
---|---|
gptkbp:instanceOf |
gptkb:academic_journal
|
gptkbp:author |
gptkb:Xuezhi_Wang
gptkb:Aakanksha_Chowdhery gptkb:Jiawei_Han gptkb:Andrew_M._Dai gptkb:Yanping_Huang gptkb:Zoubin_Ghahramani gptkb:Zhenzhong_Lan gptkb:Hieu_Pham gptkb:Sharan_Narang gptkb:Quoc_V._Le gptkb:Jeffrey_Dean gptkb:Yonghui_Wu Jeffrey Shlens Le Hou Nan Du |
gptkbp:describes |
gptkb:GLaM
|
gptkbp:focusesOn |
Mixture-of-Experts architecture
efficient scaling of language models |
https://www.w3.org/2000/01/rdf-schema#label |
GLaM: Efficient Scaling of Language Models with Mixture-of-Experts (Du et al., 2022)
|
gptkbp:proposedBy |
GLaM model
|
gptkbp:publicationYear |
2022
|
gptkbp:publishedIn |
gptkb:arXiv
|
gptkbp:url |
https://arxiv.org/abs/2112.06905
|
gptkbp:bfsParent |
gptkb:Mixture_of_Experts
|
gptkbp:bfsLayer |
6
|