GLaM: Efficient Scaling of Language Models with Mixture-of-Experts (Du et al., 2022)
GPTKB entity
Statements (26)
| Predicate | Object |
|---|---|
| gptkbp:instanceOf |
gptkb:academic_journal
|
| gptkbp:author |
gptkb:Xuezhi_Wang
gptkb:Aakanksha_Chowdhery gptkb:Jiawei_Han gptkb:Andrew_M._Dai gptkb:Yanping_Huang gptkb:Zoubin_Ghahramani gptkb:Zhenzhong_Lan gptkb:Hieu_Pham gptkb:Sharan_Narang gptkb:Quoc_V._Le gptkb:Jeffrey_Dean gptkb:Yonghui_Wu Jeffrey Shlens Le Hou Nan Du |
| gptkbp:describes |
gptkb:GLaM
|
| gptkbp:focusesOn |
Mixture-of-Experts architecture
efficient scaling of language models |
| gptkbp:proposedBy |
GLaM model
|
| gptkbp:publicationYear |
2022
|
| gptkbp:publishedIn |
gptkb:arXiv
|
| gptkbp:url |
https://arxiv.org/abs/2112.06905
|
| gptkbp:bfsParent |
gptkb:Mixture_of_Experts
|
| gptkbp:bfsLayer |
8
|
| https://www.w3.org/2000/01/rdf-schema#label |
GLaM: Efficient Scaling of Language Models with Mixture-of-Experts (Du et al., 2022)
|