Attention Is All You Need (paper)

GPTKB entity

Statements (43)
Predicate Object
gptkbp:instanceOf gptkb:academic_journal
gptkbp:affiliation gptkb:Google_Research
gptkb:Google_Brain
gptkbp:application machine translation
question answering
text generation
language modeling
text summarization
gptkbp:arXivID 1706.03762
gptkbp:author gptkb:Illia_Polosukhin
gptkb:Łukasz_Kaiser
gptkb:Aidan_N._Gomez
gptkb:Ashish_Vaswani
gptkb:Jakob_Uszkoreit
gptkb:Llion_Jones
gptkb:Niki_Parmar
gptkb:Noam_Shazeer
gptkbp:citation over 100,000
gptkbp:contribution positional encoding
improved translation quality
parallelization of training
introduction of multi-head attention
elimination of recurrence in sequence transduction models
gptkbp:doi 10.48550/arXiv.1706.03762
gptkbp:field gptkb:machine_learning
deep learning
natural language processing
https://www.w3.org/2000/01/rdf-schema#label Attention Is All You Need (paper)
gptkbp:impact foundation for large language models
inspired BERT
inspired GPT
inspired T5
gptkbp:introduced Transformer architecture
self-attention mechanism
gptkbp:language English
gptkbp:openAccess true
gptkbp:pages 11
gptkbp:publicationYear 2017
gptkbp:publishedIn gptkb:NeurIPS_2017
gptkbp:title gptkb:Attention_Is_All_You_Need
gptkbp:url https://arxiv.org/abs/1706.03762
gptkbp:bfsParent gptkb:Noam_Shazeer
gptkbp:bfsLayer 6