GPTKB
Browse
Query
Compare
Download
Publications
Contributors
Search
Attention Is All You Need (paper)
URI:
https://gptkb.org/entity/Attention_Is_All_You_Need_(paper)
GPTKB entity
Statements (43)
Predicate
Object
gptkbp:instanceOf
gptkb:academic_journal
gptkbp:affiliation
gptkb:Google_Research
gptkb:Google_Brain
gptkbp:application
machine translation
question answering
text generation
language modeling
text summarization
gptkbp:arXivID
1706.03762
gptkbp:author
gptkb:Illia_Polosukhin
gptkb:Łukasz_Kaiser
gptkb:Aidan_N._Gomez
gptkb:Ashish_Vaswani
gptkb:Jakob_Uszkoreit
gptkb:Llion_Jones
gptkb:Niki_Parmar
gptkb:Noam_Shazeer
gptkbp:citation
over 100,000
gptkbp:contribution
positional encoding
improved translation quality
parallelization of training
introduction of multi-head attention
elimination of recurrence in sequence transduction models
gptkbp:doi
10.48550/arXiv.1706.03762
gptkbp:field
gptkb:machine_learning
deep learning
natural language processing
https://www.w3.org/2000/01/rdf-schema#label
Attention Is All You Need (paper)
gptkbp:impact
foundation for large language models
inspired BERT
inspired GPT
inspired T5
gptkbp:introduced
Transformer architecture
self-attention mechanism
gptkbp:language
English
gptkbp:openAccess
true
gptkbp:pages
11
gptkbp:publicationYear
2017
gptkbp:publishedIn
gptkb:NeurIPS_2017
gptkbp:title
gptkb:Attention_Is_All_You_Need
gptkbp:url
https://arxiv.org/abs/1706.03762
gptkbp:bfsParent
gptkb:Noam_Shazeer
gptkbp:bfsLayer
6