Attention is All You Need

GPTKB entity

Statements (54)
Predicate Object
gptkbp:instance_of gptkb:Research_Institute
gptkbp:addresses long-range dependencies
gptkbp:applies_to text generation
language modeling
text summarization
gptkbp:author Vaswani et al.
gptkbp:contains encoder-decoder architecture
gptkbp:enhances contextual understanding
gptkbp:field gptkb:software
gptkbp:focuses_on self-attention mechanism
gptkbp:has_programs gptkb:Research_Institute
dialog systems
question answering
chatbots
virtual assistants
text classification
https://www.w3.org/2000/01/rdf-schema#label Attention is All You Need
gptkbp:impact machine translation
gptkbp:improves training efficiency
gptkbp:influenced gptkb:BERT
gptkb:GPT
gptkb:T5
gptkbp:introduced gptkb:Transformers_character
gptkbp:is_available_on gptkb:ar_Xiv
gptkbp:is_cited_in gptkb:ACM_Digital_Library
gptkb:Google_Scholar
gptkb:IEEE_Xplore
gptkb:Research_Gate
gptkb:Semantic_Scholar
thousands of papers
gptkbp:is_compared_to RN Ns and CN Ns
gptkbp:is_described_as multi-head attention
gptkbp:is_discussed_in gptkb:academic_conferences
workshops
seminars
gptkbp:is_influential_in AI community
machine learning community
NLP community
gptkbp:is_part_of AI curriculum
NLP courses
NLP literature
deep learning courses
gptkbp:is_recognized_by seminal work
gptkbp:is_used_in transformer-based models
gptkbp:proposed_by attention scores
gptkbp:provides parallelization benefits
gptkbp:published_by gptkb:2017
gptkbp:reduces computational complexity
gptkbp:reign_end deep learning architectures
gptkbp:translated_into multiple languages
gptkbp:utilizes positional encoding
gptkbp:bfsParent gptkb:GPT-3
gptkb:ICLR
gptkbp:bfsLayer 4