gptkbp:instance_of
|
gptkb:Research_Institute
|
gptkbp:alliance
|
1706.03762
|
gptkbp:application
|
machine translation
question answering
text generation
text summarization
|
gptkbp:author
|
gptkb:Ashish_Vaswani
gptkb:Noam_Shazeer
gptkb:Aidan_N._Gomez
gptkb:Taku_Kudo
gptkb:Jakob_Uszkoreit
gptkb:Clifford_Young
gptkb:Lukasz_Kaiser
gptkb:Niki_Parmar
Llion Jones
|
gptkbp:contribution
|
self-attention mechanism
|
gptkbp:doi
|
10.48550/ar Xiv.1706.03762
|
gptkbp:field
|
gptkb:software
|
https://www.w3.org/2000/01/rdf-schema#label
|
Attention Is All You Need
|
gptkbp:impact
|
gptkb:significant
|
gptkbp:influenced
|
gptkb:BERT
gptkb:GPT
gptkb:T5
gptkb:Distil_BERT
gptkb:ELECTRA
gptkb:Transformer-XL
ALBERT
ERNIE
Ro BER Ta
XL Net
De BER Ta
|
gptkbp:introduced
|
gptkb:Transformers_character
|
gptkbp:is_cited_in
|
high
|
gptkbp:key
|
scalability
contextual embeddings
parallelization
multi-head attention
positional encoding
layer normalization
|
gptkbp:language
|
English
|
gptkbp:published_by
|
gptkb:Neur_IPS_2017
|
gptkbp:related_works
|
CN Ns
RN Ns
Seq2 Seq models
|
gptkbp:bfsParent
|
gptkb:Google_Brain
gptkb:Ilya_Sutskever
|
gptkbp:bfsLayer
|
4
|