BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
GPTKB entity
Statements (27)
Predicate | Object |
---|---|
gptkbp:instanceOf |
gptkb:academic_journal
|
gptkbp:application |
machine translation
natural language generation text comprehension |
gptkbp:arXivID |
1910.13461
|
gptkbp:author |
gptkb:Mike_Lewis
gptkb:Omer_Levy gptkb:Yinhan_Liu gptkb:Veselin_Stoyanov gptkb:Abdelrahman_Mohamed gptkb:Marjan_Ghazvininejad gptkb:Luke_Zettlemoyer gptkb:Naman_Goyal |
gptkbp:citation |
gptkb:transformation
gptkb:BERT gptkb:GPT |
gptkbp:focusesOn |
denoising autoencoder
sequence-to-sequence pre-training |
https://www.w3.org/2000/01/rdf-schema#label |
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
|
gptkbp:proposedBy |
BART model
|
gptkbp:publicationYear |
2019
|
gptkbp:publishedIn |
gptkb:arXiv
|
gptkbp:url |
https://arxiv.org/abs/1910.13461
|
gptkbp:bfsParent |
gptkb:Yinhan_Liu
gptkb:Naman_Goyal gptkb:Lewis_et_al.,_2020 |
gptkbp:bfsLayer |
7
|