BERT

GPTKB entity

Statements (63)
Predicate Object
gptkbp:instanceOf large language model
gptkbp:architecture gptkb:transformation
gptkbp:attentionMechanism self-attention
gptkbp:author gptkb:Jacob_Devlin
gptkb:Kenton_Lee
gptkb:Ming-Wei_Chang
gptkb:Kristina_Toutanova
gptkbp:citation over 50,000 (as of 2023)
gptkbp:developedBy gptkb:Google
gptkbp:format WordPiece tokens
gptkbp:fullName gptkb:Bidirectional_Encoder_Representations_from_Transformers
gptkbp:github https://github.com/google-research/bert
gptkbp:hasVariant gptkb:BERT_Base
gptkb:BERT_Large
gptkb:Chinese_BERT
gptkb:Multilingual_BERT
gptkbp:headOfState 12 (BERT Base)
16 (BERT Large)
gptkbp:hiddenSize 1024 (BERT Large)
768 (BERT Base)
https://www.w3.org/2000/01/rdf-schema#label BERT
gptkbp:influenced gptkb:ERNIE
gptkb:ALBERT
gptkb:DistilBERT
gptkb:RoBERTa
gptkb:XLNet
gptkbp:introducedIn 2018
gptkbp:language English
gptkbp:layer 12 (BERT Base)
24 (BERT Large)
gptkbp:license Apache 2.0
gptkbp:maxSequenceLength 512
gptkbp:notableFor improving state-of-the-art on NLP tasks
gptkbp:notablePublication gptkb:BERT:_Pre-training_of_Deep_Bidirectional_Transformers_for_Language_Understanding
gptkbp:openSource true
gptkbp:parameter 110 million (BERT Base)
340 million (BERT Large)
gptkbp:pretrainingTask masked language modeling
next sentence prediction
gptkbp:programmingLanguage gptkb:Python
gptkbp:publishedIn gptkb:NAACL_2019
gptkbp:relatedTo gptkb:NLP
deep learning
transformers
gptkbp:supports transfer learning
fine-tuning
gptkbp:tokenizer gptkb:WordPiece
gptkbp:trainer gptkb:English_Wikipedia
gptkb:BooksCorpus
gptkbp:usedFor natural language processing
question answering
sentiment analysis
text classification
named entity recognition
gptkbp:bfsParent gptkb:transformation
gptkb:convolutional_neural_network
gptkb:GPT-3
gptkb:Zhenzhong_Lan
gptkb:ERNIE
gptkb:Google_AI
gptkb:LaMDA
gptkb:Large_Language_Models
gptkbp:bfsLayer 5