BERT-large

GPTKB entity

Statements (50)
Predicate Object
gptkbp:instanceOf large language model
gptkbp:architecture gptkb:transformation
gptkbp:author gptkb:Jacob_Devlin
gptkb:Kenton_Lee
gptkb:Ming-Wei_Chang
gptkb:Kristina_Toutanova
gptkbp:availableOn gptkb:Hugging_Face_Model_Hub
gptkb:TensorFlow_Hub
gptkbp:bidirectional true
gptkbp:citation over 50,000
gptkbp:developedBy gptkb:Google_AI
gptkbp:fineTunedWith true
gptkbp:hiddenSize 1024
https://www.w3.org/2000/01/rdf-schema#label BERT-large
gptkbp:impact state-of-the-art on GLUE
state-of-the-art on SQuAD
gptkbp:input gptkb:text
gptkbp:inputLengthLimit 512 tokens
gptkbp:introducedIn 2018
gptkbp:language English
gptkbp:level 24
gptkbp:license Apache 2.0
gptkbp:memiliki_tugas question answering
sentiment analysis
named entity recognition
natural language inference
gptkbp:notableFor deep bidirectional context
transformer-based pretraining
gptkbp:notablePublication gptkb:BERT:_Pre-training_of_Deep_Bidirectional_Transformers_for_Language_Understanding
gptkbp:numberOfAttentionHeads 16
gptkbp:openSource true
gptkbp:output contextual embeddings
gptkbp:parameter 340 million
gptkbp:pretrainingObjective masked language modeling
next sentence prediction
gptkbp:pretrainingTime several days on 16 TPU chips
gptkbp:relatedTo gptkb:BERT-base
gptkb:DistilBERT
gptkb:RoBERTa
gptkbp:tokenizerType gptkb:WordPiece
gptkbp:trainer gptkb:English_Wikipedia
gptkb:BooksCorpus
gptkbp:usedIn information retrieval
search engines
chatbots
text classification
machine reading comprehension
gptkbp:usesSelfAttention true
gptkbp:bfsParent gptkb:TinyBERT
gptkbp:bfsLayer 6