Longformer

GPTKB entity

Statements (57)
Predicate Object
gptkbp:instanceOf Transformer model
gptkbp:completed state-of-the-art results
gptkbp:designedFor long documents
gptkbp:developedBy gptkb:Facebook_AI_Research
gptkbp:hasRelatedPatent question answering
language modeling
summarization
text classification
gptkbp:hasVariants Longformer-Decoder
Longformer-Encoder
https://www.w3.org/2000/01/rdf-schema#label Longformer
gptkbp:improves scalability
gptkbp:introduced 2020
gptkbp:isAttendedBy industry
academia
gptkbp:isAvailableIn Hugging Face Model Hub
gptkbp:isBasedOn BERT architecture
gptkbp:isCompatibleWith gptkb:PyTorch
TensorFlow
gptkbp:isConsidered a_breakthrough_in_NLP
gptkbp:isDiscussedIn workshops
AI conferences
gptkbp:isDocumentedIn research papers
technical blogs
gptkbp:isEvaluatedBy gptkb:GLUE_benchmark
SQuAD benchmark
Long Range Arena
gptkbp:isInfluencedBy Attention is All You Need
gptkbp:isKnownFor enhancing performance
reducing computational cost
handling long-range dependencies
gptkbp:isLocatedIn Python
gptkbp:isOpenTo true
gptkbp:isOptimizedFor memory efficiency
gptkbp:isPartOf Big Bird
Reformer
AI research community
standard Transformers
transformer family
gptkbp:isSupportedBy community contributions
gptkbp:isTrainedIn large datasets
gptkbp:isUsedBy developers
researchers
gptkbp:isUsedIn natural language processing tasks
gptkbp:isUtilizedIn data analysis
language translation
dialog systems
sentiment analysis
chatbots
information retrieval
text mining
virtual assistants
content generation
knowledge extraction
gptkbp:mayHave up to 4096 tokens
gptkbp:supports self-attention
gptkbp:uses sparse attention mechanism