KL divergence

GPTKB entity

Statements (32)
Predicate Object
gptkbp:instanceOf statistical analysis
gptkbp:alsoKnownAs gptkb:Kullback–Leibler_divergence
gptkbp:category divergence measures
information theory concepts
gptkbp:definedIn continuous probability distributions
discrete probability distributions
gptkbp:describes difference between two probability distributions
gptkbp:field gptkb:information_theory
gptkb:machine_learning
statistics
gptkbp:form KL(P||Q) = Σ P(x) log(P(x)/Q(x))
gptkbp:generalizes cross-entropy
gptkbp:heldBy asymmetric
non-negative
non-symmetric
https://www.w3.org/2000/01/rdf-schema#label KL divergence
gptkbp:minimumAt P = Q
gptkbp:minimumPressure 0
gptkbp:namedAfter gptkb:Richard_Leibler
gptkb:Solomon_Kullback
gptkbp:relatedTo gptkb:organization
gptkb:Jensen–Shannon_divergence
mutual information
cross-entropy
gptkbp:usedFor relative entropy
model selection
variational inference
measuring information loss
gptkbp:bfsParent gptkb:Kullback–Leibler_divergence
gptkb:Kullback-Leibler_divergence
gptkb:convolutional_neural_network
gptkbp:bfsLayer 5