Skip to content

KullbackLeiblerDivergence

Stephen Crowley edited this page Mar 22, 2023 · 1 revision

The Kullback-Leibler (KL) divergence, also known as the relative entropy, is a measure of the difference between two probability distributions. Given two probability distributions $P$ and $Q$ over the same event space, the KL divergence from $Q$ to $P$ is defined as:

$$D_{\text{KL}}(P|Q) = \sum_{x\in\mathcal{X}} P(x)\log\frac{P(x)}{Q(x)}$$

where $\mathcal{X}$ is the event space, $P(x)$ and $Q(x)$ are the probabilities of the event $x$ under the distributions $P$ and $Q$, respectively, and the logarithm is usually taken with respect to some base, such as 2 or $e$.

The KL divergence measures the amount of information lost when approximating the true distribution $P$ with the approximating distribution $Q$. It is a non-negative quantity, and it is zero if and only if $P$ and $Q$ are identical. The KL divergence is asymmetric, which means that $D_{\text{KL}}(P|Q)\neq D_{\text{KL}}(Q|P)$ in general.

The KL divergence has many applications in statistics, machine learning, and information theory. In particular, it is often used in model selection and model comparison, where it can be used to quantify the distance between a true data-generating process and a model. The KL divergence can also be used to design loss functions for training models, or to measure the similarity between two probability distributions in clustering or classification problems.

In summary, the Kullback-Leibler divergence is a measure of the difference between two probability distributions, and it quantifies the amount of information lost when approximating one distribution with another. The KL divergence has many applications in statistics, machine learning, and information theory, and it is often used in model selection, loss function design, and clustering or classification problems.

Clone this wiki locally