Sanov's theorem

(Learn how and when to remove this message)

In mathematics and information theory, Sanov's theorem gives a bound on the probability of observing an atypical sequence of samples from a given probability distribution. In the language of large deviations theory, Sanov's theorem identifies the rate function for large deviations of the empirical measure of a sequence of i.i.d. random variables.

Let A be a set of probability distributions over an alphabet X, and let q be an arbitrary distribution over X (where q may or may not be in A). Suppose we draw n i.i.d. samples from q, represented by the vector x n = x 1 , x 2 , , x n {\displaystyle x^{n}=x_{1},x_{2},\ldots ,x_{n}} . Then, we have the following bound on the probability that the empirical measure p ^ x n {\displaystyle {\hat {p}}_{x^{n}}} of the samples falls within the set A:

q n ( p ^ x n A ) ( n + 1 ) | X | 2 n D K L ( p | | q ) {\displaystyle q^{n}({\hat {p}}_{x^{n}}\in A)\leq (n+1)^{|X|}2^{-nD_{\mathrm {KL} }(p^{*}||q)}} ,

where

In words, the probability of drawing an atypical distribution is bounded by a function of the KL divergence from the true distribution to the atypical one; in the case that we consider a set of possible atypical distributions, there is a dominant atypical distribution, given by the information projection.

Furthermore, if A is a closed set, then

lim n 1 n log q n ( p ^ x n A ) = D K L ( p | | q ) . {\displaystyle \lim _{n\to \infty }{\frac {1}{n}}\log q^{n}({\hat {p}}_{x^{n}}\in A)=-D_{\mathrm {KL} }(p^{*}||q).}

Technical statement

Define:

{ ( a 1 / n , , a | Σ | / n ) : i a i = n , a i N } {\displaystyle \{(a_{1}/n,\dots ,a_{|\Sigma |}/n):\sum _{i}a_{i}=n,a_{i}\in \mathbb {N} \}}
Then, Sanov's theorem states:[1]

Here, i n t ( S ) {\displaystyle int(S)} means the interior, and c l ( S ) {\displaystyle cl(S)} means the closure.

References

  1. ^ Dembo, Amir; Zeitouni, Ofer (2010). "Large Deviations Techniques and Applications". Stochastic Modelling and Applied Probability. 38: 16–17. doi:10.1007/978-3-642-03311-7. ISBN 978-3-642-03310-0. ISSN 0172-4568.


  • v
  • t
  • e