Jump to content

Sanov's theorem

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by Benni678 (talk | contribs) at 20:59, 28 April 2011 (added citation). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

In information theory, Sanov's theorem gives a bound on the probability of observing an atypical sequence of samples from a given probability distribution.

Let A be a set of probability distributions over an alphabet X, and let q be an arbitrary distribution over X (where q may or may not be in A). Suppose we draw n i.i.d. samples from q, represented by the vector . Then

,

where

  • is shorthand for and is shorthand for ,
  • is the empirical distribution of the sample , and
  • is the information projection of q onto A.

Furthermore, if A is a closed set,

References

Cover, Thomas M.; Thomas, Joy A. (2006). Elements of Information Theory (2 ed.). Hoboken, New Jersey: Wiley Interscience. p. 362.