Hoeffding's inequality

From Wikipedia, the free encyclopedia
Jump to navigation Jump to search

In probability theory, Hoeffding's inequality provides an upper bound on the probability that the sum of bounded independent random variables deviates from its expected value by more than a certain amount. Hoeffding's inequality was proven by Wassily Hoeffding in 1963.[1]

Hoeffding's inequality is a generalization of the Chernoff bound, which applies only to Bernoulli random variables,[2] and a special case of the Azuma–Hoeffding inequality and the McDiarmid's inequality. It is similar to, but incomparable with, the Bernstein inequality, proved by Sergei Bernstein in 1923.

Special case of Bernoulli random variables[edit]

Hoeffding's inequality can be applied to the important special case of identically distributed Bernoulli random variables, and this is how the inequality is often used in combinatorics and computer science. We consider a coin that shows heads with probability p and tails with probability 1 − p. We toss the coin n times. The expected number of times the coin comes up heads is pn. Furthermore, the probability that the coin comes up heads at most k times can be exactly quantified by the following expression:

where H(n) is the number of heads in n coin tosses.

When k = (pε)n for some ε > 0, Hoeffding's inequality bounds this probability by a term that is exponentially small in ε2n:

Similarly, when k = (p + ε)n for some ε > 0, Hoeffding's inequality bounds the probability that we see at least εn more tosses that show heads than we would expect:

Hence Hoeffding's inequality implies that the number of heads that we see is concentrated around its mean, with exponentially small tail.

For example, taking gives:

General case of bounded random variables[edit]

Let X1, ..., Xn be independent random variables bounded by the interval [0, 1]: 0 ≤ Xi ≤ 1. We define the empirical mean of these variables by

One of the inequalities in Theorem 1 of Hoeffding (1963) states

where .

Theorem 2 of Hoeffding (1963) is a generalization of the above inequality when it is known that Xi are strictly bounded by the intervals [ai, bi]:

which are valid for positive values of t. Here E[X] is the expected value of X. The inequalities can be also stated in terms of the sum

of the random variables:

Note that the inequalities also hold when the Xi have been obtained using sampling without replacement; in this case the random variables are not independent anymore. A proof of this statement can be found in Hoeffding's paper. For slightly better bounds in the case of sampling without replacement, see for instance the paper by Serfling (1974).

General case of sub-Gaussian random variables[edit]

A random variable X is called sub-Gaussian,[3] if

for some c>0. For a random variable X, the following norm is finite if and only if it is sub-Gaussian:

Then let X1, ..., Xn be zero-mean independent sub-Gaussian random variables, the general version of the Hoeffding's inequality states that:

where c > 0 is an absolute constant. See Theorem 2.6.2 of Vershynin (2018) for details.


In this section, we give a proof of Hoeffding's inequality.[4] The proof uses Hoeffding's Lemma:

Suppose X is a real random variable such that . Then

Using this lemma, we can prove Hoeffding's inequality. Suppose X1, ..., Xn are n independent random variables such that


Then for s, t > 0, Markov's inequality and the independence of Xi implies:

To get the best possible upper bound, we find the minimum of the right hand side of the last inequality as a function of s. Define

Note that g is a quadratic function and achieves its minimum at

Thus we get


Confidence intervals[edit]

Hoeffding's inequality is useful to analyse the number of required samples needed to obtain a confidence interval by solving the inequality in Theorem 1:

The inequality states that the probability that the estimated and true values differ by more than t is bounded by e−2nt2. Symmetrically, the inequality is also valid for another side of the difference:

By adding them both up, we can obtain two-sided variant of this inequality:

This probability can be interpreted as the level of significance (probability of making an error) for a confidence interval around of size 2t:

Solving the above for n gives us the following:

Therefore, we require at least samples to acquire -confidence interval .

Hence, the cost of acquiring the confidence interval is sublinear in terms of confidence level and quadratic in terms of precision.

Note that this inequality is the most conservative of the three in Theorem 1, and there are more efficient methods of estimating a confidence interval.

See also[edit]


  1. ^ Hoeffding (1963)
  2. ^ Nowak (2009); for a more intuitive proof, see this note
  3. ^ Kahane (1960)
  4. ^ Nowak (2009); for a more intuitive proof, see this note


  • Serfling, Robert J. (1974). "Probability Inequalities for the Sum in Sampling without Replacement". The Annals of Statistics. 2 (1): 39–48. doi:10.1214/aos/1176342611. MR 0420967.CS1 maint: ref=harv (link)
  • Hoeffding, Wassily (1963). "Probability inequalities for sums of bounded random variables" (PDF). Journal of the American Statistical Association. 58 (301): 13–30. doi:10.1080/01621459.1963.10500830. JSTOR 2282952. MR 0144363.CS1 maint: ref=harv (link)
  • Nowak, Robert (2009). "Lecture 7: Chernoff's Bound and Hoeffding's Inequality" (PDF). ECE 901 (Summer '09) : Statistical Learning Theory Lecture Notes. University of Wisconsin-Madison. Retrieved May 16, 2014.
  • Vershynin, Roman (2018). High-Dimensional Probability. Cambridge University Press. ISBN 9781108415194.
  • Kahane, J.P. (1960). "Propriétés locales des fonctions à séries de Fourier aléatoires". Stud. Math. 19. pp. 1–25. [1].