In probability theory, Bennett's inequality provides an upper bound on the probability that the sum of independent random variables deviates from its expected value by more than any specified amount. Bennett's inequality was proved by George Bennett of the University of New South Wales in 1962.
Then for any t ≥ 0,
where h(u) = (1 + u)log(1 + u) – u.
Comparison to other bounds
Hoeffding's inequality only assumes the summands are bounded almost surely, while Bennett's inequality offers some improvement when the variances of the summands are small compared to their almost sure bounds. In both inequalities, unlike some other inequalities or limit theorems, there is no requirement that the component variables have identical or similar distributions.
- Concentration inequality - a summary of tail-bounds on random variables.
- Bennett, G. (1962). "Probability Inequalities for the Sum of Independent Random Variables". Journal of the American Statistical Association. 57 (297): 33–45. doi:10.2307/2282438. JSTOR 2282438.
- Devroye, Luc; Lugosi, Gábor (2001). Combinatorial methods in density estimation. Springer. p. 11. ISBN 978-0-387-95117-1.
- Freedman, D. A. (1975). "On tail probabilities for martingales.". 3. The Annals of Probability: 100–118.
- Fan, X.; Grama, I.; Liu, Q. (2012). "Hoeffding's inequality for supermartingales". Stochastic Processes and their Applications. 122: 3545–3559. doi:10.1016/j.spa.2012.06.009.
|This probability-related article is a stub. You can help Wikipedia by expanding it.|