Not to be confused either with the Chebychev's inequalities on the size of the number-theoretic function .
In probability theory, Chebyshev's inequality (also spelled as Tchebysheff's inequality, Нера́венство Чебышева) guarantees that in any probability distribution,"nearly all" values are close to the mean — the precise statement being that no more than 1/k2 of the distribution's values can be more than k standard deviations away from the mean (or equivalently, at least 1 - 1/k2 of the distribution's values are within k standard deviations of the mean). The inequality has great utility because it can be applied to completely arbitrary distributions (unknown except for mean and variance), for example it can be used to prove the weak law of large numbers.
In practical usage, in contrast to the empirical rule, which applies to normal distributions, under Chebyshev's Inequality just 75% of values lie within two standard deviations of the mean and 89% of values within three standard deviations.
The term Chebyshev's inequality may also refer to the Markov's inequality, especially in the context of analysis.
The theorem is named after Russian mathematician Pafnuty Chebyshev, although it was first formulated by his friend and colleague Irénée-Jules Bienaymé.:98 The theorem was first stated without proof by Bienaymé in 1853 and later proved by Chebyshev in 1867. His student Andrey Markov provided another proof in his 1884 PhD thesis.
Probabilistic statement 
Only the case k > 1 provides useful information. When k < 1 the right-hand side is greater than one, so the inequality becomes vacuous, as the probability of any event cannot be greater than one. When k = 1 it just says the probability is less than or equal to one, which is always true.
As an example, using k = √ shows that at least half of the values lie in the interval (μ − √σ, μ + √σ).
Because it can be applied to completely arbitrary distributions (unknown except for mean and variance), the inequality generally gives a poor bound compared to what might be possible if something is known about the distribution involved.
|k||Min % within k standard deviations of mean||Max % beyond k standard deviations from mean|
Measure-theoretic statement 
More generally, if g is an extended real-valued measurable function, nonnegative and nondecreasing on the range of f, then
The previous statement then follows by defining as if and otherwise, and taking instead of .
Suppose we randomly select a journal article from a source with an average of 1000 words per article, with a standard deviation of 200 words. We can then infer that the probability that it has between 600 and 1400 words (i.e. within k = 2 SDs of the mean) must be more than 75%, because there is less than 1⁄k2
= 1 chance to be outside that range, by Chebyshev's inequality. But if we additionally know that the distribution is normal, we can say that is a 75% chance the word count is between 770 and 1230 (which is an even tighter bound).
This example should be treated with caution as the inequality is only stated for probability distributions rather than for finite sample sizes. The inequality has since been extended to apply to finite sample sizes (see below).
Sharpness of bounds 
As shown in the example above, the theorem will typically provide rather loose bounds. However, the bounds provided by Chebyshev's inequality cannot, in general (remaining sound for variables of arbitrary distribution), be improved upon. For example, for any k ≥ 1, the following example meets the bounds exactly.
For this distribution, mean μ = 0 and standard deviation σ = 1, so
Equality holds only for distributions that are a linear transformation of this one.
Proof (of the two-sided version) 
Probabilistic proof 
Markov's inequality states that for any non-negative random variable Y and any positive number a, we have Pr(|Y| > a) ≤ E(|Y|)/a. One way to prove Chebyshev's inequality is to apply Markov's inequality to the random variable Y = (X − μ)2 with a = (σk)2.
It can also be proved directly. For any event A, let IA be the indicator random variable of A, i.e. IA equals 1 if A occurs and 0 otherwise. Then
The direct proof shows why the bounds are quite loose in typical cases: the number 1 to the left of "≥" is replaced by [(X − μ)/(kσ)]2 to the right of "≥" whenever the latter exceeds 1. In some cases it exceeds 1 by a very wide margin.
Measure-theoretic proof 
Fix and let be defined as , and let be the indicator function of the set . Then, it is easy to check that, for any ,
since g is nondecreasing on the range of f, and therefore,
The desired inequality follows from dividing the above inequality by g(t).
Several extensions of Chebyshev's inequality have been developed.
Asymmetric two-sided case 
An asymmetric two-sided version of this inequality is also known.
When the distribution is asymmetric or is unknown
where σ2 is the variance and μ is the mean.
Bivariate case 
A version for the bivariate case is known.
Let X1 and X2 be two random variables with means and finite variances of μ1 and μ2 and σ1 and σ2 respectively. Then
where for i = 1,2,
Berge derived an inequality for two correlated variables X1 and X2. Let ρ be the correlation coefficient between X1 and X2 and let σi2 be the variance of Xi. Then
Lal later obtained an alternative bound
Isii derived a further generalisation. Let
with 0 < k1 ≤ k2.
There are now three cases.
Case A: If and where
Case B: If the conditions in case A are not met but k1k2 ≥ 1 and
Case C: If the conditions in cases A or B are not met there is no universal bound other than 1.
Multivariate case 
The general case is known as the Birnbaum–Raymond–Zuckerman inequality after the authors who proved it for two dimensions.
where Xi is the ith random variable, μi is the ith mean and σi2 is the ith variance.
If the variables are independent this inequality can be sharpened.
Olkin and Pratt derived an inequality for n correlated variables.
where the sum is taken over the n variables and
where ρij is the correlation between Xi and Xj
Olkin and Pratt's inequality was subsequently generalised by Godwin.
Vector version 
where YT is the transpose of Y.
Higher moments 
An extension to higher moments is also possible
where k > 0 and n ≥ 2.
Exponential version 
A related inequality sometimes known as the exponential Chebyshev's inequality is the inequality
where t > 0.
Let K( x, t ) be the cumulant generating function,
This inequality may be used to obtain exponential inequalities for unbounded variables.
Inequalities for bounded variables 
The second of these inequalities with r = 2 is the Chebyshev bound. The first provides a lower bound for the value of P(x).
Sharp bounds for a bounded variate have been derived by Niemitalo
Let 0 ≤ X ≤ M where M > 0. Then
- Case 1
- Case 2
- Case 3
Finite samples 
Saw et al extended Chebyshev's inequality to cases where the population mean and variance are not known but are instead replaced by their sample estimates.
where N is the sample size, m is the sample mean, k is a constant and s is the sample standard deviation. g(x) is defined as follows:
Let x ≥ 1, Q = N + 1, and R be the greatest integer less than Q / x. Let
This inequality holds when the population moments do not exist and when the sample is weakly exchangeably distributed.
Kabán gives a somewhat less complex version of this inequality.
If the standard deviation is a multiple of the mean then a further inequality can be derived,
A table of values for the Saw–Yang–Mo inequality for finite sample sizes (n < 100) has been determined by Konijn.
For fixed N and large m the Saw–Yang–Mo inequality is approximately
Beasley et al have suggested a modification of this inequality
In empirical testing this modification is conservative but appears to have low statistical power. Its theoretical basis currently remains unexplored.
Dependence of sample size 
The bounds these inequalities give on a finite sample are less tight than those the Chebyshev inequality gives for a distribution. To illustrate this let the sample size n = 100 and let k = 3. Chebyshev's inequality states that approximately 11.11% of the distribution will lie outside these limits. Kabán's version of the inequality for a finite sample states that approximately 12.05% of the sample lies outside these limits. The dependence of the confidence intervals on sample size is further illustrated below.
For N = 10, the 95% confidence interval is approximately ±13.5789 standard deviations.
For N = 100 the 95% confidence interval is approximately ±4.9595 standard deviations; the 99% confidence interval is approximately ±140.0 standard deviations.
For N = 500 the 95% confidence interval is approximately ±4.5574 standard deviations; the 99% confidence interval is approximately ±11.1620 standard deviations.
For N = 1000 the 95% and 99% confidence intervals are approximately ±4.5141 and approximately ±10.5330 standard deviations respectively.
The Chebyshev inequality for the distribution gives 95% and 99% confidence intervals of approximately ±4.472 standard deviations and ±10 standard deviations respectively.
Comparative bounds 
Although Chebyshev's inequality is the best possible bound for an arbitrary distribution, this is not necessarily true for finite samples. Samuelson's inequality states that all values of a sample will lie within √(N − 1) standard deviations of the mean. Chebyshev's bound improves as the sample size increases.
When N = 10, Samuelson's inequality states that all members of the sample lie within 3 standard deviations of the mean: in contrast Chebyshev's states that 95% of the sample lies within 13.5789 standard deviations of the mean.
When N = 100, Samuelson's inequality states that all members of the sample lie within approximately 9.9499 standard deviations of the mean: Chebyshev's states that 99% of the sample lies within 140.0 standard deviations of the mean.
When N = 500, Samuelson's inequality states that all members of the sample lie within approximately 22.3383 standard deviations of the mean: Chebyshev's states that 99% of the sample lies within 11.1620 standard deviations of the mean.
It is likely that better bounds for finite samples than these exist.
Sharpened bounds 
Chebyshev's inequality is important because of its applicability to any distribution. As a result of its generality it may not (and usually does not) provide as sharp a bound as alternative methods that can be used if the distribution of the random variable is known. To improve the sharpness of the bounds provided by Chebyshev's inequality a number of methods have been developed.
Standardised variables 
Sharpened bounds can be derived by first standardising the random variable.
Let X be a random variable with finite variance Var(x). Let Z be the standardised form defined as
Cantelli's lemma is then
This inequality is sharp and is attained by k and −1/k with probability 1/(1 + k2) and k2/(1 + k2) respectively.
If k > 1 and the distribution of X is symmetric then we have
Equality holds if and only if Z = −k, 0 or k with probabilities 1 / 2 k2, 1 − 1 / k2 and 1 / 2 k2 respectively. An extension to a two-sided inequality is also possible.
Let u, v > 0. Then we have
An alternative method of obtaining sharper bounds is through the use of semivariances (partial moments). The upper (σ+2) and lower (σ−2) semivariances are defined
where m is the arithmetic mean of the sample, n is the number of elements in the sample and the sum for the upper (lower) semivariance is taken over the elements greater (less) than the mean.
The variance of the sample is the sum of the two semivariances
In terms of the lower semivariance Chebyshev's inequality can be written
Chebyshev's inequality can now be written
A similar result can also be derived for the upper semivariance.
If we put
Chebyshev's inequality can be written
Because σu2 ≤ σ2, use of the semivariance sharpens the original inequality.
If the distribution is known to be symmetric, then
This result agrees with that derived using standardised variables.
- The inequality with the lower semivariance has been found to be of use in estimating downside risk in finance and agriculture.
Selberg's inequality 
Selberg derived an inequality for P(x) when a ≤ x ≤ b. To simplify the notation let
The result of this linear transformation is to make P(a ≤ X ≤ b) equal to P(|Y| ≤ k).
The mean (μX) and variance (σX) of X are related to the mean (μY) and variance (σY) of Y:
With this notation Selberg's inequality states that
These are known to be the best possible bounds.
Cantelli's inequality 
where a ≥ 0.
This inequality can be used to prove a one tailed variant of Chebyshev's inequality with k > 0
The bound on the one tailed variant is known to be sharp. To see this consider the random variable X that takes the values
- with probability
- with probability
Then E(X) = 0 and E(X2) = σ2 and P(X < 1) = 1 / (1 + σ2).
- An application – distance between the mean and the median
The one-sided variant can be used to prove the proposition that for probability distributions having an expected value and a median, the mean and the median can never differ from each other by more than one standard deviation. To express this in symbols let μ, ν, and σ be respectively the mean, the median, and the standard deviation. Then
There is no need to assume that the variance is finite because this inequality is trivially true if the variance is infinite.
The proof is as follows. Setting k = 1 in the statement for the one-sided inequality gives:
Changing the sign of X and of μ, we get
Thus the median is within one standard deviation of the mean.
For a proof using Jensen's inequality see An inequality relating means and medians.
Bhattacharyya's inequality 
Bhattacharyya extended Cantelli's inequality using the third and fourth moments of the distribution.
Let μ = 0 and σ2 be the variance. Let γ = E(X3) / σ3 and κ = E(X4) / σ4.
If k2 − kγ − 1 > 0 then
The necessity of k2 − kγ − 1 > 0 requires that k be reasonably large.
Mitzenmacher and Upfal's inequality 
Mitzenmacher and Upfal note that
for any real k > 0 and that
is the kth central moment. They then show that for t > 0
For k = 2 we obtain Chebyshev's inequality. For t ≥ 1, k > 2 and assuming that the kth moment exists, this bound is tighter than Chebyshev's inequality.
Related inequalities 
Several other related inequalities are also known.
Zelen's inequality 
Zelen has shown that
where Mm is the Mth moment and σ is the standard deviation.
He, Zhang and Zhang's inequality 
For any collection of n nonnegative independent random variables Xi 
Hoeffding’s lemma 
Let X be a random variable with a ≤ X ≤ b and E[ X ] = 0, then for any s > 0, we have
van Zuijlen's bound 
van Zuijlen has proved the following result.
Let Xi be a set of independent Rademacher random variables: P( Xi = 1 ) = P( Xi = −1 ) = 0.5. Then
The bound is sharp and better than that which can be derived from the normal distribution (approximately P > 0.31).
Unimodal distributions 
If the second condition holds then the second bound is always less than or equal to the first.
If the mode (ν) is not zero and the mean (μ) and standard deviation (σ) are both finite then denoting the root mean square deviation from the mode by ω, we have
Winkler in 1866 extended Gauss' inequality to rth moments  where r > 0 and the distribution is unimodal with a mode of zero:
Gauss' bound has been subsequently sharpened and extended to apply to departures from the mean rather than the mode: see the Vysochanskiï–Petunin inequality for details.
The Vysochanskiï–Petunin inequality has been extended by Dharmadhikari and Joag-Dev
where s is a constant satisfying both s > r + 1 and s(s − r − 1) = rr and r > 0.
It can be shown that these inequalities are the best possible and that further sharpening of the bounds requires that additional restrictions be placed on the distributions.
Unimodal symmetrical distributions 
The bounds on this inequality can also be sharpened if the distribution is both unimodal and symmetrical. An empirical distribution can be tested for symmetry with a number of tests including McWilliam's R*. It is known that the variance of a unimodal symmetrical distribution with finite support [a, b] is less than or equal to ( b − a )2 / 12.
Let the distribution be supported on the finite interval [ −N, N ] and the variance be finite. Let the mode of the distribution be zero and rescale the variance to 1. Let k > 0 and assume k < 2N/3. Then
If 0 < k ≤ 2 / √3 the bounds are reached with the density
If 2 / √3 < k ≤ 2N / 3 the bounds are attained by the distribution
where βk = 4 / 3k2, δ0 is the Dirac delta function and where
The existence of these densities shows that the bounds are optimal. Since N is arbitrary these bounds apply to any value of N.
The Camp–Meidell's inequality is a related inequality. For an absolutely continuous unimodal and symmetrical distribution
The second of these inequality is the same as the Vysochanskiï–Petunin inequality.
DasGupta has shown that if the distribution is known to be normal
- Effects of symmetry and unimodality
Symmetry of the distribution decreases the inequality's bounds by a factor of 2 while unimodality sharpens the bounds by a factor of 4/9.
- Unimodal distributions
Because the mean and the mode in a unimodal distribution differ by at most √3 standard deviations at most 5% of a symmetrical unimodal distribution lies outside (2√10 + 3√3)/3 standard deviations of the mean (approximately 3.840 standard deviations). This is sharper than the bounds provided by the Chebyshev inequality (approximately 4.472 standard deviations).
These bounds on the mean are less sharp than those that can be derived from symmetry of the distribution alone which shows that at most 5% of the distribution lies outside approximately 3.162 standard deviations of the mean. The Vysochanskiï–Petunin inequality further sharpens this bound by showing that for such a distribution that at most 5% of the distribution lies outside 4√5/3 (approximately 2.981) standard deviations of the mean.
- Symmetrical unimodal distributions
For any symmetrical unimodal distribution:
- approximately 5.784% of the distribution lies outside 1.96 standard deviations of the mode
- 5% of the distribution lies outside 2√10/3 (approximately 2.11) standard deviations of the mode
DasGupta's inequality states that for a normal distribution at least 95% lies within approximately 2.582 standard deviations of the mean. This is less sharp than the true figure (approximately 1.96 standard deviations of the mean).
Bounds for specific distributions 
Zero means 
When the mean (μ) is zero Chebyshev's inequality takes a simple form. Let σ2 be the variance. Then
With the same conditions Cantelli's inequality takes the form
Unit variance 
If in addition E( X2 ) = 1 and E( X4 ) = ψ then for any 0 ≤ ε ≤ 1
The first inequality is sharp.
It is also known that for a random variable obeying the above conditions that
It is also known that
The value of C0 is optimal and the bounds are sharp if
then the sharp bound is
Integral Chebyshev inequality 
There is a second (less well known) inequality also named after Chebyshev
If f and g are of opposite monotonicity, then the above inequality works in the reverse way.
Other inequalities 
There are also a number of other inequalities associated with Chebyshev
Haldane's transformation 
One use of Chebyshev's inequality in applications is to create confidence intervals for variates with an unknown distribution. Haldane noted, using an equation derived by Kendall, that if a variate (x) has a zero mean, unit variance and both finite skewness (γ) and kurtosis (κ) then the variate can be converted to a normally distributed standard score (z):
This transformation may be useful as an alternative to Chebyshev's inequality or as an adjunct to it for deriving confidence intervals for variates with unknown distributions.
While this transformation may be useful for moderately skewed and/or kurtotic distributions, it performs poorly when the distribution is markedly skewed and/or kurtotic.
Chernoff bounds 
If the random variables may also be assumed to be independently distributed it is possible to obtain sharper bounds. Let δ > 0. Then
With this inequality it can be shown that
where μ is the mean of the distribution. Further discussion may be found in the article on Chernoff bounds
Caution concerning use of Chebyshev's inequality 
The Environmental Protection Agency has suggested best practices for the use of Chebyshev's inequality for estimating confidence intervals. This caution appears to be justified as its use in this context may be seriously misleading 
See also 
- Chernoff bound — a bound on sums of random variables
- Cornish–Fisher expansion
- Eaton's inequality
- Hoeffding's inequality — an exponential bound on the sum of a series of random variables
- Kolmogorov's inequality
- Proof of the weak law of large numbers using Chebyshev's inequality
- Le Cam's theorem
- Multidimensional Chebyshev's inequality
- Paley–Zygmund inequality
- Vysochanskiï–Petunin inequality — a stronger result applicable to unimodal probability distributions
- Kvanli, Alan H.; Pavur, Robert J.; Keeling, Kellie B. (2006). Concise Managerial Statistics. cEngage Learning. pp. 81–82. ISBN 9780324223880.
- Chernick, Michael R. (2011). The Essentials of Biostatistics for Physicians, Nurses, and Clinicians. John Wiley & Sons. pp. 49–50. ISBN 9780470641859.
- Knuth, Donald (1997). The Art of Computer Programming: Fundamental Algorithms, Volume 1 (3rd ed.). Reading, Massachusetts: Addison–Wesley. ISBN 0-201-89683-4. Retrieved 1 October 2012.
- Bienaymé LJ (1853) Considérations àl'appui de la découverte de Laplace. Comptes Rendus de l'Académie des Sciences 37: 309–324
- Tchebichef, P. (1867). "Des valeurs moyennes". Journal de mathématiques pures et appliquées. 2 12: 177–184.
- Markov A (1884) On certain applications of algebraic continued fractions, Ph.D. thesis, St Petersburg
- Steliga, Katarzyna; Szynal, Dominik (2010). "On Markov-Type Inequalities". International Journal of Pure and Applied Mathematics 58 (2): 137–152. ISSN 1311-8080. Retrieved 10 October 2012.
- Ferentinos K (1982) "On Tchebycheﬀ type inequalities". Trabajos Estadıst Investigacion Oper, 33: 125–132
- Berge PO (1938) A note on a form of Tchebycheff's theorem for two variables. Biometrika 29, 405–406
- Lal DN (1955) A note on a form of Tchebycheﬀ's inequality for two or more variables. Sankhya 15(3):317–320
- Isii K (1959) On a method for generalizations of Tchebycheff's inequality. Ann Inst Stat Math 10: 65–88
- Birnbaum, Z. W.; Raymond, J.; Zuckerman, H. S. (1947). "A Generalization of Tshebyshev's Inequality to Two Dimensions". The Annals of Mathematical Statistics 18 (1): 70–79. doi:10.1214/aoms/1177730493. ISSN 0003-4851. MR 19849. Zbl 0032.03402. Retrieved 7 October 2012.
- Kotz, Samuel; Balakrishnan, N.; Johnson, Norman L. (2000). Continuous Multivariate Distributions, Volume 1, Models and Applications (2nd ed.). Boston [u.a.]: Houghton Mifflin. ISBN 978-0-471-18387-7. Retrieved 7 October 2012.
- Olkin, Ingram; Pratt, John W. (1958). "A Multivariate Tchebycheff Inequality". The Annals of Mathematical Statistics 29 (1): 226–234. doi:10.1214/aoms/1177706720. MR MR93865. Zbl 0085.35204. Retrieved 2 October 2012.
- Godwin HJ (1964) Inequalities on distribution functions. New York, Hafner Pub Co
- Xinjia Chen (2007). "A New Generalization of Chebyshev Inequality for Random Vectors". arXiv:0707.0805v2.
- Section 2.1
- Baranoski, Gladimir V. G.; Rokne, Jon G.; Guangwu Xu (15 May 2001). "Applying the exponential Chebyshev inequality to the nondeterministic computation of form factors". Journal of Quantitative Spectroscopy and Radiative Transfer 69 (4): 199–200. doi:10.1016/S0022-4073(00)00095-9. Retrieved 2 October 2012. (the references for this article are corrected by Baranoski, Gladimir V. G.; Rokne, Jon G.; Guangwu Xu (15 January 2002). "Corrigendum to: 'Applying the exponential Chebyshev inequality to the nondeterministic computation of form factors'". Journal of Quantitative Spectroscopy and Radiative Transfer 72 (2): 199–200. doi:10.1016/S0022-4073(01)00171-6. Retrieved 2 October 2012.)
- Dufour (2003) Properties of moments of random variables
- Niemitalo O (2012) One-sided Chebyshev-type inequalities for bounded probability distributions.
- Saw, John G.; Mark C. K. Yang; Tse Chin Mo (1984). "Chebyshev Inequality with Estimated Mean and Variance". The American Statistician 38 (2): 130–132. Retrieved 1 October 2012. Unknown parameter
- Kabán, A. (2012). "Non-parametric detection of meaningless distances in high dimensional data". Statistics and Computing 22 (2): 375–385. doi:10.1007/s11222-011-9229-0. Retrieved 1 October 2012.
- Konijn, Hendrik S. (February 1987). "Distribution-Free and Other Prediction Intervals". The American Statistician (American Statistical Association) 41 (1): 11–15. JSTOR 2684311.
- Beasley, T. Mark; Page, Grier P.; Brand, Jaap P. L.; Gadbury, Gary L.; Mountz, John D.; Allison, David B. (January 2004). "Chebyshev's inequality for nonparametric testing with small N and α in microarray research". Journal of the Royal Statistical Society. C (Applied Statistics) 53 (1): 95–108. doi:10.1111/j.1467-9876.2004.00428.x. ISSN 1467-9876. Retrieved 3 October 2012.
- Ion, Roxana Alice (2001). "Chapter 4: Sharp Chebyshev-type inequalities". Nonparametric Statistical Process Control. Universiteit van Amsterdam. ISBN 9057760762. Retrieved 1 October 2012.
- Berck, Peter; Hihn, Jairus M. (May 1982). "Using the Semivariance to Estimate Safety-First Rules". American Journal of Agricultural Economics (Oxford University Press) 64 (2): 298–300. doi:10.2307/1241139. ISSN 0002-9092. Retrieved 8 October 2012.
- Nantell, Timothy J.; Price, Barbara (June 1979). "An Analytical Comparison of Variance and Semivariance Capital Market Theories". The Journal of Financial and Quantitative Analysis (University of Washington School of Business Administration) 14 (2): 221–242. doi:10.2307/2330500. JSTOR 2330500. Retrieved 8 October 2012.
- Neave EH, Ross MN, Yang J (2008) Distinguishing upside potential from downside risk. Management Research News. 32 (1) 26–36
- Selberg, Henrik L. (1940). "Zwei Ungleichungen sur Ergänzung des Tchebycheffschen Lemmas" [Two Inequalities Supplementing the Tchebycheff Lemma]. Skandinavisk Aktuarietidskrift (Scandinavian Actuarial Journal) (in German) 1940 (3–4): 121–125. doi:10.1080/03461238.1940.10404804. ISSN 0346-1238. OCLC 610399869. Retrieved 7 October 2012.
- Conlon, J.; Dulá, J. H. A geometric derivation and interpretation of Tchebyscheff's Inequality. Retrieved 2 October 2012.
- Cantelli F (1910)Intorno ad un teorema fondamentale della teoria del rischio. Bolletino dell Associazione degli Attuari Italiani
- Grimmett and Stirzaker, problem 7.11.9. Several proofs of this result can be found in Chebyshev's Inequalities by A. G. McDowell.
- Bhattacharyya, B. B. (27 June 2007). "One-sided chebyshev inequality when the first four moments are known". Communications in Statistics - Theory and Methods 16 (9): 2789–2791. doi:10.1080/03610928708829540. ISSN 0361-0926. Retrieved 6 October 2012.
- Mitzenmacher, Michael; Upfal, Eli (January 2005). Probability and Computing: Randomized Algorithms and Probabilistic Analysis (Repr. ed.). Cambridge [u.a.]: Cambridge Univ. Press. ISBN 9780521835404. Retrieved 6 October 2012.
- Zelen M (1954) Bounds on a distribution function that are functions of moments to order four. J Res Nat Bur Stand 53:377-381
- He S, Zhang J, Zhang S (2010) Bounding probability of small deviation: A fourth moment approach. Mathematics of operations research 35 (1) 208-232. doi: 10.1287/moor.1090.0438
- van Zuijlen Martien CA (2011) On a conjecture concerning the sum of independent Rademacher random variables. http://arxiv.org/abs/1112.4988
- Feller, William (1966). An Introduction to Probability Theory and Its Applications, Volume 2 (2 ed.). Wiley. p. 155. Retrieved 6 October 2012.
- Hartigan J.A., Hartigan P.M. (1985) "The dip test of unimodality". Annals of Statistics 13 (1) 70–84 doi:10.1214/aos/1176346577 MR773153
- Gauss C.F. Theoria Combinationis Observationum Erroribus Minimis Obnoxiae. Pars Prior. Pars Posterior. Supplementum. Theory of the Combination of Observations Least Subject to Errors. Part One. Part Two. Supplement. 1995. Translated by G.W. Stewart. Classics in Applied Mathematics Series, Society for Industrial and Applied Mathematics, Philadelphia
- Winkler A (1886) Math-Natur theorie Kl. Akad. Wiss Wien Zweite Abt 53, 6-41
- Dharmadhikari SW, Joag-Dev K(1985) The Gauss–Tchebyshev inequality for unimodal distributions. Teor Veroyatnost i Primenen 30(4) 817–820
- Clarkson, E.; Denny, J. L.; Shepp, L. (2009). "ROC and the bounds on tail probabilities via theorems of Dubins and F. Riesz". The Annals of Applied Probability 19 (1): 467–476. doi:10.1214/08-AAP536. PMC 2828638. PMID 20191100.
- McWilliams T.P. (1990) "A distribution-free test for symmetry based on a runs statistic".Journal of the American Statistical Association 85 (412)1130–1133 JSTOR 2289611
- Seaman, John W., Jr.; Young, Dean M.; Odell, Patrick. L. (1987). "Improving small sample variance estimators for bounded random variables". Industrial Mathematics 37: 65–75. ISSN 0019-8528. Zbl 0637.62024.
- Bickel, Peter J.; Krieger, Abba M. (1992). "Extensions of Chebyshev's Inequality with Applications". Probability and Mathematical Statistics (in English) (Wydawnictwo Uniwersytetu Wrocławskiego) 13 (2): 293–310. ISSN 0208-4147. Retrieved 6 October 2012.
- DasGupta A (2000) Best constants in Chebychev inequalities with various applications. Metrika 5 (1) 185–200
- "More thoughts on a one tailed version of Chebyshev's inequality – by Henry Bottomley". se16.info. Retrieved 2012-06-12.
- Godwin HJ (1964) Inequalities on distribution functions. (Chapter 3) New York, Hafner Pub Co
- Lesley FD, Rotar VI (2003) Some remarks on lower bounds of Chebyshev's type for half-lines. J Inequalities Pure Appl Math 4 (5) Art 96
- Fink, A. M.; Jodeit, Max, Jr. (1984). "On Chebyshev's Other Inequality". In Tong, Y. L.; Gupta, Shanti S. Institute of Mathematical Statistics Lecture Notes - Monograph Series. Inequalities in Statistics and Probability: Proceedings of the Symposium on Inequalities in Statistics and Probability, October 27–30, 1982, Lincoln, Nebraska 5: 115–120. doi:10.1214/lnms/1215465637. ISBN 0-940600-04-8. MR 789242. Retrieved 7 October 2012.
- Niculescu, Constantin P. (2001). "An extension of Chebyshev's inequality and its connection with Jensen's inequality". Journal of Inequalities and Applications 6 (4): 451–462. doi:10.1155/S1025583401000273. ISSN 1025-5834. Retrieved 6 October 2012.
- Niculescu, Constantin P.; Pečarić, Josip (2010). "The Equivalence of Chebyshev's Inequality to the Hermite–Hadamard Inequality". Mathematical Reports (in English) (Publishing House of the Romanian Academy) 12 (62): 145–156. ISSN 1582-3067. Retrieved 6 October 2012.
- Malamud, S. M. (15 February 2001). "Some complements to the Jensen and Chebyshev inequalities and a problem of W. Walter". Proceedings of the American Mathematical Society 129 (9): 2671–2678. doi:10.1090/S0002-9939-01-05849-X. ISSN 0002-9939. MR 1838791. Retrieved 7 October 2012.
- Haldane JB (1952) Simple tests for bimodality and bitangentiality. Annals of Eugenics 16(4):359–364 doi:10.1111/j.1469-1809.1951.tb02488.x
- Kendall MG (1943) The Advanced Theory of Statistics, 1. London
- Chernoff, H. (1952). "A Measure of Asymptotic Efficiency for Tests of a Hypothesis Based on the sum of Observations". Annals of Mathematical Statistics 23 (4): 493–507. doi:10.1214/aoms/1177729330. JSTOR 2236576. MR 57518. Zbl 0048.11804.
- (PDF) Calculating Upper Confidence Limits for Exposure Point Concentrations at hazardous Waste Sites (Report). Office of Emergency and Remedial Response of the U.S. Environmental Protection Agency. December 2002. http://www.epa.gov/oswer/riskassessment/pdf/ucl.pdf. Retrieved 1 October 2012.
Further reading 
- A. Papoulis (1991), Probability, Random Variables, and Stochastic Processes, 3rd ed. McGraw–Hill. ISBN 0-07-100870-5. pp. 113–114.
- G. Grimmett and D. Stirzaker (2001), Probability and Random Processes, 3rd ed. Oxford. ISBN 0-19-857222-0. Section 7.3.
|Wikimedia Commons has media related to: Chebyshev's inequality|