Jump to content

De Moivre–Laplace theorem

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by 195.194.120.132 (talk) at 13:50, 14 December 2015 (Proof). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Within a system whose bins are filled according to the binomial distribution (such as Galton's "bean machine", shown here), given a sufficient number of trials (here the rows of pins, each of which causes a dropped "bean" to fall toward the left or right), a shape representing the probability distribution of k successes in n trials (see bottom of Fig. 7) matches approximately the Gaussian distribution with mean np and variance np(1−p), assuming the trials are independent and successes occur with probability p.
Consider tossing a set of n coins a very large number of times and counting the number of "heads" that result each time. The possible number of heads on each toss, k, runs from 0 to n along the horizontal axis, while the vertical axis represents the relative frequency of occurrence of the outcome k heads. The height of each dot is thus the probability of observing k heads when tossing n coins (a binomial distribution based on n trials). According to the de Moivre–Laplace theorem, as n grows large, the shape of the discrete distribution converges to the continuous Gaussian curve of the normal distribution.

In probability theory, the de Moivre–Laplace theorem, which is a special case of the central limit theorem, states that the normal distribution may be used as an approximation to the binomial distribution under certain conditions. In particular, the theorem shows that the probability mass function of the random number of "successes" observed in a series of n independent Bernoulli trials, each having probability p of success (a binomial distribution with n trials), converges to the probability density function of the normal distribution with mean np and standard deviation np(1-p), as n grows large, assuming p is not 0 or 1.

The theorem appeared in the second edition of The Doctrine of Chances by Abraham de Moivre, published in 1738. Although de Moivre did not use the term "Bernoulli trials", he wrote about the probability distribution of the number of times "heads" appears when a coin is tossed 3600 times.[1]

This is one derivation of the particular Gaussian function used in the normal distribution.

Theorem

As n grows large, for k in the neighborhood of np we can approximate[2][3]

in the sense that the ratio of the left-hand side to the right-hand side converges to 1 as n → ∞.

Proof

Note that k cannot be fixed or it would quickly fall outside the range of interest as n → ∞. What is needed is to let k vary but always be a fixed number of standard deviations from the mean, so that it is always associated with the same point on the standard normal distribution. We can do this by defining

for some fixed x. Then, for example, when x = 1, k will always be 1 standard deviation from the mean. From this definition we have the approximations knp and as n → ∞.

However, the left-hand side requires that k be an integer. Keeping the notation but assuming that k is the nearest integer given by the definition, this is seen to be inconsequential in the limit by noting that as n → ∞ the change in x required to make k an integer becomes small and successive integer values of k produce converging values on the right-hand side:

The proof consists of transforming the left-hand side to the right-hand side by three approximations.

First, according to Stirling's formula, we can replace the factorial of a large number n with the approximation:

Thus

Next, use the approximation to match the root above to the desired root on the right-hand side.

Finally, rewrite the expression as an exponential and use the Taylor Series approximation for ln(1+x):

Then

See also

  • Poisson distribution is an alternative approximation of the binomial distribution for large values of n.

Notes

  1. ^ Walker, Helen M (1985). "De Moivre on the law of normal probability". In Smith, David Eugene (ed.). A source book in mathematics. Dover. p. 78. ISBN 0-486-64690-4. But altho' the taking an infinite number of Experiments be not practicable, yet the preceding Conclusions may very well be applied to finite numbers, provided they be great, for Instance, if 3600 Experiments be taken, make n = 3600, hence ½n will be = 1800, and ½√n 30, then the Probability of the Event's neither appearing oftner than 1830 times, nor more rarely than 1770, will be 0.682688. {{cite book}}: External link in |chapterurl= (help); Unknown parameter |chapterurl= ignored (|chapter-url= suggested) (help)
  2. ^ Papoulis, Pillai, "Probability, Random Variables, and Stochastic Processes", 4th Edition
  3. ^ Feller, W. (1968) An Introduction to Probability Theory and Its Applications (Volume 1). Wiley. ISBN 0-471-25708-7. Section VII.3