Neyman–Pearson lemma

From Wikipedia, the free encyclopedia
  (Redirected from Neyman-Pearson lemma)
Jump to: navigation, search

In statistics, the Neyman–Pearson lemma, named after Jerzy Neyman and Egon Pearson, states that when performing a hypothesis test between two simple hypotheses H0θ = θ0 and H1θ = θ1, the likelihood-ratio test which rejects H0 in favour of H1 when

\Lambda(x)=\frac{ L( \theta _0 \mid x)}{ L (\theta _1 \mid x)} \leq \eta


P(\Lambda(X)\leq \eta\mid H_0)=\alpha

is the most powerful test at significance level α for a threshold η. If the test is most powerful for all \theta_1 \in \Theta_1, it is said to be uniformly most powerful (UMP) for alternatives in the set \Theta_1 \, .

In practice, the likelihood ratio is often used directly to construct tests — see Likelihood-ratio test. However it can also be used to suggest particular test-statistics that might be of interest or to suggest simplified tests — for this, one considers algebraic manipulation of the ratio to see if there are key statistics in it related to the size of the ratio (i.e. whether a large statistic corresponds to a small ratio or to a large one).


Define the rejection region of the null hypothesis for the NP test as

R_{NP}=\left\{ x: \frac{L(\theta_{0}|x)}{L(\theta_{1}|x)} \leq \eta\right\}

where \eta is chosen so that P(R_{NP},\theta_0)=\alpha\,.

Any other test will have a different rejection region that we define as R_A. Furthermore, define the probability of the data falling in region R, given parameter \theta as

P(R,\theta)=\int_R L(\theta|x)\, dx,

For the test with critical region R_A to have level \alpha, it must be true that \alpha \ge P(R_A, \theta_0) , hence

\alpha= P(R_{NP}, \theta_0) \ge P(R_A, \theta_0) \,.

It will be useful to break these down into integrals over distinct regions:

P(R_{NP},\theta) = P(R_{NP} \cap R_A, \theta) + P(R_{NP} \cap R_A^c, \theta),


P(R_A,\theta) = P(R_{NP} \cap R_A, \theta) + P(R_{NP}^c \cap R_A, \theta).

Setting \theta=\theta_0, these two expressions and the above inequality yield that

P(R_{NP} \cap R_A^c, \theta_0) \ge  P(R_{NP}^c \cap R_A, \theta_0).

Comparing the powers of the two tests, P(R_{NP},\theta_1) and P(R_A,\theta_1), one can see that

P(R_{NP},\theta_1) \geq P(R_A,\theta_1) \iff
P(R_{NP} \cap R_A^c, \theta_1) \geq P(R_{NP}^c \cap R_A, \theta_1).

Now by the definition of R_{NP},

 P(R_{NP} \cap R_A^c, \theta_1)= \int_{R_{NP}\cap R_A^c} L(\theta_{1}|x)\,dx \geq \frac{1}{\eta} \int_{R_{NP}\cap R_A^c} L(\theta_0|x)\,dx = \frac{1}{\eta}P(R_{NP} \cap R_A^c, \theta_0)
 \ge \frac{1}{\eta}P(R_{NP}^c \cap R_A, \theta_0) = \frac{1}{\eta}\int_{R_{NP}^c \cap R_A} L(\theta_{0}|x)\,dx \geq \int_{R_{NP}^c\cap R_A} L(\theta_{1}|x)dx  = P(R_{NP}^c \cap R_A, \theta_1).

Hence the inequality holds.


Let X_1,\dots,X_n be a random sample from the \mathcal{N}(\mu,\sigma^2) distribution where the mean \mu is known, and suppose that we wish to test for H_0:\sigma^2=\sigma_0^2 against H_1:\sigma^2=\sigma_1^2. The likelihood for this set of normally distributed data is

L\left(\sigma^2;\mathbf{x}\right)\propto \left(\sigma^2\right)^{-n/2} \exp\left\{-\frac{\sum_{i=1}^n \left(x_i-\mu\right)^2}{2\sigma^2}\right\}.

We can compute the likelihood ratio to find the key statistic in this test and its effect on the test's outcome:

\Lambda(\mathbf{x}) = \frac{L\left(\sigma_0^2;\mathbf{x}\right)}{L\left(\sigma_1^2;\mathbf{x}\right)} =
\left(\frac{\sigma_0^2}{\sigma_1^2}\right)^{-n/2}\exp\left\{-\frac{1}{2}(\sigma_0^{-2}-\sigma_1^{-2})\sum_{i=1}^n \left(x_i-\mu\right)^2\right\}.

This ratio only depends on the data through \sum_{i=1}^n \left(x_i-\mu\right)^2. Therefore, by the Neyman–Pearson lemma, the most powerful test of this type of hypothesis for this data will depend only on \sum_{i=1}^n \left(x_i-\mu\right)^2. Also, by inspection, we can see that if \sigma_1^2>\sigma_0^2, then \Lambda(\mathbf{x}) is a decreasing function of \sum_{i=1}^n \left(x_i-\mu\right)^2. So we should reject H_0 if \sum_{i=1}^n \left(x_i-\mu\right)^2 is sufficiently large. The rejection threshold depends on the size of the test. In this example, the test statistic can be shown to be a scaled Chi-square distributed random variable and an exact critical value can be obtained.

Application in economics[edit]

A variant of the Neyman–Pearson lemma has found an application in the seemingly-unrelated domain of economy with land. One of the fundamental problems in consumer theory is calculating the demand function of the consumer given the prices. In particular, given a heterogeneous land-estate, a price measure over the land, and a subjective utility measure over the land, the consumer's problem is to calculate the best land parcel that he can buy – i.e, the land parcel with the largest utility, whose price is at most his budget. It turns out that this problem is very similar to the problem of finding the most powerful stastistical test, and so the Neyman–Pearson lemma can be used.[1]

See also[edit]


External links[edit]

  • Cosma Shalizi, a professor of statistics at Carnegie Mellon University, gives an intuitive derivation of the Neyman–Pearson Lemma using ideas from economics
  1. ^ Berliant, M. (1984). "A characterization of the demand for land". Journal of Economic Theory 33 (2): 289. doi:10.1016/0022-0531(84)90091-7.