Basu's theorem

From Wikipedia, the free encyclopedia
Jump to: navigation, search

In statistics, Basu's theorem states that any boundedly complete sufficient statistic is independent of any ancillary statistic. This is a 1955 result of Debabrata Basu.[1]

It is often used in statistics as a tool to prove independence of two statistics, by first demonstrating one is complete sufficient and the other is ancillary, then appealing to the theorem.[citation needed] An example of this is to show that the sample mean and sample variance of a normal distribution are independent statistics, which is done in the Examples section below. This property (independence of sample mean and sample variance) characterizes normal distributions.

Statement[edit]

Let Pθ be a family of distributions on a measurable space (X, Σ). Then if T is a boundedly complete sufficient statistic for θ, and A is ancillary to θ, then T is independent of A.

Proof[edit]

Let PθT and PθA be the marginal distributions of T and A respectively.

P_\theta^A(B) = P_\theta (A^{-1} B) = \int_{T(X)} P_\theta(A^{-1}B | T=t) \  P_\theta^T (dt) \,

The PθA does not depend on θ because A is ancillary. Likewise, Pθ(·|T = t) does not depend on θ because T is sufficient. Therefore:

 \int_{T(X)} \big[ P(A^{-1}B | T=t) - P^A(B)  \big] \ P_\theta^T (dt) = 0 \,

Note the integrand (the function inside the intergal) is a function of t and not θ. Therefore, since T is boundedly complete:

P(A^{-1}B | T=t) = P^A(B) \quad \text{for all }t\,

Therefore, A is independent of T.

Example[edit]

Independence of sample mean and sample variance of a normal distribution[edit]

Let X1, X2, ..., Xn be independent, identically distributed normal random variables with mean μ and variance σ2.

Then with respect to the parameter μ, one can show that

\widehat{\mu}=\frac{\sum X_i}{n},\,

the sample mean, is a complete sufficient statistic – it is all the information one can derive to estimate μ, and no more – and

\widehat{\sigma}^2=\frac{\sum \left(X_i-\bar{X}\right)^2}{n-1},\,

the sample variance, is an ancillary statistic – its distribution does not depend on μ.

Therefore, from Basu's theorem it follows that these statistics are independent.

This independence result can also be proven by Cochran's theorem.

Further, this property (that the sample mean and sample variance of the normal distribution are independent) characterizes the normal distribution – no other distribution has this property.[2]

Notes[edit]

  1. ^ Basu (1955)
  2. ^ Geary, R.C. (1936). "The Distribution of the "Student's" Ratio for the Non-Normal Samples". Supplement to the Journal of the Royal Statistical Society 3 (2): 178–184. doi:10.2307/2983669. JFM 63.1090.03. JSTOR 2983669. 

References[edit]