Brown–Forsythe test

From Wikipedia, the free encyclopedia
(Redirected from Brown-Forsythe)

The Brown–Forsythe test is a statistical test for the equality of group variances based on performing an Analysis of Variance (ANOVA) on a transformation of the response variable. When a one-way ANOVA is performed, samples are assumed to have been drawn from distributions with equal variance. If this assumption is not valid, the resulting F-test is invalid. The Brown–Forsythe test statistic is the F statistic resulting from an ordinary one-way analysis of variance on the absolute deviations of the groups or treatments data from their individual medians.[1]

Transformation[edit]

The transformed response variable is constructed to measure the spread in each group. Let

where is the median of group j. The Brown–Forsythe test statistic is the model F statistic from a one way ANOVA on zij:

where p is the number of groups, nj is the number of observations in group j, and N is the total number of observations. Also are the group means of the and is the overall mean of the . This F-statistic follows the F-distribution with degrees of freedom and under the null hypothesis.

If the variances are indeed heterogeneous, techniques that allow for this (such as the Welch one-way ANOVA) may be used instead of the usual ANOVA.

Good, noting that the deviations are linearly dependent, has modified the test so as to drop the redundant deviations.[2]

Comparison with Levene's test[edit]

Levene's test uses the mean instead of the median. Although the optimal choice depends on the underlying distribution, the definition based on the median is recommended as the choice that provides good robustness against many types of non-normal data while retaining good statistical power.[3] If one has knowledge of the underlying distribution of the data, this may indicate using one of the other choices. Brown and Forsythe[4] performed Monte Carlo studies that indicated that using the trimmed mean performed best when the underlying data followed a Cauchy distribution (a heavy-tailed distribution) and the median performed best when the underlying data followed a χ2 distribution with four degrees of freedom (a sharply skewed distribution). Using the mean provided the best power for symmetric, moderate-tailed, distributions. O'Brien tested several ways of using the traditional analysis of variance to test heterogeneity of spread in factorial designs with equal or unequal sample sizes. The jackknife pseudovalues of s2 and the absolute deviations from the cell median are shown to be robust and relatively powerful.[5]

See also[edit]

References[edit]

  1. ^ "plot.hov function | R Documentation". www.rdocumentation.org. DataCamp.
  2. ^ Good, P. I. (2005). Permutation, Parametric, and Bootstrap Tests of Hypotheses (3rd ed.). New York: Springer.
  3. ^ Derrick, B; Ruck, A; Toher, D; White, P (2018). "Tests for equality of variances between two samples which contain both paired observations and independent observations" (PDF). Journal of Applied Quantitative Methods. 13 (2): 36–47.
  4. ^ Brown, Morton B.; Forsythe, Alan B. (1974). "Robust tests for the equality of variances". Journal of the American Statistical Association. 69 (346): 364–367. doi:10.1080/01621459.1974.10482955. JSTOR 2285659.
  5. ^ O'Brien, R. G. (1978). "Robust techniques for testing heterogeneity of variance effects in factorial designs". Psychometrika. 43 (3): 327–342. doi:10.1007/BF02293643.

External links[edit]

Public Domain This article incorporates public domain material from the National Institute of Standards and Technology