# Median absolute deviation

For a broader coverage related to this topic, see Average absolute deviation.

In statistics, the median absolute deviation (MAD) is a robust measure of the variability of a univariate sample of quantitative data. It can also refer to the population parameter that is estimated by the MAD calculated from a sample.

For a univariate data set X1X2, ..., Xn, the MAD is defined as the median of the absolute deviations from the data's median:

${\displaystyle \operatorname {MAD} =\operatorname {median} \left(\ \left|X_{i}-\operatorname {median} (X)\right|\ \right),\,}$

that is, starting with the residuals (deviations) from the data's median, the MAD is the median of their absolute values.

## Example

Consider the data (1, 1, 2, 2, 4, 6, 9). It has a median value of 2. The absolute deviations about 2 are (1, 1, 0, 0, 2, 4, 7) which in turn have a median value of 1 (because the sorted absolute deviations are (0, 0, 1, 1, 2, 4, 7)). So the median absolute deviation for this data is 1.

## Uses

The median absolute deviation is a measure of statistical dispersion. Moreover, the MAD is a robust statistic, being more resilient to outliers in a data set than the standard deviation. In the standard deviation, the distances from the mean are squared, so large deviations are weighted more heavily, and thus outliers can heavily influence it. In the MAD, the deviations of a small number of outliers are irrelevant.

Because the MAD is a more robust estimator of scale than the sample variance or standard deviation, it works better with distributions without a mean or variance, such as the Cauchy distribution.

## Relation to standard deviation

In order to use the MAD as a consistent estimator for the estimation of the standard deviation σ, one takes

${\displaystyle {\hat {\sigma }}=k\cdot \operatorname {MAD} ,\,}$

where k is a constant scale factor, which depends on the distribution.[1]

For normally distributed data k is taken to be:

${\displaystyle k=1/\left(\Phi ^{-1}(3/4)\right)\approx 1.4826}$,

i.e., the reciprocal of the quantile function ${\displaystyle \Phi ^{-1}}$ (also known as the inverse of the cumulative distribution function) for the standard normal distribution Z = X/σ.[2][3] The argument 3/4 is such that ${\displaystyle \pm MAD}$ covers 50% (between 1/4 and 3/4) of the standard normal cumulative distribution function, i.e.:

${\displaystyle {\frac {1}{2}}=P(|X-\mu |\leq \operatorname {MAD} )=P\left(\left|{\frac {X-\mu }{\sigma }}\right|\leq {\frac {\operatorname {MAD} }{\sigma }}\right)=P\left(|Z|\leq {\frac {\operatorname {MAD} }{\sigma }}\right).}$

Therefore, we must have that:

${\displaystyle \Phi \left(\operatorname {MAD} /\sigma \right)-\Phi \left(-\operatorname {MAD} /\sigma \right)=1/2}$.

Noticing that:

${\displaystyle \Phi \left(-\operatorname {MAD} /\sigma \right)=1-\Phi \left(\operatorname {MAD} /\sigma \right)}$

we have that ${\displaystyle \operatorname {MAD} /\sigma =\Phi ^{-1}\left(3/4\right)=0.67449}$ from which we obtain the scale factor ${\displaystyle k=1/\Phi ^{-1}(3/4)=1.4826}$.

Another way of establishing the relationship is noting that MAD equals the half-normal distribution median:

${\displaystyle MAD=\sigma {\sqrt {2}}{\mbox{erf}}^{-1}(1/2)\approx 0.67449\sigma }$.

This form is used in, e.g., the probable error.

The population MAD is defined analogously to the sample MAD, but is based on the complete distribution rather than on a sample. For a symmetric distribution with zero mean, the population MAD is the 75th percentile of the distribution.

Unlike the variance, which may be infinite or undefined, the population MAD is always a finite number. For example, the standard Cauchy distribution has undefined variance, but its MAD is 1.

The earliest known mention of the concept of the MAD occurred in 1816, in a paper by Carl Friedrich Gauss on the determination of the accuracy of numerical observations.[4][5]

## Notes

1. ^ Rousseeuw, P. J.; Croux, C. (1993). "Alternatives to the median absolute deviation". Journal of the American Statistical Association. 88 (424): 1273–1283. doi:10.1080/01621459.1993.10476408.
2. ^ Ruppert, D. (2010). Statistics and Data Analysis for Financial Engineering. Springer. p. 118. ISBN 9781441977878. Retrieved 2015-08-27.
3. ^ Leys, C.; et al. (2013). "Detecting outliers: Do not use standard deviation around the mean, use absolute deviation around the median". Journal of Experimental Social Psychology. 49: 764–766. doi:10.1016/j.jesp.2013.03.013.
4. ^ Gauss, Carl Friedrich (1816). "Bestimmung der Genauigkeit der Beobachtungen". Zeitschrift für Astronomie und verwandte Wissenschaften. 1: 187–197.
5. ^ Walker, Helen (1931). Studies in the History of the Statistical Method. Baltimore, MD: Williams & Wilkins Co. pp. 24–25.

## References

• Hoaglin, David C.; Frederick Mosteller; John W. Tukey (1983). Understanding Robust and Exploratory Data Analysis. John Wiley & Sons. pp. 404–414. ISBN 0-471-09777-2.
• Russell, Roberta S.; Bernard W. Taylor III. (2006). Operations Management. John Wiley & Sons. pp. 497–498. ISBN 0-471-69209-3.
• Venables, W.N.; B.D. Ripley (1999). Modern Applied Statistics with S-PLUS. Springer. p. 128. ISBN 0-387-98825-4.