Coefficient of variation
In probability theory and statistics, the coefficient of variation (CV) is a normalized measure of dispersion of a probability distribution. It is also known as unitized risk or the variation coefficient. The absolute value of the CV is sometimes known as relative standard deviation (RSD), which is expressed as a percentage.
Definition
The coefficient of variation (CV) is defined as the ratio of the standard deviation to the mean :
which is the inverse of the signal-to-noise ratio. It shows the extent of variability in relation to mean of the population.
The coefficient of variation should be computed only for data measured on a ratio scale, as these are measurements that can only take non-negative values. The coefficient of variation may not have any meaning for data on an interval scale.[1] For example, most temperature scales are interval scales (e.g., Celsius, Fahrenheit etc.), they can take both positive and negative values. The Kelvin scale has an absolute null value, and no negative values can naturally occur. Hence, the Kelvin scale is a ratio scale. While the standard deviation (SD) can be derived on both the Kelvin and the Celsius scale (with both leading to the same SDs), the CV is only relevant as a measure of relative variability for the Kelvin scale.
Often, laboratory values that are measured based on chromatographic methods are log-normally distributed. In this case, the CV would be constant over a large range of measurements, while SDs would vary depending on typical values that are being measured.
Estimation
When only a sample of data from a population is available, the population CV can be estimated using the ratio of the sample standard deviation to the sample mean :
But this estimator, when applied to a small or moderately sized sample, tends to be too low: it is a biased estimator. For normally distributed data, an unbiased estimator[2] for a sample of size n is:
If it is assumed that the data are log-normally distributed, a more accurate estimate, derived from the properties of the log-normal distribution,[3][4][5] is defined as:
where is the sample standard deviation of the data after a natural log transformation. (In the event that measurements are recorded using any other logarithmic base, b, their standard deviation is converted to base e using , and the formula for remains the same.[6]) This estimate is sometimes referred to as the “geometric coefficient of variation”[7] in order to distinguish it from the simple estimate above. However, "geometric coefficient of variation" has also been defined[8] as:
This term was intended to be analogous to the coefficient of variation, for describing multiplicative variation in log-normal data, but this definition of GCV has no theoretical basis as an estimate of itself.
For many practical purposes (such as sample size determination and calculation of confidence intervals) it is which is of most use in the context of log-normally distributed data. If necessary, this can be derived from an estimate of or GCV by inverting the corresponding formula.
Comparison to standard deviation
Advantages
The coefficient of variation is useful because the standard deviation of data must always be understood in the context of the mean of the data. In contrast, the actual value of the CV is independent of the unit in which the measurement has been taken, so it is a dimensionless number. For comparison between data sets with different units or widely different means, one should use the coefficient of variation instead of the standard deviation.
Disadvantages
- When the mean value is close to zero, the coefficient of variation will approach infinity and is therefore sensitive to small changes in the mean. This is often the case if the values do not originate from a ratio scale.
- Unlike the standard deviation, it cannot be used directly to construct confidence intervals for the mean.
Applications
The coefficient of variation is also common in applied probability fields such as renewal theory, queueing theory, and reliability theory. In these fields, the exponential distribution is often more important than the normal distribution. The standard deviation of an exponential distribution is equal to its mean, so its coefficient of variation is equal to 1. Distributions with CV < 1 (such as an Erlang distribution) are considered low-variance, while those with CV > 1 (such as a hyper-exponential distribution) are considered high-variance. Some formulas in these fields are expressed using the squared coefficient of variation, often abbreviated SCV. In modeling, a variation of the CV is the CV(RMSD). Essentially the CV(RMSD) replaces the standard deviation term with the Root Mean Square Deviation (RMSD).
Distribution
Provided that negative and small positive values of the sample mean occur with negligible frequency, the probability distribution of the coefficient of variation for a sample of size n has been shown by Hendricks and Robey[9] to be
where the symbol indicates that the summation is over only even values of n-1-i, i.e, if n is odd, sum over even values of i and if n is even, sum only over odd values of i.
This is useful, for instance, in the construction of hypothesis tests or confidence intervals.
Similar ratios
Standardized moments are similar ratios, , which are also dimensionless and scale invariant. The variance-to-mean ratio, , is another similar ratio, but is not dimensionless, and hence not scale invariant. See Normalization (statistics) for further ratios.
In signal processing, particularly image processing, the reciprocal ratio is referred to as the signal to noise ratio.
- Relative standard deviation,
- Standardized moment,
- Variance-to-mean ratio,
- Fano factor, (windowed VMR)
- Signal-to-noise ratio, (in signal processing)
This article includes a list of general references, but it lacks sufficient corresponding inline citations. (November 2010) |
References
- ^ "What is the difference between ordinal, interval and ratio variables? Why should I care?". GraphPad Software Inc. Retrieved 2008-02-22.
- ^ Sokal RR & Rohlf FJ. Biometry (3rd Ed). New York: Freeman, 1995. p. 58. ISBN 0-7167-2411-1
- ^ Koopmans, L. H.; Owen, D. B.; Rosenblatt, J. I. (1964). "Confidence intervals for the coefficient of variation for the normal and log normal distributions". Biometrika. 51: 25. doi:10.1093/biomet/51.1-2.25.
- ^ Diletti, E; Hauschke, D; Steinijans, VW (1992). "Sample size determination for bioequivalence assessment by means of confidence intervals". International journal of clinical pharmacology, therapy, and toxicology. 30 Suppl 1: S51–8. PMID 1601532.
- ^ Julious, Steven A.; Debarnot, Camille A. M. (2000). "Why Are Pharmacokinetic Data Summarized by Arithmetic Means?". Journal of Biopharmaceutical Statistics. 10 (1): 55–71. doi:10.1081/BIP-100101013. PMID 10709801.
- ^ Reed JF, Lynn F, Meade BD. (2002) "Use of Coefficient of Variation in Assessing Variability of Quantitative Assays". Clin Diagn Lab Immunol. 9(6): 1235–1239. doi:10.1128/CDLI.9.6.1235-1239.2002
- ^ Sawant,S.; Mohan, N. (2011) "FAQ: Issues with Efficacy Analysis of Clinical Trial Data Using SAS", PharmaSUG2011, Paper PO08
- ^ Kirkwood, TBL (1979). "Geometric means and measures of dispersion". Biometrics. 35: 908–9.
- ^ Hendricks, Walter A.; Robey, Kate W. (1936). "The Sampling Distribution of the Coefficient of Variation". The Annals of Mathematical Statistics. 7 (3): 129–32. doi:10.1214/aoms/1177732503. JSTOR 2957564.