Nonparametric regression

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by CBM (talk | contribs) at 19:05, 21 February 2010 (Removing stub tag(s)). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Nonparametric regression is a form of regression analysis in which the predictor does not take a predetermined form but is constructed according to information derived from the data. Nonparametric regression requires larger sample sizes than regression based on parametric models because the data must supply the model structure as well as the model estimates.

Kernel regression

Kernel regression estimates the continuous dependent variable from a limited set of data points by convolving the data points' locations with a kernel function - approximately speaking, the kernel function specifies how to "blur" the influence of the data points so that their values can be used to predict the value for nearby locations.


Nonparametric multiplicative regression

Nonparametric multiplicative regression (NPMR) is a form of nonparametric regression based on multiplicative kernel estimation. This is a smoothing technique that can be cross-validated and applied in a predictive way. Many other smoothing techniques are well known, for example smoothing splines and wavelets. Optimum choice of a smoothing method depends on the specific application. NPMR is useful for habitat modeling. The multidimensionality is provided multiplicatively – this automatically and parsimoniously models the complex interactions among predictors in much the same way that organisms integrate the numerous factors affecting their performance [1]. Optimizing the selection of predictors and their smoothing parameters in a multiplicative model is computationally intensive. NPMR can be applied to either presence-absence or quantitative response data, with either categorical or quantitative predictors.

NPMR can be applied with a local mean estimator, a local linear estimator, or a local logistic estimator. In each case the weights can be extended multiplicatively to m dimensions. In words, the estimate of the response is a local estimate (for example a local mean) of the observed values, each value weighted by its proximity to the target point in the predictor space, the weights being the product of weights for individual predictors. The model allows interactions, because weights for individual predictors are combined by multiplication rather than addition. A key biological feature of the model is that failure of a population with respect to any single dimension of the predictor space results in failure at that point, because the product of the weights for the point is zero or near zero if any of the individual weights are zero or near zero.

Regression trees

Decision tree learning algorithms can be applied to learn to predict a dependent variable from data[2]. Although the original CART formulation applied only to predicting univariate data, the framework can be used to predict multivariate data including time series[3].

See also

References

  1. ^ McCune, B. (2006). "Non-parametric habitat models with automatic interactions". Journal of Vegetation Science. 17: 819–830. doi:10.1658/1100-9233(2006)17[819:NHMWAI]2.0.CO;2.
  2. ^ Breiman, Leo (1984), Classification and regression trees, Monterey, CA: Wadsworth & Brooks/Cole Advanced Books & Software, ISBN 978-0412048418 {{citation}}: Unknown parameter |coauthors= ignored (|author= suggested) (help)
  3. ^ Segal, M.R. (1992), "Tree-structured methods for longitudinal data", Journal of the American Statistical Association, vol. 87, no. 418, pp. 407--418
  • McCune, B. and M. J. Mefford. 2004. HyperNiche. Nonparametric Multiplicative Habitat Modeling. MjM Software, Gleneden Beach, Oregon, U.S.A.

External links