| Part of a series on |
| Regression analysis |
|---|
| Models |
| Estimation |
| Background |
Nonparametric regression is a form ofregression analysis where the predictor does not take a predetermined form but is completely constructed using information derived from the data. That is, noparametric equation is assumed for the relationship betweenpredictors and dependent variable. A largersample size is needed to build a nonparametric model having the same level ofuncertainty as aparametric model because the data must supply both the model structure and the parameter estimates.
Nonparametric regression assumes the following relationship, given the random variables and:
where is some deterministic function.Linear regression is a restricted case of nonparametric regression where is assumed to be a linear function of the data.Sometimes a slightly stronger assumption of additive noise is used:
where the random variable is the `noise term', with mean 0.Without the assumption that belongs to a specific parametric family of functions it is impossible to get an unbiased estimate for, however most estimators areconsistent under suitable conditions.
This is a non-exhaustive list of non-parametric models for regression.
In Gaussian process regression, also known as Kriging, a Gaussian prior is assumed for the regression curve. The errors are assumed to have amultivariate normal distribution and the regression curve is estimated by itsposterior mode. The Gaussian prior may depend on unknown hyperparameters, which are usually estimated viaempirical Bayes. The hyperparameters typically specify a prior covariance kernel. In case the kernel should also be inferred nonparametrically from the data, thecritical filter can be used.
Smoothing splines have an interpretation as the posterior mode of a Gaussian process regression.

This sectiondoes notcite anysources. Please helpimprove this section byadding citations to reliable sources. Unsourced material may be challenged andremoved.(August 2020) (Learn how and when to remove this message) |
Kernel regression estimates the continuous dependent variable from a limited set of data points byconvolving the data points' locations with akernel function—approximately speaking, the kernel function specifies how to "blur" the influence of the data points so that their values can be used to predict the value for nearby locations.
Decision tree learning algorithms can be applied to learn to predict a dependent variable from data.[2] Although the original Classification And Regression Tree (CART) formulation applied only to predicting univariate data, the framework can be used to predict multivariate data, including time series.[3]