Skip to content
Surf Wiki
Save to docs
general/nonparametric-regression

From Surf Wiki (app.surf) — the open knowledge base

Kernel regression

Technique in statistics

Kernel regression

Technique in statistics

In statistics, kernel regression is a non-parametric technique to estimate the conditional expectation of a random variable. The objective is to find a non-linear relation between a pair of random variables X and Y.

In any nonparametric regression, the conditional expectation of a variable Y relative to a variable X may be written:

: \operatorname{E}(Y \mid X) = m(X)

where m is an unknown function.

Nadaraya–Watson kernel regression

Nadaraya and Watson, both in 1964, proposed to estimate m as a locally weighted average, using a kernel as a weighting function.{{cite journal

: \widehat{m}h(x)=\frac{\sum{i=1}^n K_h(x-x_i) y_i}{\sum_{i=1}^nK_h(x-x_i)}

where K_h(t) = \frac{1}{h}K\left(\frac{t}{h}\right) is a kernel with a bandwidth h such that K(\cdot) is of order at least 1, that is \int_{-\infty}^\infty u K(u) , du = 0.

Derivation

Starting with the definition of conditional expectation, : \operatorname{E}(Y \mid X=x) = \int y f(y\mid x) , dy = \int y \frac{f(x,y)}{f(x)} , dy

we estimate the joint distributions f(x,y) and f(x) using kernel density estimation with a kernel K:

: \hat{f}(x,y) = \frac{1}{n}\sum_{i=1}^n K_h(x-x_i) K_h(y-y_i), : \hat{f}(x) = \frac{1}{n} \sum_{i=1}^n K_h(x-x_i),

We get:

: \begin{align} \operatorname{\hat E}(Y \mid X=x) &= \int y \frac{\hat{f}(x,y)}{\hat{f}(x)} , dy,\[6pt] &= \int y \frac{\sum_{i=1}^n K_h(x-x_i) K_h(y-y_i)}{\sum_{j=1}^n K_h(x-x_j)} ,dy,\[6pt] &= \frac{\sum_{i=1}^n K_h(x-x_i) \int y , K_h(y-y_i) ,dy}{\sum_{j=1}^n K_h(x-x_j)},\[6pt] &= \frac{\sum_{i=1}^n K_h(x-x_i) y_i}{\sum_{j=1}^n K_h(x-x_j)}, \end{align}

which is the Nadaraya–Watson estimator.

Priestley–Chao kernel estimator

: \widehat{m}{PC}(x) = h^{-1} \sum{i=2}^n (x_i - x_{i-1}) K\left(\frac{x-x_i}{h}\right) y_i

where h is the bandwidth (or smoothing parameter).

Gasser–Müller kernel estimator

: \widehat{m}{GM}(x) = h^{-1} \sum{i=1}^n \left[\int_{s_{i-1}}^{s_i} K\left(\frac{x-u}{h}\right) , du\right] y_i

where s_i = \frac{x_{i-1} + x_i}{2}.{{cite book

Example

Estimated regression function.

This example is based upon Canadian cross-section wage data consisting of a random sample taken from the 1971 Canadian Census Public Use Tapes for male individuals having common education (grade 13). There are 205 observations in total.

The figure to the right shows the estimated regression function using a second order Gaussian kernel along with asymptotic variability bounds.

Script for example

The following commands of the R programming language use the npreg() function to deliver optimal smoothing and to create the figure given above. These commands can be entered at the command prompt via cut and paste.

install.packages("np")
library(np) # non parametric library
data(cps71)
attach(cps71)

m <- npreg(logwage~age)

plot(m, plot.errors.method="asymptotic",
     plot.errors.style="band",
     ylim=c(11, 15.2))

points(age, logwage, cex=.25)
detach(cps71)

Statistical implementation

  • GNU Octave mathematical program package
  • Julia: KernelEstimator.jl
  • MATLAB: A free MATLAB toolbox with implementation of kernel regression, kernel density estimation, kernel estimation of hazard function and many others is available on these pages (this toolbox is a part of the book ).
  • Python: the [KernelReg](http://www.statsmodels.org/stable/generated/statsmodels.nonparametric.kernel_regression.KernelReg.html) class for mixed data types in the [statsmodels.nonparametric](http://www.statsmodels.org/stable/nonparametric.html) sub-package (includes other kernel density related classes), the package kernel_regression as an extension of scikit-learn (inefficient memory-wise, useful only for small datasets)
  • R: the function npreg of the np package can perform kernel regression.
  • Stata: npregress, kernreg2

References

References

  1. Watson, G. S.. (1964). "Smooth regression analysis". Sankhyā: The Indian Journal of Statistics, Series A.
  2. Bierens, Herman J.. (1994). "Topics in Advanced Econometrics". Cambridge University Press.
  3. Salsburg, D.. (2002). "[[The Lady Tasting Tea". W.H. Freeman.
  4. (2012). "Kernel Smoothing in MATLAB: Theory and Practice of Kernel Smoothing". World Scientific Publishing.
  5. [https://cran.r-project.org/web/packages/np/index.html ''np'': Nonparametric kernel smoothing methods for mixed data types]
  6. (2014). "Nonparametric Statistical Methods Using R". CRC Press.
Info: Wikipedia Source

This article was imported from Wikipedia and is available under the Creative Commons Attribution-ShareAlike 4.0 License. Content has been adapted to SurfDoc format. Original contributors can be found on the article history page.

Want to explore this topic further?

Ask Mako anything about Kernel regression — get instant answers, deeper analysis, and related topics.

Research with Mako

Free with your Surf account

Content sourced from Wikipedia, available under CC BY-SA 4.0.

This content may have been generated or modified by AI. CloudSurf Software LLC is not responsible for the accuracy, completeness, or reliability of AI-generated content. Always verify important information from primary sources.

Report