1992
DOI: 10.1080/00031305.1992.10475879
|View full text |Cite
|
Sign up to set email alerts
|

An Introduction to Kernel and Nearest-Neighbor Nonparametric Regression

Abstract: Nonparametric regressiOn 1s a set of techniques for estimating a regression curve without making strong assumptions about the shape of the true regression function. These techniques are therefore useful for building and checking parametric models, as well as for data description. Kernel and nearest neighbor regression estimators are local versions of univariate location estimators, and so they can readily be introduced to beginning students, and consulting clients who are familiar with such summaries as the sa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
803
0
11

Year Published

2015
2015
2021
2021

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 3,498 publications
(815 citation statements)
references
References 29 publications
1
803
0
11
Order By: Relevance
“…These algorithms were chosen based on their potential ability to determine the relationship between inputs (V WE , V AE , T ) and outputs ([SO 2 ]) without needing to know the functional form of the relationship itself. The methods examined were: ridge regression (RR), which attempts to reduce standard error by introducing bias to reduce multicollinearity among independent variables (Rifkin, 2007); least absolute shrinkage and selection operator (LASSO) regression, which similarly reduces covariance and overfitting by eliminating similar features and imposing an absolute limit on the sum of the coefficients (Tibshirani, 1996); classification and decision trees (CART), which forms a collection of rules based in a recursive fashion by selecting data that differentiate observations based on the dependent variable (Breiman et al, 1984); and k nearest neighbors regression (kNN), which estimates the regression curve without making assumptions about the structure of the model (Altman, 1992). The kNN approach, which was found to have the best performance (see Results, below), involves mapping input variables from the training data (V WE , V AE , T ) to the output variable (SO 2 mixing ratio) in an ndimensional vector space.…”
Section: Nonparametric Calibration Approachesmentioning
confidence: 99%
“…These algorithms were chosen based on their potential ability to determine the relationship between inputs (V WE , V AE , T ) and outputs ([SO 2 ]) without needing to know the functional form of the relationship itself. The methods examined were: ridge regression (RR), which attempts to reduce standard error by introducing bias to reduce multicollinearity among independent variables (Rifkin, 2007); least absolute shrinkage and selection operator (LASSO) regression, which similarly reduces covariance and overfitting by eliminating similar features and imposing an absolute limit on the sum of the coefficients (Tibshirani, 1996); classification and decision trees (CART), which forms a collection of rules based in a recursive fashion by selecting data that differentiate observations based on the dependent variable (Breiman et al, 1984); and k nearest neighbors regression (kNN), which estimates the regression curve without making assumptions about the structure of the model (Altman, 1992). The kNN approach, which was found to have the best performance (see Results, below), involves mapping input variables from the training data (V WE , V AE , T ) to the output variable (SO 2 mixing ratio) in an ndimensional vector space.…”
Section: Nonparametric Calibration Approachesmentioning
confidence: 99%
“…In example acknowledgment, the k-closest neighbors calculation (k-NN) is a non-parametric technique utilized for order and relapse. In both cases, the info comprises of the k nearest preparing cases in the component space [3]. The yield relies on upon whether k-NN is utilized for arrangement or relapse:…”
Section: E Genetic Algorithmsmentioning
confidence: 99%
“…The tenets are checked, and the ones that fit the information best are kept, the guidelines that don't fit the information are discarded. [3] The principles that were kept are then changed, and duplicated to make new rules. This procedure repeats as essential with a specific end goal to deliver decide that matches the dataset as nearly as possible.…”
Section: E Genetic Algorithmsmentioning
confidence: 99%
“…ML does not make any assumptions about the right structure of the data model, allowing the construction of complex non-linear models. There are many different paradigms in ML: lazy methods such as K-Nearest Neighbours (KNN) (Altman, 1992) methods, based on tree construction, as, for instance, C4.5 (Quinlan, 1993) or Neural or Bayesian networks (Mitchell, 1997). All of them have been successfully used in many different domains.…”
Section: Introductionmentioning
confidence: 99%
“…When the variable to predict is continuous, ML methods more commonly used are CART (Breiman, 2001), M5 (Quinlan, 1992), M5-Prime (Wang & Witten, 1997), KNN (Altman, 1992), or support vector regression (SVR, see Basak et al, 2007).…”
Section: Introductionmentioning
confidence: 99%