1999
DOI: 10.1109/72.788641
|View full text |Cite
|
Sign up to set email alerts
|

Input space versus feature space in kernel-based methods

Abstract: This paper collects some ideas targeted at advancing our understanding of the feature spaces associated with support vector (SV) kernel functions. We first discuss the geometry of feature space. In particular, we review what is known about the shape of the image of input space under the feature space map, and how this influences the capacity of SV methods. Following this, we describe how the metric governing the intrinsic geometry of the mapped surface can be computed in terms of the kernel, using the example … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
657
0
9

Year Published

2003
2003
2015
2015

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 1,035 publications
(666 citation statements)
references
References 21 publications
0
657
0
9
Order By: Relevance
“…16,37 Finally, performing kernel-based nonlinear mapping is shown to perform well but, thus far, it is not used to retrieve physicochemical information. Reconstructing the mapping according to Schölkopf et al, 38 for example, and knowing what kind of mapping is preferred for specific features (e.g., spectral bands) can increase the knowledge of the problem. This in turn can give further directions to interpret and improve the results.…”
Section: Discussionmentioning
confidence: 99%
“…16,37 Finally, performing kernel-based nonlinear mapping is shown to perform well but, thus far, it is not used to retrieve physicochemical information. Reconstructing the mapping according to Schölkopf et al, 38 for example, and knowing what kind of mapping is preferred for specific features (e.g., spectral bands) can increase the knowledge of the problem. This in turn can give further directions to interpret and improve the results.…”
Section: Discussionmentioning
confidence: 99%
“…that maps (usually non-linearly) the input feature vectors x i to a possibly high dimensional space F , called feature space, which usually has the structure of a Hilbert space [32], [33], where the data are supposed to be linearly or near linearly separable. The exact form of the mapping function is not required to be known, since all required subsequent operations of the learning algorithm are expressed in terms of dot products between the input vectors in the Hilbert space performed by the kernel trick [34].…”
Section: Non-linear Maximum Margin Projectionsmentioning
confidence: 99%
“…Thus, the new point x (τ +1) is the conditional mean of the mixture under the current point x (τ ) . This is formally akin to clustering by deterministic annealing (Rose, 1998), to algorithms for finding pre-images in kernel-based methods (Schölkopf et al, 1999) and to mean-shift algorithms (section 5.2).…”
Section: Particular Casesmentioning
confidence: 99%