1997
DOI: 10.1109/72.554199
|View full text |Cite
|
Sign up to set email alerts
|

Curvilinear component analysis: a self-organizing neural network for nonlinear mapping of data sets

Abstract: We present a new strategy called "curvilinear component analysis" (CCA) for dimensionality reduction and representation of multidimensional data sets. The principle of CCA is a self-organized neural network performing two tasks: vector quantization (VQ) of the submanifold in the data set (input space); and nonlinear projection (P) of these quantizing vectors toward an output space, providing a revealing unfolding of the submanifold. After learning, the network has the ability to continuously map any new point … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
318
0
4

Year Published

1999
1999
2016
2016

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 533 publications
(334 citation statements)
references
References 9 publications
0
318
0
4
Order By: Relevance
“…In particular, we mention the following seven techniques: (1) Sammon mapping (Sammon, 1969), (2) curvilinear components analysis (CCA; Demartines and Hérault, 1997), (3) Stochastic Neighbor Embedding (SNE; Hinton and Roweis, 2002), (4) Isomap (Tenenbaum et al, 2000), (5) Maximum Variance Unfolding (MVU; Weinberger et al, 2004), (6) Locally Linear Embedding (LLE; Roweis and Saul, 2000), and (7) Laplacian Eigenmaps (Belkin and Niyogi, 2002). Despite the strong performance of these techniques on artificial data sets, they are often not very successful at visualizing real, high-dimensional data.…”
Section: Introductionmentioning
confidence: 99%
“…In particular, we mention the following seven techniques: (1) Sammon mapping (Sammon, 1969), (2) curvilinear components analysis (CCA; Demartines and Hérault, 1997), (3) Stochastic Neighbor Embedding (SNE; Hinton and Roweis, 2002), (4) Isomap (Tenenbaum et al, 2000), (5) Maximum Variance Unfolding (MVU; Weinberger et al, 2004), (6) Locally Linear Embedding (LLE; Roweis and Saul, 2000), and (7) Laplacian Eigenmaps (Belkin and Niyogi, 2002). Despite the strong performance of these techniques on artificial data sets, they are often not very successful at visualizing real, high-dimensional data.…”
Section: Introductionmentioning
confidence: 99%
“…In this study, some unsupervised statistical and neural projection models, namely Principal Component Analysis (PCA) [9], Curvilinear Component Analysis (CCA) [10], and Cooperative Maximum Likelihood Hebbian Learning (CMLHL) [11] have been applied, comparing their results. PCA [9] is a standard statistical technique for compressing multidimensional data; it can be shown to give the best linear compression of the data in terms of least mean square error.…”
Section: Unsupervised Projection Modelsmentioning
confidence: 99%
“…CCA [10] is a nonlinear dimensionality reduction method. It was developed as an improvement on the Self Organizing Map (SOM) [12], trying to circumvent the limitations inherent in some linear models such as PCA.…”
Section: Unsupervised Projection Modelsmentioning
confidence: 99%
“…Another method based on curvilinear component analysis (CCA) [16] uses similar ideas to this work in term of finding mappings between data space and unfolded spaces. This task is done in two stages of learning the unfolding transformation by a neural network, and then nonlinear projection of the input space through the mapping.…”
Section: Introductionmentioning
confidence: 99%