2013
DOI: 10.1214/13-aos1097
|View full text |Cite
|
Sign up to set email alerts
|

Sparse principal component analysis and iterative thresholding

Abstract: Principal component analysis (PCA) is a classical dimension reduction method which projects data onto the principal subspace spanned by the leading eigenvectors of the covariance matrix. However, it behaves poorly when the number of features p is comparable to, or even much larger than, the sample size n. In this paper, we propose a new iterative thresholding approach for estimating principal subspaces in the setting where the leading eigenvectors are sparse. Under a spiked covariance model, we find that the n… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
288
0

Year Published

2013
2013
2023
2023

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 263 publications
(289 citation statements)
references
References 31 publications
1
288
0
Order By: Relevance
“…Birnbaum et al (2012) also proposed an estimation procedure named ASPCA which is based on a two-stage coordinate selection scheme. Both these estimators are rateoptimal, so long as the diagonal thresholding scheme is consistent (see Ma, 2013;Paul and Johnstone, 2012). A different two-stage estimation scheme, based on a regression framework, has been proposed by .…”
Section: Covariance Estimation Using Sparse Factor Modelsmentioning
confidence: 99%
See 1 more Smart Citation
“…Birnbaum et al (2012) also proposed an estimation procedure named ASPCA which is based on a two-stage coordinate selection scheme. Both these estimators are rateoptimal, so long as the diagonal thresholding scheme is consistent (see Ma, 2013;Paul and Johnstone, 2012). A different two-stage estimation scheme, based on a regression framework, has been proposed by .…”
Section: Covariance Estimation Using Sparse Factor Modelsmentioning
confidence: 99%
“…A different two-stage estimation scheme, based on a regression framework, has been proposed by . Incidentally, in order to establish upper bounds on rates of convergence of the estimators, Amini and Wainwright (2008), Ma (2013) and Paul and Johnstone (2012) all utilize results on the concentration of the extreme singular values of a rectangular matrix with i.i.d. standard Gaussian entries, which translates into finite sample probabilistic bounds on the extreme eigenvalues of a Wishart matrix.…”
Section: Covariance Estimation Using Sparse Factor Modelsmentioning
confidence: 99%
“…The separations between the common factors and idiosyncratic components are carried out by the low-rank plus sparsity decomposition. See, for example, Cai et al (2013); Candès and Recht (2009) ;Fan et al (2013); Koltchinskii et al (2011);Ma (2013); Negahban and Wainwright (2011).…”
Section: Introductionmentioning
confidence: 99%
“…The basic idea is to produce the optimal simplification of a multi variable system in which the p variable values x 1 , x 2 ,…, x p described in the data sheet of the original system re-adjust and combine, and extract m (m ≤ p) synthesize variable f 1 , f 2 , …, f m , which can maximize the overlap of the information of data object described in the original system. Dimensionality reduction and the simplification goal can then be obtained [12][13][14]. They can be represented: (10) Where m f is the principal component, and m ≤ p. Generally, the cumulative contribution of m principal components is 85%, so in keeping the most information, the goal of dimensionality reduction can be achieve.…”
Section: Principal Component Analysismentioning
confidence: 99%