2021
DOI: 10.48550/arxiv.2102.00305
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Spike and slab Bayesian sparse principal component analysis

Abstract: Sparse principal component analysis (PCA) is a popular tool for dimensional reduction of high-dimensional data. Despite its massive popularity, there is still a lack of theoretically justifiable Bayesian sparse PCA that is computationally scalable. A major challenge is choosing a suitable prior for the loadings matrix, as principal components are mutually orthogonal. We propose a spike and slab prior that meets this orthogonality constraint and show that the posterior enjoys both theoretical and computational … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 39 publications
0
4
0
Order By: Relevance
“…To tackle such problems, sparse PCA was developed by involving an automatic selection of the appropriate model dimensionality with high-dimensional data; i.e., the principal components obtained by sparse PCA are linear combinations of only a few important variables, which facilitates the interpretation of data feature selection in practice. The main methods commonly used are the addition of a regularization term [19,20] and the use of promoting sparsity of a prior distribution for the coefficients of the projection matrix [21,22]. However, the sparseness patterns of the original variables for each principal component may be different, and thus we need to understand which original variables are active for interpreting each principal component separately.…”
Section: Identification Means Drawbacksmentioning
confidence: 99%
See 3 more Smart Citations
“…To tackle such problems, sparse PCA was developed by involving an automatic selection of the appropriate model dimensionality with high-dimensional data; i.e., the principal components obtained by sparse PCA are linear combinations of only a few important variables, which facilitates the interpretation of data feature selection in practice. The main methods commonly used are the addition of a regularization term [19,20] and the use of promoting sparsity of a prior distribution for the coefficients of the projection matrix [21,22]. However, the sparseness patterns of the original variables for each principal component may be different, and thus we need to understand which original variables are active for interpreting each principal component separately.…”
Section: Identification Means Drawbacksmentioning
confidence: 99%
“…For conventional sparse PCA [21,22], its loading matrix P would be expected to have few nonzero coefficients. However, each principal component usually selects different relevant variables (data columns) in data matrix X because we do not have the same sparseness pattern, i.e., they cannot be expressed as the linear combination of the same active variables in X.…”
Section: Bayesian Pca Modelmentioning
confidence: 99%
See 2 more Smart Citations