2018
DOI: 10.1016/j.jmva.2018.06.002
|View full text |Cite
|
Sign up to set email alerts
|

Asymptotic performance of PCA for high-dimensional heteroscedastic data

Abstract: Principal Component Analysis (PCA) is a classical method for reducing the dimensionality of data by projecting them onto a subspace that captures most of their variation. Effective use of PCA in modern applications requires understanding its performance for data that are both high-dimensional and heteroscedastic. This paper analyzes the statistical performance of PCA in this setting, i.e., for high-dimensional data drawn from a low-dimensional subspace and degraded by heteroscedastic noise. We provide simplifi… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
50
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
9

Relationship

1
8

Authors

Journals

citations
Cited by 58 publications
(51 citation statements)
references
References 46 publications
1
50
0
Order By: Relevance
“…Remark 16. This noise model generalizes two previous models of heteroscedastic noise in the context of principal component analysis [21,22,23,55,39]. In both, the matrix X consists of random, i.i.d.…”
Section: 2supporting
confidence: 57%
See 1 more Smart Citation
“…Remark 16. This noise model generalizes two previous models of heteroscedastic noise in the context of principal component analysis [21,22,23,55,39]. In both, the matrix X consists of random, i.i.d.…”
Section: 2supporting
confidence: 57%
“…The model from [55,39] takes T = I n , in which case the observations are of the form Y j = X j + S 1/2 G j , where G j \sim N (0, I p ). By contrast, the papers [21,22,23] take S = I p , in which case the observations are of the form Y j =…”
Section: 2mentioning
confidence: 99%
“…e main idea of the principal component analysis (PCA) method is to transform the n-dimensional feature variable through the coordinate axis and the origin to form a new m-dimensional feature (usually, m is less than n) [11].…”
Section: Basic Theorymentioning
confidence: 99%
“…Finally, a prominent alternative approach would be to facilitate Weighted-PCA (WPCA) [18]. In principle, WPCA was developed to address data samples with varying quality, by giving less informative samples less weight [34]. One could harness this approach by viewing the trajectory length as quality-the longer the length, the higher the weight of the points in the corresponding trajectory.…”
Section: Discussionmentioning
confidence: 99%