Wiley StatsRef: Statistics Reference Online 2018
DOI: 10.1002/9781118445112.stat08122
|View full text |Cite
|
Sign up to set email alerts
|

Principal Component Analysis for Big Data

Abstract: Big data is transforming our world, revolutionizing operations and analytics everywhere, from financial engineering to biomedical sciences. The complexity of big data often makes dimension reduction techniques necessary before conducting statistical inference. Principal component analysis, commonly referred to as PCA, has become an essential tool for multivariate data analysis and unsupervised dimension reduction, the goal of which is to find a lower dimensional subspace that captures most of the variation in … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
17
0
4

Year Published

2018
2018
2024
2024

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 35 publications
(21 citation statements)
references
References 96 publications
0
17
0
4
Order By: Relevance
“…Lan Fu [17] described the Discrimination Analysis of Multivariate Statistical Analysis, Linear Dimensionality Reduction and Nonlinear Dimensionality Reduction Methods in the light of the wide range of applications of high-dimensional data. Zebin Wu [18] developed a parallel and distributed technique for hyperspectral dimensionality reduction and Principal Component Analysis (PCA), of cloud computing architectures. Marco Cavallo [19] proposed a different framework that interacts visually to improve dimensionality reduction based exploratory data analysis.…”
Section: Literature Surveymentioning
confidence: 99%
“…Lan Fu [17] described the Discrimination Analysis of Multivariate Statistical Analysis, Linear Dimensionality Reduction and Nonlinear Dimensionality Reduction Methods in the light of the wide range of applications of high-dimensional data. Zebin Wu [18] developed a parallel and distributed technique for hyperspectral dimensionality reduction and Principal Component Analysis (PCA), of cloud computing architectures. Marco Cavallo [19] proposed a different framework that interacts visually to improve dimensionality reduction based exploratory data analysis.…”
Section: Literature Surveymentioning
confidence: 99%
“…Eigenvector computation is a ubiquitous problem in data processing nowadays, such as spectral clustering [Ng et al, 2002;Xu and Ke, 2016a], pagerank computation, dimensionality reduction [Fan et al, 2018], and so on. Classic solvers from numerical algebra are power methods and Lanczos algorithms [Golub and Van Loan, 1996], based on which there has been a recent surge of interest in developing varieties of solvers [Hardt and Price, 2014;Musco and Musco, 2015;Garber et al, 2016;Balcan et al, 2016].…”
Section: Introductionmentioning
confidence: 99%
“…Complementary work on the fundamental law of active management can be found inClarke et al (2002),Buckle (2004),Ye (2008) andDing et al (2020).6 Following this definition and assuming K common factors with K < N , factor-based methods only need to estimate K(K + 1)∕2 covariance entries and thus do not suffer from the curse of dimensionality.7 This approach is widely spread in the recent financial literature, since it does not require a specific factor model, see, e.g.,Fan et al (2013),Fan et al (2018),Fan et al (2016).…”
mentioning
confidence: 99%