Algorithms that can efficiently recover principal components of high-dimensional data from compressive sensing measurements (e.g. low-dimensional random projections) of it have been an important topic of recent interest in the literature. In this paper, we show that, under certain conditions, normal principal component analysis (PCA) on such low-dimensional random projections of data actually returns the same result as PCA on the original data set would. In particular, as the number of data samples increases, the center of the randomly projected data converges to the true center of the original data (up to a known scaling factor) and the principal components converge to the true principal components of the original data as well, even if the dimension of each random subspace used is very low. Indeed, experimental results verify that this approach does estimate the original center and principal components very well for both synthetic and real-world datasets, including hyperspectral data. Its performance is even superior to that of other algorithms recently developed in the literature for this purpose.Index Terms-Compressive sensing, Principal component analysis, Random projections, Low-rank matrix recovery, Hyperspectral data 1. INTRODUCTION Principal component analysis (PCA) [1] selects the best lowdimensional linear projection of a set of data points to minimize error between the original and projected data. It can also be thought of as finding the linear subspace that maximally preserves the variance of, or in some cases, the information in, the data. PCA is frequently used for dimensionality reduction, or as a summary of interesting features of the data. It is also often used as a precursor to signal classification.To obtain the principal components (PCs) of data, one typically centers the data and then computes the eigenvectors of the data's covariance matrix, using full knowledge of all data. However, in this paper, we will show that when the PCA algorithm is instead applied to low-dimensional random projections of each data point, as are acquired in many compressive sensing (CS) measurement schemes [2], it will often return the same center (up to a known scaling factor) and principal components as it would for the original dataset.More precisely, we show that the center of the low-dimensional random projections of the data converges to the true center of the original data (up to a known scaling factor) almost surely as the number of data samples increases. We then show that under certain conditions the top d eigenvectors of the randomly projected data's covariance matrix converge to the true d principal components of the original data as the number of data samples increases. Moreover, both of the above conclusions are true regardless of how few dimensions we use in our random projections (i.e. how few CS measurements we take of each sample).Furthermore, experimentally, we find that on both synthetic and real-world examples, including hyperspectral data, normal PCA on low-dimensional random projections of the...
Compressive sensing accurately reconstructs a signal that is sparse in some basis from measurements, generally consisting of the signal's inner products with Gaussian random vectors. The number of measurements needed is based on the sparsity of the signal, allowing for signal recovery from far fewer measurements than is required by the traditional Shannon sampling theorem. In this paper, we show how to apply the kernel trick, popular in machine learning, to adapt compressive sensing to a different type of sparsity. We consider a signal to be "nonlinearly K-sparse" if the signal can be recovered as a nonlinear function of K underlying parameters. Images that lie along a low-dimensional manifold are good examples of this type of nonlinear sparsity. It has been shown that natural images are as well [1]. We show how to accurately recover these nonlinearly K-sparse signals from approximately 2K measurements, which is often far lower than the number of measurements usually required under the assumption of sparsity in an orthonormal basis (e.g. wavelets). In experimental results, we find that we can recover images far better for small numbers of compressive sensing measurements, sometimes reducing the mean square error (MSE) of the recovered image by an order of magnitude or more, with little computation. A bound on the error of our recovered signal is also proved.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2025 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.