2020
DOI: 10.48550/arxiv.2001.06970
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Finding the Sparsest Vectors in a Subspace: Theory, Algorithms, and Applications

Qing Qu,
Zhihui Zhu,
Xiao Li
et al.

Abstract: The problem of finding the sparsest vector (direction) in a low dimensional subspace can be considered as a homogeneous variant of the sparse recovery problem, which finds applications in robust subspace recovery, dictionary learning, sparse blind deconvolution, and many other problems in signal processing and machine learning. However, in contrast to the classical sparse recovery problem, the most natural formulation for finding the sparsest vector in a subspace is usually nonconvex. In this paper, we overvie… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
1

Relationship

4
1

Authors

Journals

citations
Cited by 5 publications
(9 citation statements)
references
References 115 publications
(262 reference statements)
0
9
0
Order By: Relevance
“…Moreover, from a boarder perspective our work is rooted in recent advances on global nonconvex optimization theory for signal processing and machine learning problems [52,[78][79][80][81]. In a sequence of works [78,80,[82][83][84][85][86][87][88][89][90][91][92][93][94][95], the authors showed that many problems exhibit "equivalently good" global minimizers due to symmetries and intrinsic low-dimensional structures, and the loss functions are usually strict saddles [50][51][52]. These problems include, but are not limited to, phase retrieval [86,87], low-rank matrix recovery [78,82,85,88,90], dictionary learning [80,83,84,91,96], and sparse blind deconvolution [92][93][94][95].…”
Section: Bias Term Results Constraint Weight Decaymentioning
confidence: 99%
See 2 more Smart Citations
“…Moreover, from a boarder perspective our work is rooted in recent advances on global nonconvex optimization theory for signal processing and machine learning problems [52,[78][79][80][81]. In a sequence of works [78,80,[82][83][84][85][86][87][88][89][90][91][92][93][94][95], the authors showed that many problems exhibit "equivalently good" global minimizers due to symmetries and intrinsic low-dimensional structures, and the loss functions are usually strict saddles [50][51][52]. These problems include, but are not limited to, phase retrieval [86,87], low-rank matrix recovery [78,82,85,88,90], dictionary learning [80,83,84,91,96], and sparse blind deconvolution [92][93][94][95].…”
Section: Bias Term Results Constraint Weight Decaymentioning
confidence: 99%
“…In a sequence of works [78,80,[82][83][84][85][86][87][88][89][90][91][92][93][94][95], the authors showed that many problems exhibit "equivalently good" global minimizers due to symmetries and intrinsic low-dimensional structures, and the loss functions are usually strict saddles [50][51][52]. These problems include, but are not limited to, phase retrieval [86,87], low-rank matrix recovery [78,82,85,88,90], dictionary learning [80,83,84,91,96], and sparse blind deconvolution [92][93][94][95]. As we shall see, the global minimizers (i.e., simplex ETFs) of our problem here also exhibit a similar rotational symmetry, compared to low-rank matrix recovery.…”
Section: Bias Term Results Constraint Weight Decaymentioning
confidence: 99%
See 1 more Smart Citation
“…To deal with this challenge, in the following we further investigate the global optimization landscape of (3). By leveraging recent advances on nonconvex optimization [32][33][34][39][40][41][42], we first show that our nonconvex MSE loss (3) without bias term is a strict saddle function that every non-global critical point is a saddle point with negative curvature (i.e., its Hessian has at least one negative eigenvalue).…”
Section: Characterizations Of the Benign Global Landscapementioning
confidence: 99%
“…The form in ( 4) is closely related to recent work on blind deconvolution [45,50]. More specifically, the work showed that normalizing the output z out via preconditioning eliminates bad local minimizers and dramatically improves the nonconvex optimization landscapes for learning the kernel a.…”
Section: A Warm-up Studymentioning
confidence: 95%