2015
DOI: 10.48550/arxiv.1506.02181
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The LASSO with Non-linear Measurements is Equivalent to One With Linear Measurements

Chrtistos Thrampoulidis,
Ehsan Abbasi,
Babak Hassibi

Abstract: Consider estimating an unknown, but structured (e.g. sparse, low-rank, etc.), signal x 0 ∈ R n from a vector y ∈ R m of measurements of the form y i = g i (a i T x 0 ), where the a i 's are the rows of a known measurement matrix A, and, g(•) is a (potentially unknown) nonlinear and random link-function. Such measurement functions could arise in applications where the measurement device has nonlinearities and uncertainties. It could also arise by design, e.g., g i (x) = sign(x + z i ), corresponds to noisy 1-bi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
17
0

Year Published

2015
2015
2017
2017

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(19 citation statements)
references
References 18 publications
2
17
0
Order By: Relevance
“…Thus, the theorem above shows that the projected gradient updates converge at a linear rate to a small neighborhood around the "true" solution µθ * . The radius of this neighborhood decreases with an increase in the number of samples n. The size of this radius is near-optimal and comparable to recent results [19,22,28] where the estimate is obtained by solving the convex optimization problem in (1.1) which applies only when the regularization function R is convex. Indeed, the size of this radius scales like n 0 n w 2 which up to a small constant is exactly the same as the result one would get when the model is linear of the form y = µXθ * + w. This is perhaps unexpected as it demonstrates that our nonlinear model exactly behaves like a fictitious linear model with the same effective noise!…”
Section: Note That η σsupporting
confidence: 85%
See 1 more Smart Citation
“…Thus, the theorem above shows that the projected gradient updates converge at a linear rate to a small neighborhood around the "true" solution µθ * . The radius of this neighborhood decreases with an increase in the number of samples n. The size of this radius is near-optimal and comparable to recent results [19,22,28] where the estimate is obtained by solving the convex optimization problem in (1.1) which applies only when the regularization function R is convex. Indeed, the size of this radius scales like n 0 n w 2 which up to a small constant is exactly the same as the result one would get when the model is linear of the form y = µXθ * + w. This is perhaps unexpected as it demonstrates that our nonlinear model exactly behaves like a fictitious linear model with the same effective noise!…”
Section: Note That η σsupporting
confidence: 85%
“…The plots in Figure 1a clearly show that PGD iterates applied to nonlinear observations converge quickly to an estimate which is of the same size as the effective noise induced by the nonlinearity. In this sense, PGD iterates converge quickly to a reliable solution which has exactly the same quality as the optimal results obtained in [22,28]. 3 Figure 1a also clearly demonstrates that the behavior of the PGD iterates applied to both models are essentially the same further corroborating the results of Theorem 3.1.…”
Section: Synthetic Experimentssupporting
confidence: 75%
“…In this work, we consider embedding a high-dimensional set of points K to the hamming cube in a lower dimension, which is known as binary embedding. Binary embedding is a natural problem arising from quantization of the measurements and is connected to 1-bit compressed sensing as well as locality sensitive hashing [1,3,14,20,22,25,27]. In particular, given a subset K of the unit sphere in R n , and a dimensionality reduction map A ∈ R m×n , we wish to ensure that 1 m sgn(Ax), sgn(Ay) H − ang(x, y) ≤ δ.…”
Section: Introductionmentioning
confidence: 99%
“…More recently in [12] Jacques studies a related problem in which samples Ax are uniformly quantized instead of being discretized to {+1, −1}. There is also a growing amount of literature related to binary embedding and one-bit compressed sensing [1,3,22,25]. In this work, we significantly improve the distortion dependence bounds for the binary embedding of arbitrary sets.…”
Section: Introductionmentioning
confidence: 99%
“…Neykov et al (2015) analyzed two algorithms based on Sliced Inverse Regression (Li 1991) under the assumption that X ∼ N (0, I p×p ), and demonstrated that they can uncover the support optimally in terms of the rescaled sample size. Plan & Vershynin (2015) and Thrampoulidis et al (2015) demonstrated that a constrained version of LASSO can be used to obtain an L 2 consistent estimator of β 0 . None of these procedures provide results on the performance of the LASSO algorithm in support recovery, which relates to L 2 consistency but is a fundamentally different theoretical aspect.…”
Section: Overview Of Related Workmentioning
confidence: 99%