2022
DOI: 10.48550/arxiv.2202.05258
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Hardness of Noise-Free Learning for Two-Hidden-Layer Neural Networks

Abstract: We give exponential statistical query (SQ) lower bounds for learning two-hidden-layer ReLU networks with respect to Gaussian inputs in the standard (noise-free) model. No general SQ lower bounds were known for learning ReLU networks of any depth in this setting: previous SQ lower bounds held only for adversarial noise models (agnostic learning) [KK14, GGK20, DKZ20] or restricted models such as correlational SQ [GGJ + 20, DKKZ20].Prior work hinted at the impossibility of our result: Vempala and Wilmes [VW19] sh… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
8
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(8 citation statements)
references
References 11 publications
0
8
0
Order By: Relevance
“…For example, Chen, Klivans, and Meka [19] showed fixed-parameter tractability of learning a ReLU network under some assumptions including Gaussian data and Lipschitz continuity of the network. We refer to [8,18,25,36,39,37] as a non-exhaustive list of other results about (non-)learnability of ReLU networks in different settings.…”
Section: Neural Networkmentioning
confidence: 99%
“…For example, Chen, Klivans, and Meka [19] showed fixed-parameter tractability of learning a ReLU network under some assumptions including Gaussian data and Lipschitz continuity of the network. We refer to [8,18,25,36,39,37] as a non-exhaustive list of other results about (non-)learnability of ReLU networks in different settings.…”
Section: Neural Networkmentioning
confidence: 99%
“…There are many lower bounds known in the distribution-free setting [BR92, Vu98, KS09, LSSS14, DV20], however, these do not transfer over to our (unsupervised) setting. When x is Gaussian, the aforementioned work of [CGKM22] derives hardness for learning twohidden-layer networks with polynomial size for all SQ algorithms, as well as under cryptographic assumptions (see also [DV21]). It is not hard to show (see Appendix C) that this lower bound immediately implies a lower bound for the unsupervised problem.…”
Section: Related Workmentioning
confidence: 99%
“…In fact, a recent line of work suggests that learning neural network pushforwards of Gaussians may be an inherently difficult computational task. Recent results of [DV21,CGKM22] show hardness of supervised learning from labeled Gaussian examples under cryptographic asssumptions, and the latter also demonstrates hardness for all statistical query (SQ) algorithms (see Section 1.3 for a more detailed description of related work). These naturally imply hardness in the unsupervised setting (see Appendix C).…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations