2020
DOI: 10.48550/arxiv.2006.03177
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Hardness of Learning Neural Networks with Natural Weights

Abstract: Neural networks are nowadays highly successful despite strong hardness results. The existing hardness results focus on the network architecture, and assume that the network's weights are arbitrary. A natural approach to settle the discrepancy is to assume that the network's weights are "well-behaved" and posses some generic properties that may allow efficient learning. This approach is supported by the intuition that the weights in real-world networks are not arbitrary, but exhibit some "random-like" propertie… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
6
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
6

Relationship

1
5

Authors

Journals

citations
Cited by 7 publications
(6 citation statements)
references
References 31 publications
0
6
0
Order By: Relevance
“…Hence, the results from Klivans and Sherstov [2006] and Daniely and Shalev-Shwartz [2016] imply hardness of improperly learning depth-2 neural networks with n ǫ and ω(log(n)) hidden neurons (respectively). Daniely and Vardi [2020] showed, under the assumption that refuting a random K-SAT formula is hard, that improperly learning depth-2 neural networks is hard already if its weights are drawn from some "natural" distribution or satisfy some "natural" properties. While hardness of proper learning is implied by hardness of improper learning, there are some recent works that show hardness of properly learning depth-2 networks under more standard assumptions (cf.…”
Section: Intersections Of Halfspacesmentioning
confidence: 99%
See 1 more Smart Citation
“…Hence, the results from Klivans and Sherstov [2006] and Daniely and Shalev-Shwartz [2016] imply hardness of improperly learning depth-2 neural networks with n ǫ and ω(log(n)) hidden neurons (respectively). Daniely and Vardi [2020] showed, under the assumption that refuting a random K-SAT formula is hard, that improperly learning depth-2 neural networks is hard already if its weights are drawn from some "natural" distribution or satisfy some "natural" properties. While hardness of proper learning is implied by hardness of improper learning, there are some recent works that show hardness of properly learning depth-2 networks under more standard assumptions (cf.…”
Section: Intersections Of Halfspacesmentioning
confidence: 99%
“…It does not imply the hardness result fromDaniely and Vardi [2020] for learning depth-2 neural networks whose weights are drawn from some "natural" distribution.…”
mentioning
confidence: 99%
“…All known works take additional assumptions on the unknown weights and coefficients or do not run in polynomial-time in all the important parameters. In fact, there are a number of lower bounds, both for restricted models of computation like correlational statistical queries [GGJ + 20, DKKZ20] as well as under various average-case assumptions [DV20,DV21], that suggest that a truly polynomial-time algorithm may be impossible to achieve.…”
Section: Introductionmentioning
confidence: 99%
“…Note in particular, that this problem includes problems of learning generative deep networks [16]. For models with hidden nodes there are many computational hardness results that are known to hold both in the worst-case [18] and in averagecase sense [9,10] even for two layer networks. Even for tree models with hidden nodes, without additional assumption, learning is as hard as learning parity with noise [23].…”
Section: Introductionmentioning
confidence: 99%