SciPost Phys. Core 2020
DOI: 10.21468/scipostphyscore.2.2.005
|View full text |Cite
|
Sign up to set email alerts
|

From complex to simple : hierarchical free-energy landscape renormalized in deep neural networks

Abstract: We develop a statistical mechanical approach based on the replica method to study the solution space of deep neural networks. Specifically we analyze the configuration space of the synaptic weights in a simple feed-forward perceptron network within a Gaussian approximation for two scenarios : a setting with random inputs/outputs and a teacher-student setting. By increasing the strength of constraints, i. e. increasing the number of imposed patterns, successive 2nd order glass transition (random inputs/outputs)… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 9 publications
(12 citation statements)
references
References 80 publications
0
12
0
Order By: Relevance
“…The former non-rigorousness might be resolved using the techniques developed in analyzing other convex learning problems (Mignacco et al, 2020;Gerbelot et al, 2020). For simplification of the model architecture, we might be able to extend the present analysis based on the previous analytical techniques to handle the replicated system for more complex model architectures, such as the random feature model (Gerace et al, 2020), kernel method (Canatar et al, 2021;Dietrich et al, 1999), and multi-layer neural networks (Schwarze and Hertz, 1992;Yoshino, 2020).…”
Section: Discussionmentioning
confidence: 99%
“…The former non-rigorousness might be resolved using the techniques developed in analyzing other convex learning problems (Mignacco et al, 2020;Gerbelot et al, 2020). For simplification of the model architecture, we might be able to extend the present analysis based on the previous analytical techniques to handle the replicated system for more complex model architectures, such as the random feature model (Gerace et al, 2020), kernel method (Canatar et al, 2021;Dietrich et al, 1999), and multi-layer neural networks (Schwarze and Hertz, 1992;Yoshino, 2020).…”
Section: Discussionmentioning
confidence: 99%
“…Moreover we expect that dense piecewise linear spheres, at finite temperature, will display strong Gardner phenomenology [18] upon cooling. Finally it would be interesting to see what happens for deeper models beyond the perceptron architecture [19][20][21] as well as more complex constraint satisfaction problems [22].…”
Section: Discussionmentioning
confidence: 99%
“…The phase space volume, which is called Gardner's volume, can be expressed for the present DNN as [1],…”
Section: Gardner's Volumementioning
confidence: 99%
“…For the theoretical part, we use the replica approach developed recently [1] which works on high dimensional data D = N 1. We show that it becomes exact in the dense limit N c 1 and M 1 with fixed α = M/c.…”
Section: Introductionmentioning
confidence: 99%