1999
DOI: 10.1016/s0010-4655(06)70021-2
|View full text |Cite
|
Sign up to set email alerts
|

Phase transitions in soft-committee machines

Abstract: Equilibrium statistical physics is applied to layered neural networks with differentiable activation functions. A first analysis of off-line learning in soft-committee machines with a finite number (K) of hidden units learning a perfectly matching rule is performed. Our results are exact in the limit of high training temperatures (β → 0). For K = 2 we find a second order phase transition from unspecialized to specialized student configurations at a critical size P of the training set, whereas for K ≥ 3 the tra… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

3
21
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
1

Relationship

3
2

Authors

Journals

citations
Cited by 5 publications
(24 citation statements)
references
References 16 publications
3
21
0
Order By: Relevance
“…The term Soft Committee Machine (SCM) has been coined for feedforward neural networks with sigmoidal activations in a single hidden layer and a linear output unit (see, for instance, [ 30 , 31 , 32 , 33 , 34 , 35 , 36 , 55 , 56 ]). Its structure resembles that of a (crisp) committee machine with binary threshold hidden units, where the network’s response is given by their majority vote (see [ 5 , 6 , 7 ] and references therein).…”
Section: Models and Mathematical Analysismentioning
confidence: 99%
“…The term Soft Committee Machine (SCM) has been coined for feedforward neural networks with sigmoidal activations in a single hidden layer and a linear output unit (see, for instance, [ 30 , 31 , 32 , 33 , 34 , 35 , 36 , 55 , 56 ]). Its structure resembles that of a (crisp) committee machine with binary threshold hidden units, where the network’s response is given by their majority vote (see [ 5 , 6 , 7 ] and references therein).…”
Section: Models and Mathematical Analysismentioning
confidence: 99%
“…The quantity e N s corresponds to the volume in weight space that is consistent with a given configuration of order parameters. Independent of the activation functions or other details of the learning problem, one obtains for large N [30], [31]…”
Section: Thermal Equilibrium and The High-temperature Limitmentioning
confidence: 99%
“…Omitting additive constants and assuming the normalization (6) and site-symmetry (10), the entropy term reads [30], [31]…”
Section: Thermal Equilibrium and The High-temperature Limitmentioning
confidence: 99%
See 1 more Smart Citation
“…The research field of deep learning has recently attracted considerable attention due to significant progress in performing tasks relevant to many different applications [1][2][3][4][5]. Neural networks are learning machines inspired by the structure of the human brain [6], which have been studied with methods from statistical mechanics [5,[7][8][9][10][11], starting with simpler versions such as the perceptron [12][13][14] and also including two-layer networks [15][16][17][18][19][20][21][22][23][24][25][26]. Often, learning is studied in the framework of the student-teacher scenario, in which a student has to learn the connection vectors according to which a teacher classifies input patterns [10].…”
mentioning
confidence: 99%