2019
DOI: 10.1073/pnas.1802705116
|View full text |Cite
|
Sign up to set email alerts
|

Optimal errors and phase transitions in high-dimensional generalized linear models

Abstract: Generalized linear models (GLMs) are used in high-dimensional machine learning, statistics, communications, and signal processing. In this paper we analyze GLMs when the data matrix is random, as relevant in problems such as compressed sensing, error-correcting codes, or benchmark models in neural networks. We evaluate the mutual information (or “free entropy”) from which we deduce the Bayes-optimal estimation and generalization errors. Our analysis applies to the high-dimensional limit where both the number o… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

9
324
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
7
2

Relationship

2
7

Authors

Journals

citations
Cited by 178 publications
(333 citation statements)
references
References 98 publications
(381 reference statements)
9
324
0
Order By: Relevance
“…• Since the original conference versions of this paper [1], [2], formulae for the minimum mean-squared error (MMSE) for inference in deep networks have been conjectured in [47]- [49]. As discussed in Section IV-C, these formulae are based on heuristic techniques, such as the replica method from statistical physics, and have been rigorously proven in special cases [50].…”
Section: Main Contributionsmentioning
confidence: 99%
“…• Since the original conference versions of this paper [1], [2], formulae for the minimum mean-squared error (MMSE) for inference in deep networks have been conjectured in [47]- [49]. As discussed in Section IV-C, these formulae are based on heuristic techniques, such as the replica method from statistical physics, and have been rigorously proven in special cases [50].…”
Section: Main Contributionsmentioning
confidence: 99%
“…The duality among disorder-to-order transition in statistical mechanics of disordered systems and detectability-undetectability transition in machine learning (see e.g. [6,8,9,10,12,15,19]) suggests that the knowledge of the optimal shapes stemmed from the former could play some role in the latter.…”
Section: Precisely Equation (38) Holds If and Only Ifmentioning
confidence: 99%
“…The fluctuations of this matrix are easier to control than the ones of the overlap because L is related to the λ ngradient of the free energy, which is self-averaging by hypothesis (9). The proof is a straightforward extension to the matrix case of the one found in [23], [30] and requires no new ideas. This general result does not depend on the fact that we consider optimal Bayesian inference; it is only a consequence of the perturbation, i.e., the side information coming from the channel (6).…”
Section: Proof Ideamentioning
confidence: 88%
“…the product posterior measure, and g is any bounded function. This innocent-looking key identity on which relies the whole proof follows directly from Bayes' law -thus the importance of placing ourselves in the Bayesian optimal setting-, see [23], [30]. Applied to…”
Section: Proof Ideamentioning
confidence: 99%