2022
DOI: 10.1088/1367-2630/ac395d
|View full text |Cite
|
Sign up to set email alerts
|

Gaussian information bottleneck and the non-perturbative renormalization group

Abstract: The renormalization group (RG) is a class of theoretical techniques used to explain the collective physics of interacting, many-body systems. It has been suggested that the RG formalism may be useful in finding and interpreting emergent low-dimensional structure in complex systems outside of the traditional physics context, such as in biology or computer science. In such contexts, one common dimensionality-reduction framework already in use is information bottleneck (IB), in which the goal is to compress an ``… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
21
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(21 citation statements)
references
References 50 publications
(77 reference statements)
0
21
0
Order By: Relevance
“…Mutual information between partitions of a system allowed the automatic discovery and information-based prescription of relevant features [70][71][72]. The connection between the information bottleneck and the renormalization group [21,22] suggests IB can uncover the most relevant information in a relationship. By constraining the encoder in a standard IB framework to be a linear projection, the authors of Ref.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Mutual information between partitions of a system allowed the automatic discovery and information-based prescription of relevant features [70][71][72]. The connection between the information bottleneck and the renormalization group [21,22] suggests IB can uncover the most relevant information in a relationship. By constraining the encoder in a standard IB framework to be a linear projection, the authors of Ref.…”
Section: Discussionmentioning
confidence: 99%
“…Given random variables X and Y serving as an input and an output, the IB defines a spectrum of compressed representations of X that retain only the most relevant information about Y . The potential of IB to analyze relationships was recently strengthened through connections to the renormalization group in statistical physics, one of the field's most powerful tools [21,22]. Although IB serves as a useful framework for examining the process of learning [23,24], it has limited capacity to find useful approximations through optimization, particularly when the relationship between X and Y is deterministic (or nearly so) [25,26].…”
Section: Introductionmentioning
confidence: 99%
“…This procedure is related to a variety of different data compression techniques such as the information bottleneck [37] and variational autoencoding [56]. In this way, the Bayesian renormalization scheme makes direct contact with previous insights into the relationship between model building/dimensional reduction for complex systems and the RG such as [19][20][21]36].…”
Section: Comparison With Diffusion Learningmentioning
confidence: 99%
“…Or worse yet, what if we are interested in renormalizing a model that does not have a physical interpretation at all? Such a situation presents itself in recent work which seeks to import the machinery of renormalization into data science contexts as a tool for performing data compression and improving the interpretability and performance of high dimensional models [19][20][21][22][23][24][25][26][27][28][29].…”
Section: Introductionmentioning
confidence: 99%
“…What makes these neural computers so compelling is that they are exceptional in different ways than modern-day silicon computers: the latter relies on binary representations, rapid sequential processing [6], and segregated memory and CPU [7], while the former utilizes continuum representations [8,9], parallel and distributed processing [10,11], and distributed memory [12]. To harness these distinct computational abilities, prior work has studied a vast array of different network architectures [13,14], learning algorithms [15,16], and information-theoretic frameworks [17][18][19] in both biological and artificial neural networks. Despite these significant advances, the relationship between neural computers and modern-day silicon computers remains an analogy due to our lack of a concrete and low-level programming language, thereby limiting our access to neural computation.…”
Section: Introductionmentioning
confidence: 99%