2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW) 2019
DOI: 10.1109/iccvw.2019.00099
|View full text |Cite
|
Sign up to set email alerts
|

Direct Validation of the Information Bottleneck Principle for Deep Nets

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
25
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 25 publications
(25 citation statements)
references
References 5 publications
0
25
0
Order By: Relevance
“…Another type of variational lower bound can be obtained from the Donsker-Varadhan representation of mutual information, which is the principle underlying MINE [34]. MINE was used to estimate the IPs in [12] and [15] for a 784 − 512 − 512 − 10 MLP with ReLU or tanh activation functions trained on MNIST and a VGG-16 CNN trained on CIFAR-10, respectively. Reference [11] used a NN classifier and a generative PixelCNN++ [35] as variational distributions for analysing the IP of a ResNet trained on CINIC-10.…”
Section: Variational and Neural Network-based Estimatorsmentioning
confidence: 99%
“…Another type of variational lower bound can be obtained from the Donsker-Varadhan representation of mutual information, which is the principle underlying MINE [34]. MINE was used to estimate the IPs in [12] and [15] for a 784 − 512 − 512 − 10 MLP with ReLU or tanh activation functions trained on MNIST and a VGG-16 CNN trained on CIFAR-10, respectively. Reference [11] used a NN classifier and a generative PixelCNN++ [35] as variational distributions for analysing the IP of a ResNet trained on CINIC-10.…”
Section: Variational and Neural Network-based Estimatorsmentioning
confidence: 99%
“…Recently, the IB principle has been proposed for analyzing and understanding the dynamics of learning and the generalization of DNNs [14]. [15] applies the recently proposed mutual information neural estimator (MINE) [16] to train hidden layer with IB loss, and freeze it before moving on to the next layer. Although the result corroborates partially the IB hypothesis in DNNs, some claims are still controversial [17,18].…”
Section: Ib Principle and Deep Neural Networkmentioning
confidence: 99%
“…Pairs {(x (i) , z (i) )} come from joint distribution p(x, z), while samples {ž (i) } come from marginal distribution p(z). It has been used for analyzing mutual information between layers of neural networks [32,33].…”
Section: Non-parametric Estimation Of Mutual Informationmentioning
confidence: 99%