2018
DOI: 10.3390/e20110823
|View full text |Cite
|
Sign up to set email alerts
|

Dissecting Deep Learning Networks—Visualizing Mutual Information

Abstract: Deep Learning (DL) networks are recent revolutionary developments in artificial intelligence research. Typical networks are stacked by groups of layers that are further composed of many convolutional kernels or neurons. In network design, many hyper-parameters need to be defined heuristically before training in order to achieve high cross-validation accuracies. However, accuracy evaluation from the output layer alone is not sufficient to specify the roles of the hidden units in associated networks. This result… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
13
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 12 publications
(14 citation statements)
references
References 32 publications
1
13
0
Order By: Relevance
“…We will refer to them in the following. TE has been used for the quantification of effective connectivity between neurons [10,11,12,13]. To the extent of our knowledge, the work in [14,15] represent the only attempts to use TE for improving the learning capability of neural networks.…”
Section: Arxiv:210414616v1 [Cslg] 29 Apr 2021mentioning
confidence: 99%
“…We will refer to them in the following. TE has been used for the quantification of effective connectivity between neurons [10,11,12,13]. To the extent of our knowledge, the work in [14,15] represent the only attempts to use TE for improving the learning capability of neural networks.…”
Section: Arxiv:210414616v1 [Cslg] 29 Apr 2021mentioning
confidence: 99%
“…To further understand whether our Level 1 architecture learns to differentiate the robot from the environment, we computed the Mutual Information [24] for each group's train dataset (Table I). Our objective is to measure and compare if four Level 1 trained architectures have a degree of similar knowledge that it is invariant to the training set.…”
Section: Resultsmentioning
confidence: 99%
“…This paper presents significant improvements over the previous version, significant change to the LR policy to enable it to build on the well understood convergence properties of decaying LR SGD and significantly better experimental results. The use of MI as a metric for LR adaptation is intended to lead to further work towards a deeper understanding of DNNs [ 19 ] and learning of DNNs through maximizing Mutual Information [ 16 , 20 ].…”
Section: Related Workmentioning
confidence: 99%