2020
DOI: 10.1007/s41884-020-00038-y
|View full text |Cite
|
Sign up to set email alerts
|

On the locality of the natural gradient for learning in deep Bayesian networks

Abstract: We study the natural gradient method for learning in deep Bayesian networks, including neural networks. There are two natural geometries associated with such learning systems consisting of visible and hidden units. One geometry is related to the full system, the other one to the visible sub-system. These two geometries imply different natural gradients. In a first step, we demonstrate a great simplification of the natural gradient with respect to the first geometry, due to locality properties of the Fisher inf… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2
1
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(1 citation statement)
references
References 21 publications
0
1
0
Order By: Relevance
“…This method was first proposed by Amari [1] using the geometry induced by the Fisher-Rao metric. It is an active field of study within information geometry [3,6,10] and has been shown extremely effective in many applications [4,16,17]. More recently, also other geometries on the model have been studied, such as the Wasserstein geometry [9,11].…”
Section: Introductionmentioning
confidence: 99%
“…This method was first proposed by Amari [1] using the geometry induced by the Fisher-Rao metric. It is an active field of study within information geometry [3,6,10] and has been shown extremely effective in many applications [4,16,17]. More recently, also other geometries on the model have been studied, such as the Wasserstein geometry [9,11].…”
Section: Introductionmentioning
confidence: 99%