2019
DOI: 10.48550/arxiv.1903.10567
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

General Probabilistic Surface Optimization and Log Density Estimation

Dmitry Kopitkov,
Vadim Indelman

Abstract: Probabilistic inference, such as density estimation and distribution transformation, is a fundamental and highly important problem that needs to be solved in many different domains. Recently, a lot of research was done to solve it using Deep Learning (DL) approaches, including unnormalized and energy models, as well as Generative Adversarial Networks, where DL has shown top approximation performance. In this paper we contribute a novel algorithm family, which generalizes all above, and allows to infer differen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

1
1
0

Year Published

2019
2019
2019
2019

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 48 publications
1
1
0
Order By: Relevance
“…As well, we showed various trends of the kernel dynamics as a result of the learning rate decay, accounting for which we argue may lead to a further progress in DL theory. The considered herein optimization scenario is 2D regression, yet it is consistent with our previous numerous experiments over different unsupervised losses and architecture types in [12], where considered data dimension was ranging between 20 and 100. Likewise, beyond GD similar trends were also previously observed for a stochastic gradient-descent (SGD) and Adam optimizer.…”
Section: Discussionsupporting
confidence: 72%
See 1 more Smart Citation
“…As well, we showed various trends of the kernel dynamics as a result of the learning rate decay, accounting for which we argue may lead to a further progress in DL theory. The considered herein optimization scenario is 2D regression, yet it is consistent with our previous numerous experiments over different unsupervised losses and architecture types in [12], where considered data dimension was ranging between 20 and 100. Likewise, beyond GD similar trends were also previously observed for a stochastic gradient-descent (SGD) and Adam optimizer.…”
Section: Discussionsupporting
confidence: 72%
“…(1) The above formulation can be generalized to include unsupervised learning methods in [12] by eliminating labels Y i from the equations. Further, techniques with a model f θ (X) returning multidimensional outputs are out of scope for this paper, to simplify the formulation.…”
Section: Notationsmentioning
confidence: 99%