2021
DOI: 10.48550/arxiv.2110.06021
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Embedded-model flows: Combining the inductive biases of model-free deep learning and explicit probabilistic modeling

Abstract: Normalizing flows have shown great success as general-purpose density estimators. However, many real world applications require the use of domain-specific knowledge, which normalizing flows cannot readily incorporate. We propose embedded-model flows (EMF), which alternate general-purpose transformations with structured layers that embed domain-specific inductive biases. These layers are automatically constructed by converting user-specified differentiable probabilistic models into equivalent bijective transfor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 23 publications
(27 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?