2021
DOI: 10.1609/aaai.v35i9.16997
|View full text |Cite
|
Sign up to set email alerts
|

HINT: Hierarchical Invertible Neural Transport for Density Estimation and Bayesian Inference

Abstract: Many recent invertible neural architectures are based on coupling block designs where variables are divided in two subsets which serve as inputs of an easily invertible (usually affine) triangular transformation. While such a transformation is invertible, its Jacobian is very sparse and thus may lack expressiveness. This work presents a simple remedy by noting that subdivision and (affine) coupling can be repeated recursively within the resulting subsets, leading to an efficiently invertible block with dense,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
3
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(3 citation statements)
references
References 11 publications
0
3
0
Order By: Relevance
“…Similarly, Behrmann et al [4] proposed invertible residual networks (i-ResNet), introducing a tractable estimation to the Jacobian log-determinant of a residual block. Other representative normalizing flow work includes hierarchical recursive coupling [23] for increasing flow expressiveness, Wavelet Flow [53] for scaling flow to ultra-high dimensional data, and Discrete Flow [48] for discrete data modeling. In this paper, we do not use normalizing flow for generative modeling, but for invertible feature transform.…”
Section: Related Workmentioning
confidence: 99%
“…Similarly, Behrmann et al [4] proposed invertible residual networks (i-ResNet), introducing a tractable estimation to the Jacobian log-determinant of a residual block. Other representative normalizing flow work includes hierarchical recursive coupling [23] for increasing flow expressiveness, Wavelet Flow [53] for scaling flow to ultra-high dimensional data, and Discrete Flow [48] for discrete data modeling. In this paper, we do not use normalizing flow for generative modeling, but for invertible feature transform.…”
Section: Related Workmentioning
confidence: 99%
“…The cINN architecture, as illustrated in Figure 1, avoids these problems [2,3,29,30,37]. It uses a di↵erent mapping system by considering the observations y in both the forward and inverse process as a condition c: f (x; c = y) = z, x = g(z; c = y) [3].…”
Section: Invertible Neural Networkmentioning
confidence: 99%
“…In this setting, the KL divergence D KL (ν X T ν U ) is approximated by the Monte Carlo method using samples drawn from ν X . The normalizing flow methods, e.g., [7,9,30,40], adopt a similar KL minimization problem, but the maps may not be triangular and are often parametrized by neural networks. By and large, the map-from-samples approach is flexible to implement, as it only requires a set of samples drawn from ν X .…”
mentioning
confidence: 99%