2019
DOI: 10.48550/arxiv.1909.01377
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Deep Equilibrium Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
80
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
2

Relationship

1
8

Authors

Journals

citations
Cited by 76 publications
(82 citation statements)
references
References 0 publications
2
80
0
Order By: Relevance
“…Motivated by the surprisingly recent works [1,8,9] that employ the same transformation in each layer and still achieve competitive results with the state-of-the-art, Bai et al [2] proposed a new approach to model this process and directly computed the fixed point. To leverage ideas from DEQ, Gilton et al [10] proposed DEQ for inverse problems in imaging, which corresponds to a potentially infinite number of iteration steps in the PnP reconstruction scheme.…”
Section: Deep Equilibrium Modelsmentioning
confidence: 99%
“…Motivated by the surprisingly recent works [1,8,9] that employ the same transformation in each layer and still achieve competitive results with the state-of-the-art, Bai et al [2] proposed a new approach to model this process and directly computed the fixed point. To leverage ideas from DEQ, Gilton et al [10] proposed DEQ for inverse problems in imaging, which corresponds to a potentially infinite number of iteration steps in the PnP reconstruction scheme.…”
Section: Deep Equilibrium Modelsmentioning
confidence: 99%
“…In cases where the property of the landscape that we are trying to optimize is known and can be phrased in terms of the fixed point of some dynamical system, there can often be significant benefit to computing derivatives implicitly rather than unrolling optimization explicitly Rajeswaran et al [2019], Bai et al [2019], Blondel et al [2021]. Using implicit differentiation in this way removes conditioning issues due to unrolling and improves memory cost of computing derivatives.…”
Section: Well-behaved Proxy Objectivesmentioning
confidence: 99%
“…Therefore, as an IGNN forward pass we can iterate (2) K times, where K is sufficiently large, to compute node embeddings Y (K) ≈ Y * for use within the meta-loss from (1). For the IGNN backward pass, implicit differentiation methods (Bai et al, 2019;, carefully tailored for handling graph-based models , can but used to compute gradients of Y * with respect to W p and W x . Critically, this implicit technique does not require storing each intermediate representation {Y (k) } K k=1 from the forward pass, and hence is quite memory efficient even if K is arbitrarily large.…”
Section: Overview Of Implicit Gnnsmentioning
confidence: 99%
“…To address these issues, at least in part, two distinct strategies have recently been proposed. First, the framework of implicit deep learning (Bai et al, 2019; has be applied to producing supervised node embeddings that satisfy an equilibrium criteria instantiated through a graph-dependent fixed-point equation. The resulting so-called implicit GNN (IGNN) pipeline , and related antecedents (Dai et al, 2018;Gallicchio & Micheli, 2020), mimics the behavior of a GNN model with arbitrary depth for handling long-range dependencies, but is nonetheless trainable via a fixed memory budget and robust against oversmoothing effects.…”
Section: Introductionmentioning
confidence: 99%