2021
DOI: 10.48550/arxiv.2112.06074
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Early Stopping for Deep Image Prior

Abstract: Deep image prior (DIP) and its variants have showed remarkable potential for solving inverse problems in computer vision, without any extra training data. Practical DIP models are often substantially overparameterized. During the fitting process, these models learn mostly the desired visual content first, and then pick up the potential modeling and observational noise, i.e., overfitting. Thus, the practicality of DIP often depends critically on good early stopping (ES) that captures the transition period. In t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 9 publications
(9 citation statements)
references
References 27 publications
1
8
0
Order By: Relevance
“…Similar phenomena have been empirically observed in many other nonconvex problems, where vanilla gradient descent, when coupled with small random initialization (SRI) and early stopping (ES), has good generalization performance even with overpametrization due to the algorithmic regularization effect of SRI annd ES [Woodworth et al, 2020, Ghorbani et al, 2020, Prechelt, 1998, Wang et al, 2021, Li et al, 2018, Stöger and Soltanolkotabi, 2021. This motivates us to study the following question: What is the general behavior of the gradient descent dynamic (GD-M) coupled with SRI and ES?…”
Section: Introductionsupporting
confidence: 59%
“…Similar phenomena have been empirically observed in many other nonconvex problems, where vanilla gradient descent, when coupled with small random initialization (SRI) and early stopping (ES), has good generalization performance even with overpametrization due to the algorithmic regularization effect of SRI annd ES [Woodworth et al, 2020, Ghorbani et al, 2020, Prechelt, 1998, Wang et al, 2021, Li et al, 2018, Stöger and Soltanolkotabi, 2021. This motivates us to study the following question: What is the general behavior of the gradient descent dynamic (GD-M) coupled with SRI and ES?…”
Section: Introductionsupporting
confidence: 59%
“…For the reasons just mentioned, the speed of convergence as a function of performance is essential. Recent works have attempted to study convergence speed from the point of view of the optimizer used, while others have tried to propose ad hoc Early Stopping techniques [25]. In this context, this paper aims to study the interpretability of this model through the various initializations in the literature.…”
Section: Discussionmentioning
confidence: 99%
“…DIP often uses early-stopping to deliver a satisfactory reconstruction: the update of θ is stopped early to avoid overfitting to the noise [10]. This has motivated developing automated rules for early stopping [24], [25].…”
Section: Deep Image Priormentioning
confidence: 99%
“…Jo et al [24] propose to penalize the complexity of the reconstruction using Stein's unbiased risk estimator. See also [25] for a stopping criterion based on monitoring the running variance of iterate sequence and references therein for further discussions. [6] suggests at test-time to start optimizing a randomly initialized DIP to match a reconstruction, produced by another method.…”
Section: Deep Image Priormentioning
confidence: 99%