2018
DOI: 10.1007/s10957-018-1298-1
|View full text |Cite
|
Sign up to set email alerts
|

Exact Worst-Case Convergence Rates of the Proximal Gradient Method for Composite Convex Minimization

Abstract: We study the worst-case convergence rates of the proximal gradient method for minimizing the sum of a smooth strongly convex function and a non-smooth convex function, whose proximal operator is available.We establish the exact worst-case convergence rates of the proximal gradient method in this setting for any step size and for different standard performance measures: objective function accuracy, distance to optimality and residual gradient norm.The proof methodology relies on recent developments in performan… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
61
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
6

Relationship

1
5

Authors

Journals

citations
Cited by 58 publications
(64 citation statements)
references
References 18 publications
3
61
0
Order By: Relevance
“…Because of this SSOS operation, our method is less dependent on the threshold (i.e., 20%) over the central surface of the k‐space than the method of, 58 who directly exploits the boldxLR images as sensitivity map information. Once the sensitivity maps were estimated, an accelerated proximal gradient method 59 was implemented to solve (11). The regularization parameter λ controls the tradeoff between data consistency and confidence in the sparsity prior, and this parameter was tuned manually over a discrete grid of values within the interval [10-7;10-4].…”
Section: Methodsmentioning
confidence: 99%
“…Because of this SSOS operation, our method is less dependent on the threshold (i.e., 20%) over the central surface of the k‐space than the method of, 58 who directly exploits the boldxLR images as sensitivity map information. Once the sensitivity maps were estimated, an accelerated proximal gradient method 59 was implemented to solve (11). The regularization parameter λ controls the tradeoff between data consistency and confidence in the sparsity prior, and this parameter was tuned manually over a discrete grid of values within the interval [10-7;10-4].…”
Section: Methodsmentioning
confidence: 99%
“…Among them, a systematic approach to lower bounds (which focuses on quadratic cases) is presented in by Arjevani et al in [1], a systematic use of control theory (via integral quadratic constraints) for developing upper bounds is presented by Lessard et al in [26], and the performance estimation approach, which aims at finding worst-case bounds was originally developed in [13] (see also surveys in [11] and [47]). Those methodologies are mostly presented as tools for performing worst-cases analyses (see the numerous examples in [11,19,47,48,50,51]), however, such techniques were also recently used to develop new methods with improved worst-case complexities. Among others, such an approach was used in [13,22] to devise a fixed-step method that attains the best possible worst-case performance for smooth convex minimization [12], and later in [14] to obtain a variant of Kelley's cutting plane method with the best possible worst-case guarantee for non-smooth convex minimization.…”
Section: Links With Systematic and Computer-assisted Approaches To Womentioning
confidence: 99%
“…with τ ≥ 0 being as small as possible. Note that the presented analysis allows for more general initial conditions and performance measures (see discussions in [31,Section 4], and more specifically in [51,Section 4] in the context of performance estimation), however, for the sake of simplicity we do not pursue this direction in this work. We start the analysis of GFOM with the observation that, under the assumptions discussed above, the worst-case performance of GFOM is by definition the optimal value to the following performance estimation problem (PEP):…”
Section: Estimating the Worst-case Performance Of Gfommentioning
confidence: 99%
“…There is a host of results in the literature using SDPs to analyze the convergence of first-order optimization algorithms [10,18,28,29]. The first among them is [10], in which Drori and Teboulle developed an SDP to derive analytical/numerical bounds on the worst-case performance of the unconstrained gradient method and its accelerated variant.…”
Section: Related Workmentioning
confidence: 99%
“…The first among them is [10], in which Drori and Teboulle developed an SDP to derive analytical/numerical bounds on the worst-case performance of the unconstrained gradient method and its accelerated variant. An extension of this framework to the proximal gradient method-for the case of strongly convex problems-has been recently proposed in [28]. These SDP formulations, despite being able to yield new performance bounds, are highly algorithm dependent.…”
Section: Related Workmentioning
confidence: 99%