2021
DOI: 10.48550/arxiv.2111.15645
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Survey Descent: A Multipoint Generalization of Gradient Descent for Nonsmooth Optimization

Abstract: For strongly convex objectives that are smooth, the classical theory of gradient descent ensures linear convergence relative to the number of gradient evaluations. An analogous nonsmooth theory is challenging: even when the objective is smooth at every iterate, the corresponding local models are unstable, and traditional remedies need unpredictably many cutting planes. We instead propose a multipoint generalization of the gradient descent iteration for local optimization. While designed with general objectives… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(6 citation statements)
references
References 27 publications
0
6
0
Order By: Relevance
“…A popular class of decomposable objectives arise from pointwise maxima of smooth functions that satisfy an affine independence property. For example, this class was considered in the work of Han and Lewis [25]. As an immediate corollary of Theorem 3.3, we show that such functions satisfy Assumption A.…”
Section: (Quadratic Growth)mentioning
confidence: 69%
See 4 more Smart Citations
“…A popular class of decomposable objectives arise from pointwise maxima of smooth functions that satisfy an affine independence property. For example, this class was considered in the work of Han and Lewis [25]. As an immediate corollary of Theorem 3.3, we show that such functions satisfy Assumption A.…”
Section: (Quadratic Growth)mentioning
confidence: 69%
“…The local rate holds under a key structural assumption -Assumption A -which formalizes the concept of typical structure and mirrors the structure of the simple function considered in Section 1.3.1. While we formally describe Assumption A in Section 3, for now, we mention that it holds for max-of-smooth and properly C p decomposable functions, provided the local minimizer x is in fact a strong local minimizer that satisfies a strict complementarity condition; this class that includes the max-of-smooth setting considered in [25]. Assumption A also holds for generic linear tilts of semialgebraic functions: if f is semialgebraic, then for a full Lebesgue measure set of w ∈ R d , Assumption A holds at every local minimizer x of the tilted function f w : x → f (x) + w x.…”
Section: Main Convergence Guarantees For Ntdescentmentioning
confidence: 99%
See 3 more Smart Citations