2022
DOI: 10.1007/s10957-022-02122-y
|View full text |Cite
|
Sign up to set email alerts
|

SABRINA: A Stochastic Subspace Majorization-Minimization Algorithm

Abstract: A wide class of problems involves the minimization of a coercive and differentiable function F on R N whose gradient cannot be evaluated in an exact manner. In such context, many existing convergence results from standard gradientbased optimization literature cannot be directly applied and robustness to errors in the gradient is not necessarily guaranteed. This work is dedicated to investigating the convergence of Majorization-Minimization (MM) schemes when stochastic errors affect the gradient terms. We intro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3

Relationship

1
5

Authors

Journals

citations
Cited by 8 publications
(12 citation statements)
references
References 50 publications
0
12
0
Order By: Relevance
“…This strategy has been initially introduced for full-batch MM algorithms (i.e., without any block coordinate strategy) in [17]. Convergence analysis can be found in [18,16,21,20,19] under various situations. We recently extended this strategy to cope with block coordinate updates with the form of (B2M) [33], leading to the B2MS (Block MM Subspace) scheme that we present hereafter.…”
Section: Subspace Accelerationmentioning
confidence: 99%
See 1 more Smart Citation
“…This strategy has been initially introduced for full-batch MM algorithms (i.e., without any block coordinate strategy) in [17]. Convergence analysis can be found in [18,16,21,20,19] under various situations. We recently extended this strategy to cope with block coordinate updates with the form of (B2M) [33], leading to the B2MS (Block MM Subspace) scheme that we present hereafter.…”
Section: Subspace Accelerationmentioning
confidence: 99%
“…Several choices for the subspace matrix are discussed in [17,21,19]. Intensive comparisons in the fields of inverse problems, image processing and machine learning (e.g., [34,16]), have shown the superiority of the socalled memory gradient subspace which seems to reach the best compromise between simplicity and efficiency. In the context of (B2MS), this amounts to defining, for every k P N, the memory gradient matrix D k "…”
Section: Subspace Accelerationmentioning
confidence: 99%
“…Actually, almost-sure convergence based on sequence (x k ) k∈N for stochastic gradient schemes remains scarcely studied in the non-convex case. Recent results in [25,14] proposed an alternative proof technique without KL inequality by constraining the topology of the stationary points of F .…”
Section: Sgd From [19]mentioning
confidence: 99%
“…Many effective functions satisfy the KL property, including but not limited to ∥x∥ 1 , ∥Ax − b∥ 2 , log-exp, and the logistic loss function ψ(t) = log 1 + e −t . For more examples, readers can refer to [28], [29], [30], [31], and [32](see Page 919). Therefore, the potential energy function defined in (15) can easily satisfy the KL property.…”
Section: Definitionmentioning
confidence: 99%
“…It is important to acknowledge that the implementation of the KL technique displays slight differences between stochastic and deterministic algorithms. For further details, readers are encouraged to refer to [30], [31].…”
Section: Appendix Dmentioning
confidence: 99%