2021
DOI: 10.48550/arxiv.2112.14368
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Adaptivity and Non-stationarity: Problem-dependent Dynamic Regret for Online Convex Optimization

Abstract: We investigate online convex optimization in non-stationary environments and choose the dynamic regret as the performance measure, defined as the difference between cumulative loss incurred by the online algorithm and that of any feasible comparator sequence. Let T be the time horizon and P T be the path-length that essentially reflects the non-stationarity of environments, the state-of-theart dynamic regret is O( T (1 + P T )). Although this bound is proved to be minimax optimal for convex functions, in this … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 15 publications
0
4
0
Order By: Relevance
“…Bound on dR worker * T ({u t } t∈[T ] ): The regret bound for the worker is very similar to wellknown results (c.f. or Zhao et al (2021)). However, typically it contains a variational term V T .…”
Section: Analysis Of Dynmetagradmentioning
confidence: 98%
“…Bound on dR worker * T ({u t } t∈[T ] ): The regret bound for the worker is very similar to wellknown results (c.f. or Zhao et al (2021)). However, typically it contains a variational term V T .…”
Section: Analysis Of Dynmetagradmentioning
confidence: 98%
“…, u T ∈ X are allowed to change over time. Therefore, this measure is more attractive in non-stationary online learning (Zhao et al, 2021). Notably, the static regret can be treated as its special case with a fixed comparator, i.e., u 1 = .…”
Section: Extension To Dynamic Regret Minimizationmentioning
confidence: 99%
“…Remark 8. Our algorithm design and regret analysis follow the recent work of Zhao et al (2021), where the correction term λ x t,i − x t−1,i 2 2 in the meta-algorithm (both feedback loss and optimism) plays an important role. Technically, in this two-layer structure, to cancel the additional positive term T t=2 x t − x t−1 2 2 appearing in the derivation of σ 2 1:T and Σ 2 1:T , one needs to simultaneously exploit negative terms of the regret upper bounds in both base level and meta level as well as additional negative terms introduced by the above correction term.…”
Section: Extension To Dynamic Regret Minimizationmentioning
confidence: 99%
See 1 more Smart Citation