2020
DOI: 10.1016/j.neucom.2019.06.106
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive hedging under delayed feedback

Abstract: The article is devoted to investigating the application of hedging strategies to online expert weight allocation under delayed feedback. As the main result we develop the General Hedging algorithm G based on the exponential reweighing of experts' losses. We build the artificial probabilistic framework and use it to prove the adversarial loss bounds for the algorithm G in the delayed feedback setting. The designed algorithm G can be applied to both countable and continuous sets of experts. We also show how algo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 8 publications
(6 citation statements)
references
References 27 publications
0
6
0
Order By: Relevance
“…In addition, a variety of works highlight sample complexity and topological properties [103], [104] of these distances. We further highlight that ICNNs are gaining an increasing attention by the ML community, especially in generative modeling [161], [162]. Furthermore, ICNNs represent a good case of ML for OT.…”
Section: Conclusion and Future Directions 81 Computational Optimal Tr...mentioning
confidence: 87%
“…In addition, a variety of works highlight sample complexity and topological properties [103], [104] of these distances. We further highlight that ICNNs are gaining an increasing attention by the ML community, especially in generative modeling [161], [162]. Furthermore, ICNNs represent a good case of ML for OT.…”
Section: Conclusion and Future Directions 81 Computational Optimal Tr...mentioning
confidence: 87%
“…; F is the cumulative distribution function (CDF) associated with the underlying distribution. 1-Wasserstein distance satisfies positivedefiniteness, symmetry, and triangle inequality (Nietert et al 2022;Korotin, Selikhanovych, and Burnaev 2023;Chen et al 2023c,b;Naderializadeh et al 2021). Problem Definition.…”
Section: Preliminaries and Problem Formulationmentioning
confidence: 99%
“…Among lots of existing learner's strategies for combining experts predictions [2,4,3], the Vovk's aggregating algorithm [8] (AA for short) is typically considered to be the best. For a wide class of loss functions λ (called η-mixable for some constant η > 0) it provides the way to efficiently combine experts predictions {γ n t } N n=1 to a single γ t so that learner achieves small regret bound.…”
Section: T Domentioning
confidence: 99%