2015
DOI: 10.48550/arxiv.1501.06598
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Online Nonparametric Regression with General Loss Functions

Abstract: This paper establishes minimax rates for online regression with arbitrary classes of functions and general losses. 1 We show that below a certain threshold for the complexity of the function class, the minimax rates depend on both the curvature of the loss function and the sequential complexities of the class. Above this threshold, the curvature of the loss does not affect the rates. Furthermore, for the case of square loss, our results point to the interesting phenomenon: whenever sequential and i.i.d. empiri… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
6

Relationship

1
5

Authors

Journals

citations
Cited by 7 publications
(8 citation statements)
references
References 20 publications
0
8
0
Order By: Relevance
“…A full proof is in Appendix C. Combining Theorem 3 with [HRS21, Theorem 3.2], we have a complete characterization of the statistical rates of smooth online learning. As a final remark, we note that our proof applied to the results of [RS15] immediately extends to nonconstructively show that the dependence of the minimax value on T for squared loss are the expected "fast rates" from [RST17]. Unfortunately, our efficient algorithms below do not provably achieve these rates; resolving this disparity is an interesting future direction with applications to the study of contextual bandits, as described in Appendix A.…”
Section: Minimax Valuementioning
confidence: 82%
See 1 more Smart Citation
“…A full proof is in Appendix C. Combining Theorem 3 with [HRS21, Theorem 3.2], we have a complete characterization of the statistical rates of smooth online learning. As a final remark, we note that our proof applied to the results of [RS15] immediately extends to nonconstructively show that the dependence of the minimax value on T for squared loss are the expected "fast rates" from [RST17]. Unfortunately, our efficient algorithms below do not provably achieve these rates; resolving this disparity is an interesting future direction with applications to the study of contextual bandits, as described in Appendix A.…”
Section: Minimax Valuementioning
confidence: 82%
“…This characterization was extended to the case of constrained adversaries (including the special case of smoothed adversaries) in [RST11]. Several subsequent papers have established further refined regret bounds [BDR21,RS15]. The profusion of publications relating to algorithmic questions about online learning is too large to enumerate here, but notable relevant work includes [HK16], which provides lower bounds on oracle-efficiency and [RSS12] which introduces a general framework for constructing algorithms.…”
Section: B Related Workmentioning
confidence: 99%
“…Here, the gap between the PAC-misspecified and worst-case is very easy to demonstrate (take Y i = 1{X i ≤ θ} with θ ∈ [0, 1] -the 1D-barrier -which cannot be predicted, Γ n = ∞, in the worst case, but is easy in the iid case). For increasingly more general losses, [24,21,22] show that regret can be sharply characterized by the metric-entropy type quantites (sequential Rademacher complexity). However, for the log-loss it turns out that the entropic characterization is not possible, cf.…”
Section: Motivation and Literaturementioning
confidence: 99%
“…As for the applications of the chaining idea in online learning, the log loss was studied in [OH99,CBL01], and [RS14,RST17] considered the square loss. For general loss functions, the online nonparametric regression problem was studied in [GG15,RS15], and [CBGGG17] considered general contextual bandits with full or censored feedbacks. However, a key distinguishing factor between the previous works and ours is that the loss (or reward) function in first-price auctions is not continuous, resulting in a potentially large performance gap among the children of a single internal node in the chain and rendering the previous arguments inapplicable.…”
Section: Related Workmentioning
confidence: 99%