2005
DOI: 10.1007/11503415_13
|View full text |Cite
|
Sign up to set email alerts
|

The Weak Aggregating Algorithm and Weak Mixability

Abstract: Abstract. This paper resolves the problem of predicting as well as the best expert up to an additive term o(n), where n is the length of a sequence of letters from a finite alphabet. For the bounded games the paper introduces the Weak Aggregating Algorithm that allows us to obtain additive terms of the form C √ n. A modification of the Weak Aggregating Algorithm that covers unbounded games is also described.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2005
2005
2015
2015

Publication Types

Select...
5
1

Relationship

2
4

Authors

Journals

citations
Cited by 10 publications
(14 citation statements)
references
References 4 publications
0
14
0
Order By: Relevance
“…Remark 2 In fact, the second difficulty is more apparent than real: for example, in the binary case (Y = {0, 1}) with the loss function λ(γ, y) independent of x, there are many non-trivial continuous prediction rules in the canonical form of the prediction game [45] with the prediction set redefined as the boundary of the set of superpredictions [19].…”
Section: Universal Consistency For Randomized Prediction Algorithmsmentioning
confidence: 99%
See 1 more Smart Citation
“…Remark 2 In fact, the second difficulty is more apparent than real: for example, in the binary case (Y = {0, 1}) with the loss function λ(γ, y) independent of x, there are many non-trivial continuous prediction rules in the canonical form of the prediction game [45] with the prediction set redefined as the boundary of the set of superpredictions [19].…”
Section: Universal Consistency For Randomized Prediction Algorithmsmentioning
confidence: 99%
“…The second addend on the right-hand side of (19) tends to zero by the continuity of the mapping Q ∈ P(Y) → Y f (y)Q(dy) for a continuous f ( [7], III.4.2, Proposition 6).…”
Section: Remark 4 Another Popular Notion Of the Integral For Vector-vmentioning
confidence: 99%
“…An especially important class of loss functions is that of "mixable" ones, for which the learner's loss can be made as small as the best expert's loss plus a constant (depending on the number of experts). It is known (Haussler et al, 1998;Vovk, 1998) that the optimal additive constant is attained by the "strong aggregating algorithm" proposed in Vovk (1990) (we use the adjective "strong" to distinguish it from the "weak aggregating algorithm" of Kalnishkan & Vyugin, 2005). …”
Section: Introductionmentioning
confidence: 99%
“…This concept has long 2 In [KV08] and other earlier papers it was required that for every γ0 ∈ Γ such that λ(ω * , γ0) = +∞ for some ω * ∈ Ω there should be a sequence γ1, γ2, . .…”
Section: Generalised Entropiesmentioning
confidence: 99%
“…In order to get an effective version of Proposition 1, one needs to restate results of [KV08] in an effective fashion. The procedures used in [KV08] are essentially effective (and efficient) but require certain properties of Γ and λ; otherwise the prediction space and the loss function can be distorted in such a way as to make the procedures from [KV08] unusable. Formalising these properties in a simple form appears to be a difficult task.…”
Section: Computable Gamesmentioning
confidence: 99%