2019
DOI: 10.1109/tkde.2019.2937078
|View full text |Cite
|
Sign up to set email alerts
|

Distribution-Free One-Pass Learning

Abstract: In many large-scale machine learning applications, data are accumulated with time, and thus, an appropriate model should be able to update in an online paradigm. Moreover, as the whole data volume is unknown when constructing the model, it is desired to scan each data item only once with a storage independent with the data volume. It is also noteworthy that the distribution underlying may change during the data accumulation procedure. To handle such tasks, in this paper we propose DFOP, a distribution-free one… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
3

Relationship

4
4

Authors

Journals

citations
Cited by 20 publications
(7 citation statements)
references
References 44 publications
0
7
0
Order By: Relevance
“…In our simulation, we set Γ = 1, D = 2, d = 5, T = 50000, and S = 1000. Next, we employ a real-world dataset called Sulfur recovery unit (SRU) (Zhao et al, 2021b), which is a regression dataset with slowly evolving distribution changes. There are in total 10,081 data samples representing the records of gas diffusion, where the feature consists of five different chemical and physical indexes and the label is the concentration of SO 2 .…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…In our simulation, we set Γ = 1, D = 2, d = 5, T = 50000, and S = 1000. Next, we employ a real-world dataset called Sulfur recovery unit (SRU) (Zhao et al, 2021b), which is a regression dataset with slowly evolving distribution changes. There are in total 10,081 data samples representing the records of gas diffusion, where the feature consists of five different chemical and physical indexes and the label is the concentration of SO 2 .…”
Section: Methodsmentioning
confidence: 99%
“…Dynamic regret enforces the player to compete with time-varying comparators, and thus is favored in online learning in non-stationary environments (Sugiyama and Kawanabe, 2012;Zhao et al, 2021b). The notion of dynamic regret is also referred to as tracking regret or shifting regret in the prediction with expert advice setting Warmuth, 1998, 2001).…”
Section: Dynamic Regretmentioning
confidence: 99%
“…In order to have an algorithm that adapts to non-stationary data, it is common to use a forgetting factor. For the recursive least squares, [14] analyzed the effect of the forgetting factor in terms of the tracking error covariance matrix, and [15] made the tracking error analysis with the assumptions that the noise is sub-Gaussian and the parameter follows a drifting model. However, none of the analysis mentioned is done in terms of the regret, which eliminates any noise assumption.…”
Section: Arxiv:190903118v2 [Cslg] 21 Nov 2019mentioning
confidence: 99%
“…Although equipped with rich theories, the notion of regret defined in Eqn. ( 2) is not always the right objective to minimize, especially in dynamic environments where the underlying decision function can vary over time (Zhao, Cai, and Zhou 2019;Zhao et al 2019b) and there is no single fixed decision function doing well overall. To overcome this limitation, it is natural to consider a more stringent measure, i.e., dynamic regret (Hall and Willett 2013;Besbes, Gur, and Zeevi 2015;Mokhtari et al 2016;Yang et al 2016;Zhao et al 2019a), defined as the difference between the cumulative loss of the learner and that of a sequence of local minimizers:…”
Section: Dynamic Environmentsmentioning
confidence: 99%
“…Note that the standard cross-validation in the batch learning settings is not suitable here, due to inherent temporal relationships of the streaming data. Therefore, following the setup of previous works (Gama et al 2014;Zhao et al 2019b), for the data set with T instances, we select 10 different subsets with consecutive instances starting from {T /50, T/25, . .…”
Section: Settingsmentioning
confidence: 99%