2021
DOI: 10.48550/arxiv.2111.00032
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Parallel-and-stream accelerator for computationally fast supervised learning

Abstract: Two dominant distributed computing strategies have emerged to overcome the computational bottleneck of supervised learning with big data: parallel data processing in the MapReduce paradigm and serial data processing in the online streaming paradigm. Despite the two strategies' common divide-and-combine approach, they differ in how they aggregate information, leading to different trade-offs between statistical and computational performance. In this paper, we propose a new hybrid paradigm, termed a Parallel-and-… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 14 publications
0
1
0
Order By: Relevance
“…Lin et al (2021) studied a homogenization strategy for heterogeneous streaming data. Hector et al (2021) proposed a new big data learning method by seamlessly integrating parallel data processing and online streaming paradigm. proposed an online debiased lasso method for high-dimensional generalized linear models with streaming data.…”
Section: Introductionmentioning
confidence: 99%
“…Lin et al (2021) studied a homogenization strategy for heterogeneous streaming data. Hector et al (2021) proposed a new big data learning method by seamlessly integrating parallel data processing and online streaming paradigm. proposed an online debiased lasso method for high-dimensional generalized linear models with streaming data.…”
Section: Introductionmentioning
confidence: 99%