2023
DOI: 10.26599/bdma.2022.9020014
|View full text |Cite
|
Sign up to set email alerts
|

Survey of Distributed Computing Frameworks for Supporting Big Data Analysis

Abstract: Distributed computing frameworks are the fundamental component of distributed computing systems.They provide an essential way to support the efficient processing of big data on clusters or cloud. The size of big data increases at a pace that is faster than the increase in the big data processing capacity of clusters. Thus, distributed computing frameworks based on the MapReduce computing model are not adequate to support big data analysis tasks which often require running complex analytical algorithms on extre… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(2 citation statements)
references
References 94 publications
(80 reference statements)
0
2
0
Order By: Relevance
“…Its underlying core is Hadoop Distributed File System (HDFS) and the MapReduce framework, each of which has different functions. Sun et al (2023) analyzed three challenges in a distributed computing framework based on the MapReduce computing model: 1) computational inefficiency due to high I/O and communication costs, 2) inability to scale to big data due to memory constraints, and 3) inability to scale to big data due to the fact that many serial algorithms cannot be implemented in the MapReduce programming model. They finally proposed a non-MapReduce distributed computing framework, which has the potential to overcome the challenges of big data analytics.…”
Section: Research Methodology and Solutionmentioning
confidence: 99%
“…Its underlying core is Hadoop Distributed File System (HDFS) and the MapReduce framework, each of which has different functions. Sun et al (2023) analyzed three challenges in a distributed computing framework based on the MapReduce computing model: 1) computational inefficiency due to high I/O and communication costs, 2) inability to scale to big data due to memory constraints, and 3) inability to scale to big data due to the fact that many serial algorithms cannot be implemented in the MapReduce programming model. They finally proposed a non-MapReduce distributed computing framework, which has the potential to overcome the challenges of big data analytics.…”
Section: Research Methodology and Solutionmentioning
confidence: 99%
“…That is, some are busy, and some are idle, but the distribution and merging of data are imperceptible to other editors. Borrowing the thinking of MapReduce distributed/parallel computing [6], there are other editors who can queue up to enter the editing area [7]. The concurrent editing framework for video streaming is shown in Figure 1.…”
Section: Introductionmentioning
confidence: 99%