2019
DOI: 10.1109/access.2019.2907018
|View full text |Cite
|
Sign up to set email alerts
|

An Energy Efficiency Optimization and Control Model for Hadoop Clusters

Abstract: The majority of large-scale data intensive applications designed by MapReduce model are deployed and executed on a large-scale distributed Hadoop system. Running such application on a large cluster requires a large amount of energy. Therefore improving energy efficiency and minimizing energy consumption when executing each MapReduce job is a critical concern for data centers. We propose a control model based on model prediction control (MPC) for improving energy efficiency of Hadoop cluster while satisfying pe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
8
1

Relationship

1
8

Authors

Journals

citations
Cited by 13 publications
(5 citation statements)
references
References 39 publications
0
5
0
Order By: Relevance
“…While Figure 1a is obtained by adding two real operational data points to the initial synthetic data set, Figure 1b is obtained by adding five operational data points. 4 Importing a larger number of operational data points, one would expect that the prediction curve gets closer to the expected values curve. However, comparing the two graphs, it is evident that adding new points, instead of enhancing the ML model prediction, actually makes the ML output worse.…”
Section: Background and Motivationsmentioning
confidence: 99%
See 1 more Smart Citation
“…While Figure 1a is obtained by adding two real operational data points to the initial synthetic data set, Figure 1b is obtained by adding five operational data points. 4 Importing a larger number of operational data points, one would expect that the prediction curve gets closer to the expected values curve. However, comparing the two graphs, it is evident that adding new points, instead of enhancing the ML model prediction, actually makes the ML output worse.…”
Section: Background and Motivationsmentioning
confidence: 99%
“…Most of big data applications implemented today are based on the MapReduce (MR) programming model, which is the most common adopted solution. MapReduce-based systems have risen as a scalable and cost effective solution for massively parallel data processing [3] The majority of large-scale data intensive applications designed by MR model are deployed and executed on a large-scale distributed Hadoop system [4].…”
Section: Introductionmentioning
confidence: 99%
“…Wang and Cao [23] also supported the use of naïve Bayes in their clinical experiments to understand the neuroimaging framework of humans. The research involves different stimulus phases, but the data collection and classification presented a highly intelligent insight into how data classification can be managed if performed correctly.…”
Section: Classification Approachesmentioning
confidence: 88%
“…In this subsection the real-world dataset was used to evaluate performance of HDDM. This real-world workload is from real-time computing by taxi trajectories datasets generated by 33000 taxis over a period of six months [29,30]. The taxi trajectory is a sequence of GPS points pertaining to a trip.…”
Section: E Real-world Workloads Analysismentioning
confidence: 99%