2022
DOI: 10.1016/j.future.2022.01.004
|View full text |Cite
|
Sign up to set email alerts
|

EAIS: Energy-aware adaptive scheduling for CNN inference on high-performance GPUs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 20 publications
(7 citation statements)
references
References 16 publications
0
7
0
Order By: Relevance
“…This result is reasonable since the calculations in ML inference tasks are predefined. Meanwhile, profiling results also exhibit that the inference latency grows linearly with the batch size [9], [11], [13]. As shown in Fig.…”
Section: System Modelmentioning
confidence: 85%
See 3 more Smart Citations
“…This result is reasonable since the calculations in ML inference tasks are predefined. Meanwhile, profiling results also exhibit that the inference latency grows linearly with the batch size [9], [11], [13]. As shown in Fig.…”
Section: System Modelmentioning
confidence: 85%
“…Many researchers have conducted experiments to profile the computation latency of GPU when batch processing ML inference tasks [4], [7]- [11], [13], [18], [19]. It has been examined in [7] that the coefficient of variation (CV) of the batch processing time for image recognition is near CV=0, i.e.…”
Section: System Modelmentioning
confidence: 99%
See 2 more Smart Citations
“…Note that research investigating power and energy consumption behaviors in modern clusters due to deep learning (DL) job processing is in its early stages. Yao et al [12] propose an energy-aware DL job scheduler to reduce the energy consumption of CNN inference services on GPU devices. Their scheduler coordinates batch processing and dynamic voltage frequency scaling (DVFS) to effectively respond to workload fluctuations, achieving a 28% reduction in energy consumption compared to its competitors while meeting latency service-level objectives.…”
Section: Introductionmentioning
confidence: 99%