2018 IEEE Real-Time Systems Symposium (RTSS) 2018
DOI: 10.1109/rtss.2018.00020
|View full text |Cite
|
Sign up to set email alerts
|

PredJoule: A Timing-Predictable Energy Optimization Framework for Deep Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 30 publications
(7 citation statements)
references
References 31 publications
0
7
0
Order By: Relevance
“…where, , are model parameters. Using this simple, yet effective, model we can determine the effect frequency scaling on inference time and throughput 3 . Also we estimate energy per request for a certain model models at a frequency , denoted as ( ) as:…”
Section: Model Recommendationmentioning
confidence: 99%
See 1 more Smart Citation
“…where, , are model parameters. Using this simple, yet effective, model we can determine the effect frequency scaling on inference time and throughput 3 . Also we estimate energy per request for a certain model models at a frequency , denoted as ( ) as:…”
Section: Model Recommendationmentioning
confidence: 99%
“…We assume that these devices may run applications such as mobile AR [5] or object recognition [15] that involve running deep learning inference using a deep neural network (DNN) model. The application may impose real-time latency constraints on DNN inference processing, which requires that such processing be performed on the device or at a nearby edge node (rather than in the cloud) [3,9]. We assume that the device (or edge node) has specialized hardware in the form of an embedded edge accelerator to accelerate DNN inference.…”
Section: Introductionmentioning
confidence: 99%
“…ApNet [12] applies approximation approaches to each layer of the DNN network and makes a trade-off between accuracy and latency. PredJoule [15] optimizes energy for running DNN workloads. It adjusts power configuration based on the latency of workloads.…”
Section: Related Workmentioning
confidence: 99%
“…Research in autonomous driving is an active field, with many papers published on important research points, scheduling layers of DNNs [12], [14], [15], scheduling memory allocation of DNNs [13], applying real-time scheduling [36], supported with micro-services [40], heterogeneity study [26], [39], and so on. This study builds on the many prior research efforts, but has a very different objective and level of focus.…”
Section: Introductionmentioning
confidence: 99%
“…In the context of mobile devices, Lane et al 19 proposed two runtime algorithms to decompose a DNN model across available processors with the purpose of improving performance and energy‐efficiency. Very recently, a similar purpose has been pursued by Kang and Chung, 20 and Bateni et al 21 Hong et al 22 presented an extended synchronous dataflow model aimed at explicitly expressing the parallelism of loop structures, allowing to model the computational graph of a DNN during the training phase. Casini et al 23 proposed approaches for bounding the worst‐case response time of parallel tasks implemented with thread pools, using a task model inspired by Tensorflow.…”
Section: Related Workmentioning
confidence: 99%