Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence 2018
DOI: 10.24963/ijcai.2018/302
|View full text |Cite
|
Sign up to set email alerts
|

Energy-efficient Amortized Inference with Cascaded Deep Classifiers

Abstract: Deep neural networks have been remarkable successful in various AI tasks but often cast high computation and energy cost for energy-constrained applications such as mobile sensing. We address this problem by proposing a novel framework that optimizes the prediction accuracy and energy cost simultaneously, thus enabling effective cost-accuracy trade-off at test time. In our framework, each data instance is pushed into a cascade of deep neural networks with increasing sizes, and a selection module is used to seq… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 12 publications
(10 citation statements)
references
References 6 publications
0
10
0
Order By: Relevance
“…3 (b)), where early features can be propagated to deep layers if needed. Based on this such architecture design, early exiting can be achieved according to confidence-based criteria [43], [48] or learned decision functions [44], [49], [50], [51]. Note that the confidence-based exiting policy consumes no extra computation during inference, while usually requiring tuning the threshold(s) on the validation set.…”
Section: Dynamic Depthmentioning
confidence: 99%
“…3 (b)), where early features can be propagated to deep layers if needed. Based on this such architecture design, early exiting can be achieved according to confidence-based criteria [43], [48] or learned decision functions [44], [49], [50], [51]. Note that the confidence-based exiting policy consumes no extra computation during inference, while usually requiring tuning the threshold(s) on the validation set.…”
Section: Dynamic Depthmentioning
confidence: 99%
“…In the Social Media, Video Monitoring, and TF Cascade pipelines, a subset of models are invoked based on the output of earlier models in the pipeline. This conditional evaluation pattern appears in bandit algorithms [3,20] used for model personalization as well as more general cascaded prediction pipelines [2,14,24,34].…”
Section: Background and Motivationmentioning
confidence: 96%
“…Dynamic inference (DI) predicts different samples using data-dependent architectures or parameters, thereby improving the inference efficiency or the model's representa- tion power [10]. Specifically, early existing methods allow samples (easy to classify) to be predicted using the early outputs of cascade DNNs [33] or networks with multiple intermediate classifiers [8]. Moreover, skipping methods selectively activate the model components, e.g., layers [9], branches [23], or sub-networks [2] conditioned on the sample.…”
Section: Related Workmentioning
confidence: 99%