2020 IEEE/ACM Symposium on Edge Computing (SEC) 2020
DOI: 10.1109/sec50012.2020.00014
|View full text |Cite
|
Sign up to set email alerts
|

FlexDNN: Input-Adaptive On-Device Deep Learning for Efficient Mobile Vision

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
24
0
1

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 47 publications
(25 citation statements)
references
References 23 publications
0
24
0
1
Order By: Relevance
“…Accuracy is the top-1 accuracy of the model execution and latency is measured as the end-to-end delay in processing an input (e.g., an image). We use GPU frequency to quantify the power of edge device, which is reasonable because there usually exists approximately linear correlations between power and frequency [7,35]. We use SR to represent the percentage of executions that satisfy the latency and energy constraints imposed by users.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Accuracy is the top-1 accuracy of the model execution and latency is measured as the end-to-end delay in processing an input (e.g., an image). We use GPU frequency to quantify the power of edge device, which is reasonable because there usually exists approximately linear correlations between power and frequency [7,35]. We use SR to represent the percentage of executions that satisfy the latency and energy constraints imposed by users.…”
Section: Methodsmentioning
confidence: 99%
“…Although above techniques reduce the overhead of DNN execution, they are not designed to optimize the performance of DNN in the presence of dynamics of input data and resource budget. Extending the static model compression approach, several efforts [7,8,27] proposed dynamic neural networks that allow selective execution to improve DNN compute efficiency. D2NN [27] optimizes dynamic resource-accuracy trade-offs, while its complicated network structure incurs significant memory overhead, making it ill-suited for resource-constrained platforms.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Initially, one needs to pick the architecture of the early-exit model. There are largely two avenues followed in the literature: i) hand-tuned end-to-end designed networks for early-exiting, such as MSDNet [30], and ii) vanilla backbone networks, enhanced with early exits along their depth [12,35,41,68]. This design choice is crucial as it later affects the capacity and the learning process of the network, with different architectures offering varying scalability potential and convergence dynamics.…”
Section: Designing the Architecturementioning
confidence: 99%
“…SPINN [42] Vision/Classification Partial inference offloading of EE-networks. FlexDNN [12] Vision/Classification Footprint overhead-aware design of EE-networks. DDI [76] Vision/Classification Combines layer/channel skipping with early exiting.…”
Section: Early-exiting Network-agnostic Techniquesmentioning
confidence: 99%