2022
DOI: 10.1109/tcad.2022.3197522
|View full text |Cite
|
Sign up to set email alerts
|

Energy-Efficient DNN Inference on Approximate Accelerators Through Formal Property Exploration

Abstract: Deep Neural Networks (DNNs) are being heavily utilized in modern applications and are putting energy-constraint devices to the test. To bypass high energy consumption issues, approximate computing has been employed in DNN accelerators to balance out the accuracy-energy reduction trade-off. However, the approximation-induced accuracy loss can be very high and drastically degrade the performance of the DNN. Therefore, there is a need for a fine-grain mechanism that would assign specific DNN operations to approxi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 39 publications
0
2
0
Order By: Relevance
“…In the case of channel pruning, there are various methods for retraining the pruned model to recover the lost accuracy, such as fine-tuning the model on the pruned architecture or using knowledge distillation to transfer the knowledge from the original model to the pruned model [ 27 ]. However, these methods can be computationally expensive and time-consuming, and they may not be feasible or practical in all scenarios [ 18 ]. On the contrary, the proposed method does not involve retraining or fine-tuning; therefore, to keep evaluations fair, we avoid directly comparing our results with other pruning methods that do involve such techniques.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…In the case of channel pruning, there are various methods for retraining the pruned model to recover the lost accuracy, such as fine-tuning the model on the pruned architecture or using knowledge distillation to transfer the knowledge from the original model to the pruned model [ 27 ]. However, these methods can be computationally expensive and time-consuming, and they may not be feasible or practical in all scenarios [ 18 ]. On the contrary, the proposed method does not involve retraining or fine-tuning; therefore, to keep evaluations fair, we avoid directly comparing our results with other pruning methods that do involve such techniques.…”
Section: Discussionmentioning
confidence: 99%
“…We combine the approaches presented in Section 2.1 and Section 2.2 to achieve lightweight DNN inference without compromises in terms of accuracy at run-time. We consider run-time scenarios where real-time performance is prioritized [ 18 ] and therefore target DNN inference where each image is processed independently rather than in big batches.…”
Section: Methodsmentioning
confidence: 99%