2018
DOI: 10.1587/elex.15.20180212
|View full text |Cite
|
Sign up to set email alerts
|

EERA-DNN: An energy-efficient reconfigurable architecture for DNNs with hybrid bit-width and logarithmic multiplier

Abstract: Abstract:This paper proposes an energy-efficient reconfigurable architecture for deep neural networks (EERA-DNN) with hybrid bitwidth and logarithmic multiplier. To speed up the computing and achieve high energy efficiency, we first propose an efficient network compression method with hybrid bit-width weights scheme, that saves the memory storage of network LeNet, AlexNet and EESEN by 7x-8x with negligible accuracy loss. Then, we propose an approximate unfolded logarithmic multiplier to process the multiplicat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
1

Relationship

2
4

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 21 publications
0
5
0
Order By: Relevance
“…Therefore, we can use approximate computing units with reduced power consumption to replace the traditional standard computing units adopted in DNNs. In our previous work [10], [11] and [13], we have proposed three digital approximate multiplication unit architectures to reduce the DNN computing power consumption. The approximate multiplication units can be dynamically reconfigured and adaptive to different accuracy requirements.…”
Section: B Approximate Computing For Dnnsmentioning
confidence: 99%
“…Therefore, we can use approximate computing units with reduced power consumption to replace the traditional standard computing units adopted in DNNs. In our previous work [10], [11] and [13], we have proposed three digital approximate multiplication unit architectures to reduce the DNN computing power consumption. The approximate multiplication units can be dynamically reconfigured and adaptive to different accuracy requirements.…”
Section: B Approximate Computing For Dnnsmentioning
confidence: 99%
“…The conventional DNN optimization methods are pruning, encoding and quantization, which are discussed in work [8]- [10]. In our previous work [11] and [12], we proposed several compression methods with hybrid bit-width weights scheme, which can save the memory storage of the typical DNN networks, LeNet, AlexNet and EESEN by 7x∼8x. However, for KWS systems, where the adopted DNNs are typically compact networks customized for specific scenarios, these conventional network compression approaches with pruning and encoding, are likely to cause great accuracy loss.…”
Section: Preliminaries a Network Optimization Approaches For Low Powe...mentioning
confidence: 99%
“…Thus approximate multiplication units are required to be adopted in DNN processing because they can significantly improve energy efficient with little cost in accuracy loss. In our previous work [18] and [11], we have proposed two digital approximate multiplication unit architectures to reduce the DNN computing power consumption. These two approximate multiplication units are customized for DNNs based on the iterative logarithmic multiplication principle [19].…”
Section: B Energy Efficient Approximate Computing For Customized Dnnsmentioning
confidence: 99%
“…However, NMS is a greedy algorithm that is computationally intensive and has a complexity of O(N 2 ), leading to increased processing time for a large number of detected targets. Recent many FPGA-based and ASIC edge neural network acceleration chips [7,8,9,10,11,12,13,14] such as UNPU [11], Eyeriss [12], and CASSANN-v2 [13], have been proposed to target general neural network operations (i.e., convolution). However, when deploying object detection neural networks, these chips often offload the NMS algorithm to the on-chip embedded CPU, significantly increasing the end-to-end inference time of object detection neural networks at the edge.Therefore, it is vital to develop a customized circuit to reduce the computation time of the NMS algorithm at the edge.…”
Section: Introductionmentioning
confidence: 99%