2018
DOI: 10.1109/jssc.2017.2778702
|View full text |Cite
|
Sign up to set email alerts
|

BRein Memory: A Single-Chip Binary/Ternary Reconfigurable in-Memory Deep Neural Network Accelerator Achieving 1.4 TOPS at 0.6 W

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
62
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 136 publications
(62 citation statements)
references
References 10 publications
0
62
0
Order By: Relevance
“…YodaNN [34] targets binary-weight networks and reaches energy efficiency up to 61 Top/s/W. Other accelerators exploit extreme quantization for the deployment of binary neural networks on silicon using in-or near-memory computing techniques (e.g., Brein [35], Conv-RAM [36]) with energy efficiencies in the range 20-55 Top/s/W. Such high energy efficiency and throughput achievable using ASIC accelerators are counterbalanced by limited flexibility, being application specific, which makes them unattractive to satisfy fully the flexibility demand of IoT edge nodes.…”
Section: Related Workmentioning
confidence: 99%
“…YodaNN [34] targets binary-weight networks and reaches energy efficiency up to 61 Top/s/W. Other accelerators exploit extreme quantization for the deployment of binary neural networks on silicon using in-or near-memory computing techniques (e.g., Brein [35], Conv-RAM [36]) with energy efficiencies in the range 20-55 Top/s/W. Such high energy efficiency and throughput achievable using ASIC accelerators are counterbalanced by limited flexibility, being application specific, which makes them unattractive to satisfy fully the flexibility demand of IoT edge nodes.…”
Section: Related Workmentioning
confidence: 99%
“…One of the first architectures to exploit these peculiarities has been FINN [28], which is able to reach more than 200 Gop/s/W on a Xilinx FPGA, vastly outperforming the state-of-the-art for FPGA-based deep inference accelerators. Recent efforts for the deployment of binary neural networks on silicon, such as BRein [29], XNOR-POP [30], Conv-RAM [31] and Khwa et al [32] have mainly targeted in-memory computing, with energy efficiencies in the range 20-55 Top/s/W. However, the advantage of this methodology is not yet clear, as more "traditional" ASICs such as UNPU [33] and XNORBIN [34] can reach a similar level of efficiency of 50-100 Top/s/W.…”
Section: Related Workmentioning
confidence: 99%
“…In Table IV, we compare the 256 × 256 PPAC with existing hardware accelerators that have been specialized for binarized neural network (BNN) inference and support fully-connected layers [6], [10], [19], [23], [24]. We compare against these designs as their operation closely resembles that of PPAC's 1-bit {±1} MVP operation mode.…”
Section: B Comparison With Existing Acceleratorsmentioning
confidence: 99%