2021
DOI: 10.48550/arxiv.2109.01262
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On the Accuracy of Analog Neural Network Inference Accelerators

Abstract: Specialized accelerators have recently garnered attention as a method to reduce the power consumption of neural network inference. A promising category of accelerators utilizes nonvolatile memory arrays to both store weights and perform in situ analog computation inside the array. While prior work has explored the design space of analog accelerators to optimize performance and energy efficiency, there is seldom a rigorous evaluation of the accuracy of these accelerators. This work shows how architectural desig… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 44 publications
(105 reference statements)
0
1
0
Order By: Relevance
“…As most neural network weights are low significance (low-conductance), the hardware mapping for this scheme is natural and could lead to low overall error profiles at inference stage. [22] Meanwhile, the range clipping would not be a major detriment to learning performance, since a compressed range of 6-8 bits writable space is more than sufficient for most online learning applications using emerging non-volatile memory devices [23] even when considering write noise in the loop. [24] Endurance: The channel conductance can be repeatedly cycled using 10 6 switching pulses without altering switching characteristics (Section S5, Supporting Information).…”
Section: Long-term Synaptic Plasticitymentioning
confidence: 99%
“…As most neural network weights are low significance (low-conductance), the hardware mapping for this scheme is natural and could lead to low overall error profiles at inference stage. [22] Meanwhile, the range clipping would not be a major detriment to learning performance, since a compressed range of 6-8 bits writable space is more than sufficient for most online learning applications using emerging non-volatile memory devices [23] even when considering write noise in the loop. [24] Endurance: The channel conductance can be repeatedly cycled using 10 6 switching pulses without altering switching characteristics (Section S5, Supporting Information).…”
Section: Long-term Synaptic Plasticitymentioning
confidence: 99%