Proceedings of the 52nd Annual IEEE/ACM International Symposium on Microarchitecture 2019
DOI: 10.1145/3352460.3358309
|View full text |Cite
|
Sign up to set email alerts
|

Neuron-Level Fuzzy Memoization in RNNs

Abstract: Recurrent Neural Networks (RNNs) are a key technology for applications such as automatic speech recognition or machine translation. Unlike conventional feed-forward DNNs, RNNs remember past information to improve the accuracy of future predictions and, therefore, they are very effective for sequence processing problems. For each application run, each recurrent layer is executed many times for processing a potentially large sequence of inputs (words, images, audio frames, etc.). In this paper, we make the obser… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
16
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 17 publications
(18 citation statements)
references
References 26 publications
2
16
0
Order By: Relevance
“…Even though most neurons exhibit a high correlation, a significant number of neurons have moderate or even low correlations. This observation is consistent with the observations made by Anderson et al [18] and more recently, Silfa et al [9]. A predictor based on 1-bit weights for a neuron with a low self-correlation coefficient is expected to make frequent mistakes, and consequently, reduce the overall accuracy of the DNN.…”
Section: Exploiting Self-correlationsupporting
confidence: 92%
See 2 more Smart Citations
“…Even though most neurons exhibit a high correlation, a significant number of neurons have moderate or even low correlations. This observation is consistent with the observations made by Anderson et al [18] and more recently, Silfa et al [9]. A predictor based on 1-bit weights for a neuron with a low self-correlation coefficient is expected to make frequent mistakes, and consequently, reduce the overall accuracy of the DNN.…”
Section: Exploiting Self-correlationsupporting
confidence: 92%
“…First, the dot-product between 1-bit valued vectors does not require multipliers, simplifying the hardware by a large extent. Second, since the 1-bit weights are obtained from the sign bits of the full precision weights [9], they do not incur in any memory footprint overhead since they do not have to be stored separately, but can be obtained directly from the full precision weights.…”
Section: Exploiting Self-correlationmentioning
confidence: 99%
See 1 more Smart Citation
“…The memoization technique has been studied in many fields. Franyell et al [71] propose a fuzzy memoization scheme that avoids more than 24.2% of computation for RNN training. Liu et al [44] replace a long sequence of instructions with a two-level lookup table.…”
Section: Related Workmentioning
confidence: 99%
“…DeltaRNN [20] and the work in [21] exploit temporal coherency of the LSTM data to reuse computations and avoid redundant memory accesses. Work in [22] improves RNN energy efficiency by skipping computations, since previous calculations are memoized and reused in future evaluations. We focus on changing the bit-width at runtime, and computations are never skipped.…”
Section: Related Workmentioning
confidence: 99%