2018 ACM/IEEE 45th Annual International Symposium on Computer Architecture (ISCA) 2018
DOI: 10.1109/isca.2018.00016
|View full text |Cite
|
Sign up to set email alerts
|

Computation Reuse in DNNs by Exploiting Input Similarity

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
48
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 95 publications
(49 citation statements)
references
References 28 publications
1
48
0
Order By: Relevance
“…An analysis of the output results reveals that many neurons produce very similar outputs for consecutive elements in the input sequence. On average, the relative difference between the current and previous output of a neuron is smaller than 23% in our set of RNNs, whereas previous work in [28] has reported similar results. Since RNNs are inherently error tolerant [36], we propose to exploit the aforementioned property to save computations by using a neuron-level fuzzy memoization scheme.…”
Section: Introductionsupporting
confidence: 60%
See 1 more Smart Citation
“…An analysis of the output results reveals that many neurons produce very similar outputs for consecutive elements in the input sequence. On average, the relative difference between the current and previous output of a neuron is smaller than 23% in our set of RNNs, whereas previous work in [28] has reported similar results. Since RNNs are inherently error tolerant [36], we propose to exploit the aforementioned property to save computations by using a neuron-level fuzzy memoization scheme.…”
Section: Introductionsupporting
confidence: 60%
“…Note that RNNs are used in sequence processing problems such as speech recognition or video processing, where RNN inputs in consecutive time steps tend to be extremely similar. Prior work in [28] reports high similarity across consecutive frames of audio or video. Not surprisingly, our own numbers for our set of RNNs also support this claim.…”
Section: Rnns Redundancymentioning
confidence: 99%
“…Cross-frame optimization. It is a new research direction to reduce CNN computation by exploiting temporal redundancy [8,67] or input similarity [50] across video or audio frames. Moreover, this concept has also been applied to compensate the unreliability brought by pruning [60].…”
Section: Related Workmentioning
confidence: 99%
“…The work by Riera et al leverages the error tolerance of deep neural networks (DNN) on a hardware implementation of a reuse‐based DNN accelerator. By using the proposed accelerator, an improvement on energy utilization was observed.…”
Section: Related Workmentioning
confidence: 99%