2020 53rd Annual IEEE/ACM International Symposium on Microarchitecture (MICRO) 2020
DOI: 10.1109/micro50266.2020.00032
|View full text |Cite
|
Sign up to set email alerts
|

Non-Blocking Simultaneous Multithreading: Embracing the Resiliency of Deep Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3

Relationship

2
5

Authors

Journals

citations
Cited by 11 publications
(11 citation statements)
references
References 42 publications
0
11
0
Order By: Relevance
“…SySMT [28] leverages sparsity in quantization of both activations and weights to 4 bits. Their method incurs relatively high area overheads, since the quantization logic has to be scaled with the number of processing units.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…SySMT [28] leverages sparsity in quantization of both activations and weights to 4 bits. Their method incurs relatively high area overheads, since the quantization logic has to be scaled with the number of processing units.…”
Section: Related Workmentioning
confidence: 99%
“…Also, consider a single MAC unit that computes a single activation-weight multiplication per cycle. vSPARQ, similar to [28,30], groups activations in pairs, to leverage the dynamic and unstructured activation sparsity. That is, the DP calculations can be formulated as:…”
Section: Vsparq: Leveraging Sparsity With Pairs Of Activationsmentioning
confidence: 99%
See 1 more Smart Citation
“…Inspired by conventional SMT, Shomron and Weiser [18] propose non-blocking SMT (NB-SMT) designated for deep neural networks (DNNs). NB-SMT mitigates hardware underutiliation of DNNs caused by unstructured sparsity.…”
Section: Introductionmentioning
confidence: 99%
“…The importance of recalibrating the BatchNorm layers due to an external corrupted dataset or internal perturbations in activations and/or weights has been discussed in other works as well. Schneider et al [16] show how BatchNorm recalibration can improve model robustness of vision models to image corruptions (e.g., blurring and compression artifacts); Tsai et al [25] propose to recalibrate the BatchNorm layers in the scenario of noise in analog accelerators; Shomron et al [20] recalibrate the BatchNorm layers to redeem some of the accuracy degradation due to zero-valued activation mispredictions; and Hubara et al [9], as well as Shomron and Weiser [18], suggest post-quantization BatchNorm recalibration. Other works [4,10,21,22] tackle the problem of mitigating BatchNorm training and inference discrepancy due to relatively small batch sizes.…”
Section: Introductionmentioning
confidence: 99%