Proceedings of the 2016 International Conference on Supercomputing 2016
DOI: 10.1145/2925426.2926294
|View full text |Cite
|
Sign up to set email alerts
|

Proteus

Abstract: This work exploits the tolerance of Deep Neural Networks (DNNs) to reduced precision numerical representations and specifically, their recently demonstrated ability to tolerate representations of di↵erent precision per layer while maintaining accuracy. This flexibility enables improvements over conventional DNN implementations that use a single, uniform representation. This work proposes Proteus, which reduces the data tra c and storage footprint needed by DNNs, resulting in reduced energy and improved area ef… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
3
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 56 publications
(3 citation statements)
references
References 26 publications
0
3
0
Order By: Relevance
“…Even though VS shares the benefits from recent advances in ISP [6,9,14,23,31,33,35,40,67,75,76,82,85,89,96] and neardata processing [2,17,20,39,48,50,63,80,84], these frameworks need the mechanisms that VS offers in order to execute approximate computing applications efficiently. And while using approximate computing in channel encoding [38,62] and memory controller [30] can achieve an effect similar to that of VS in terms of reducing data-movement overhead, VS is independent of these projects and requires no changes in hardware.…”
Section: Other Related Workmentioning
confidence: 99%
“…Even though VS shares the benefits from recent advances in ISP [6,9,14,23,31,33,35,40,67,75,76,82,85,89,96] and neardata processing [2,17,20,39,48,50,63,80,84], these frameworks need the mechanisms that VS offers in order to execute approximate computing applications efficiently. And while using approximate computing in channel encoding [38,62] and memory controller [30] can achieve an effect similar to that of VS in terms of reducing data-movement overhead, VS is independent of these projects and requires no changes in hardware.…”
Section: Other Related Workmentioning
confidence: 99%
“…[1] proposes a CNN accelerator design that can skip computations on input values that are zeros. [14,21] reduce an accelerator's bandwidth and buffer use. [21] uses per-layer data quantization and matrix-decomposition, whereas [14] uses perlayer numerical precision reduction.…”
Section: Related Workmentioning
confidence: 99%
“…[14,21] reduce an accelerator's bandwidth and buffer use. [21] uses per-layer data quantization and matrix-decomposition, whereas [14] uses perlayer numerical precision reduction. [2] uses a fused-layer technique to reduce bandwidth use of convolutional layers.…”
Section: Related Workmentioning
confidence: 99%