2022
DOI: 10.1145/3527156
|View full text |Cite
|
Sign up to set email alerts
|

Hardware Approximate Techniques for Deep Neural Network Accelerators: A Survey

Abstract: Deep Neural Networks (DNNs) are very popular because of their high performance in various cognitive tasks in Machine Learning (ML). Recent advancements in DNNs have brought beyond human accuracy in many tasks, but at the cost of high computational complexity. To enable efficient execution of DNN inference, more and more research works, therefore, exploit the inherent error resilience of DNNs and employ Approximate Computing (AC) principles to address the elevated energy demands of DNN accelerators. This articl… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
14
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1

Relationship

1
7

Authors

Journals

citations
Cited by 60 publications
(14 citation statements)
references
References 111 publications
0
14
0
Order By: Relevance
“…Another aggressive resource reduction technique impacting accuracy is the use of approximate computing units to perform the required computation [282].…”
Section: ) Arithmetic Unitsmentioning
confidence: 99%
“…Another aggressive resource reduction technique impacting accuracy is the use of approximate computing units to perform the required computation [282].…”
Section: ) Arithmetic Unitsmentioning
confidence: 99%
“…Approximate computing has been heavily utilized on DNN inference [8]. Many works present mapping methodologies that balance out the computation accuracy-power consumption trade-off, and recent research has focused on the design and utilization of approximate multipliers on DNN inference.…”
Section: Related Workmentioning
confidence: 99%
“…A great amount of DNN operations can tolerate some degree of approximation [6], [7], and since the majority of DNN inference is spent on convolution and matrix multiplication operations, the design of approximate MAC units has attracted significant interest. Particularly, the majority of research has been focused on the design of approximate multipliers [8], as they are the most complex components of the MAC units and dominate energy consumption inside the unit. However, such multipliers are not application specific.…”
Section: Introductionmentioning
confidence: 99%
“…Consequently, their deployment as DNN accelerators in smart nanoscale applications becomes very challenging [7] e.g., complex DNN analytics in resource-constrained edge devices, where safety and energy are critical considerations, can lead to an unexpected energy outage which can jeopardize human lives. This problem can be solved by leveraging approximate computing [8], [9] -an inexact computing method that exploits the inherent error resilience of onboard applications for energy efficiency in DNN accelerators. However, approximate hardware acceleration is deemed to be inherently less reliable [10].…”
mentioning
confidence: 99%
“…Approximate computing-based deep neural network (AxDNN) accelerators are designed by incorporating inexact arithmetic units [11] [12], computation skipping [13], memory skipping [14], etc. Their fabrication at the nanoscale follows a sophisticated manufacturing process whose imperfections may result in manufacturing defects, such as process variations and permanent faults (stuck-at faults) [9]. As discussed in this paper, the permanent faults affect the compute units of AxDNN accelerators in every execution cycle and their presence as unmasked faults leads to serious failures in the whole system.…”
mentioning
confidence: 99%