2021
DOI: 10.4218/etrij.2020-0370
|View full text |Cite
|
Sign up to set email alerts
|

A low‐cost compensated approximate multiplier for Bfloat16 data processing on convolutional neural network inference

Abstract: This paper presents a low‐cost two‐stage approximate multiplier for bfloat16 (brain floating‐point) data processing. For cost‐efficient approximate multiplication, the first stage implements Mitchell's algorithm that performs the approximate multiplication using only two adders. The second stage adopts the exact multiplication to compensate for the error from the first stage by multiplying error terms and adding its truncated result to the final output. In our design, the low‐cost multiplications in both stage… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 34 publications
0
3
0
Order By: Relevance
“…Lightweight depthwise convolutions are used in the intermediate expansion layer as a source of nonlinearity to filter features. The MobileNetV2 architecture includes a 32‐filter with an initial fully convolution layer as well as 19 residual bottleneck layers [40, 41]. DenseNet201: DenseNet is based on the notion that convolutional networks may be trained to be significantly deeper, more precise, and more effective if they have shorter connections between layers that are close to the input and those that are close to the output.…”
Section: Proposed Methodology and Workflow Of The Frameworkmentioning
confidence: 99%
See 1 more Smart Citation
“…Lightweight depthwise convolutions are used in the intermediate expansion layer as a source of nonlinearity to filter features. The MobileNetV2 architecture includes a 32‐filter with an initial fully convolution layer as well as 19 residual bottleneck layers [40, 41]. DenseNet201: DenseNet is based on the notion that convolutional networks may be trained to be significantly deeper, more precise, and more effective if they have shorter connections between layers that are close to the input and those that are close to the output.…”
Section: Proposed Methodology and Workflow Of The Frameworkmentioning
confidence: 99%
“…The most used deep learning model for image classification is CNN and transfer-learning models [38,39]. The two parts of the presented MLCNN-COV framework are presented in Figure 5.…”
Section: Proposed Frameworkmentioning
confidence: 99%
“…In precision scaling [1], we use fewer bits to represent numeric values rather than executing all the required mathematical operations with the full representation. Several standards for the floating-point presentation recently appeared: IEEE 754-2019 for half-precision [2], posit format with dynamic range and mantissa [3], and Google's bfloat16, targeting the machine-learning workloads [4]. Storing the numeric values with fewer bits reduces the size of arithmetic circuits and their complexity.…”
Section: Introductionmentioning
confidence: 99%