2018 International Conference on Information and Communication Technology Convergence (ICTC) 2018
DOI: 10.1109/ictc.2018.8539530
|View full text |Cite
|
Sign up to set email alerts
|

Reducing MAC operation in convolutional neural network with sign prediction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 15 publications
(5 citation statements)
references
References 7 publications
0
5
0
Order By: Relevance
“…On the other hand, in the HW-oriented approach (Figure 4.b), the developers are mainly focusing on designing enhanced hardware platforms that are optimized for embedded applications in order to run current and future state-of-the-art ML algorithms. This often involves investigating the bottlenecks in an existing architecture with regard to computations within a ML framework, like neural networks, and the design of hardware accelerator modules to improve throughput and consumption: e.g., reducing computational complexity in convolution layers [49], [50], efficient, low-power and feature-rich perceptrons [51], enhanced data caches [52]. In other cases, the developers design new hardware platforms optimized for embedded applications with extended digital signal processing capabilities already integrated [53].…”
Section: Tinyml Workflowsmentioning
confidence: 99%
“…On the other hand, in the HW-oriented approach (Figure 4.b), the developers are mainly focusing on designing enhanced hardware platforms that are optimized for embedded applications in order to run current and future state-of-the-art ML algorithms. This often involves investigating the bottlenecks in an existing architecture with regard to computations within a ML framework, like neural networks, and the design of hardware accelerator modules to improve throughput and consumption: e.g., reducing computational complexity in convolution layers [49], [50], efficient, low-power and feature-rich perceptrons [51], enhanced data caches [52]. In other cases, the developers design new hardware platforms optimized for embedded applications with extended digital signal processing capabilities already integrated [53].…”
Section: Tinyml Workflowsmentioning
confidence: 99%
“…As compare to the single stream, our model has added some additional cost, we show that the existing architecture is highly able to augment multiple feature spaces using single stream. We calculate and compare the number of parameters and computational complexity (MACs) [74] of HAFS with compare to the existing ResNext101 and BESS in Fig. 11.…”
Section: Complexity Analysis Of Bess and Hafsmentioning
confidence: 99%
“…In the case of single stream HAFS has increased the number of parameters which is less and comparable. Comparative computa-tional complexity is measure using the popular metric called "Multiply-and-accumulates (MACs) per frames" [74]. We notice that HAFS has nearly similar MACs (38.58GM ACs) with compare to the existing ResNext101(Basic*) and BESS (38.57GM ACs).…”
Section: Complexity Analysis Of Bess and Hafsmentioning
confidence: 99%
“…Akhlaghi et al [1] predict during the convolution computation whether the convolution results will end up negative. Song et al [42], Lin et al [30], and Chang et al [4] predict whether an entire convolution result is negative according to a partial result yielded by the input MSB bits. Huan et al [19] avoid convolution multiplications by predicting and skipping near-zero valued data, given certain thresholds.…”
Section: Related Workmentioning
confidence: 99%