2021
DOI: 10.1109/access.2021.3077596
|View full text |Cite
|
Sign up to set email alerts
|

Compressing Neural Networks With Inter Prediction and Linear Transformation

Abstract: Because of resource-constrained environments, network compression has become an essential part of deep neural networks research. In this paper, we found a mutual relationship between kernel weights termed as Inter-Layer Kernel Correlation (ILKC). The kernel weights between two different convolution layers share a substantial similarity in shapes and values. Based on this relationship, we propose a new compression method, Inter-Layer Kernel Prediction (ILKP), which represents convolutional kernels with fewer bi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
3
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 24 publications
0
3
0
Order By: Relevance
“…From there, AGB as the AXI master generates signals for AXI4 burst transmission according to Eq. ( 7)- (9), where Burst len is 1, . .…”
Section: Accelerator Architecturementioning
confidence: 99%
See 2 more Smart Citations
“…From there, AGB as the AXI master generates signals for AXI4 burst transmission according to Eq. ( 7)- (9), where Burst len is 1, . .…”
Section: Accelerator Architecturementioning
confidence: 99%
“…Most existing FPGA-based CNN inference implementations use DDR, which has a narrow bandwidth, for the offchip memory. Compression mechanisms [9], [10] or quantization [11]- [13] have been applied for low numerical precision to reduce the off-chip memory bandwidth pressure. In addition, using on-chip buffers (e.g., [14]- [16]) is a popular solution for limiting bandwidth memory.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation