2017 IEEE International Conference on Computer Vision (ICCV) 2017
DOI: 10.1109/iccv.2017.282
|View full text |Cite
|
Sign up to set email alerts
|

Performance Guaranteed Network Acceleration via High-Order Residual Quantization

Abstract: Input binarization has shown to be an effective way for network acceleration. However, previous binarization scheme could be regarded as simple pixel-wise thresholding operations (i.e., order-one approximation) and suffers a big accuracy loss. In this paper, we propose a high-order binarization scheme, which achieves more accurate approximation while still possesses the advantage of binary operation. In particular, the proposed scheme recursively performs residual quantization and yields a series of binary inp… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
70
0
1

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 100 publications
(71 citation statements)
references
References 14 publications
0
70
0
1
Order By: Relevance
“…BNNs [23,40] propose to constrain both weights and activations to binary values (i.e., +1 and -1), where the multiply-accumulations can be replaced by purely xnor(·) and popcount(·) operations. To make a trade-off between accuracy and complexity, [13,15,29,48] propose to recursively perform residual quantization and yield a series of binary tensors with decreasing magnitude scales. However, multiple binarizations are sequential process which cannot be paralleled.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…BNNs [23,40] propose to constrain both weights and activations to binary values (i.e., +1 and -1), where the multiply-accumulations can be replaced by purely xnor(·) and popcount(·) operations. To make a trade-off between accuracy and complexity, [13,15,29,48] propose to recursively perform residual quantization and yield a series of binary tensors with decreasing magnitude scales. However, multiple binarizations are sequential process which cannot be paralleled.…”
Section: Related Workmentioning
confidence: 99%
“…We explore the difference between layer-wise and group-wise design strategies in approach can be treated as a kind of tensor approximation which has similarities with multiple binarizations methods in [13,15,29,30,48] and the differences are described in Sec. 4.…”
Section: Layer-wise Vs Group-wise Binary Decompositionmentioning
confidence: 99%
“…Binarization and Convolution Process of XNOR-NetTo further reduce the quantization error, High-Order Residual Quantization (HORQ)[70] adopts a recursive approximation to the full-precision activation based on the quantized residual, instead of one-step approximation used in XNOR-Net. It generates the final quantized activation by a linear combination of the approximation in each recursive step.…”
mentioning
confidence: 99%
“…Specifically, the quantized weights are exactly equal to αM T Z when Eq.3 has the optimal solution. In contrast, the re-training strategy with the cluster regularization brings less quantization error than the reconstruction-based methods [11,15], since the weight W remains in a highly clustered state after re-training. To further reduce the effects of quantization error to the classification loss, we fine-tune the re-trained model for several epochs.…”
Section: The Whole Quantization Frameworkmentioning
confidence: 99%
“…Rastegari et al presented XNOR-Network [11] that approximated the full-precision weights by introducing a scaling factor during binarization. For pursuing higher accuracy, High-Order Residual Quantization (HORQ) [15] sought to compensate the information loss of binary quantization by conducting convolutional operations on inputs in different scales and then combined the results. The Ternary Weight Network (TWN) [12] introduced zero as a third quantized value and was the first method that achieved decent results on the ILSVRC-12 dataset.…”
Section: Related Workmentioning
confidence: 99%