2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.01156
|View full text |Cite
|
Sign up to set email alerts
|

Building Efficient Deep Neural Networks With Unitary Group Convolutions

Abstract: We propose unitary group convolutions (UGConvs), a building block for CNNs which compose a group convolution with unitary transforms in feature space to learn a richer set of representations than group convolution alone. UGConvs generalize two disparate ideas in CNN architecture, channel shuffling (i.e. ShuffleNet [29]) and blockcirculant networks (i.e. CirCNN [6]), and provide unifying insights that lead to a deeper understanding of each technique. We experimentally demonstrate that dense unitary transforms c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
55
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 41 publications
(55 citation statements)
references
References 22 publications
0
55
0
Order By: Relevance
“…For ResNet-18, we can reduce 83% parameters and 95% FLOPS with an increase of 14.5% in test error. It is not promising regarding the high test error, but this performance is on par with [50], even if they have higher budget in both training and inference phases: their ResNet-18 example is trained from scratch and the Hadamard transform adds overhead. We also compare ResNet-34 to [32], which uses lowrank approximation to perform GConv pruning.…”
Section: Resultsmentioning
confidence: 99%
See 3 more Smart Citations
“…For ResNet-18, we can reduce 83% parameters and 95% FLOPS with an increase of 14.5% in test error. It is not promising regarding the high test error, but this performance is on par with [50], even if they have higher budget in both training and inference phases: their ResNet-18 example is trained from scratch and the Hadamard transform adds overhead. We also compare ResNet-34 to [32], which uses lowrank approximation to perform GConv pruning.…”
Section: Resultsmentioning
confidence: 99%
“…Additionally, it cannot deal with 1 × 1 group convolution, which is critical since recent efficient CNN heavily rely on them [10,36,49]. [50] applies block Hadamard transform, which is more efficient but still requires extra computation. On the other hand, permuting channels is a much simpler way to mingle groups ( Figure 3b) since neither additional FMA nor parameter is required.…”
Section: Background and Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…The second direction is to enhance the hardware implementation efficiency by deriving an effective tradeoff between accuracy and pruning rate, e.g., the energyaware pruning [17], and structure-aware pruning [18], [10]. FPGA hardware accelerators [19], [20] have also been investigated to accommodate pruned CNNs, by leveraging the reconfigurability in on-chip resources. Recently, the authors of [14] have developed a systematic weight pruning framework based on the powerful optimization tool ADMM (Alternating Direction Method of Multipliers) [21].…”
Section: B Cnn Weight Pruningmentioning
confidence: 99%