2018
DOI: 10.1007/978-3-030-01264-9_8
|View full text |Cite
|
Sign up to set email alerts
|

ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design

Abstract: Currently, the neural network architecture design is mostly guided by the indirect metric of computation complexity, i.e., FLOPs. However, the direct metric, e.g., speed, also depends on the other factors such as memory access cost and platform characterics. Thus, this work proposes to evaluate the direct metric on the target platform, beyond only considering FLOPs. Based on a series of controlled experiments, this work derives several practical guidelines for efficient network design. Accordingly, a new archi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

8
2,381
0
8

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 4,332 publications
(2,830 citation statements)
references
References 49 publications
8
2,381
0
8
Order By: Relevance
“…In contrast to the inverted bottleneck in MobileNetv2, the GE Layer has one more 3 × 3 convolution. However, this layer is also friendly to the computation cost and memory access cost (Ma et al, 2018;Sandler et al, 2018), because the 3 × 3 convolution is specially optimized in the CUDNN library (Chetlur et al, 2014;Ma et al, 2018). Meanwhile, because of this layer, the GE Layer has higher feature expression ability than the inverted bottleneck.…”
Section: Gather-and-expansion Layermentioning
confidence: 99%
“…In contrast to the inverted bottleneck in MobileNetv2, the GE Layer has one more 3 × 3 convolution. However, this layer is also friendly to the computation cost and memory access cost (Ma et al, 2018;Sandler et al, 2018), because the 3 × 3 convolution is specially optimized in the CUDNN library (Chetlur et al, 2014;Ma et al, 2018). Meanwhile, because of this layer, the GE Layer has higher feature expression ability than the inverted bottleneck.…”
Section: Gather-and-expansion Layermentioning
confidence: 99%
“…They visualize the feature maps extracted by different filters and view each filter as a visual unit focusing on different visual components.of the ResNet-50 [28], and meanwhile save more than 75% of parameters and 50% computational time. In the literature, approaches for compressing the deep networks can be classified into five categories: parameter pruning [26,29,30,31], parameter quantizing [32,33,34,35,36,37,38,39,40,41], low-rank parameter factorization [42,43,44,45,46], transferred/compact convolutional filters [47,48,49,50], and knowledge distillation [51,52,53,54,55,56]. The parameter pruning and quantizing mainly focus on eliminating the redundancy in the model parameters respectively by removing the redundant/uncritical ones or compressing the parameter space (e.g.…”
mentioning
confidence: 99%
“…Thus, the Depthwise-STFT separable layer has a lower space-time complexity when com-arXiv:2001.09912v1 [cs.CV] 27 Jan 2020 pared to the depthwise separable convolutions. Furthermore, we show experimentally that the proposed layer achieves better performance compared to the many state-of-the-art depthwise separable based models such as MobileNet [6,7] and ShuffleNet [8,9].…”
Section: Introductionmentioning
confidence: 93%
“…Recently, there has been a growing interest into developing space-time efficient neural networks for real time and resource restricted applications [10,6,7,9,8,11,12]. Depthwise Separable Convolutions.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation