2018
DOI: 10.48550/arxiv.1807.11164
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
145
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 143 publications
(145 citation statements)
references
References 19 publications
0
145
0
Order By: Relevance
“…Non expansion. According to the paper of ShuffleNetV2 [16], we can learn that if the number of input feature maps and the number of output feature maps of convolutional layer are the same, computing speed will be maximized. Therefore, we don't expand the number of feature maps in our DFSEB Block.…”
Section: A Dfseb Blockmentioning
confidence: 99%
“…Non expansion. According to the paper of ShuffleNetV2 [16], we can learn that if the number of input feature maps and the number of output feature maps of convolutional layer are the same, computing speed will be maximized. Therefore, we don't expand the number of feature maps in our DFSEB Block.…”
Section: A Dfseb Blockmentioning
confidence: 99%
“…Group convolution [39] is an efficient convolution operation widely used in many efficient networks, it divides the input into independent groups and the kernels of each group share the same weight in order to reduce the number of parameters. Other efficient networks benefit from depth-wise convolution [22]- [25], [48] which is extreme case of group convolution. Recent works such as BiSeNet [36] and ICNet [37] also have better trade-off between accuracy and speed, but they are not easy to deploy and difficulty in migrating to other tasks and areas due to their complex structures.…”
Section: Related Workmentioning
confidence: 99%
“…1) Depth-wise Separable Convolutions: The depth-wise separable convolution is considered as key-module in recent efficient networks [22]- [25], it splits the full convolution operations into two independent operations, depth-wise convolution and point-wise convolution. In depth-wise convolution, the number of groups is equal to the number of feature maps, it means each kernel has single feature map in and single feature map out, and the shared weight kernels make the depth-wise Where 'DW' is depth-wise convolution, 'concat' is concatenation operation, 'MP' is Map-Pooling layer and 's' is layer stride.…”
Section: A Our Core Modulementioning
confidence: 99%
See 2 more Smart Citations