2019
DOI: 10.48550/arxiv.1906.03951
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

SCAN: A Scalable Neural Networks Framework Towards Compact and Efficient Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
7
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(7 citation statements)
references
References 0 publications
0
7
0
Order By: Relevance
“…Therefore, they performed binarization on the k × k convolution kernels to cut down parameters. Meanwhile, [150] introduced scalable neural networks, which achieve neural network compression and acceleration simultaneously. Moreover, Li et al [151] designed an intensely-inverted residual block unit, which introduces inverted residual structure and multi-scale low-redundancy convolution kernels.…”
Section: B Shortcut Connectionsmentioning
confidence: 99%
“…Therefore, they performed binarization on the k × k convolution kernels to cut down parameters. Meanwhile, [150] introduced scalable neural networks, which achieve neural network compression and acceleration simultaneously. Moreover, Li et al [151] designed an intensely-inverted residual block unit, which introduces inverted residual structure and multi-scale low-redundancy convolution kernels.…”
Section: B Shortcut Connectionsmentioning
confidence: 99%
“…The researchers who developed ResNet [13] first took the multigrid methods as evidence to support what is known as a residual representation for the interpretation of ResNet. Further, [20,8,44] adopted multi-resolution ideas to improve the performance of their networks. Additionally, a CNN model with a structure similar to that of the V-cycle multigrid is proposed to address volumetric medical image segmentation and biomedical image segmentation in [31,29].…”
Section: Related Workmentioning
confidence: 99%
“…For VGG-16, [5] achieves a 1.98x Ops reduction for a 2% decrease in accuracy on ImageNet and show that they outperform both [10] and [21] on the same metric. [37] splits the network into multiple sections and learns classifiers that allow for early exit through the network depending on the input image processed. They achieve on average 2.17x reduction in Ops across networks on CIFAR-100 for no accuracy loss, and 1.99x reduction in Ops on ImageNet also for no accuracy loss.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Frameworks -The various pruning techniques discussed above each have a unique set of hyperparameters that relate to filter ranking metrics as well as the manner in which the models are re-trained. For instance, [31] sequentially prunes and retrains on a per layer basis, while works such as [37] have to add many auxiliary layers on top of the chosen architecture in order to create and train their early exit classifiers. Distiller [41] and Mayo [39] are two state-of-the-art open-source frameworks that allow for experimentation with such pruning techniques.…”
Section: Background and Related Workmentioning
confidence: 99%