2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018
DOI: 10.1109/cvpr.2018.00368
|View full text |Cite
|
Sign up to set email alerts
|

Learning Time/Memory-Efficient Deep Architectures with Budgeted Super Networks

Abstract: We propose to focus on the problem of discovering neural network architectures efficient in terms of both prediction quality and cost. For instance, our approach is able to solve the following tasks: learn a neural network able to predict well in less than 100 milliseconds or learn an efficient model that fits in a 50 Mb memory. Our contribution is a novel family of models called Budgeted Super Networks (BSN). They are learned using gradient descent techniques applied on a budgeted learning objective function … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
64
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 75 publications
(64 citation statements)
references
References 8 publications
0
64
0
Order By: Relevance
“…Macro search algorithms aim to directly discover the entire neural networks [5,4,38,46,21]. To search convolutional neural networks (CNNs) [20], typical approaches apply RL to optimize the searching policy to discover architectures [1,5,46,31].…”
Section: Related Workmentioning
confidence: 99%
“…Macro search algorithms aim to directly discover the entire neural networks [5,4,38,46,21]. To search convolutional neural networks (CNNs) [20], typical approaches apply RL to optimize the searching policy to discover architectures [1,5,46,31].…”
Section: Related Workmentioning
confidence: 99%
“…As discussed, our model allows to dynamically dynamically set the trade-off between accuracy and inference time with no additional cost. ferent state-of-the-art CNN acceleration strategies [17,19,22,38,56]. We consider methods applying pruning at different levels, such as independent filters (Network slimming [38]), groups of weights (CondenseNet) [19], connections in multi-branch architectures (SuperNet) [56], or a combination of them (SSS) [22].…”
Section: Comparison With the State Of The Artmentioning
confidence: 99%
“…Our work is motivated by differentiable architecture search [34,37,26,1], which is based on the continuous relaxation of the architecture representation, allowing efficient search with gradient descent. [34,37] propose a gridlike network as the search space, while [26] relax the search space to be continuous and search the space by solving a bilevel optimization problem.…”
Section: Architecture Searchmentioning
confidence: 99%
“…Our work is motivated by differentiable architecture search [34,37,26,1], which is based on the continuous relaxation of the architecture representation, allowing efficient search with gradient descent. [34,37] propose a gridlike network as the search space, while [26] relax the search space to be continuous and search the space by solving a bilevel optimization problem. Other works in architecture search employ reinforcement learning [4,49], evolutionary algorithms [31,39,25], and sequential model-based optimization [29,24] to search the discrete space.…”
Section: Architecture Searchmentioning
confidence: 99%