2022
DOI: 10.48550/arxiv.2207.11719
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Gradient-based Bi-level Optimization for Deep Learning: A Survey

Abstract: Bi-level optimization, especially the gradient-based category, has been widely used in the deep learning community including hyperparameter optimization and meta-knowledge extraction. Bi-level optimization embeds one problem within another and the gradient-based category solves the outer level task by computing the hypergradient, which is much more efficient than classical methods such as the evolutionary algorithm. In this survey we first give a formal definition of the gradient-based bi-level optimization. S… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 18 publications
0
5
0
Order By: Relevance
“…Given a good choice of γ, the produced X h is expected to have a high score f aux (X h ), based on which we can choose γ. To make the search for γ more efficient, we can formulate this process as a bi-level optimization problem (Chen et al, 2022a;2022c):…”
Section: Deep Linearization For Bidirectional Learningmentioning
confidence: 99%
“…Given a good choice of γ, the produced X h is expected to have a high score f aux (X h ), based on which we can choose γ. To make the search for γ more efficient, we can formulate this process as a bi-level optimization problem (Chen et al, 2022a;2022c):…”
Section: Deep Linearization For Bidirectional Learningmentioning
confidence: 99%
“…In the meta-training phase, meta-learning occurs via casting the problem as a bilevel optimization [6], where the inner optimizaiton uses a set of inner learning algorithms to solve a set of related tasks, minimizing some inner objective. During metalearning, an outer algorithm updates the inner learning algorithm's inductive biases such that the models learn to improve on some pre-determined outer objective.…”
Section: Background and Related Workmentioning
confidence: 99%
“…( 5) may lead to a trivial solution where all hierarchical weights reduce to zeros. In this work, we propose to formulate hierarchical weight learning as bi-level optimization Luca et al (2018); Chen et al (2022a); where the hierarchical weight is decided by an outer level learning task:…”
Section: Hierarchical Weight Learningmentioning
confidence: 99%