2019
DOI: 10.48550/arxiv.1904.00310
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learn to Grow: A Continual Structure Learning Framework for Overcoming Catastrophic Forgetting

Abstract: Addressing catastrophic forgetting is one of the key challenges in continual learning where machine learning systems are trained with sequential or streaming tasks. Despite recent remarkable progress in state-of-the-art deep learning, deep neural networks (DNNs) are still plagued with the catastrophic forgetting problem. This paper presents a conceptually simple yet general and effective framework for handling catastrophic forgetting in continual learning with DNNs. The proposed method consists of two componen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
22
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 28 publications
(22 citation statements)
references
References 20 publications
0
22
0
Order By: Relevance
“…A few works also look at continual learning from the perspectives of the loss landscape and dynamics of optimization [Mirzadeh et al, 2020, Mirzadeh et al, 2020b. Modularity-based methods allocate different subsets of the parameters to each task [Rusu et al, 2016, Yoon et al, 2018, Jerfel et al, 2019, Li et al, 2019, Wortsman et al, 2020, Mirzadeh et al, 2020a.…”
Section: Related Workmentioning
confidence: 99%
“…A few works also look at continual learning from the perspectives of the loss landscape and dynamics of optimization [Mirzadeh et al, 2020, Mirzadeh et al, 2020b. Modularity-based methods allocate different subsets of the parameters to each task [Rusu et al, 2016, Yoon et al, 2018, Jerfel et al, 2019, Li et al, 2019, Wortsman et al, 2020, Mirzadeh et al, 2020a.…”
Section: Related Workmentioning
confidence: 99%
“…Microscopically, existing methods dynamically expand networks using thresholds on loss functions over new tasks and retrain the selected weights to prevent semantic drift [68]. Reinforced continual learning [65] employs a controller to define a strategy that expands the architecture of a given network while the learn-to-grow model [32] relies on neural architecture search [76] to define optimal architectures on new tasks. Other models [7], inspired by the process of adult neurogenesis in the hippocampus, combine architecture expansion with pseudo-rehearsal using auto-encoders.…”
Section: Related Workmentioning
confidence: 99%
“…Experiments show that Firefly efficiently learns accurate and resource-efficient networks in various settings. In particular, for continual learning, our method learns more accurate and smaller networks that can better prevent catastrophic forgetting, outperforming state-of-the-art methods such as Learn-to-Grow (Li et al, 2019) and Compact-Pick-Grow (Hung et al, 2019a).…”
Section: Introductionmentioning
confidence: 98%
“…In addition, dynamically growing neural network has also been proposed as a promising approach for preventing the challenging catastrophic forgetting problem in continual learning (Rusu et al, 2016;Yoon et al, 2017;Rosenfeld & Tsotsos, 2018;Li et al, 2019).…”
Section: Introductionmentioning
confidence: 99%