2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.01158
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Remember: A Synaptic Plasticity Driven Framework for Continual Learning

Abstract: Models trained in the context of continual learning (CL) should be able to learn from a stream of data over an indefinite period of time. The main challenges herein are: 1) maintaining old knowledge while simultaneously benefiting from it when learning new tasks, and 2) guaranteeing model scalability with a growing amount of data to learn from. In order to tackle these challenges, we introduce Dynamic Generative Memory (DGM) -synaptic plasticity driven framework for continual learning. DGM relies on conditiona… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
178
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 235 publications
(178 citation statements)
references
References 24 publications
0
178
0
Order By: Relevance
“…DGM was proposed in [ 96 ], which relies on conditional generative adversarial networks. It trained a sparse binary mask for each layer of the generator.…”
Section: Methods Descriptionmentioning
confidence: 99%
See 2 more Smart Citations
“…DGM was proposed in [ 96 ], which relies on conditional generative adversarial networks. It trained a sparse binary mask for each layer of the generator.…”
Section: Methods Descriptionmentioning
confidence: 99%
“…All images are gray level with size of 32 × 32 pixels. MNIST is adopted in [ 15 , 22 , 33 , 47 , 52 , 54 , 75 , 92 , 96 ].…”
Section: Datasetsmentioning
confidence: 99%
See 1 more Smart Citation
“…Expanding the model capacities during training has been mainly discussed in the lifelong learning area, which tries to fit a model to the data sequence without catastrophic forgetting [9,17,22,27]. One of the common approaches is assigning an expanded capacity for the incoming data while limiting the parameters' change to the proximity of the trained parameters by the previous data [6,21,27].…”
Section: Model Expansionmentioning
confidence: 99%
“…According to recent works [28,2,34] aligning domain shift through batch normalization (BN) [23] layers, although in online condition we never have access to full target data, we are inspired to smoothly adjust model statistics through online data stream of target videos. Besides the statistics, works [20,37,41,40] on multi-domain or incremental learning inspire us to selectively tune a small subset of learned basic parameters while keeping all the other fixed. In our setting this is meant to ensure that visual appearance variations arising from scene changes will never influence the networks weights encoding the reliable knowledge.…”
Section: Introductionmentioning
confidence: 99%