2021
DOI: 10.1002/ail2.44
|View full text |Cite
|
Sign up to set email alerts
|

Remembering for the right reasons: Explanations reduce catastrophic forgetting

Abstract: The goal of continual learning (CL) is to learn a sequence of tasks without suffering from the phenomenon of catastrophic forgetting. Previous work has shown that leveraging memory in the form of a replay buffer can reduce performance degradation on prior tasks. We hypothesize that forgetting can be further reduced when the model is encouraged to remember the evidence for previously made decisions. As a first step towards exploring this hypothesis, we propose a simple novel training paradigm, called Rememberin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
16
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 25 publications
(16 citation statements)
references
References 30 publications
0
16
0
Order By: Relevance
“…Since later layers in the convolutional neural network capture high-level semantics (Mahendran and Vedaldi 2016), taking gradients of a model output with respect to the feature map activations from one such layers identifies which high-level semantics are important for the model prediction. In our analysis, we select this layer and refer to as target layer (Ebrahimi et al 2021). List of target layer for different experiments is given in Table 3.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…Since later layers in the convolutional neural network capture high-level semantics (Mahendran and Vedaldi 2016), taking gradients of a model output with respect to the feature map activations from one such layers identifies which high-level semantics are important for the model prediction. In our analysis, we select this layer and refer to as target layer (Ebrahimi et al 2021). List of target layer for different experiments is given in Table 3.…”
Section: Discussionmentioning
confidence: 99%
“…Guo et al (2020) improved such methods by proposing a loss balancing update rules in MEGA. Ebrahimi et al (2021) stores a saliency map corresponding to each episodic memory sample and complements replay with a regularization objective so that model explanations for the past tasks have minimal drift. Our method, EPR, also uses episodic memory for experience replay.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Finally, our system performs a kind of continual learning (Parisi et al, 2019;Carlson et al, 2010). While recent work in this area has focused on dynamic update of model parameters, e.g., (Ebrahimi et al, 2021), our work leaves the model fixed, and seeks improvement in the broader system in which the model is embedded, exploring an alternative and potentially more interpretable architecture towards this goal.…”
Section: Related Workmentioning
confidence: 99%
“…Previous works have tackled the problem of CF using different strategies. Memory-based methods mitigate CF inserting data from past tasks into the training process of new tasks, continuously retraining previous tasks [30], either with raw samples [26,31], or minimizing gradient interference [28,32]. Other works such as [33,34] train generator functions (GANs) or autoencoders [35] to Similarly, [36,37] seek to be memory-efficient by saving feature vectors of instances from previous tasks, while learning a transformation from the feature space of past tasks to current ones.…”
Section: Related Workmentioning
confidence: 99%