2022
DOI: 10.48550/arxiv.2202.00155
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Fortuitous Forgetting in Connectionist Networks

Abstract: Forgetting is often seen as an unwanted characteristic in both human and machine learning. However, we propose that forgetting can in fact be favorable to learning. We introduce forget-and-relearn as a powerful paradigm for shaping the learning trajectories of artificial neural networks. In this process, the forgetting step selectively removes undesirable information from the model, and the relearning step reinforces features that are consistently useful under different conditions. The forget-and-relearn frame… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 34 publications
0
1
0
Order By: Relevance
“…Forgetting The use of probabilistic models such as GMMs has another interesting consequence, namely the ability to control forgetting. Following Zhou et al (2022), forgetting can be a beneficial functionality, which is simple to control in GMMs by eliminating certain components without impacting the remaining ones.…”
Section: Ar Assumptionsmentioning
confidence: 99%
“…Forgetting The use of probabilistic models such as GMMs has another interesting consequence, namely the ability to control forgetting. Following Zhou et al (2022), forgetting can be a beneficial functionality, which is simple to control in GMMs by eliminating certain components without impacting the remaining ones.…”
Section: Ar Assumptionsmentioning
confidence: 99%
“…Regularization in Self-supervised Learning. The concept of regularizing a specific subset of the network is relatively unexplored in self-supervised learning but finds motivation in recent findings from supervised settings, such as addressing minority examples (Hooker et al, 2019), out-ofdistribution generalization , late-layer regularizations through head weight-decay (Abnar et al, 2021), and initialization (Zhou et al, 2022). Additionally, Lee et al (2022a) propose surgically fine-tuning specific layers of the network to handle distribution shifts in particular categories.…”
Section: Related Workmentioning
confidence: 99%