Encyclopedia of Cognitive Science 2006
DOI: 10.1002/0470018860.s00096
|View full text |Cite
|
Sign up to set email alerts
|

Catastrophic Forgetting in Connectionist Networks

Abstract: Unlike human brains, connectionist networks can forget previously learned information suddenly and completely (‘catastrophically’) when learning new information. Various solutions to this problem have been proposed.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
144
0

Year Published

2006
2006
2018
2018

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 116 publications
(144 citation statements)
references
References 12 publications
0
144
0
Order By: Relevance
“…The rapid learning that we have observed is not typically associated with networks trained using back propagation, which often exhibit a trade-off between the speed of new learning and the stability of previously acquired knowledge (French, 1999;McClelland, McNaughton, & O'Reilly, 1995;Page, 2000). However, other supervised learning algorithms exist, including those in which sparse or localist representations mediate between the speech input and lexical output, and these might be capable of simulating a more rapid learning process.…”
Section: Implications For Models Of Speech Perceptionmentioning
confidence: 91%
“…The rapid learning that we have observed is not typically associated with networks trained using back propagation, which often exhibit a trade-off between the speed of new learning and the stability of previously acquired knowledge (French, 1999;McClelland, McNaughton, & O'Reilly, 1995;Page, 2000). However, other supervised learning algorithms exist, including those in which sparse or localist representations mediate between the speech input and lexical output, and these might be capable of simulating a more rapid learning process.…”
Section: Implications For Models Of Speech Perceptionmentioning
confidence: 91%
“…This is in contrast to the behavior of, e.g., multilayer perceptrons [17], where retraining with slightly different input statistics can lead to a complete reorganization of the hidden layer structure, and therefore to a loss of already learned capabilities.…”
Section: Avoidance Of Catastrophic Forgettingmentioning
confidence: 75%
“…While this will certainly generate an induced representation, this solution is unsuitable because the hidden layer, while manifestly two-dimensional irrespectively of input data, would lack any topological organization if the MLP is trained using a back-propagation learning algorithm [15]. Furthermore, issues of catastrophic forgetting [17] would complicate the use of MLP still further. As an alternative, PCA produces, for each input, a set coordinates in the space of principal components which are not in any way topologically organized.…”
Section: Critical Examination and Justification Of Used Methodsmentioning
confidence: 99%
“…For the shape categories the SLP network architecture is only superior at earlier learning epochs, but is worse if the learning process is continued. Overall the SLP performance is surprisingly good, which is in contrast to classification tasks with a one-out-of-n class selection, where the SLP approach is known for the "catastrophic forgetting effect" (French, 1999). For our categorization task this effect is only slightly visible for the shape categories, where the performance increase for newly presented objects is distinctly less than for all other tested approaches.…”
Section: Color and Parts-based Featuresmentioning
confidence: 89%
“…Therefore in this paper we are particularly interested in incremental learning of representations under the condition, where a particular training vector can only be accessed for a limited time period. As a consequence the training with such a changing data ensemble typically causes the well-known "catastrophic forgetting effect" (French, 1999): With the incorporation of newly acquired knowledge, the previously learned knowledge is quickly fading out. Closely related to this effect is the term "catastrophic interference" (McCloskey & Cohen, 1989): Patterns of different categories which are similar in feature space, confuse the learning and overwrite earlier presented patterns.…”
Section: Introductionmentioning
confidence: 99%