The 2006 IEEE International Joint Conference on Neural Network Proceedings 2006
DOI: 10.1109/ijcnn.2006.246618
|View full text |Cite
|
Sign up to set email alerts
|

Adaptation of Artificial Neural Networks Avoiding Catastrophic Forgetting

Abstract: In connectionist learning, one relevant problem is "catastrophic forgetting" that may occur when a network, trained with a large set of patterns, has to learn new input patterns, or has to be adapted to a different environment. The risk of catastrophic forgetting is particularly high when a network is adapted with new data that do not adequately represent the knowledge included in the original training data.Two original solutions are proposed to reduce the risk that the network focuses on new data only, loosin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2013
2013
2023
2023

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(3 citation statements)
references
References 14 publications
0
3
0
Order By: Relevance
“…Another popular category of adaptation techniques is conservative training (CT) [22]. CT can be achieved by adding regularizations to the adaptation criterion.…”
Section: Conservative Trainingmentioning
confidence: 99%
“…Another popular category of adaptation techniques is conservative training (CT) [22]. CT can be achieved by adding regularizations to the adaptation criterion.…”
Section: Conservative Trainingmentioning
confidence: 99%
“…Feature‐based adaptation methods extract features by minimizing an individual speaker's characteristics using feature transformation methods , whereas a model‐adaptation is a method for transforming a speaker‐independent (SI) model into a more speaker‐dependent (SD) version. Model‐adaptation methods include regularization methods that decrease the overfitting by preventing many deviations from the SI model; learning speaker information such as a speaker code , bases , and i‐vectors that can effectively combine with general features and other adaptation methods; and transformation‐based techniques that convert the DNN layer into a speaker adapted layer. In particular, whereas feature adaptation affects all senones, by applying relatively small transformation parameters to the input layer, a model adaptation can be applied to all other layers.…”
Section: Introductionmentioning
confidence: 99%
“…Linear transformation can be applied at different levels of the DNN system: to the input features, as in linear input network transformation (LIN) [6] or featurespace discriminative linear regression (fDLR); to the activations of hidden layers, as in linear hidden network transformation (LHN) [6]; or to the softmax layer, as in LON or in output-feature discriminative linear regression. The second type of adaptation consists in re-training the entire network or only a part of it using special regularization techniques for improving generalization, such as L2prior regularization [7], Kullback-Leibler divergence regularization [8], conservative training [9]. The concept of multi-task learning (MTL) has recently been applied to the task of speaker adaptation and has been shown to improve the performance of different model-based DNN adaptation techniques, such as LHN and learning speaker-specific hidden unit contributions [10].…”
Section: Introductionmentioning
confidence: 99%