2020
DOI: 10.1016/j.neucom.2020.01.093
|View full text |Cite
|
Sign up to set email alerts
|

Efficient continual learning in neural networks with embedding regularization

Abstract: Continual learning of deep neural networks is a key requirement for scaling them up to more complex applicative scenarios and for achieving real lifelong learning of these architectures. Previous approaches to the problem have considered either the progressive increase in the size of the networks, or have tried to regularize the network behavior to equalize it with respect to previously observed tasks. In the latter case, it is essential to understand what type of information best represents this past behavior… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
29
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 38 publications
(29 citation statements)
references
References 25 publications
0
29
0
Order By: Relevance
“…In a recent study, the authors of [38] developed a deep learning framework to ensure a more realistic learning analysis. According to the authors, CL is a new, simple, and efficient method proven to be valid as an alternative to standard regularization techniques.…”
Section: Identification Strategymentioning
confidence: 99%
See 1 more Smart Citation
“…In a recent study, the authors of [38] developed a deep learning framework to ensure a more realistic learning analysis. According to the authors, CL is a new, simple, and efficient method proven to be valid as an alternative to standard regularization techniques.…”
Section: Identification Strategymentioning
confidence: 99%
“…While Ref. [38] adopted CL in the context of neural networks, we used this method in the broad context of machine learning. The idea of CL is to allow a refined treatment of covariates based on a bias-variance trade-off argument since it consists of a two-step approach that performs a bias-variance decomposition.…”
Section: Identification Strategymentioning
confidence: 99%
“…Usually, rehearsal and progressive strategies, performs very well but always declines as the number of tasks increase, and might require a high computational power. With some differences from the first two approaches, the implementation of regularization strategy is quite simple, they require little memory, but it performance might not be up to that of rehearsal methods [7]. One main problem encountered when applying regularization strategy is determining what task best represents the behaviour of the network and, this can lead to the form of regularization penalty that would be taken [7].…”
Section: Wherementioning
confidence: 99%
“…With remarkable successes accomplish over the past few years in AI, deep network applications are however restricted to sole, distinct problem. Where every single network has to be trained and re-trained from the beginning every single time a new task is fed into the network and as a result their training remains very challenging to deal with particularly in real-world settings and in situations where data are scarce and/or computation is costly [7]. Furthermore, the sequence of tasks may not be clearly labelled tasks and they may switch randomly, leading to an individual task recuring in long time intervals.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation