2021
DOI: 10.48550/arxiv.2103.07492
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Continual Learning for Recurrent Neural Networks: an Empirical Evaluation

Andrea Cossu,
Antonio Carta,
Vincenzo Lomonaco
et al.

Abstract: Learning continuously during all model lifetime is fundamental to deploy machine learning solutions robust to drifts in the data distribution. Advances in Continual Learning (CL) with recurrent neural networks could pave the way to a large number of applications where incoming data is non stationary, like natural language processing and robotics. However, the existing body of work on the topic is still fragmented, with approaches which are application-specific and whose assessment is based on heterogeneous lea… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
3

Relationship

2
1

Authors

Journals

citations
Cited by 3 publications
(7 citation statements)
references
References 52 publications
0
7
0
Order By: Relevance
“…However, as noted by Cossu et al [11], this does not capture the full breadth of potential scenarios. For example, newly encountered datasets may introduce a mix of domain shifted instances of old classes, new classes, or novel combinations of classes.…”
Section: A Continual Learning Scenariosmentioning
confidence: 94%
See 3 more Smart Citations
“…However, as noted by Cossu et al [11], this does not capture the full breadth of potential scenarios. For example, newly encountered datasets may introduce a mix of domain shifted instances of old classes, new classes, or novel combinations of classes.…”
Section: A Continual Learning Scenariosmentioning
confidence: 94%
“…Strategy specific hyperparameters were tuned independently for each method. Hyperparameters were sampled from a range of reasonable values determined from the literature [36,11]. Where methods shared identical or analogous parameters, the search-space was also shared to ensure fair comparison (for example, regularization strength in EWC, SI, and LwF).…”
Section: B Ontology Of Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…to Robotics [29,39,41], from Reinforcement Learning [26,32] to Sequence Learning [8], among others. However, continual learning algorithms today are often designed and implemented from scratch with different assumptions, settings, and benchmarks that make them diffi-3595 cult to compare among each other or even port to slightly different contexts.…”
Section: Introductionmentioning
confidence: 99%