Findings of the Association for Computational Linguistics: EMNLP 2021 2021
DOI: 10.18653/v1/2021.findings-emnlp.319
|View full text |Cite
|
Sign up to set email alerts
|

Towards Realistic Single-Task Continuous Learning Research for NER

Abstract: There is an increasing interest in continuous learning (CL), as data privacy is becoming a priority for real-world machine learning applications. Meanwhile, there is still a lack of academic NLP benchmarks that are applicable for realistic CL settings, which is a major challenge for the advancement of the field. In this paper we discuss some of the unrealistic data characteristics of public datasets, study the challenges of realistic single-task continuous learning as well as the effectiveness of data rehearsa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
4
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 7 publications
0
4
0
Order By: Relevance
“…Note that we select this setup to account for a production scenario when a model is trained in several iterations within a fixed release cycle (e.g., a period of several weeks); for simplicity, we assume that after each iteration the model is fully retrained on all data to avoid any model drift-related effects (which are out of scope of this work). This would also correspond to a 100% data replay strategy in continuous learning approaches (Payan et al, 2021).…”
Section: Gradually Adding New Featuresmentioning
confidence: 99%
See 1 more Smart Citation
“…Note that we select this setup to account for a production scenario when a model is trained in several iterations within a fixed release cycle (e.g., a period of several weeks); for simplicity, we assume that after each iteration the model is fully retrained on all data to avoid any model drift-related effects (which are out of scope of this work). This would also correspond to a 100% data replay strategy in continuous learning approaches (Payan et al, 2021).…”
Section: Gradually Adding New Featuresmentioning
confidence: 99%
“…Many previous works on continual learning have focused on learning from a continuous stream of data (Biesialska et al, 2020) or on an incremental learning of new tasks (Kanwatchara et al, 2021) and languages (Castellucci et al, 2021). Payan et al (2021) discuss a single-task continual learning setup and simulated a passive data extension scenario where new examples are coming in for all output classes on a public dataset. Similarly, Ash and Adams (2020) evaluate a batch-learning setup, where each model iteration is warm-started from the previous step and the whole training data is always available, while some new data is added across all output classes in each batch.…”
Section: Introductionmentioning
confidence: 99%
“…Note that we select this setup to account for a production scenario when a model is trained in several iterations within a fixed release cycle (e.g., a period of several weeks); for simplicity, we assume that after each iteration the model is fully retrained on all data to avoid any model drift-related effects (which are out of scope of this work). This would also correspond to a 100% data replay strategy in continuous learning approaches (Payan et al, 2021).…”
Section: Gradually Adding New Featuresmentioning
confidence: 99%
“…Many previous works on continual learning have focused on learning from a continuous stream of data (Biesialska et al, 2020) or on an incremental learning of new tasks (Kanwatchara et al, 2021) and languages (Castellucci et al, 2021). Payan et al (2021) discuss a single-task continual learning setup and simulated a passive data extension scenario where new examples are coming in for all output classes on a public dataset. Similarly, Ash and Adams (2020) evaluate a batch-learning setup, where each model iteration is warm-started from the previous step and the whole training data is always available, while some new data is added across all output classes in each batch.…”
Section: Introductionmentioning
confidence: 99%