2021
DOI: 10.48550/arxiv.2110.08611
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Deep Active Learning by Leveraging Training Dynamics

Abstract: Active learning theories and methods have been extensively studied in classical statistical learning settings. However, deep active learning, i.e., active learning with deep learning models, is usually based on empirical criteria without solid theoretical justification, thus suffering from heavy doubts when some of those fail to provide benefits in applications. In this paper, by exploring the connection between the generalization performance and the training dynamics, we propose a theory-driven deep active le… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 29 publications
0
2
0
Order By: Relevance
“…Margin algorithms are flexible and can be adapted to both streaming and pool settings. In the pool setting, a line of works utilize the neural networks in active learning to improve the empirical performance [38,44,6,18,32,47,49,56,5]. However, they do not provide performance guarantee for NN-based active learning algorithms.…”
Section: Related Workmentioning
confidence: 99%
“…Margin algorithms are flexible and can be adapted to both streaming and pool settings. In the pool setting, a line of works utilize the neural networks in active learning to improve the empirical performance [38,44,6,18,32,47,49,56,5]. However, they do not provide performance guarantee for NN-based active learning algorithms.…”
Section: Related Workmentioning
confidence: 99%
“…Diversity sampling algorithms label examples that are most different from each other, based on similarity metrics such as distances in penultimate layer representations (Sener & Savarese, 2017;Geifman & El-Yaniv, 2017;Citovsky et al, 2021) or discriminator networks (Gissin & Shalev-Shwartz, 2019). Lastly, gradient embeddings, which encode both softmax likelihood and penultimate layer representation, have become widely adopted in recent approaches (Ash et al, 2019;Wang et al, 2021;Elenter et al, 2022;Mohamadi et al, 2022). As an example, Ash et al ( 2019) uses a k-means++ algorithm to query a diverse set of examples in the gradient embedding space.…”
Section: Related Workmentioning
confidence: 99%