DOI: 10.29007/thws
|View full text |Cite
|
Sign up to set email alerts
|

On the Robustness of Active Learning

Abstract: Active Learning is concerned with the question of how to identify the most useful samples for a Machine Learning algorithm to be trained with. When applied correctly, it can be a very powerful tool to counteract the immense data requirements of Artificial Neural Networks. However, we find that it is often applied with not enough care and domain knowledge. As a consequence, unrealistic hopes are raised and transfer of the experimental results from one dataset to another becomes unnecessarily hard.In this work w… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

2
1
0

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 14 publications
2
1
0
Order By: Relevance
“…While two of three baselines, i.e., MC-Dropout and Ensemble, beat the random acquisition, CoreSet does not reach a comparable classification accuracy within the labeling budget. This finding is in line with previous research (Sinha et al, 2019;Hahn et al, 2019a) and can be attributed to a weakness of the utilized p-norm distance metric regarding high-dimensional data, called the distance concentration phenomenon.…”
Section: Policy Trainingsupporting
confidence: 92%
See 1 more Smart Citation
“…While two of three baselines, i.e., MC-Dropout and Ensemble, beat the random acquisition, CoreSet does not reach a comparable classification accuracy within the labeling budget. This finding is in line with previous research (Sinha et al, 2019;Hahn et al, 2019a) and can be attributed to a weakness of the utilized p-norm distance metric regarding high-dimensional data, called the distance concentration phenomenon.…”
Section: Policy Trainingsupporting
confidence: 92%
“…Our method consistently outperforms the other approaches on both datasets. On FMNIST our method is the only method that is actually able to beat a Random Sampling strategy (similar findings have previously been reported by Hahn et al (2019b)). While our method is consistently 1 − 3% better than the Random Sampling strategy on FMNIST, on the harder KMNIST dataset we are 7.9% ahead on average.…”
Section: Policy Transfersupporting
confidence: 86%
“…At the heart of an AL method is the so-called query strategy that decides which data to present to the annotator / oracle next. In general, uncertainty sampling is one of the most common query strategies [10,38,1,13,29], besides that there also exist approaches on expected model change [36] and reinforcement learning based AL [2].…”
Section: Introductionmentioning
confidence: 99%