2014
DOI: 10.1007/978-3-662-44845-8_43
|View full text |Cite
|
Sign up to set email alerts
|

Active Learning Is Planning: Nonmyopic ε-Bayes-Optimal Active Learning of Gaussian Processes

Abstract: Abstract. A fundamental issue in active learning of Gaussian processes is that of the exploration-exploitation trade-off. This paper presents a novel nonmyopic -Bayes-optimal active learning ( -BAL) approach [4] that jointly optimizes the trade-off. In contrast, existing works have primarily developed greedy algorithms or performed exploration and exploitation separately. To perform active learning in real time, we then propose an anytime algorithm [4] based on -BAL with performance guarantee and empirically d… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
43
0

Year Published

2014
2014
2018
2018

Publication Types

Select...
3
2

Relationship

2
3

Authors

Journals

citations
Cited by 29 publications
(43 citation statements)
references
References 7 publications
0
43
0
Order By: Relevance
“…To address this trade-off, one principled approach is to frame active sensing as a sequential decision problem that jointly optimizes the above exploration-exploitation trade-off while maintaining a Bayesian belief over the model parameters. Solving this problem then results in an induced policy that is guaranteed to be optimal in the expected active sensing performance [13]. Unfortunately, such a nonmyopic Bayes-optimal active learning (BAL) policy cannot be derived exactly due to an uncountable set of candidate observations and unknown model parameters.…”
Section: Includementioning
confidence: 99%
See 4 more Smart Citations
“…To address this trade-off, one principled approach is to frame active sensing as a sequential decision problem that jointly optimizes the above exploration-exploitation trade-off while maintaining a Bayesian belief over the model parameters. Solving this problem then results in an induced policy that is guaranteed to be optimal in the expected active sensing performance [13]. Unfortunately, such a nonmyopic Bayes-optimal active learning (BAL) policy cannot be derived exactly due to an uncountable set of candidate observations and unknown model parameters.…”
Section: Includementioning
confidence: 99%
“…As a result, existing works advocate using greedy policies [24] or performing exploration and exploitation separately [15] to sidestep the difficulty of solving for the exact BAL policy. But, these algorithms are sub-optimal in the presence of budget constraints due to their imbalance between exploration and exploitation [13].…”
Section: Includementioning
confidence: 99%
See 3 more Smart Citations