2009
DOI: 10.1007/978-3-642-04174-7_20
|View full text |Cite
|
Sign up to set email alerts
|

Boosting Active Learning to Optimality: A Tractable Monte-Carlo, Billiard-Based Algorithm

Abstract: Abstract. This paper focuses on Active Learning with a limited number of queries; in application domains such as Numerical Engineering, the size of the training set might be limited to a few dozen or hundred examples due to computational constraints. Active Learning under bounded resources is formalized as a finite horizon Reinforcement Learning problem, where the sampling strategy aims at minimizing the expectation of the generalization error. A tractable approximation of the optimal (intractable) policy is p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
12
0

Year Published

2009
2009
2015
2015

Publication Types

Select...
4
2
1

Relationship

3
4

Authors

Journals

citations
Cited by 15 publications
(12 citation statements)
references
References 28 publications
0
12
0
Order By: Relevance
“…Usually these schedules restrict the number of children to grow as the logarithm or some root of the number of simulations -examples of each case can be found in [12] and [27] respectively. Unlike FPU, this approach could be very poor if legal actions for expansion are selected entirely randomly: even if initial actions look poor, the schedule prevents further exploration.…”
Section: Tree Policies For Large Branching Factorsmentioning
confidence: 99%
See 1 more Smart Citation
“…Usually these schedules restrict the number of children to grow as the logarithm or some root of the number of simulations -examples of each case can be found in [12] and [27] respectively. Unlike FPU, this approach could be very poor if legal actions for expansion are selected entirely randomly: even if initial actions look poor, the schedule prevents further exploration.…”
Section: Tree Policies For Large Branching Factorsmentioning
confidence: 99%
“…Unlike FPU, this approach could be very poor if legal actions for expansion are selected entirely randomly: even if initial actions look poor, the schedule prevents further exploration. For this reason, progressive widening orders the legal actions based on some quality heuristic [27] (such as an evaluation function), and expands them in decreasing order of the heuristic.…”
Section: Tree Policies For Large Branching Factorsmentioning
confidence: 99%
“…Indeed, MCTS has been recently used by Gaudel and Sebag (2010) in their FUSE (Feature Uct SElection) system to perform feature selection, and by Rolet et al (2009) in BAAL (Banditbased Active Learner) for active learning with small training sets. Gaudel and Sebag (2010) firstly formalize feature selection as a Reinforcement Learning (RL) problem and then provide an approximation of the optimal policy by casting the RL problem as a one-player game whose states are all possible subsets of features and whose actions consist of choosing a feature and adding it to a subset.…”
Section: Related Workmentioning
confidence: 99%
“…The problem is then solved with the UCT approach leading to the FUSE algorithm. Rolet et al (2009) focus on Active Learning (AL) with a limited number of queries. The authors formalized AL under bounded resources as a finite horizon RL problem.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation