2018
DOI: 10.48550/arxiv.1806.05631
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning in POMDPs with Monte Carlo Tree Search

Abstract: The POMDP is a powerful framework for reasoning under outcome and information uncertainty, but constructing an accurate POMDP model is difficult. Bayes-Adaptive Partially Observable Markov Decision Processes (BA-POMDPs) extend POMDPs to allow the model to be learned during execution. BA-POMDPs are a Bayesian RL approach that, in principle, allows for an optimal trade-off between exploitation and exploration. Unfortunately, BA-POMDPs are currently impractical to solve for any non-trivial domain. In this paper, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2019
2019
2019
2019

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(3 citation statements)
references
References 0 publications
0
3
0
Order By: Relevance
“…One restriction of this framework is that it assumes a fixed reactive probabilistic model of the opponent, implying stationary behavior without rationality. To mitigate performance degradation due to modeling uncertainty, existing approaches include Bayesian-Adaptive POMDP (BA-POMDP) [6], [7], robust POMDP [8], Chance-constrained POMDP (CC-POMDP) [9], and Interactive-POMDP (I-POMDP) [10].…”
Section: A Pomdp Frameworkmentioning
confidence: 99%
See 2 more Smart Citations
“…One restriction of this framework is that it assumes a fixed reactive probabilistic model of the opponent, implying stationary behavior without rationality. To mitigate performance degradation due to modeling uncertainty, existing approaches include Bayesian-Adaptive POMDP (BA-POMDP) [6], [7], robust POMDP [8], Chance-constrained POMDP (CC-POMDP) [9], and Interactive-POMDP (I-POMDP) [10].…”
Section: A Pomdp Frameworkmentioning
confidence: 99%
“…BA-POMDP augments the state space with a state transition count and a state observation count variables as additional hidden states [6], [7]. It maintains a belief over the augmented state space, resulting in an optimal trade-off between model learning and reward collecting.…”
Section: A Pomdp Frameworkmentioning
confidence: 99%
See 1 more Smart Citation