2018
DOI: 10.48550/arxiv.1811.05612
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Bayesian Reinforcement Learning in Factored POMDPs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 0 publications
0
1
0
Order By: Relevance
“…With our natural definition of regret, their algorithm suffers linear regret. Other learning algorithms for POMDPs either consider linear dynamics (Lale et al, 2020b;Tsiamis & Pappas, 2020) or do not consider regret (Shani et al, 2005;Ross et al, 2007;Poupart & Vlassis, 2008;Cai et al, 2009;Liu et al, 2011;Doshi-Velez et al, 2013;Katt et al, 2018;Azizzade-nesheli et al, 2018) and are not directly comparable to our setting.…”
Section: Related Literaturementioning
confidence: 88%
“…With our natural definition of regret, their algorithm suffers linear regret. Other learning algorithms for POMDPs either consider linear dynamics (Lale et al, 2020b;Tsiamis & Pappas, 2020) or do not consider regret (Shani et al, 2005;Ross et al, 2007;Poupart & Vlassis, 2008;Cai et al, 2009;Liu et al, 2011;Doshi-Velez et al, 2013;Katt et al, 2018;Azizzade-nesheli et al, 2018) and are not directly comparable to our setting.…”
Section: Related Literaturementioning
confidence: 88%