2021
DOI: 10.48550/arxiv.2109.11792
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Regularization Guarantees Generalization in Bayesian Reinforcement Learning through Algorithmic Stability

Abstract: In the Bayesian reinforcement learning (RL) setting, a prior distribution over the unknown problem parameters -the rewards and transitions -is assumed, and a policy that optimizes the (posterior) expected return is sought. A common approximation, which has been recently popularized as meta-RL, is to train the agent on a sample of N problem instances from the prior, with the hope that for large enough N , good generalization behavior to an unseen test instance will be obtained. In this work, we study generaliza… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 17 publications
0
1
0
Order By: Relevance
“…The supplementary material can be found in the full technical report(Tamar, Soudry, and Zisselman 2021) 3 The linear programming formulation is not suitable for establishing stability in our BRL setting, as changing the prior would change the constraints in the linear program.…”
mentioning
confidence: 99%
“…The supplementary material can be found in the full technical report(Tamar, Soudry, and Zisselman 2021) 3 The linear programming formulation is not suitable for establishing stability in our BRL setting, as changing the prior would change the constraints in the linear program.…”
mentioning
confidence: 99%