2021
DOI: 10.48550/arxiv.2107.00602
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Importance Sampling based Exploration in Q Learning

Vijay Kumar,
Mort Webster

Abstract: Approximate Dynamic Programming (ADP) is a methodology to solve multi-stage stochastic optimization problems in multi-dimensional discrete or continuous spaces. ADP approximates the optimal value function by adaptively sampling both action and state space. It provides a tractable approach to very large problems, but can suffer from the exploration-exploitation dilemma. We propose a novel approach for selecting actions using importance sampling weighted by the value function approximation in continuous decision… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 17 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?