2022
DOI: 10.1109/access.2022.3170582
|View full text |Cite
|
Sign up to set email alerts
|

Off-Policy Meta-Reinforcement Learning With Belief-Based Task Inference

Abstract: Meta-reinforcement learning (RL) addresses the problem of sample inefficiency in deep RL by using experience obtained in past tasks for solving a new task. However, most existing meta-RL methods require partially or fully on-policy data, which hinders the improvement of sample efficiency. To alleviate this problem, we propose a novel off-policy meta-RL method, embedding learning and uncertainty evaluation (ELUE). An ELUE agent is characterized by the learning of what we call a task embedding space, an embeddin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(7 citation statements)
references
References 49 publications
0
7
0
Order By: Relevance
“…Some methods use both PPG and black-box components [252,187]. In particular, even when training a fully black-box method, the policy or inner-loop can be fine-tuned with policy gradients at meta-test time [116,252,97].…”
Section: Black Box Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…Some methods use both PPG and black-box components [252,187]. In particular, even when training a fully black-box method, the policy or inner-loop can be fine-tuned with policy gradients at meta-test time [116,252,97].…”
Section: Black Box Methodsmentioning
confidence: 99%
“…Finally, we conclude with a discussion of the trade-offs concerning task inference methods. Table 2 summarizes these categories and task inference methods [95,185,104,213,269,281,64,130,282,17,97].…”
Section: Task Inference Methodsmentioning
confidence: 99%
See 3 more Smart Citations