2021
DOI: 10.48550/arxiv.2111.13119
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Interesting Object, Curious Agent: Learning Task-Agnostic Exploration

Abstract: Common approaches for task-agnostic exploration learn tabula-rasa -the agent assumes isolated environments and no prior knowledge or experience. However, in the real world, agents learn in many environments and always come with prior experiences as they explore new ones. Exploration is a lifelong process. In this paper, we propose a paradigm change in the formulation and evaluation of taskagnostic exploration. In this setup, the agent first learns to explore across many environments without any extrinsic goal … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(6 citation statements)
references
References 23 publications
0
6
0
Order By: Relevance
“…Baseline . To establish a baseline for the navigation tasks, we compare our method against: C-BET [ 16 ], an RL algorithm combining model-based planning with uncertainty estimation for efficient exploration and decision-making. Random network distillation (RND) [ 58 ], integrates intrinsic curiosity-driven exploration to incentivise the agent’s visitation of novel states, meant to foster a deeper understanding of the environment.…”
Section: Resultsmentioning
confidence: 99%
See 4 more Smart Citations
“…Baseline . To establish a baseline for the navigation tasks, we compare our method against: C-BET [ 16 ], an RL algorithm combining model-based planning with uncertainty estimation for efficient exploration and decision-making. Random network distillation (RND) [ 58 ], integrates intrinsic curiosity-driven exploration to incentivise the agent’s visitation of novel states, meant to foster a deeper understanding of the environment.…”
Section: Resultsmentioning
confidence: 99%
“…C-BET [ 16 ], an RL algorithm combining model-based planning with uncertainty estimation for efficient exploration and decision-making.…”
Section: Resultsmentioning
confidence: 99%
See 3 more Smart Citations