2023
DOI: 10.48550/arxiv.2301.12050
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Do Embodied Agents Dream of Pixelated Sheep?: Embodied Decision Making using Language Guided World Modelling

Abstract: Reinforcement learning (RL) agents typically learn tabula rasa, without prior knowledge of the world, which makes learning complex tasks with sparse rewards difficult. If initialized with knowledge of high-level subgoals and transitions between subgoals, RL agents could utilize this Abstract World Model (AWM) for planning and exploration. We propose using few-shot large language models (LLMs) to hypothesize an AWM, that is tested and verified during exploration, to improve sample efficiency in embodied RL agen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 18 publications
0
5
0
Order By: Relevance
“…Therefore, large language models are often combined with agent decision-making, usually introducing the text knowledge of large language models to improve agent training and generalization [32]. In cases of vast state spaces, large language models are used to directly make decisions for agents [33]. Unlike all previous research, in this paper, we focus on using large language models for trajectory selection to enhance the training speed of reinforcement learning agents.…”
Section: Related Workmentioning
confidence: 99%
“…Therefore, large language models are often combined with agent decision-making, usually introducing the text knowledge of large language models to improve agent training and generalization [32]. In cases of vast state spaces, large language models are used to directly make decisions for agents [33]. Unlike all previous research, in this paper, we focus on using large language models for trajectory selection to enhance the training speed of reinforcement learning agents.…”
Section: Related Workmentioning
confidence: 99%
“…Dialogue for information gathering as in [160] can accelerate training. Other works such as [106,143] employ autonomous agents for embodied decision-making and exploration guided by internal world models.…”
Section: Engineeringmentioning
confidence: 99%
“…SmolModels [44], DemoGPT [41] Aerospace Engineering IELLM [107] Industrial Automation GPT4IA [144] Robotics & Embodied AI Planner-Actor-Reporter [25], Dialogue Shaping [160], DECKARD [106], TaPA [143],…”
Section: General Autonomous Ai Agentmentioning
confidence: 99%
“…• Source of Inspiration, Suggestions: LLMs can serve as an invaluable source of inspiration and suggestions, helping users brainstorm ideas [602], create content [603], and make decisions [604]. • Copilots Over Autonomous Agents: Given its limitations, LLMs are better suited as a 'copilot' that provides assistance and suggestions, rather than an autonomous agent that acts without human input or oversight [605], [606].…”
Section: B Ethical Considerationsmentioning
confidence: 99%