2022
DOI: 10.48550/arxiv.2205.10712
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Housekeep: Tidying Virtual Households using Commonsense Reasoning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 0 publications
0
5
0
Order By: Relevance
“…While there is a challenge in transferring results from simulated to real environments, simulated environments are more accessible, less expensive, and allow for the testing of technologies that may not be sufficiently safe for use in the real world (Savva et al, 2019). Additionally, while simulated environments can be used for tasks that do not require the use of language (Anderson et al, 2018a;Batra et al, 2020;Gan et al, 2020;Kant et al, 2022), they play a particularly valuable role in developing language understanding and reasoning capabilities over actions that are currently difficult for physical robots to complete, but we hope it will become a reality in the future (Kolve et al, 2017). Much of the work in language understanding for embodied AI happens using vision and language navigation, where an agent must learn to navigate through a previously unseen environment purely based on natural language route instructions (Anderson et al, 2018b;.…”
Section: B Additional Related Workmentioning
confidence: 99%
“…While there is a challenge in transferring results from simulated to real environments, simulated environments are more accessible, less expensive, and allow for the testing of technologies that may not be sufficiently safe for use in the real world (Savva et al, 2019). Additionally, while simulated environments can be used for tasks that do not require the use of language (Anderson et al, 2018a;Batra et al, 2020;Gan et al, 2020;Kant et al, 2022), they play a particularly valuable role in developing language understanding and reasoning capabilities over actions that are currently difficult for physical robots to complete, but we hope it will become a reality in the future (Kolve et al, 2017). Much of the work in language understanding for embodied AI happens using vision and language navigation, where an agent must learn to navigate through a previously unseen environment purely based on natural language route instructions (Anderson et al, 2018b;.…”
Section: B Additional Related Workmentioning
confidence: 99%
“…Another stream of research attempts to leverage the Large Language Model (LLM) or large Visual Language Model (VLM) for goal specification. [24], [25] notice the necessity of automatic goal inference for tying rooms and exploit the commonsense knowledge from LLM or memex graph to infer rearrangements goals when the goal is unspecified. TidyBot [26] also leverages an LLM to summarize the rearrangement preference from a few examples provided by the user.…”
Section: A Object Rearrangement With Functional Requirementsmentioning
confidence: 99%
“…As a result, the robotics community has started to integrate the pre-trained large models into the robot learning workflow [33], [34], [35], [36], [37], [38], [7], [24], [25], [26]. In navigation, VLMaps [33] and NavGPT [34] leverage the LLM to translate natural language instructions into explicit goals or actions.…”
Section: B Leveraging Large Models For Robot Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…Many LLMs have been developed in recent years, such as BERT [24], GPT-3 [12], ChatGPT [13], CodeX [25], and OPT [26]. These LLMs can encode a large amount of common sense [14] and have been applied to robot task planning [27]- [32]. For instance, the work of Huang et.…”
Section: Robot Planning With Large Language Modelsmentioning
confidence: 99%