2021
DOI: 10.1109/lra.2021.3062303
|View full text |Cite
|
Sign up to set email alerts
|

Bi-Directional Domain Adaptation for Sim2Real Transfer of Embodied Navigation Agents

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
19
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 51 publications
(19 citation statements)
references
References 15 publications
0
19
0
Order By: Relevance
“…Education Because simulation provides the ability to train a large number of robots in parallel and provides rich data, Truong et al [353] used educational simulations before deploying the robots. They proposed bidirectional domain adaptation (BDA), an approach that connects the sim-vs-real gap in both directions for point goal navigation.…”
Section: ) Metaverse Applicationsmentioning
confidence: 99%
See 1 more Smart Citation
“…Education Because simulation provides the ability to train a large number of robots in parallel and provides rich data, Truong et al [353] used educational simulations before deploying the robots. They proposed bidirectional domain adaptation (BDA), an approach that connects the sim-vs-real gap in both directions for point goal navigation.…”
Section: ) Metaverse Applicationsmentioning
confidence: 99%
“…SocialOnaolapo et al[346], Bailey et al[347], Luria and Foulds[348], Do et al[349] Marketing Fernanda[350], Avadhanula et al[351], Sinha et al[352] Education Truong et al[353] …”
mentioning
confidence: 99%
“…But simulated locomotion is hard to replicate on a real platform, even after hand-tuning, which emphasizes the importance of learning a transferable policy that is the focus of this work. Domain randomization [37], [38], [39] is a possible way to close the sim2real gap. Previous domain randomization efforts [37] often assume the system parameters are within a Gaussian distribution and use a non-differentiable physics engine.…”
Section: Related Workmentioning
confidence: 99%
“…Significant progress has been made in the problem of adapting policies trained with RL from a source to a target environment. Unlike RO-BUSTNAV, major assumptions involved in such transfer settings are either access to task-supervision in the target environment [24] or access to paired data from the source and target environments [23,54]. Domain Randomization (DR) [2,46,38,42] is another common approach to train policies robust to various environmental factors.…”
Section: Related Workmentioning
confidence: 99%