2021
DOI: 10.1609/aaai.v35i12.17251
|View full text |Cite
|
Sign up to set email alerts
|

Domain Adaptation In Reinforcement Learning Via Latent Unified State Representation

Abstract: Despite the recent success of deep reinforcement learning (RL), domain adaptation remains an open problem. Although the generalization ability of RL agents is critical for the real-world applicability of Deep RL, zero-shot policy transfer is still a challenging problem since even minor visual changes could make the trained agent completely fail in the new task. To address this issue, we propose a two-stage RL agent that first learns a latent unified state representation (LUSR) which is consistent across multip… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
13
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 23 publications
(13 citation statements)
references
References 12 publications
0
13
0
Order By: Relevance
“…Most domain adaptation approaches are designed to deal with the changes to the observation distribution. Current domain adaptation methods can be roughly divided into three categories: domain randomization (Tobin et al, 2017;Sadeghi & Levine, 2017;James et al, 2019), image-to-image translation (Gamrian & Goldberg, 2018a;Tzeng et al, 2015;You et al, 2017;Zhang et al, 2018), and adaptation via aligned representations (Xing et al, 2021;Higgins et al, 2017;.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…Most domain adaptation approaches are designed to deal with the changes to the observation distribution. Current domain adaptation methods can be roughly divided into three categories: domain randomization (Tobin et al, 2017;Sadeghi & Levine, 2017;James et al, 2019), image-to-image translation (Gamrian & Goldberg, 2018a;Tzeng et al, 2015;You et al, 2017;Zhang et al, 2018), and adaptation via aligned representations (Xing et al, 2021;Higgins et al, 2017;.…”
Section: Related Workmentioning
confidence: 99%
“…Ideally, representations in this latent space can share consistent semantic meanings no matter which domain they come from. For example, (Xing et al, 2021) explicitly splits the latent representations into domain-specific and domain-general features and then builds policy on the domain-general features to ignore domain-specific variations.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…TL has played an important role in accelerating singleagent RL by adapting learnt knowledge from past relevant tasks. 10,16,17 Inspired by this scenario, TL in MARL [18][19][20][21] is also studied with respect to transferring knowledge across multi-agent tasks to help improve the learning performance. The above work has two main directions: knowledge transfer across tasks and transfer among agents.…”
Section: Related Workmentioning
confidence: 99%