2022
DOI: 10.48550/arxiv.2201.00248
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Transfer RL across Observation Feature Spaces via Model-Based Regularization

Abstract: In many reinforcement learning (RL) applications, the observation space is specified by human developers and restricted by physical realizations, and may thus be subject to dramatic changes over time (e.g. increased number of observable features). However, when the observation space changes, the previous policy will likely fail due to the mismatch of input features, and another policy must be trained from scratch, which is inefficient in terms of computation and sample complexity. Following theoretical insight… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4

Citation Types

0
4
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 21 publications
0
4
0
Order By: Relevance
“…In transfer RL, most previous work [22,13] focuses on transferring under certain prior knowledge between the two observation spaces. Very recent work from Sun et al [21] is closest to the settings considered in this paper with a different solution than ours. Sun et al [21] tackled drastic changes in observation spaces and proposed an algorithm transferring the policy from source domain via learning a sufficient representation.…”
Section: Related Workmentioning
confidence: 86%
See 2 more Smart Citations
“…In transfer RL, most previous work [22,13] focuses on transferring under certain prior knowledge between the two observation spaces. Very recent work from Sun et al [21] is closest to the settings considered in this paper with a different solution than ours. Sun et al [21] tackled drastic changes in observation spaces and proposed an algorithm transferring the policy from source domain via learning a sufficient representation.…”
Section: Related Workmentioning
confidence: 86%
“…Very recent work from Sun et al [21] is closest to the settings considered in this paper with a different solution than ours. Sun et al [21] tackled drastic changes in observation spaces and proposed an algorithm transferring the policy from source domain via learning a sufficient representation. It provide an asymptotic guarantee of the policy learning given some representation condition, and empirical validation of the algorithm.…”
Section: Related Workmentioning
confidence: 86%
See 1 more Smart Citation
“…Since collective behavior data of simulated boids, simulated robots and real robots, have the same set of features, but with different observation space ranges and distributions (Abpeikar et al, 2022b ), the proposed method in this paper focuses on feature-based (observation space) transfer learning. Some feature-based transfer learning methods applied to RL are based on distribution similarity (Zhong et al, 2018 ), model-based regularization (Sun et al, 2022 ), and feature-space re-mapping (Feuz and Cook, 2015 ). The transfer learning on observation space used in this paper is based on using the Kullback-Leibler Divergence (KLD) method described by Zhong et al ( 2018 ).…”
Section: Background and Related Workmentioning
confidence: 99%