2021
DOI: 10.48550/arxiv.2106.08053
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On the Power of Multitask Representation Learning in Linear MDP

Rui Lu,
Gao Huang,
Simon S. Du

Abstract: While multitask representation learning has become a popular approach in reinforcement learning (RL), theoretical understanding of why and when it works remains limited. This paper presents analyses for the statistical benefit of multitask representation learning in linear Markov Decision Process (MDP) under a generative model. In this paper, we consider an agent to learn a representation function φ out of a function class Φ from T source tasks with N data per task, and then use the learned φ to reduce the req… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

1
24
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 9 publications
(25 citation statements)
references
References 20 publications
1
24
0
Order By: Relevance
“…In contrast, REP-UCB carefully trades exploration versus exploitation by combining the reward signal and exploration bonus (constructed using the latest learned representation), and enables data sharing across all time steps. Our sample complexity nearly matches the ones from those computationally inefficient algorithms (Jiang et al, 2017;Sun et al, 2019;Du et al, 2021). We summarize the comparison with the prior works that study representation learning in Table 1.…”
Section: Introductionsupporting
confidence: 65%
See 3 more Smart Citations
“…In contrast, REP-UCB carefully trades exploration versus exploitation by combining the reward signal and exploration bonus (constructed using the latest learned representation), and enables data sharing across all time steps. Our sample complexity nearly matches the ones from those computationally inefficient algorithms (Jiang et al, 2017;Sun et al, 2019;Du et al, 2021). We summarize the comparison with the prior works that study representation learning in Table 1.…”
Section: Introductionsupporting
confidence: 65%
“…OLIVE (Jiang et al, 2017), Witness rank (Sun et al, 2019) and Bilinear-UCB (Du et al, 2021), when specialized to low-rank MDPs, have slightly tighter dependence on d (e.g., O(d 2 / 2 )). But these algorithms are computationally inefficient as they are version space algorithms.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Recently, Du et al (2020) and Tripuraneni et al (2021) consider the multi-task regression problem and show that under a shared low-dimensional representation, improved statistical rates can be obtained if the intrinsic dimension is small. Further, Hu et al (2021) and Lu et al (2021) show similar results for linear contextual bandits and Markov decision processes.…”
Section: Introductionmentioning
confidence: 61%