2023
DOI: 10.48550/arxiv.2301.03044
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Survey on Transformers in Reinforcement Learning

Abstract: Transformer has been considered the dominating neural architecture in NLP and CV, mostly under a supervised setting. Recently, a similar surge of using Transformers has appeared in the domain of reinforcement learning (RL), but it is faced with unique design choices and challenges brought by the nature of RL. However, the evolution of Transformers in RL has not yet been well unraveled. Hence, in this paper, we seek to systematically review motivations and progress on using Transformers in RL, provide a taxonom… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
1
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 35 publications
0
3
0
Order By: Relevance
“…In recent years, the transformer architecture [21] has revolutionized the learning paradigm across numerous fields of artificial intelligence and proven superior performance over the convolutional neural network and RNN. Inspired by the success of transformers in other domains (mainly natural language processing), there has been a surge of interest in applying these scalable, attention-based models to RL [22].…”
Section: Transformers In Reinforcement Learningmentioning
confidence: 99%
“…In recent years, the transformer architecture [21] has revolutionized the learning paradigm across numerous fields of artificial intelligence and proven superior performance over the convolutional neural network and RNN. Inspired by the success of transformers in other domains (mainly natural language processing), there has been a surge of interest in applying these scalable, attention-based models to RL [22].…”
Section: Transformers In Reinforcement Learningmentioning
confidence: 99%
“…Indeed, the limitations of RCSL are not confined to stochastic settings. RCSL methods may exhibit suboptimal performance even in deterministic environments when subopti-mal data is prevalent (Li et al 2023). In deterministic environments, the uncertainty and approximation errors within the behavior policy introduce a form of stochasticity that resembles environmental stochasticity.…”
Section: Stitching In Offline Rlmentioning
confidence: 99%
“…However, despite the rapid adoption of the transformer architecture in reinforcement learning (RL) after the release of Decision Transformer (DT) (Chen et al, 2021;Hu et al, 2022;Agarwal et al, 2023;Li et al, 2023), models with in-context learning capabilities appeared only recently. This delay is caused by a number of reasons.…”
Section: Introductionmentioning
confidence: 99%