2023
DOI: 10.3233/faia230359
|View full text |Cite
|
Sign up to set email alerts
|

Pretraining the Vision Transformer Using Self-Supervised Methods for Vision Based Deep Reinforcement Learning

Manuel Goulão,
Arlindo L. Oliveira

Abstract: The Vision Transformer architecture has shown to be competitive in the computer vision (CV) space where it has dethroned convolution-based networks in several benchmarks. Nevertheless, convolutional neural networks (CNN) remain the preferential architecture for the representation module in reinforcement learning. In this work, we study pretraining a Vision Transformer using several state-of-the-art self-supervised methods and assess the quality of the learned representations. To show the importance of the temp… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
references
References 39 publications
0
0
0
Order By: Relevance