2022
DOI: 10.48550/arxiv.2201.10801
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

When Shift Operation Meets Vision Transformer: An Extremely Simple Alternative to Attention Mechanism

Abstract: Attention mechanism has been widely believed as the key to success of vision transformers (ViTs), since it provides a flexible and powerful way to model spatial relationships. However, is the attention mechanism truly an indispensable part of ViT? Can it be replaced by some other alternatives? To demystify the role of attention mechanism, we simplify it into an extremely simple case: ZERO FLOP and ZERO parameter. Concretely, we revisit the shift operation. It does not contain any parameter or arithmetic calcul… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 21 publications
0
1
0
Order By: Relevance
“…This model is designed with compromising the accuracy performance of small-sized datasets. Similarly, the Shift-ViT model introduced by Wang et al [27] replaced attention with zero parameter shift operation. The model classification was performed by linear layer.…”
Section: Related Workmentioning
confidence: 99%
“…This model is designed with compromising the accuracy performance of small-sized datasets. Similarly, the Shift-ViT model introduced by Wang et al [27] replaced attention with zero parameter shift operation. The model classification was performed by linear layer.…”
Section: Related Workmentioning
confidence: 99%