2020
DOI: 10.48550/arxiv.2006.07872
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Explicitly Modeled Attention Maps for Image Classification

Abstract: Self-attention networks have shown remarkable progress in computer vision tasks such as image classification. The main benefit of the self-attention mechanism is the ability to capture long-range feature interactions in attention-maps. However, the computation of attention-maps requires a learnable key, query, and positional encoding, whose usage is often not intuitive and computationally expensive. To mitigate this problem, we propose a novel self-attention module with explicitly modeled attention-maps using … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 15 publications
0
1
0
Order By: Relevance
“…However, none of these methods can escape the matrix multiplication framework, which has two problems -instability of the attention map and high computational complexity. Tan et al [35] describes the problem that the attention map is not intuitive. Their visualization results of the attention map show that content-dependent key and query play a minor role in the final attention-maps.…”
Section: Introductionmentioning
confidence: 99%
“…However, none of these methods can escape the matrix multiplication framework, which has two problems -instability of the attention map and high computational complexity. Tan et al [35] describes the problem that the attention map is not intuitive. Their visualization results of the attention map show that content-dependent key and query play a minor role in the final attention-maps.…”
Section: Introductionmentioning
confidence: 99%