2023
DOI: 10.2139/ssrn.4398261
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Constituent Attention for Vision Transformers

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 0 publications
0
1
0
Order By: Relevance
“…In addition, building a hybrid model combining CNN and Transformer has stronger information aggregation capabilities. The combination of the attention mechanism and hybrid model is the most important method of current high‐efficiency models [35–38]. Among these methods, on the basis of the research, MobileXT uses XCA [25] based on cross‐covariance matrix computation to reduce the time complexity.…”
Section: Related Workmentioning
confidence: 99%
“…In addition, building a hybrid model combining CNN and Transformer has stronger information aggregation capabilities. The combination of the attention mechanism and hybrid model is the most important method of current high‐efficiency models [35–38]. Among these methods, on the basis of the research, MobileXT uses XCA [25] based on cross‐covariance matrix computation to reduce the time complexity.…”
Section: Related Workmentioning
confidence: 99%