2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW) 2021
DOI: 10.1109/iccvw54120.2021.00303
|View full text |Cite
|
Sign up to set email alerts
|

Learning Tracking Representations via Dual-Branch Fully Transformer Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
27
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 49 publications
(27 citation statements)
references
References 39 publications
0
27
0
Order By: Relevance
“…STARK [40] employs the original transformer structure by concatenating the template and search region features. Du-alTFR and SwinTrack [38,22] use transformer as the backbone network. In this work, we develop an efficient feature fusion network based on transformer.…”
Section: Transformer In Trackingmentioning
confidence: 99%
“…STARK [40] employs the original transformer structure by concatenating the template and search region features. Du-alTFR and SwinTrack [38,22] use transformer as the backbone network. In this work, we develop an efficient feature fusion network based on transformer.…”
Section: Transformer In Trackingmentioning
confidence: 99%
“…To improve this, our discriminative target-dependent features can greatly lighten the burden for the online DCF. Recent rising Transformer-based methods [5,42,48,52,55] exploit the long-range modelling of Transformer to effectively fuse the features. Thus, they can track robustly without online learning.…”
Section: Related Workmentioning
confidence: 99%
“…The training datasets include the train subsets of LaSOT [13], GOT-10K [19], COCO2017 [26], and TrackingNet [30]. Other settings are the same with [5,48]. Details are in supplement.…”
Section: Implementation Detailsmentioning
confidence: 99%
See 1 more Smart Citation
“…Efficient Transformers. Though ViT has achieved great success in vision tasks [1,5,11,14,17,26,47,49], it needs huge computational resources to achieve comparable performance to ResNet [16] with a similar model size trained on ImageNet-1K. Therefore, how to build a more efficient Transformer draws researchers' interest.…”
Section: Related Workmentioning
confidence: 99%