2022
DOI: 10.1609/aaai.v36i2.20016
|View full text |Cite
|
Sign up to set email alerts
|

Cross-Modal Object Tracking: Modality-Aware Representations and a Unified Benchmark

Abstract: In many visual systems, visual tracking often bases on RGB image sequences, in which some targets are invalid in low-light conditions, and tracking performance is thus affected significantly. Introducing other modalities such as depth and infrared data is an effective way to handle imaging limitations of individual sources, but multi-modal imaging platforms usually require elaborate designs and cannot be applied in many real-world applications at present. Near-infrared (NIR) imaging becomes an essential part o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
21
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 7 publications
(21 citation statements)
references
References 26 publications
0
21
0
Order By: Relevance
“…In Fig. 3, we present the results of adaptive weight prediction for a typical sequence, comparing them with the tracking results of the baseline tracker DiMP [29] and MArMOT [12]. It is evident that our method demonstrates high accuracy in both fusion weight prediction and tracking results, thereby validating the effectiveness of our approach.…”
Section: A Modality-aware Fusion Modulementioning
confidence: 62%
See 4 more Smart Citations
“…In Fig. 3, we present the results of adaptive weight prediction for a typical sequence, comparing them with the tracking results of the baseline tracker DiMP [29] and MArMOT [12]. It is evident that our method demonstrates high accuracy in both fusion weight prediction and tracking results, thereby validating the effectiveness of our approach.…”
Section: A Modality-aware Fusion Modulementioning
confidence: 62%
“…Then, we embed MAFM to bridge the appearance gap between RGB and NIR modalities by learning modality-aware fusion target representations through the proposed adaptive weighting mechanism. Finally, [29] and MArMOT [12]. the modality-aware fusion feature is sent to the classification branch and regression branch to perform target localization.…”
Section: B Integration With Tracking Frameworkmentioning
confidence: 99%
See 3 more Smart Citations