2022
DOI: 10.1109/tip.2021.3130533
|View full text |Cite
|
Sign up to set email alerts
|

LasHeR: A Large-Scale High-Diversity Benchmark for RGBT Tracking

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
55
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
2

Relationship

1
8

Authors

Journals

citations
Cited by 117 publications
(61 citation statements)
references
References 64 publications
2
55
0
Order By: Relevance
“…In this section, we evaluate our algorithm by comparing the tracking performance with some state-of-the-art trackers on three RGBT tracking benchmarks including GTOT (Li et al 2016), RGBT234 (Li et al 2019a) and LasHeR (Li et al 2021) to validate the effectiveness of proposed method and analyze effectiveness of each major component in the algorithm. In this part, we will introduce the details of the datasets, the evaluation metrics, and the implementation details of training and testing.…”
Section: Methodsmentioning
confidence: 99%
“…In this section, we evaluate our algorithm by comparing the tracking performance with some state-of-the-art trackers on three RGBT tracking benchmarks including GTOT (Li et al 2016), RGBT234 (Li et al 2019a) and LasHeR (Li et al 2021) to validate the effectiveness of proposed method and analyze effectiveness of each major component in the algorithm. In this part, we will introduce the details of the datasets, the evaluation metrics, and the implementation details of training and testing.…”
Section: Methodsmentioning
confidence: 99%
“…It is considered a positive sample when the ratio of overlap rate (IoU) between the sample and the ground truth is [0.7, 1], while a sample in the range of [0, 0.5] is considered a negative sample. Notably, we train MSIFNet with the GTOT [ 43 ] dataset and test it on RGBT234 [ 44 ] and LasHeR [ 45 ] datasets. When testing on GTOT, we randomly selected 50 video sequences on RGBT234 for training.…”
Section: Methodsmentioning
confidence: 99%
“…(i) For RGB-D tracking, we evaluate our tracker on Depthtrack [47] and VOT22-RGBD [16]. (ii) For RGB-T tracking, we provide the comparison results on RGBT234 [23] and LasHeR [25]. (iii) For RGB-E tracking, we report the experimental results on the largest VisEvent [43].…”
Section: Downstream Tasksmentioning
confidence: 99%
“…For example, the widely used RGB-based tracking datasets, GOT-10k [13], Track-ingNet [36], and LaSOT [9], contain 9.3K, 30.1K, and 1.1K sequences, corresponding to 1.4M, 14M, and 2.8M frames for training. Whereas the largest training datasets in multi-modal tracking, DepthTrack [47], LasHeR [25], VisEvent [43], contain 150, 979, 500 training sequences, corresponding to 0.22M, 0.51M, 0.21M annotated frame pairs, which is at least an order of magnitude less than the former. Accounting for the above limitation, multi-modal tracking methods [43,47,61] usually utilize pre-trained RGB-based trackers and perform fine-tuning on their taskoriented training sets (as shown in Figure 1 (a)→(b)).…”
Section: Introductionmentioning
confidence: 99%