2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW) 2021
DOI: 10.1109/iccvw54120.2021.00317
|View full text |Cite
|
Sign up to set email alerts
|

VisDrone-CC2021: The Vision Meets Drone Crowd Counting Challenge Results

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 20 publications
(11 citation statements)
references
References 21 publications
0
11
0
Order By: Relevance
“…Compared with the best result of Bayesian Loss [70], our model has improved MAE and MSE by 0.54 and 0.53 on the RGBT-CC dataset. Similarly, our method is compared with multiple best-performing multimodal fusion models of UCNet [66], HDFNet [67], and BBSNet [68], and compared with the best-performing BBSNet [7], it is found that our model has 1.4 and 0.34 improvement in MAE and MSE on the RGBT-CC dataset. In the literature of the Sun Yat-Sen University team, the integration of the IADM "early fusion" mechanism into the classical counting model networks such as MCNN [33], SANet [69], CSRNet [33], and Bayesian Loss [70] can improve the performance of the model.…”
Section: Density Regression Methodmentioning
confidence: 98%
“…Compared with the best result of Bayesian Loss [70], our model has improved MAE and MSE by 0.54 and 0.53 on the RGBT-CC dataset. Similarly, our method is compared with multiple best-performing multimodal fusion models of UCNet [66], HDFNet [67], and BBSNet [68], and compared with the best-performing BBSNet [7], it is found that our model has 1.4 and 0.34 improvement in MAE and MSE on the RGBT-CC dataset. In the literature of the Sun Yat-Sen University team, the integration of the IADM "early fusion" mechanism into the classical counting model networks such as MCNN [33], SANet [69], CSRNet [33], and Bayesian Loss [70] can improve the performance of the model.…”
Section: Density Regression Methodmentioning
confidence: 98%
“…How to fine-tune deep neural architectures to achieve an optimal balance between precision and performance is an active research area. The VisDrone Crowd Counting challenge was introduced to encourage research in this direction [8,9]; nevertheless, the solutions proposed by the participants in the challenge are not always focused on efficiency but rather on effectiveness, as the goal is only to obtain a low error in counting people. The lowest error was obtained with TransCrowd [18], based on the increasingly popular Vision Transformer [19].…”
Section: Related Workmentioning
confidence: 99%
“…This was done in our previous preliminary work [7] but, while effective, this approach proved too demanding from a computational point of view. The method was tested on the recently proposed Crowd Counting 2020 [8] and 2021 [9] datasets, used annually for the international VisDrone challenge. The peculiarity of these datasets is that they are not characterized by still images but actually by frames of video sequences that are used here to perform the crowd flow detection task.…”
Section: Introductionmentioning
confidence: 99%
“…The data set we used is VisDrone2021-counting [ 26 ], from the 2021 Vision Meets Drones: A Challenge. The data set is divided into two parts: train- and test-challenge, including 1807 and 912 RGB images, respectively.…”
Section: Experiments and Evaluationmentioning
confidence: 99%