2018
DOI: 10.1371/journal.pone.0192246
|View full text |Cite|
|
Sign up to set email alerts
|

A parallel spatiotemporal saliency and discriminative online learning method for visual target tracking in aerial videos

Abstract: Visual tracking in aerial videos is a challenging task in computer vision and remote sensing technologies due to appearance variation difficulties. Appearance variations are caused by camera and target motion, low resolution noisy images, scale changes, and pose variations. Various approaches have been proposed to deal with appearance variation difficulties in aerial videos, and amongst these methods, the spatiotemporal saliency detection approach reported promising results in the context of moving target dete… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
7
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 20 publications
(7 citation statements)
references
References 47 publications
0
7
0
Order By: Relevance
“…ii) Although the proposed spatio-temporal processing scheme leads to average F-score improvement for all the involved automatic detection algorithms, the minor loss of TPs and the remaining FPs prevented the achievement of an average F-score higher than 0.9, so this should receive further attention. iii) All seven aerial video datasets contain a limited number of image frames, so larger databases should be established so that adequate training data is available to implement object detection models using some machine learning and deep convolutional neural network (CNN) related schemes [34]- [44]; iv) GPU-based parallel computing has great potential for reducing computation time.…”
Section: Conclusion and Further Workmentioning
confidence: 99%
See 2 more Smart Citations
“…ii) Although the proposed spatio-temporal processing scheme leads to average F-score improvement for all the involved automatic detection algorithms, the minor loss of TPs and the remaining FPs prevented the achievement of an average F-score higher than 0.9, so this should receive further attention. iii) All seven aerial video datasets contain a limited number of image frames, so larger databases should be established so that adequate training data is available to implement object detection models using some machine learning and deep convolutional neural network (CNN) related schemes [34]- [44]; iv) GPU-based parallel computing has great potential for reducing computation time.…”
Section: Conclusion and Further Workmentioning
confidence: 99%
“…As future work, we plan to consider a parallel design for spatio-temporal processing via a high-performance computer [34] for better utilization of memory and, hence, decreased computation time. We also suggest improving the adapted MMA algorithm by adding a frame differencing method as a pre-processing step, then exploiting better multi-scale feature selection to improve average F-score with reference to some of the latest models [35]- [44] on object detection and their post-processing schemes [45].…”
Section: Conclusion and Further Workmentioning
confidence: 99%
See 1 more Smart Citation
“…This shift is primarily attributed to the remarkable capabilities of deep neural networks in handling complex image data. In the subsequent sections [10], [11], it will delve into the current limitations and challenges associated with deep learning-based fruit detection methods, highlighting the need for further research and innovation.…”
Section: Introductionmentioning
confidence: 99%
“…Saliency detection aims at finding the most distinctive objects in an image which are consistent with human visual perception. It is commonly utilized as a preliminary processing step to facilitate a wide range of applications such as object recognition [1], person re-identification [2], image retrieval [3], semantic segmentation [4], scene classification [5], visual tracking [6], video summarization [7] and so on.…”
Section: Introductionmentioning
confidence: 99%