Search citation statements
Paper Sections
Citation Types
Year Published
Publication Types
Relationship
Authors
Journals
Currently, with the widespread application of embedded technology and the continuous improvement of computational power in mobile terminals, the efficient deployment of algorithms on embedded devices, while maintaining high accuracy and minimizing model size, has become a research hotspot. This paper addresses the challenges of deploying the YOLOv8 algorithm on embedded devices and proposes a novel lightweight object detection algorithm focusing on small object detection. We optimize the model through two key strategies, aiming to achieve lightweight deployment and improve the accuracy of small object detection. Firstly, GhostNet is introduced as the backbone network for YOLOv8 in order to achieve lightweight deployment. By using some cost effective operations to generate redundant feature maps, we not only reduce the number of model parameters while ensuring better detection results, but also improve the speed of the model. Secondly, a new multi-scale attention module is designed to enhance the network's acquisition of crucial information for small targets, which includes a multi-scale fusion attention mechanism and the Soft-NMS algorithm. The multi-scale fusion attention mechanism captures key features of discriminative small targets in the feature map tensor from both spatial and channel dimensions, suppressing non-key information, reducing the impact of complex and unimportant information in the image, enhancing the network model's learning ability for important features of small targets. The Soft-NMS method improves accuracy by significantly reduces false positives in the detection results. To validate the performance of our proposed method, we conducted validation experiments on the PASCAL VOC dataset and evaluated the model's generalization ability on the MS COCO dataset. The experiments results demonstrate that our model achieves a significant improvement in small object detection, with a 5.41% increase in detection accuracy compared to the existing YOLOv8. Meanwhile, FLOPs are reduced by 49.62%, and the number of model parameters is reduced by 48.66%. These results fully confirm the effectiveness of our innovative method in achieving both lightweight deployment and significant efficacy in small object detection tasks.
Currently, with the widespread application of embedded technology and the continuous improvement of computational power in mobile terminals, the efficient deployment of algorithms on embedded devices, while maintaining high accuracy and minimizing model size, has become a research hotspot. This paper addresses the challenges of deploying the YOLOv8 algorithm on embedded devices and proposes a novel lightweight object detection algorithm focusing on small object detection. We optimize the model through two key strategies, aiming to achieve lightweight deployment and improve the accuracy of small object detection. Firstly, GhostNet is introduced as the backbone network for YOLOv8 in order to achieve lightweight deployment. By using some cost effective operations to generate redundant feature maps, we not only reduce the number of model parameters while ensuring better detection results, but also improve the speed of the model. Secondly, a new multi-scale attention module is designed to enhance the network's acquisition of crucial information for small targets, which includes a multi-scale fusion attention mechanism and the Soft-NMS algorithm. The multi-scale fusion attention mechanism captures key features of discriminative small targets in the feature map tensor from both spatial and channel dimensions, suppressing non-key information, reducing the impact of complex and unimportant information in the image, enhancing the network model's learning ability for important features of small targets. The Soft-NMS method improves accuracy by significantly reduces false positives in the detection results. To validate the performance of our proposed method, we conducted validation experiments on the PASCAL VOC dataset and evaluated the model's generalization ability on the MS COCO dataset. The experiments results demonstrate that our model achieves a significant improvement in small object detection, with a 5.41% increase in detection accuracy compared to the existing YOLOv8. Meanwhile, FLOPs are reduced by 49.62%, and the number of model parameters is reduced by 48.66%. These results fully confirm the effectiveness of our innovative method in achieving both lightweight deployment and significant efficacy in small object detection tasks.
Instance segmentation in Remote Sensing Images (RSI) poses significant challenges due to the diverse scales of targets, scene complexity, and a high number of targets, making most methods struggle with suboptimal performance and time-consuming computations. To solve those problems, a fast and accurate RSI instance segmentation model (named DCTC) is designed in this paper. DCTC transforms classification problem into regression problem to improve the reference speed. DCTC contains two parallel branches. The contour branch performs iterative regression on contours, extracting precise contour information to improve boundary accuracy. Meanwhile, the DCT branch refines mask predictions and supplements instance context information, which particularly benefits the segmentation of small targets. DCT encoding is employed in the DCT branch to convert the mask representation into DCT format, aligning the outputs of the contour and DCT branches. Three innovative modules are introduced in the DCT branch: the Coarse Result Generation module (CRG), Iteratively Deform and Regress module (IDR), and Contour and DCT Fusion module (CDF). The CRG module generates coarse DCT vectors and contour coordinates, facilitating information exchange between the contour and DCT branches. The IDR module iteratively refines DCT vectors, enabling DCTC to focus more on small targets and instance details. The CDF module merges DCT vectors and contour coordinates, ensuring effective interaction between boundary and context information, thereby enhancing performance. Extensive experiments demonstrate the superiority of DCTC, which achieves 67.7, 36.3, 67.4, 55.1AP on NWPU VHR-10, iSAID, SAR Ship Detection Dataset (SSDD), and High-Resolution SAR Images Dataset (HRSID), and ranks first among state-of-the-art methods while maintaining real-time processing capability. Furthermore, DCTC exhibits strong performance on both optical and SAR images, and the designed DCT branch can be simply plug into any contour-based method to improve the network performance.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.