Pine wilt nematode disease is a devastating forest disease that spreads rapidly. Using drone remote sensing to monitor pine wilt nematode trees promptly is an effective way to control the spread of pine wilt nematode disease. In this study, the YOLOv4 algorithm was used to automatically identify abnormally discolored wilt from pine wilt nematode disease on UAV remote sensing images. Because the network structure of YOLOv4 is too complex, although the detection accuracy is high, the detection speed is relatively low. To solve this problem, the lightweight deep learning network MobileNetv2 is used to optimize the backbone feature extraction network. Furthermore, the YOLOv4 algorithm was improved by improving the backbone network part, adding CBAM attention, and adding the Inceptionv2 structure to reduce the number of model parameters and improve the accuracy and efficiency of identification. The speed and accuracy of the Faster R-CNN, YOLOv4, SSD, YOLOv5, and the improved MobileNetv2-YOLOv4 algorithm were compared, and the detection effects of the Faster R-CNN, YOLOv4, SSD, YOLOv5 and the improved MobileNetv2-YOLOv4 algorithm on trees with pine wilt nematode were analyzed. The experimental results show that the average precision of the improved MobileNetv2-YOLOv4 algorithm is 86.85%, the training time of each iteration cycle is 156 s, the parameter size is 39.23 MB, and the test time of a single image is 15 ms, which is better than Faster R-CNN, YOLOv4, and SSD, but comparable to YOLOv5. Compared with the advantages and disadvantages, comprehensively comparing these four indicators, the improved algorithm has a more balanced performance in the detection speed, the parameter size, and the average precision. The F1 score of the improved algorithm (95.60%) was higher than that of Faster R-CNN (90.80%), YOLOv4 (94.56%), and SSD (92.14%), which met the monitoring requirements of pine wilt nematode trees. Faster R-CNN and SSD pine-wilt-nematode tree detection models are not ideal in practical applications. Compared with the YOLOv4 pine-wilt-nematode tree detection model, the improved MobileNetv2-YOLOv4 algorithm satisfies the condition of maintaining a lower model parameter quantity to obtain higher detection accuracy; therefore, it is more suitable for practical application scenarios of embedded devices. It can be used for the rapid detection of pine wilt nematode diseased trees.
Scene text detection refers to locating text regions in a scene image and marking them out with text boxes. With the rapid development of the mobile Internet and the increasing popularity of mobile terminal devices such as smartphones, the research on scene text detection technology has been highly valued and widely applied. In recent years, with the rise of deep learning represented by convolutional neural networks, research on scene text detection has made new developments. However, scene text detection is still a very challenging task due to the following two factors. Firstly, images in natural scenes often have complex backgrounds, which can easily interfere with the detection process. Secondly, the text in natural scenes is very diverse, with horizontal, skewed, straight, and curved text, all of which may be present in the same scene. As convolutional neural networks extract features, the convolutional layer with limited perceptual field cannot model the global semantic information well. Therefore, this paper further proposes a scene text detection algorithm based on dual-branch feature extraction. This paper enlarges the receptive field by means of a residual correction branch (RCB), to obtain contextual information with a larger receptive field. At the same time, in order to improve the efficiency of using the features, a two-branch attentional feature fusion (TB-AFF) module is proposed based on FPN, to combine global and local attention to pinpoint text regions, enhance the sensitivity of the network to text regions, and accurately detect the text location in natural scenes. In this paper, several sets of comparative experiments were conducted and compared with the current mainstream text detection methods, all of which achieved better results, thus verifying the effectiveness of the improved proposed method.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.