The occurrence of forest fires causes serious damage to ecological diversity and the safety of people’s property and life. However, due to the complex forest environment, the changeable shape of forest fires, and the uncertainty of flame color and texture, forest fire detection becomes very difficult. Traditional image processing methods rely heavily on artificial features and are not generally applicable to different forest fire scenes. In order to solve the problem of inaccurate forest fire recognition caused by the manual extraction of features, some scholars use deep learning technology to adaptively learn and extract forest fire features, but they often use a single target detection model, and their lack of learning and perception makes it difficult for them to accurately identify forest fires in a complex forest fire environment. Therefore, in order to overcome the shortcomings of the manual extraction of features and achieve a higher accuracy of forest fire recognition, this paper proposes an algorithm based on weighted fusion to identify forest fire sources in different scenarios, fuses two independent weakly supervised models Yolov5 and EfficientDet, completes the training and prediction of data sets in parallel, and uses the weighted boxes fusion algorithm (WBF) to process the prediction results to obtain the fusion frame. Finally, the model is evaluated by Microsoft COCO standard. Experimental results show that compared with Yolov5 and EfficientDet, the proposed Y4SED improves the detection performance by 2.5% to 4.5%. The fused algorithm proposed in this paper has better feature extraction ability, can extract more forest fire feature information, and better balances the recognition accuracy and complexity of the model, which provides a reference for forest fire target detection in the real environment.
The frequent occurrence of forest fires in recent years has not only seriously damaged the forests’ ecological environments but also threatened the safety of public life and property. Smoke, as the main manifestation of the flame before it is produced, has the advantage of a wide diffusion range that is not easily obscured. Therefore, timely detection of forest fire smoke with better real-time detection for early warnings of forest fires wins valuable time for timely firefighting and also has great significance and applications for the development of forest fire detection systems. However, existing forest fire smoke detection methods still have problems, such as low detection accuracy, slow detection speed, and difficulty detecting smoke from small targets. In order to solve the aforementioned problems and further achieve higher accuracy in detection, this paper proposes an improved, new, high-accuracy forest fire detection model, the OBDS. Firstly, to address the problem of insufficient extraction of effective features of forest fire smoke in complex forest environments, this paper introduces the SimAM attention mechanism, which makes the model pay more attention to the feature information of forest fire smoke and suppresses the interference of non-targeted background information. Moreover, this paper introduces Omni-Dimensional Dynamic Convolution instead of static convolution and adaptively and dynamically adjusts the weights of the convolution kernel, which enables the network to better extract the key features of forest fire smoke of different shapes and sizes. In addition, to address the problem that traditional convolutional neural networks are not capable of capturing global forest fire smoke feature information, this paper introduces the Bottleneck Transformer Net (BoTNet) to fully extract global feature information and local feature information of forest fire smoke images while improving the accuracy of small target forest fire target detection of smoke, effectively reducing the model’s computation, and improving the detection speed of model forest fire smoke. Finally, this paper introduces the decoupling head to further improve the detection accuracy of forest fire smoke and speed up the convergence of the model. Our experimental results show that the model OBDS for forest fire smoke detection proposed in this paper is significantly better than the mainstream model, with a computational complexity of 21.5 GFLOPs (giga floating-point operations per second), an improvement of 4.31% compared with the YOLOv5 (YOLO, you only look once) model mAP@0.5, reaching 92.10%, and an FPS (frames per second) of 54, which is conducive to the realization of early warning of forest fires.
Colorectal cancer (CRC) is one of the significant threats to public health and the sustainable healthcare system during urbanization. As the primary method of screening, colonoscopy can effectively detect polyps before they evolve into cancerous growths. However, the current visual inspection by endoscopists is insufficient in providing consistently reliable polyp detection for colonoscopy videos and images in CRC screening. Artificial Intelligent (AI) based object detection is considered as a potent solution to overcome visual inspection limitations and mitigate human errors in colonoscopy. This study implemented a YOLOv5 object detection model to investigate the performance of mainstream one-stage approaches in colorectal polyp detection. Meanwhile, a variety of training datasets and model structure configurations are employed to identify the determinative factors in practical applications. The designed experiments show that the model yields acceptable results assisted by transfer learning, and highlight that the primary constraint in implementing deep learning polyp detection comes from the scarcity of training data. The model performance was improved by 15.6% in terms of average precision (AP) when the original training dataset was expanded. Furthermore, the experimental results were analysed from a clinical perspective to identify potential causes of false positives. Besides, the quality management framework is proposed for future dataset preparation and model development in AI-driven polyp detection tasks for smart healthcare solutions.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.