Classifying birds accurately is essential for ecological monitoring. In recent years, bird image classification has become an emerging method for bird recognition. However, the bird image classification task needs to face the challenges of high intraclass variance and low inter-class variance among birds, as well as low model efficiency. In this paper, we propose a fine-grained bird classification method based on attention and decoupled knowledge distillation. First of all, we propose an attention-guided data augmentation method. Specifically, the method obtains images of the object’s key part regions through attention. It enables the model to learn and distinguish fine features. At the same time, based on the localization–recognition method, the bird category is predicted using the object image with finer features, which reduces the influence of background noise. In addition, we propose a model compression method of decoupled knowledge distillation. We distill the target and nontarget class knowledge separately to eliminate the influence of the target class prediction results on the transfer of the nontarget class knowledge. This approach achieves efficient model compression. With 67% fewer parameters and only 1.2 G of computation, the model proposed in this paper still has a 87.6% success rate, while improving the model inference speed.
The use of neural networks for plant disease identification is a hot topic of current research. However, unlike the classification of ordinary objects, the features of plant diseases frequently vary, resulting in substantial intra-class variation; in addition, the complex environmental noise makes it more challenging for the model to categorize the diseases. In this paper, an attention and multidimensional feature fusion neural network (AMDFNet) is proposed for Camellia oleifera disease classification network based on multidimensional feature fusion and attentional mechanism, which improves the classification ability of the model by fusing features to each layer of the Inception structure and enhancing the fused features with attentional enhancement. The model was compared with the classical convolutional neural networks GoogLeNet, Inception V3, ResNet50, and DenseNet121 and the latest disease image classification network DICNN in a self-built camellia disease dataset. The experimental results show that the recognition accuracy of the new model reaches 86.78% under the same experimental conditions, which is 2.3% higher than that of GoogLeNet with a simple Inception structure, and the number of parameters is reduced to one-fourth compared to large models such as ResNet50. The method proposed in this paper can be run on mobile with higher identification accuracy and a smaller model parameter number.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.