Discriminative feature learning is the key to remote sensing scene classification. Previous research has found that most of the existing convolutional neural networks (CNN) focus on the global semantic features and ignore shallower features (low-level and middle-level features). This study proposes a novel Lie Group deep learning model for remote sensing scene classification to solve the above-mentioned challenges. Firstly, we extract shallower and higher-level features from images based on Lie Group machine learning (LGML) and deep learning to improve the feature representation ability of the model. In addition, a parallel dilated convolution, a kernel decomposition, and a Lie Group kernel function are adopted to reduce the model’s parameters to prevent model degradation and over-fitting caused by the deepening of the model. Then, the spatial attention mechanism can enhance local semantic features and suppress irrelevant feature information. Finally, feature-level fusion is adopted to reduce redundant features and improve computational performance, and cross-entropy loss function based on label smoothing is used to improve the classification accuracy of the model. Comparative experiments on three public and challenging large-scale remote-sensing datasets show that our model improves the discriminative ability of features and achieves competitive accuracy against other state-of-the-art methods.
Recent years, the research on object classification based on threedimensional (3D) point cloud pays more attention to extract the features from point sets directly. PointNet++ is the latest network structure for 3D classification which has achieved acceptable results. Although it has achieved acceptable results, there are still two problems: (i) The farthest point sampling (FPS) algorithm in PointNet++ ignores the fact that the feature of each point contributes differently to the classification and segmentation tasks. Therefore, FPS cannot guarantee that the selected point sets can correctly represent the main features of the object. (ii) The multi-scale grouping and multi-resolution grouping in PointNet++ do not consider the features between different levels of the same point. In order to solve these problems, the authors propose the point-selection structure which can calculate the importance of each point's feature. In addition, multi-level-point-feature fusion module is proposed to extract the features of the point at all levels and fuse them as a new feature of that point. In this Letter, they make some experiments on ModelNet40 and ScanNet datasets, which achieves better results compared to the state-of-the-art methods.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.