The classification of remote sensing scene objects has been the subject of extensive studies in recent years due to the quick advancement of earth observation and remote sensing technology. More concentration, Hyperspectral images, and LiDAR data are complementary, and their combined use for data fusion can better mine the multi-dimensional features of ground objects in remote sensing scenes, which can effectively improve the classification accuracy and reliability of ground objects in remote sensing scenes. Single-modal remote sensing data frequently cannot fully meet the needs of ground feature classification due to the increasingly complex types of ground features. In order to solve this problem, we have developed two distinct multi-source fusion classification approaches using LiDAR and hyperspectral data and deep learning techniques. They are the reconstructed multi-layer perceptron network based on the variational auto-Encoder (encoding-decoding form) and the two-stream input convolutional neural network based on the cross-channel reconstruction mechanism. These two approaches can help us find more effective and deeper feature extraction methods and feature fusion methods in this research direction and design training. We adopted the network architecture model used in this area of research and used experimental data to demonstrate the effectiveness and superiority of the suggested network model.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.