To solve the problems in two-step processing of image fusion and Super-Resolution Reconstruction (SRR), we propose a joint framework of image Fusion and Super-Resolution (FSR) based on multicomponent analysis and residual compensation. Inspired by the idea of multicomponent analysis, we design a new structure-texture decomposition model to realize multicomponent dictionary learning for the above task. To depict the relationship between low-resolution image and its corresponding high-resolution image, the correlation of their sparse coding coefficients is introduced in the model. To compensate the information loss during the SRR, we propose a reconstruction residue compensation mechanism, in which the reconstruction residual error is compensated into the initial result of FSR to improve the quality of the final processing result. In addition, we propose different fusion algorithms for structure and texture components. For structure components, the fusion rule with the maximum absolute value is adopted, and for texture components, we design a new saliency measure to construct fusion results. Experiment results show that the proposed method can well retain the brightness and detail information in the original image, and is superior to other methods in subjective and objective evaluation.
In recent years, remote-sensing image super-resolution (RSISR) methods based on convolutional neural networks (CNNs) have achieved significant progress. However, the limited receptive field of the convolutional kernel in CNNs hinders the network's ability to effectively capture long-range features in images, thus limiting further improvements in model performance. Additionally, the deployment of existing RSISR models to terminal devices is challenging due to their high computational complexity and large number of parameters. To address these issues, we propose a Context-Aware Lightweight Super-Resolution Network (CALSRN) for remote-sensing images. The proposed network primarily consists of Context-Aware Transformer Blocks (CATBs), which incorporate a Local Context Extraction Branch (LCEB) and a Global Context Extraction Branch (GCEB) to explore both local and global image features. Furthermore, a Dynamic Weight Generation Branch (DWGB) is designed to generate aggregation weights for global and local features, enabling dynamic adjustment of the aggregation process. Specifically, the GCEB employs a Swin Transformer-based structure to obtain global information, while the LCEB utilizes a CNN-based cross-attention mechanism to extract local information. Ultimately, global and local features are aggregated using the weights acquired from the DWGB, capturing the global and local dependencies of the image and enhancing the quality of super-resolution reconstruction. The experimental results demonstrate that the proposed method is capable of reconstructing high-quality images with fewer parameters and less computational complexity compared with existing methods.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.