Recently, deep convolutional neural networks (CNNs) have been successfully applied to the single-image superresolution (SISR) task with great improvement in terms of both peak signal-to-noise ratio (PSNR) and structural similarity (SSIM). However, most of the existing CNN-based SR models require high computing power, which considerably limits their real-world applications. In addition, most CNN-based methods rarely explore the intermediate features that are helpful for final image recovery. To address these issues, in this article, we propose a dense lightweight network, called MADNet, for stronger multiscale feature expression and feature correlation learning. Specifically, a residual multiscale module with an attention mechanism (RMAM) is developed to enhance the informative multiscale feature representation ability. Furthermore, we present a dual residual-path block (DRPB) that utilizes the hierarchical features from original low-resolution images. To take advantage of the multilevel features, dense connections are employed among blocks. The comparative results demonstrate the superior performance of our MADNet model while employing considerably fewer multiadds and parameters.
A new meta-heuristic method is proposed to enhance current meta-heuristic methods for global optimization and test scheduling for three-dimensional (3D) stacked system-on-chip (SoC) by hybridizing grey wolf optimization with differential evolution (HGWO). Because basic grey wolf optimization (GWO) is easy to fall into stagnation when it carries out the operation of attacking prey, and differential evolution (DE) is integrated into GWO to update the previous best position of grey wolf Alpha, Beta and Delta, in order to force GWO to jump out of the stagnation with DE's strong searching ability. The proposed algorithm can accelerate the convergence speed of GWO and improve its performance. Twenty-three well-known benchmark functions and an NP hard problem of test scheduling for 3D SoC are employed to verify the performance of the proposed algorithm. Experimental results show the superior performance of the proposed algorithm for exploiting the optimum and it has advantages in terms of exploration.
Deep convolutional neural networks (CNNs) have contributed to the significant progress of the single-image super-resolution (SISR) field. However, the majority of existing CNN-based models maintain high performance with massive parameters and exceedingly deeper structures. Moreover, several algorithms essentially have underused the low-level features, thus causing relatively low performance. In this article, we address these problems by exploring two strategies based on novel local wider residual blocks (LWRBs) to effectively extract the image features for SISR. We propose a cascading residual network (CRN) that contains several locally sharing groups (LSGs), in which the cascading mechanism not only promotes the propagation of features and the gradient but also eases the model training. Besides, we present another enhanced residual network (ERN) for image resolution enhancement. ERN employs a dual global pathway structure that incorporates nonlocal operations to catch long-distance spatial features from the the original low-resolution (LR) input. To obtain the feature representation of the input at different scales, we further introduce a multiscale block (MSB) to directly detect low-level features from the LR image. The experimental results on four benchmark datasets have demonstrated that our models outperform most of the advanced methods while still retaining a reasonable number of parameters.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.